Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
17,598
23,424,466,934
IssuesEvent
2022-08-14 07:09:06
Battle-s/battle-school-backend
https://api.github.com/repos/Battle-s/battle-school-backend
closed
[FEAT] ํ† ํฐ ์žฌ๋ฐœ๊ธ‰ ๋กœ์ง ๋ฐ ์‹œํ๋ฆฌํ‹ฐ ์ ์šฉ
feature :computer: processing :hourglass_flowing_sand:
## ์„ค๋ช… ## ์ฒดํฌ์‚ฌํ•ญ - [x] ํ† ํฐ ์žฌ๋ฐœ๊ธ‰ ๋กœ์ง ์„ค๊ณ„ ๋ฐ ๋ฉ”์†Œ๋“œ ์ž‘์„ฑ - [x] ํ† ํฐ ์ ์šฉ ์‹œํ๋ฆฌํ‹ฐ ์ˆ˜์ • - [x] ํ† ๊ทผ ์žฌ๋ฐœ๊ทผ ํ•˜๋Š” api ์ƒ์„ฑ - [x] api์— ์ ์šฉ ## ์ฐธ๊ณ ์ž๋ฃŒ ## ๊ด€๋ จ ๋…ผ์˜
1.0
[FEAT] ํ† ํฐ ์žฌ๋ฐœ๊ธ‰ ๋กœ์ง ๋ฐ ์‹œํ๋ฆฌํ‹ฐ ์ ์šฉ - ## ์„ค๋ช… ## ์ฒดํฌ์‚ฌํ•ญ - [x] ํ† ํฐ ์žฌ๋ฐœ๊ธ‰ ๋กœ์ง ์„ค๊ณ„ ๋ฐ ๋ฉ”์†Œ๋“œ ์ž‘์„ฑ - [x] ํ† ํฐ ์ ์šฉ ์‹œํ๋ฆฌํ‹ฐ ์ˆ˜์ • - [x] ํ† ๊ทผ ์žฌ๋ฐœ๊ทผ ํ•˜๋Š” api ์ƒ์„ฑ - [x] api์— ์ ์šฉ ## ์ฐธ๊ณ ์ž๋ฃŒ ## ๊ด€๋ จ ๋…ผ์˜
process
ํ† ํฐ ์žฌ๋ฐœ๊ธ‰ ๋กœ์ง ๋ฐ ์‹œํ๋ฆฌํ‹ฐ ์ ์šฉ ์„ค๋ช… ์ฒดํฌ์‚ฌํ•ญ ํ† ํฐ ์žฌ๋ฐœ๊ธ‰ ๋กœ์ง ์„ค๊ณ„ ๋ฐ ๋ฉ”์†Œ๋“œ ์ž‘์„ฑ ํ† ํฐ ์ ์šฉ ์‹œํ๋ฆฌํ‹ฐ ์ˆ˜์ • ํ† ๊ทผ ์žฌ๋ฐœ๊ทผ ํ•˜๋Š” api ์ƒ์„ฑ api์— ์ ์šฉ ์ฐธ๊ณ ์ž๋ฃŒ ๊ด€๋ จ ๋…ผ์˜
1
437,513
12,598,816,132
IssuesEvent
2020-06-11 04:05:08
jenkins-x/jx
https://api.github.com/repos/jenkins-x/jx
closed
`jx add app` uses PR message that doesn't match semantic PR rules
area/apps kind/bug lifecycle/rotten priority/important-longterm
Ran `jx add app athens` and the PR raised to the dev environment git repo was.. "Add jx-app-athens 0.0.17" which the Semantic Pull Request check doesn't like. We needed to edit the PR message to have "feat: " prefix.
1.0
`jx add app` uses PR message that doesn't match semantic PR rules - Ran `jx add app athens` and the PR raised to the dev environment git repo was.. "Add jx-app-athens 0.0.17" which the Semantic Pull Request check doesn't like. We needed to edit the PR message to have "feat: " prefix.
non_process
jx add app uses pr message that doesn t match semantic pr rules ran jx add app athens and the pr raised to the dev environment git repo was add jx app athens which the semantic pull request check doesn t like we needed to edit the pr message to have feat prefix
0
16,311
20,963,049,258
IssuesEvent
2022-03-28 01:42:53
huutho77/CNPMNC_ThayAi
https://api.github.com/repos/huutho77/CNPMNC_ThayAi
opened
Design UI for Home page, Login page and Register page
processing
Design UIs related Home page, Login page and Register Account page. Goal: - Minimal (Layout and Color) - Fully features related on page - With Home page, need show products and pagination with 10-20 (user optional) products per page - Register Account Page need require related information as required fields (additional: appears error message)
1.0
Design UI for Home page, Login page and Register page - Design UIs related Home page, Login page and Register Account page. Goal: - Minimal (Layout and Color) - Fully features related on page - With Home page, need show products and pagination with 10-20 (user optional) products per page - Register Account Page need require related information as required fields (additional: appears error message)
process
design ui for home page login page and register page design uis related home page login page and register account page goal minimal layout and color fully features related on page with home page need show products and pagination with user optional products per page register account page need require related information as required fields additional appears error message
1
7,795
10,949,565,942
IssuesEvent
2019-11-26 11:06:04
Open-EO/openeo-processes
https://api.github.com/repos/Open-EO/openeo-processes
closed
Rename dimension values?
accepted new process work in progress
In a discussion in #43, we discussed a use case where two cubes are merged. The issue is: What happens if there are collisions e.g. for the band names? Both data cubes could have a band called "B1" and in this case it should be possible to rename one of them before so that there is no collision, right? Could be part of the bigger discussion in #30. Potential names: rename_dimension_value or rename_label.
1.0
Rename dimension values? - In a discussion in #43, we discussed a use case where two cubes are merged. The issue is: What happens if there are collisions e.g. for the band names? Both data cubes could have a band called "B1" and in this case it should be possible to rename one of them before so that there is no collision, right? Could be part of the bigger discussion in #30. Potential names: rename_dimension_value or rename_label.
process
rename dimension values in a discussion in we discussed a use case where two cubes are merged the issue is what happens if there are collisions e g for the band names both data cubes could have a band called and in this case it should be possible to rename one of them before so that there is no collision right could be part of the bigger discussion in potential names rename dimension value or rename label
1
4,608
7,452,882,775
IssuesEvent
2018-03-29 09:53:21
rivine/rivine
https://api.github.com/repos/rivine/rivine
closed
add oneBlockStake config to chain constants
process_wontfix type_bug wontfix
such that a block stake can be split up in the future should it be needed
1.0
add oneBlockStake config to chain constants - such that a block stake can be split up in the future should it be needed
process
add oneblockstake config to chain constants such that a block stake can be split up in the future should it be needed
1
14,453
17,533,227,472
IssuesEvent
2021-08-12 01:46:22
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
spatial join: non-matching features in joined layer despite "discard records which could not be joined"
Feedback stale Processing Bug
Hi, There is an issue with the "Join attributes by location" algorithm (SpatialJoin.py) regarding the non-matching features from the input layer. The option "Discard records which could not be joined" is not handled properly. When I enable the option "Discard records which could not be joined" and set the output of the "unjoinable features from the first layer" to "skip output" then the "joined layer" does not contain the non-matching features as I would expect. But if I enable the "discard..." option and set the output of the "unjoinable features from the first layer" to create a temporary layer then the "joined layer" does also contain the non-matching features as does the output for "unjoinable features". So I had a look in de SpatialJoin.py file (for QGIS 3.10.7) and the problem is within the if-block at line 257, which handles the output of non-matching features. The condition in the if-statement at line 262 is the culprit since it is always true (assuming the "joined layer" output is not skipped). In my local installation I changed the if statement at line 262 from "if sink is not None:โ€ to "if not discard_nomatch and sink is not None:". Now the result is what I expect it to be. Can someone please correct this bug? I'm not comfortable enough with Python and Git to do it myself. Regards, Peter
1.0
spatial join: non-matching features in joined layer despite "discard records which could not be joined" - Hi, There is an issue with the "Join attributes by location" algorithm (SpatialJoin.py) regarding the non-matching features from the input layer. The option "Discard records which could not be joined" is not handled properly. When I enable the option "Discard records which could not be joined" and set the output of the "unjoinable features from the first layer" to "skip output" then the "joined layer" does not contain the non-matching features as I would expect. But if I enable the "discard..." option and set the output of the "unjoinable features from the first layer" to create a temporary layer then the "joined layer" does also contain the non-matching features as does the output for "unjoinable features". So I had a look in de SpatialJoin.py file (for QGIS 3.10.7) and the problem is within the if-block at line 257, which handles the output of non-matching features. The condition in the if-statement at line 262 is the culprit since it is always true (assuming the "joined layer" output is not skipped). In my local installation I changed the if statement at line 262 from "if sink is not None:โ€ to "if not discard_nomatch and sink is not None:". Now the result is what I expect it to be. Can someone please correct this bug? I'm not comfortable enough with Python and Git to do it myself. Regards, Peter
process
spatial join non matching features in joined layer despite discard records which could not be joined hi there is an issue with the join attributes by location algorithm spatialjoin py regarding the non matching features from the input layer the option discard records which could not be joined is not handled properly when i enable the option discard records which could not be joined and set the output of the unjoinable features from the first layer to skip output then the joined layer does not contain the non matching features as i would expect but if i enable the discard option and set the output of the unjoinable features from the first layer to create a temporary layer then the joined layer does also contain the non matching features as does the output for unjoinable features so i had a look in de spatialjoin py file for qgis and the problem is within the if block at line which handles the output of non matching features the condition in the if statement at line is the culprit since it is always true assuming the joined layer output is not skipped in my local installation i changed the if statement at line from if sink is not none โ€ to if not discard nomatch and sink is not none now the result is what i expect it to be can someone please correct this bug i m not comfortable enough with python and git to do it myself regards peter
1
119,544
17,620,316,536
IssuesEvent
2021-08-18 14:36:52
jgeraigery/experian-java
https://api.github.com/repos/jgeraigery/experian-java
closed
CVE-2018-11307 (High) detected in jackson-databind-2.9.2.jar - autoclosed
security vulnerability
## CVE-2018-11307 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9ade2a959068cca30ecfdbb254939af6f67affb1">9ade2a959068cca30ecfdbb254939af6f67affb1</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.5. Use of Jackson default typing along with a gadget class from iBatis allows exfiltration of content. Fixed in 2.7.9.4, 2.8.11.2, and 2.9.6. <p>Publish Date: 2019-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11307>CVE-2018-11307</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2032">https://github.com/FasterXML/jackson-databind/issues/2032</a></p> <p>Release Date: 2019-03-17</p> <p>Fix Resolution: jackson-databind-2.9.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jackson-databind-2.9.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-11307","vulnerabilityDetails":"An issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.5. Use of Jackson default typing along with a gadget class from iBatis allows exfiltration of content. Fixed in 2.7.9.4, 2.8.11.2, and 2.9.6.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11307","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-11307 (High) detected in jackson-databind-2.9.2.jar - autoclosed - ## CVE-2018-11307 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9ade2a959068cca30ecfdbb254939af6f67affb1">9ade2a959068cca30ecfdbb254939af6f67affb1</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.5. Use of Jackson default typing along with a gadget class from iBatis allows exfiltration of content. Fixed in 2.7.9.4, 2.8.11.2, and 2.9.6. <p>Publish Date: 2019-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11307>CVE-2018-11307</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2032">https://github.com/FasterXML/jackson-databind/issues/2032</a></p> <p>Release Date: 2019-03-17</p> <p>Fix Resolution: jackson-databind-2.9.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jackson-databind-2.9.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-11307","vulnerabilityDetails":"An issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.5. Use of Jackson default typing along with a gadget class from iBatis allows exfiltration of content. Fixed in 2.7.9.4, 2.8.11.2, and 2.9.6.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11307","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file experian java mavenworkspace bis services lib bis services base pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in fasterxml jackson databind through use of jackson default typing along with a gadget class from ibatis allows exfiltration of content fixed in and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails an issue was discovered in fasterxml jackson databind through use of jackson default typing along with a gadget class from ibatis allows exfiltration of content fixed in and vulnerabilityurl
0
19,857
26,263,615,482
IssuesEvent
2023-01-06 10:20:03
CGAL/cgal
https://api.github.com/repos/CGAL/cgal
closed
โ€œOpenGRโ€: is not member of โ€œCGALโ€
not a bug Pkg::Point_set_processing_3
## Issue Details I have installed CGAL 5.5.1 through vcpkg when I use compile example "registration_with_OpenGR.cpp", ocurrs erro "โ€œOpenGRโ€: is not member of โ€œCGALโ€" ## Source Code ``` #include <CGAL/Simple_cartesian.h> #include <CGAL/IO/read_points.h> #include <CGAL/IO/write_points.h> #include <CGAL/property_map.h> #include <CGAL/OpenGR/compute_registration_transformation.h> #include <CGAL/OpenGR/register_point_sets.h> #include <fstream> #include <iostream> #include <utility> typedef CGAL::Simple_cartesian<double> K; typedef K::Point_3 Point_3; typedef K::Vector_3 Vector_3; typedef std::pair<Point_3, Vector_3> Pwn; typedef CGAL::First_of_pair_property_map<Pwn> Point_map; typedef CGAL::Second_of_pair_property_map<Pwn> Normal_map; namespace params = CGAL::parameters; int main(int argc, const char** argv) { const std::string fname1 = (argc>1) ? argv[1] : CGAL::data_file_path("points_3/hippo1.ply"); const std::string fname2 = (argc>2) ? argv[2] : CGAL::data_file_path("points_3/hippo2.ply"); std::vector<Pwn> pwns1, pwns2; if(!CGAL::IO::read_points(fname1, std::back_inserter(pwns1), CGAL::parameters::point_map(CGAL::First_of_pair_property_map<Pwn>()) .normal_map(Normal_map()))) { std::cerr << "Error: cannot read file " << fname1 << std::endl; return EXIT_FAILURE; } if(!CGAL::IO::read_points(fname2, std::back_inserter(pwns2), CGAL::parameters::point_map(Point_map()) .normal_map(Normal_map()))) { std::cerr << "Error: cannot read file " << fname2 << std::endl; return EXIT_FAILURE; } // EITHER call the registration method Super4PCS from OpenGR to get the transformation to apply to pwns2 // std::pair<K::Aff_transformation_3, double> res = CGAL::OpenGR::compute_registration_transformation(pwns1, pwns2, params::point_map(Point_map()) .normal_map(Normal_map()) .number_of_samples(200) .maximum_running_time(60) .accuracy(0.01), params::point_map(Point_map()) .normal_map(Normal_map())); // OR call the registration method Super4PCS from OpenGR and apply the transformation to pwn2 double score = CGAL::OpenGR::register_point_sets(pwns1, pwns2, params::point_map(Point_map()) .normal_map(Normal_map()) .number_of_samples(200) .maximum_running_time(60) .accuracy(0.01), params::point_map(Point_map()) .normal_map(Normal_map())); if(!CGAL::IO::write_points("pwns2_aligned.ply", pwns2, CGAL::parameters::point_map(Point_map()) .normal_map(Normal_map()) .stream_precision(17))) return EXIT_FAILURE; std::cout << "Registration score: " << score << ".\n" << "Transformed version of " << fname2 << " written to pwn2_aligned.ply.\n"; return EXIT_SUCCESS; } ``` ## Environment * Operating system (Windows/64 bits): * Compiler: vs2017 * Release or debug mode:release mode * Specific flags used (if any): * CGAL version:5.5.1 * Boost version:1_80 * Other libraries versions if used (Eigen, TBB, etc.):Eigen 3.4.9 GMP and MPFR
1.0
โ€œOpenGRโ€: is not member of โ€œCGALโ€ - ## Issue Details I have installed CGAL 5.5.1 through vcpkg when I use compile example "registration_with_OpenGR.cpp", ocurrs erro "โ€œOpenGRโ€: is not member of โ€œCGALโ€" ## Source Code ``` #include <CGAL/Simple_cartesian.h> #include <CGAL/IO/read_points.h> #include <CGAL/IO/write_points.h> #include <CGAL/property_map.h> #include <CGAL/OpenGR/compute_registration_transformation.h> #include <CGAL/OpenGR/register_point_sets.h> #include <fstream> #include <iostream> #include <utility> typedef CGAL::Simple_cartesian<double> K; typedef K::Point_3 Point_3; typedef K::Vector_3 Vector_3; typedef std::pair<Point_3, Vector_3> Pwn; typedef CGAL::First_of_pair_property_map<Pwn> Point_map; typedef CGAL::Second_of_pair_property_map<Pwn> Normal_map; namespace params = CGAL::parameters; int main(int argc, const char** argv) { const std::string fname1 = (argc>1) ? argv[1] : CGAL::data_file_path("points_3/hippo1.ply"); const std::string fname2 = (argc>2) ? argv[2] : CGAL::data_file_path("points_3/hippo2.ply"); std::vector<Pwn> pwns1, pwns2; if(!CGAL::IO::read_points(fname1, std::back_inserter(pwns1), CGAL::parameters::point_map(CGAL::First_of_pair_property_map<Pwn>()) .normal_map(Normal_map()))) { std::cerr << "Error: cannot read file " << fname1 << std::endl; return EXIT_FAILURE; } if(!CGAL::IO::read_points(fname2, std::back_inserter(pwns2), CGAL::parameters::point_map(Point_map()) .normal_map(Normal_map()))) { std::cerr << "Error: cannot read file " << fname2 << std::endl; return EXIT_FAILURE; } // EITHER call the registration method Super4PCS from OpenGR to get the transformation to apply to pwns2 // std::pair<K::Aff_transformation_3, double> res = CGAL::OpenGR::compute_registration_transformation(pwns1, pwns2, params::point_map(Point_map()) .normal_map(Normal_map()) .number_of_samples(200) .maximum_running_time(60) .accuracy(0.01), params::point_map(Point_map()) .normal_map(Normal_map())); // OR call the registration method Super4PCS from OpenGR and apply the transformation to pwn2 double score = CGAL::OpenGR::register_point_sets(pwns1, pwns2, params::point_map(Point_map()) .normal_map(Normal_map()) .number_of_samples(200) .maximum_running_time(60) .accuracy(0.01), params::point_map(Point_map()) .normal_map(Normal_map())); if(!CGAL::IO::write_points("pwns2_aligned.ply", pwns2, CGAL::parameters::point_map(Point_map()) .normal_map(Normal_map()) .stream_precision(17))) return EXIT_FAILURE; std::cout << "Registration score: " << score << ".\n" << "Transformed version of " << fname2 << " written to pwn2_aligned.ply.\n"; return EXIT_SUCCESS; } ``` ## Environment * Operating system (Windows/64 bits): * Compiler: vs2017 * Release or debug mode:release mode * Specific flags used (if any): * CGAL version:5.5.1 * Boost version:1_80 * Other libraries versions if used (Eigen, TBB, etc.):Eigen 3.4.9 GMP and MPFR
process
โ€œopengrโ€ is not member of โ€œcgalโ€ issue details i have installed cgal through vcpkg when i use compile example registration with opengr cpp ocurrs erro โ€œopengrโ€ is not member of โ€œcgalโ€ source code include include include include include include include include include typedef cgal simple cartesian k typedef k point point typedef k vector vector typedef std pair pwn typedef cgal first of pair property map point map typedef cgal second of pair property map normal map namespace params cgal parameters int main int argc const char argv const std string argc argv cgal data file path points ply const std string argc argv cgal data file path points ply std vector if cgal io read points std back inserter cgal parameters point map cgal first of pair property map normal map normal map std cerr error cannot read file std endl return exit failure if cgal io read points std back inserter cgal parameters point map point map normal map normal map std cerr error cannot read file std endl return exit failure either call the registration method from opengr to get the transformation to apply to std pair res cgal opengr compute registration transformation params point map point map normal map normal map number of samples maximum running time accuracy params point map point map normal map normal map or call the registration method from opengr and apply the transformation to double score cgal opengr register point sets params point map point map normal map normal map number of samples maximum running time accuracy params point map point map normal map normal map if cgal io write points aligned ply cgal parameters point map point map normal map normal map stream precision return exit failure std cout registration score score n transformed version of written to aligned ply n return exit success environment operating system windows bits compiler release or debug mode release mode specific flags used if any cgal version boost version other libraries versions if used eigen tbb etc eigen gmp and mpfr
1
9,538
12,506,661,206
IssuesEvent
2020-06-02 12:57:39
GoogleCloudPlatform/dotnet-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples
opened
[Storage] Deactivate GoogleCloudSamples.StorageTest.TestDownloadObjectRequesterPays because of flakiness.
api: storage priority: p1 type: process
It's not clear why it's failing. Have deactivate in PR (will update when PR is in)
1.0
[Storage] Deactivate GoogleCloudSamples.StorageTest.TestDownloadObjectRequesterPays because of flakiness. - It's not clear why it's failing. Have deactivate in PR (will update when PR is in)
process
deactivate googlecloudsamples storagetest testdownloadobjectrequesterpays because of flakiness it s not clear why it s failing have deactivate in pr will update when pr is in
1
9,845
12,836,894,113
IssuesEvent
2020-07-07 14:59:29
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
closed
Unittest framework
process + tools
Which unittest framework do we want to use? [Pytest](https://docs.pytest.org/en/stable/) seems popular - anything newer/better we should be aware of/use?
1.0
Unittest framework - Which unittest framework do we want to use? [Pytest](https://docs.pytest.org/en/stable/) seems popular - anything newer/better we should be aware of/use?
process
unittest framework which unittest framework do we want to use seems popular anything newer better we should be aware of use
1
85,755
15,755,229,271
IssuesEvent
2021-03-31 01:24:37
akshat702/map-ionic
https://api.github.com/repos/akshat702/map-ionic
opened
CVE-2020-28502 (High) detected in xmlhttprequest-ssl-1.5.5.tgz
security vulnerability
## CVE-2020-28502 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmlhttprequest-ssl-1.5.5.tgz</b></p></summary> <p>XMLHttpRequest for Node</p> <p>Library home page: <a href="https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.5.tgz">https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.5.tgz</a></p> <p>Path to dependency file: map-ionic/dynamic/package.json</p> <p>Path to vulnerable library: map-ionic/e2e/node_modules/xmlhttprequest-ssl/package.json,map-ionic/e2e/node_modules/xmlhttprequest-ssl/package.json,map-ionic/e2e/node_modules/xmlhttprequest-ssl/package.json,map-ionic/e2e/node_modules/xmlhttprequest-ssl/package.json,map-ionic/e2e/node_modules/xmlhttprequest-ssl/package.json</p> <p> Dependency Hierarchy: - karma-3.1.4.tgz (Root Library) - socket.io-2.1.1.tgz - socket.io-client-2.1.1.tgz - engine.io-client-3.2.1.tgz - :x: **xmlhttprequest-ssl-1.5.5.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package xmlhttprequest before 1.7.0; all versions of package xmlhttprequest-ssl. Provided requests are sent synchronously (async=False on xhr.open), malicious user input flowing into xhr.send could result in arbitrary code being injected and run. <p>Publish Date: 2021-03-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28502>CVE-2020-28502</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28502">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28502</a></p> <p>Release Date: 2021-03-05</p> <p>Fix Resolution: xmlhttprequest - 1.7.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-28502 (High) detected in xmlhttprequest-ssl-1.5.5.tgz - ## CVE-2020-28502 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmlhttprequest-ssl-1.5.5.tgz</b></p></summary> <p>XMLHttpRequest for Node</p> <p>Library home page: <a href="https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.5.tgz">https://registry.npmjs.org/xmlhttprequest-ssl/-/xmlhttprequest-ssl-1.5.5.tgz</a></p> <p>Path to dependency file: map-ionic/dynamic/package.json</p> <p>Path to vulnerable library: map-ionic/e2e/node_modules/xmlhttprequest-ssl/package.json,map-ionic/e2e/node_modules/xmlhttprequest-ssl/package.json,map-ionic/e2e/node_modules/xmlhttprequest-ssl/package.json,map-ionic/e2e/node_modules/xmlhttprequest-ssl/package.json,map-ionic/e2e/node_modules/xmlhttprequest-ssl/package.json</p> <p> Dependency Hierarchy: - karma-3.1.4.tgz (Root Library) - socket.io-2.1.1.tgz - socket.io-client-2.1.1.tgz - engine.io-client-3.2.1.tgz - :x: **xmlhttprequest-ssl-1.5.5.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package xmlhttprequest before 1.7.0; all versions of package xmlhttprequest-ssl. Provided requests are sent synchronously (async=False on xhr.open), malicious user input flowing into xhr.send could result in arbitrary code being injected and run. <p>Publish Date: 2021-03-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28502>CVE-2020-28502</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28502">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28502</a></p> <p>Release Date: 2021-03-05</p> <p>Fix Resolution: xmlhttprequest - 1.7.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in xmlhttprequest ssl tgz cve high severity vulnerability vulnerable library xmlhttprequest ssl tgz xmlhttprequest for node library home page a href path to dependency file map ionic dynamic package json path to vulnerable library map ionic node modules xmlhttprequest ssl package json map ionic node modules xmlhttprequest ssl package json map ionic node modules xmlhttprequest ssl package json map ionic node modules xmlhttprequest ssl package json map ionic node modules xmlhttprequest ssl package json dependency hierarchy karma tgz root library socket io tgz socket io client tgz engine io client tgz x xmlhttprequest ssl tgz vulnerable library vulnerability details this affects the package xmlhttprequest before all versions of package xmlhttprequest ssl provided requests are sent synchronously async false on xhr open malicious user input flowing into xhr send could result in arbitrary code being injected and run publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution xmlhttprequest step up your open source security game with whitesource
0
3,217
12,300,623,840
IssuesEvent
2020-05-11 14:15:28
short-d/short
https://api.github.com/repos/short-d/short
closed
[Refactor] Move detailed configurations into separate files
maintainability
**What is frustrating you?** There is too much content in the README. Most of them is not related to the first time environment setup. This is driving away developers who are interested in trying out Short on their local machine. **Your solution** Have only one type of sign in setup in `Getting Started` section. Move the individual setups to new markdown files.
True
[Refactor] Move detailed configurations into separate files - **What is frustrating you?** There is too much content in the README. Most of them is not related to the first time environment setup. This is driving away developers who are interested in trying out Short on their local machine. **Your solution** Have only one type of sign in setup in `Getting Started` section. Move the individual setups to new markdown files.
non_process
move detailed configurations into separate files what is frustrating you there is too much content in the readme most of them is not related to the first time environment setup this is driving away developers who are interested in trying out short on their local machine your solution have only one type of sign in setup in getting started section move the individual setups to new markdown files
0
7,437
10,550,482,132
IssuesEvent
2019-10-03 11:10:09
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Create a resource menu has been changed
Pri2 automation/svc cxp process-automation/subsvc triaged
There is no more "Management Tools" menu item on the "Create a resource" section. --- #### Document Details โš  *Do not edit this section. It is required for docs.microsoft.com โžŸ GitHub issue linking.* * ID: d7c2cc34-ba4a-1181-ccda-88dd901e0212 * Version Independent ID: 9b4a3b68-03fc-387a-ee5d-e7f73ee3c567 * Content: [Azure Quickstart - Create an Azure Automation account](https://docs.microsoft.com/en-us/azure/automation/automation-quickstart-create-account#feedback) * Content Source: [articles/automation/automation-quickstart-create-account.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-quickstart-create-account.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @csand-msft * Microsoft Alias: **csand**
1.0
Create a resource menu has been changed - There is no more "Management Tools" menu item on the "Create a resource" section. --- #### Document Details โš  *Do not edit this section. It is required for docs.microsoft.com โžŸ GitHub issue linking.* * ID: d7c2cc34-ba4a-1181-ccda-88dd901e0212 * Version Independent ID: 9b4a3b68-03fc-387a-ee5d-e7f73ee3c567 * Content: [Azure Quickstart - Create an Azure Automation account](https://docs.microsoft.com/en-us/azure/automation/automation-quickstart-create-account#feedback) * Content Source: [articles/automation/automation-quickstart-create-account.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-quickstart-create-account.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @csand-msft * Microsoft Alias: **csand**
process
create a resource menu has been changed there is no more management tools menu item on the create a resource section document details โš  do not edit this section it is required for docs microsoft com โžŸ github issue linking id ccda version independent id content content source service automation sub service process automation github login csand msft microsoft alias csand
1
635,579
20,406,619,659
IssuesEvent
2022-02-23 06:40:52
kubesphere/console
https://api.github.com/repos/kubesphere/console
closed
Edit federated service tag selector error
kind/bug kind/need-to-verify priority/low
**Describe the bug** 1. click more,click Edit Config Template 2. delete the key, click save,click to confirm 3. The update is successful, check that the resource status has not changed ![image](https://user-images.githubusercontent.com/88183150/135220190-ca1f00ee-1ed4-4f94-9be5-944baf958e2b.png) ![image](https://user-images.githubusercontent.com/88183150/135221417-2af39457-f79e-42b0-a04a-5c7465e771c4.png) **Versions used(KubeSphere/Kubernetes)** KubeSphere: nightly-20210927 /kind bug /@kubesphere/sig-console /priority low
1.0
Edit federated service tag selector error - **Describe the bug** 1. click more,click Edit Config Template 2. delete the key, click save,click to confirm 3. The update is successful, check that the resource status has not changed ![image](https://user-images.githubusercontent.com/88183150/135220190-ca1f00ee-1ed4-4f94-9be5-944baf958e2b.png) ![image](https://user-images.githubusercontent.com/88183150/135221417-2af39457-f79e-42b0-a04a-5c7465e771c4.png) **Versions used(KubeSphere/Kubernetes)** KubeSphere: nightly-20210927 /kind bug /@kubesphere/sig-console /priority low
non_process
edit federated service tag selector error describe the bug click more click edit config template delete the key click save click to confirm the update is successful check that the resource status has not changed versions used kubesphere kubernetes kubesphere nightly kind bug kubesphere sig console priority low
0
8,552
11,727,298,142
IssuesEvent
2020-03-10 15:44:44
MHRA/products
https://api.github.com/repos/MHRA/products
closed
AUTOMATIC BATCH PROCESS - Create service retrieves files from Sentinel
EPIC - Auto Batch Process :oncoming_automobile: HIGH PRIORITY :arrow_double_up: TASK :rescue_worker_helmet:
### User want As a user I want to see up to date documents on the products website So I can make informed decisions **Customer acceptance criteria** **Technical acceptance criteria** - [x] Create message retrieved (prerequisite on #421) - [ ] SFTP server credentials are read from environment variables - [ ] Create service retrieves file at path given in create message from configured SFTP server **Data acceptance criteria** **Testing acceptance criteria** - [ ] Integration tests use a local SFTP server - [ ] Integration tests cover process from calling document manager endpoint to retrieving the file from SFTP **Size** L **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
1.0
AUTOMATIC BATCH PROCESS - Create service retrieves files from Sentinel - ### User want As a user I want to see up to date documents on the products website So I can make informed decisions **Customer acceptance criteria** **Technical acceptance criteria** - [x] Create message retrieved (prerequisite on #421) - [ ] SFTP server credentials are read from environment variables - [ ] Create service retrieves file at path given in create message from configured SFTP server **Data acceptance criteria** **Testing acceptance criteria** - [ ] Integration tests use a local SFTP server - [ ] Integration tests cover process from calling document manager endpoint to retrieving the file from SFTP **Size** L **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
process
automatic batch process create service retrieves files from sentinel user want as a user i want to see up to date documents on the products website so i can make informed decisions customer acceptance criteria technical acceptance criteria create message retrieved prerequisite on sftp server credentials are read from environment variables create service retrieves file at path given in create message from configured sftp server data acceptance criteria testing acceptance criteria integration tests use a local sftp server integration tests cover process from calling document manager endpoint to retrieving the file from sftp size l value effort exit criteria met backlog discovery duxd development quality assurance release and validate
1
420,562
12,239,712,478
IssuesEvent
2020-05-04 22:16:33
ooni/ooni.org
https://api.github.com/repos/ooni/ooni.org
closed
Contribute to KeepItOn report
effort/M priority/high report
This requires potentially some data analysis to extract relevant measurements.
1.0
Contribute to KeepItOn report - This requires potentially some data analysis to extract relevant measurements.
non_process
contribute to keepiton report this requires potentially some data analysis to extract relevant measurements
0
267,270
28,502,086,389
IssuesEvent
2023-04-18 18:10:26
RG4421/ampere-centos-kernel
https://api.github.com/repos/RG4421/ampere-centos-kernel
reopened
CVE-2019-19063 (Medium) detected in linuxv5.2
Mend: dependency security vulnerability
## CVE-2019-19063 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Two memory leaks in the rtl_usb_probe() function in drivers/net/wireless/realtek/rtlwifi/usb.c in the Linux kernel through 5.3.11 allow attackers to cause a denial of service (memory consumption), aka CID-3f9361695113. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19063>CVE-2019-19063</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-19063">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-19063</a></p> <p>Release Date: 2019-11-18</p> <p>Fix Resolution: v5.5-rc1</p> </p> </details> <p></p>
True
CVE-2019-19063 (Medium) detected in linuxv5.2 - ## CVE-2019-19063 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Two memory leaks in the rtl_usb_probe() function in drivers/net/wireless/realtek/rtlwifi/usb.c in the Linux kernel through 5.3.11 allow attackers to cause a denial of service (memory consumption), aka CID-3f9361695113. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19063>CVE-2019-19063</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-19063">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2019-19063</a></p> <p>Release Date: 2019-11-18</p> <p>Fix Resolution: v5.5-rc1</p> </p> </details> <p></p>
non_process
cve medium detected in cve medium severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files vulnerability details two memory leaks in the rtl usb probe function in drivers net wireless realtek rtlwifi usb c in the linux kernel through allow attackers to cause a denial of service memory consumption aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
4,127
10,592,183,894
IssuesEvent
2019-10-09 12:35:59
dotnet/docs
https://api.github.com/repos/dotnet/docs
closed
How much information should go in an Integration Event?
:book: guide - .NET Microservices :books: Area - .NET Architecture Guide Source - Docs.ms
Should an integration event simply hold the AggregateID and the deltas of what changed in this event? Or should it publish the entirety of aggregate? In the example of `ProductPriceChangedIntegrationEvent` it looks like just the delta is published. I could see the benefit of publishing the whole aggregate as it would allow any downstream consumers to take what ever information it needs from the event. If there are any view stores that may care about displaying data from the aggregate of a different service it could simply overwrite it's current projection of that aggregate. Perhaps adding an aggregate version # would make sure that you're not overwriting newer data with an older event? --- #### Document Details โš  *Do not edit this section. It is required for docs.microsoft.com โžŸ GitHub issue linking.* * ID: b037f198-3ee2-28f4-4d93-0999e5d3d3ef * Version Independent ID: 9bd96308-c726-fcee-850e-a38fd22ff555 * Content: [Implementing event-based communication between microservices (integration events)](https://docs.microsoft.com/en-us/dotnet/architecture/microservices/multi-container-microservice-net-applications/integration-event-based-microservice-communications#feedback) * Content Source: [docs/architecture/microservices/multi-container-microservice-net-applications/integration-event-based-microservice-communications.md](https://github.com/dotnet/docs/blob/master/docs/architecture/microservices/multi-container-microservice-net-applications/integration-event-based-microservice-communications.md) * Product: **dotnet** * Technology: **dotnet-ebooks** * GitHub Login: @nishanil * Microsoft Alias: **nanil**
1.0
How much information should go in an Integration Event? - Should an integration event simply hold the AggregateID and the deltas of what changed in this event? Or should it publish the entirety of aggregate? In the example of `ProductPriceChangedIntegrationEvent` it looks like just the delta is published. I could see the benefit of publishing the whole aggregate as it would allow any downstream consumers to take what ever information it needs from the event. If there are any view stores that may care about displaying data from the aggregate of a different service it could simply overwrite it's current projection of that aggregate. Perhaps adding an aggregate version # would make sure that you're not overwriting newer data with an older event? --- #### Document Details โš  *Do not edit this section. It is required for docs.microsoft.com โžŸ GitHub issue linking.* * ID: b037f198-3ee2-28f4-4d93-0999e5d3d3ef * Version Independent ID: 9bd96308-c726-fcee-850e-a38fd22ff555 * Content: [Implementing event-based communication between microservices (integration events)](https://docs.microsoft.com/en-us/dotnet/architecture/microservices/multi-container-microservice-net-applications/integration-event-based-microservice-communications#feedback) * Content Source: [docs/architecture/microservices/multi-container-microservice-net-applications/integration-event-based-microservice-communications.md](https://github.com/dotnet/docs/blob/master/docs/architecture/microservices/multi-container-microservice-net-applications/integration-event-based-microservice-communications.md) * Product: **dotnet** * Technology: **dotnet-ebooks** * GitHub Login: @nishanil * Microsoft Alias: **nanil**
non_process
how much information should go in an integration event should an integration event simply hold the aggregateid and the deltas of what changed in this event or should it publish the entirety of aggregate in the example of productpricechangedintegrationevent it looks like just the delta is published i could see the benefit of publishing the whole aggregate as it would allow any downstream consumers to take what ever information it needs from the event if there are any view stores that may care about displaying data from the aggregate of a different service it could simply overwrite it s current projection of that aggregate perhaps adding an aggregate version would make sure that you re not overwriting newer data with an older event document details โš  do not edit this section it is required for docs microsoft com โžŸ github issue linking id version independent id fcee content content source product dotnet technology dotnet ebooks github login nishanil microsoft alias nanil
0
9,822
12,827,548,705
IssuesEvent
2020-07-06 18:42:39
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
Docker as preferred driver on macOS/Windows
co/docker-driver kind/process priority/important-soon
At the moment, we default to: - Docker on Linux - Hyperkit on macOS - Hyper-V on Windows - On macOS, Docker is better at VPN handling. - On Windows, Docker is better at privilege handling. This suggests that for most users, we could make the default Docker for all platforms, and fall-back to Hyperkit/Hyper-V if Docker is not available.
1.0
Docker as preferred driver on macOS/Windows - At the moment, we default to: - Docker on Linux - Hyperkit on macOS - Hyper-V on Windows - On macOS, Docker is better at VPN handling. - On Windows, Docker is better at privilege handling. This suggests that for most users, we could make the default Docker for all platforms, and fall-back to Hyperkit/Hyper-V if Docker is not available.
process
docker as preferred driver on macos windows at the moment we default to docker on linux hyperkit on macos hyper v on windows on macos docker is better at vpn handling on windows docker is better at privilege handling this suggests that for most users we could make the default docker for all platforms and fall back to hyperkit hyper v if docker is not available
1
98,401
29,828,897,466
IssuesEvent
2023-06-18 02:21:56
bazelbuild/bazel-buildfarm
https://api.github.com/repos/bazelbuild/bazel-buildfarm
closed
Bazel client error: FMB size limit exceeded
bug help wanted java Build Failure
Any call/construction of FMB request on server should be size limited. ``` (18:37:35) ERROR: <redacted>:11:8: Testing <redacted> failed: (Exit 34): Remote Execution Failure: Unknown: FINDMISSINGBLOBS IS TOO LARGE: 80650 digests are required in one request! java.io.IOException: com.google.devtools.build.lib.remote.ExecutionStatusException: UNKNOWN: FINDMISSINGBLOBS IS TOO LARGE: 80650 digests are required in one request! at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.executeRemotely(GrpcRemoteExecutor.java:235) at com.google.devtools.build.lib.remote.RemoteExecutionService.executeRemotely(RemoteExecutionService.java:1496) at com.google.devtools.build.lib.remote.RemoteSpawnRunner.lambda$exec$2(RemoteSpawnRunner.java:269) at com.google.devtools.build.lib.remote.Retrier.execute(Retrier.java:244) at com.google.devtools.build.lib.remote.RemoteRetrier.execute(RemoteRetrier.java:127) at com.google.devtools.build.lib.remote.RemoteRetrier.execute(RemoteRetrier.java:116) at com.google.devtools.build.lib.remote.RemoteSpawnRunner.exec(RemoteSpawnRunner.java:244) at com.google.devtools.build.lib.exec.SpawnRunner.execAsync(SpawnRunner.java:301) at com.google.devtools.build.lib.exec.AbstractSpawnStrategy.exec(AbstractSpawnStrategy.java:152) at com.google.devtools.build.lib.exec.AbstractSpawnStrategy.exec(AbstractSpawnStrategy.java:112) at com.google.devtools.build.lib.actions.SpawnStrategy.beginExecution(SpawnStrategy.java:47) at com.google.devtools.build.lib.exec.SpawnStrategyResolver.beginExecution(SpawnStrategyResolver.java:64) at com.google.devtools.build.lib.exec.StandaloneTestStrategy.beginTestAttempt(StandaloneTestStrategy.java:323) at com.google.devtools.build.lib.exec.StandaloneTestStrategy$StandaloneTestRunnerSpawn.beginExecution(StandaloneTestStrategy.java:599) at com.google.devtools.build.lib.analysis.test.TestRunnerAction.beginIfNotCancelled(TestRunnerAction.java:991) at com.google.devtools.build.lib.analysis.test.TestRunnerAction.beginExecution(TestRunnerAction.java:958) at com.google.devtools.build.lib.analysis.test.TestRunnerAction.execute(TestRunnerAction.java:1015) at com.google.devtools.build.lib.analysis.test.TestRunnerAction.execute(TestRunnerAction.java:1007) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$5.execute(SkyframeActionExecutor.java:957) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.continueAction(SkyframeActionExecutor.java:1124) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.run(SkyframeActionExecutor.java:1082) at com.google.devtools.build.lib.skyframe.ActionExecutionState.runStateMachine(ActionExecutionState.java:160) at com.google.devtools.build.lib.skyframe.ActionExecutionState.getResultOrDependOnFuture(ActionExecutionState.java:93) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor.executeAction(SkyframeActionExecutor.java:516) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.checkCacheAndExecuteIfNeeded(ActionExecutionFunction.java:827) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.computeInternal(ActionExecutionFunction.java:323) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.compute(ActionExecutionFunction.java:161) at com.google.devtools.build.skyframe.AbstractParallelEvaluator$Evaluate.run(AbstractParallelEvaluator.java:571) at com.google.devtools.build.lib.concurrent.AbstractQueueVisitor$WrappedRunnable.run(AbstractQueueVisitor.java:382) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) Caused by: com.google.devtools.build.lib.remote.ExecutionStatusException: UNKNOWN: FINDMISSINGBLOBS IS TOO LARGE: 80650 digests are required in one request! at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.handleStatus(GrpcRemoteExecutor.java:71) at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.getOperationResponse(GrpcRemoteExecutor.java:83) at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.lambda$executeRemotely$2(GrpcRemoteExecutor.java:194) at com.google.devtools.build.lib.remote.Retrier.execute(Retrier.java:244) at com.google.devtools.build.lib.remote.RemoteRetrier.execute(RemoteRetrier.java:127) at com.google.devtools.build.lib.remote.RemoteRetrier.execute(RemoteRetrier.java:116) at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.lambda$executeRemotely$3(GrpcRemoteExecutor.java:140) at com.google.devtools.build.lib.remote.util.Utils.refreshIfUnauthenticated(Utils.java:523) at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.executeRemotely(GrpcRemoteExecutor.java:138) ... 31 more ```
1.0
Bazel client error: FMB size limit exceeded - Any call/construction of FMB request on server should be size limited. ``` (18:37:35) ERROR: <redacted>:11:8: Testing <redacted> failed: (Exit 34): Remote Execution Failure: Unknown: FINDMISSINGBLOBS IS TOO LARGE: 80650 digests are required in one request! java.io.IOException: com.google.devtools.build.lib.remote.ExecutionStatusException: UNKNOWN: FINDMISSINGBLOBS IS TOO LARGE: 80650 digests are required in one request! at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.executeRemotely(GrpcRemoteExecutor.java:235) at com.google.devtools.build.lib.remote.RemoteExecutionService.executeRemotely(RemoteExecutionService.java:1496) at com.google.devtools.build.lib.remote.RemoteSpawnRunner.lambda$exec$2(RemoteSpawnRunner.java:269) at com.google.devtools.build.lib.remote.Retrier.execute(Retrier.java:244) at com.google.devtools.build.lib.remote.RemoteRetrier.execute(RemoteRetrier.java:127) at com.google.devtools.build.lib.remote.RemoteRetrier.execute(RemoteRetrier.java:116) at com.google.devtools.build.lib.remote.RemoteSpawnRunner.exec(RemoteSpawnRunner.java:244) at com.google.devtools.build.lib.exec.SpawnRunner.execAsync(SpawnRunner.java:301) at com.google.devtools.build.lib.exec.AbstractSpawnStrategy.exec(AbstractSpawnStrategy.java:152) at com.google.devtools.build.lib.exec.AbstractSpawnStrategy.exec(AbstractSpawnStrategy.java:112) at com.google.devtools.build.lib.actions.SpawnStrategy.beginExecution(SpawnStrategy.java:47) at com.google.devtools.build.lib.exec.SpawnStrategyResolver.beginExecution(SpawnStrategyResolver.java:64) at com.google.devtools.build.lib.exec.StandaloneTestStrategy.beginTestAttempt(StandaloneTestStrategy.java:323) at com.google.devtools.build.lib.exec.StandaloneTestStrategy$StandaloneTestRunnerSpawn.beginExecution(StandaloneTestStrategy.java:599) at com.google.devtools.build.lib.analysis.test.TestRunnerAction.beginIfNotCancelled(TestRunnerAction.java:991) at com.google.devtools.build.lib.analysis.test.TestRunnerAction.beginExecution(TestRunnerAction.java:958) at com.google.devtools.build.lib.analysis.test.TestRunnerAction.execute(TestRunnerAction.java:1015) at com.google.devtools.build.lib.analysis.test.TestRunnerAction.execute(TestRunnerAction.java:1007) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$5.execute(SkyframeActionExecutor.java:957) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.continueAction(SkyframeActionExecutor.java:1124) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor$ActionRunner.run(SkyframeActionExecutor.java:1082) at com.google.devtools.build.lib.skyframe.ActionExecutionState.runStateMachine(ActionExecutionState.java:160) at com.google.devtools.build.lib.skyframe.ActionExecutionState.getResultOrDependOnFuture(ActionExecutionState.java:93) at com.google.devtools.build.lib.skyframe.SkyframeActionExecutor.executeAction(SkyframeActionExecutor.java:516) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.checkCacheAndExecuteIfNeeded(ActionExecutionFunction.java:827) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.computeInternal(ActionExecutionFunction.java:323) at com.google.devtools.build.lib.skyframe.ActionExecutionFunction.compute(ActionExecutionFunction.java:161) at com.google.devtools.build.skyframe.AbstractParallelEvaluator$Evaluate.run(AbstractParallelEvaluator.java:571) at com.google.devtools.build.lib.concurrent.AbstractQueueVisitor$WrappedRunnable.run(AbstractQueueVisitor.java:382) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.base/java.lang.Thread.run(Unknown Source) Caused by: com.google.devtools.build.lib.remote.ExecutionStatusException: UNKNOWN: FINDMISSINGBLOBS IS TOO LARGE: 80650 digests are required in one request! at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.handleStatus(GrpcRemoteExecutor.java:71) at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.getOperationResponse(GrpcRemoteExecutor.java:83) at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.lambda$executeRemotely$2(GrpcRemoteExecutor.java:194) at com.google.devtools.build.lib.remote.Retrier.execute(Retrier.java:244) at com.google.devtools.build.lib.remote.RemoteRetrier.execute(RemoteRetrier.java:127) at com.google.devtools.build.lib.remote.RemoteRetrier.execute(RemoteRetrier.java:116) at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.lambda$executeRemotely$3(GrpcRemoteExecutor.java:140) at com.google.devtools.build.lib.remote.util.Utils.refreshIfUnauthenticated(Utils.java:523) at com.google.devtools.build.lib.remote.GrpcRemoteExecutor.executeRemotely(GrpcRemoteExecutor.java:138) ... 31 more ```
non_process
bazel client error fmb size limit exceeded any call construction of fmb request on server should be size limited error testing failed exit remote execution failure unknown findmissingblobs is too large digests are required in one request java io ioexception com google devtools build lib remote executionstatusexception unknown findmissingblobs is too large digests are required in one request at com google devtools build lib remote grpcremoteexecutor executeremotely grpcremoteexecutor java at com google devtools build lib remote remoteexecutionservice executeremotely remoteexecutionservice java at com google devtools build lib remote remotespawnrunner lambda exec remotespawnrunner java at com google devtools build lib remote retrier execute retrier java at com google devtools build lib remote remoteretrier execute remoteretrier java at com google devtools build lib remote remoteretrier execute remoteretrier java at com google devtools build lib remote remotespawnrunner exec remotespawnrunner java at com google devtools build lib exec spawnrunner execasync spawnrunner java at com google devtools build lib exec abstractspawnstrategy exec abstractspawnstrategy java at com google devtools build lib exec abstractspawnstrategy exec abstractspawnstrategy java at com google devtools build lib actions spawnstrategy beginexecution spawnstrategy java at com google devtools build lib exec spawnstrategyresolver beginexecution spawnstrategyresolver java at com google devtools build lib exec standaloneteststrategy begintestattempt standaloneteststrategy java at com google devtools build lib exec standaloneteststrategy standalonetestrunnerspawn beginexecution standaloneteststrategy java at com google devtools build lib analysis test testrunneraction beginifnotcancelled testrunneraction java at com google devtools build lib analysis test testrunneraction beginexecution testrunneraction java at com google devtools build lib analysis test testrunneraction execute testrunneraction java at com google devtools build lib analysis test testrunneraction execute testrunneraction java at com google devtools build lib skyframe skyframeactionexecutor execute skyframeactionexecutor java at com google devtools build lib skyframe skyframeactionexecutor actionrunner continueaction skyframeactionexecutor java at com google devtools build lib skyframe skyframeactionexecutor actionrunner run skyframeactionexecutor java at com google devtools build lib skyframe actionexecutionstate runstatemachine actionexecutionstate java at com google devtools build lib skyframe actionexecutionstate getresultordependonfuture actionexecutionstate java at com google devtools build lib skyframe skyframeactionexecutor executeaction skyframeactionexecutor java at com google devtools build lib skyframe actionexecutionfunction checkcacheandexecuteifneeded actionexecutionfunction java at com google devtools build lib skyframe actionexecutionfunction computeinternal actionexecutionfunction java at com google devtools build lib skyframe actionexecutionfunction compute actionexecutionfunction java at com google devtools build skyframe abstractparallelevaluator evaluate run abstractparallelevaluator java at com google devtools build lib concurrent abstractqueuevisitor wrappedrunnable run abstractqueuevisitor java at java base java util concurrent threadpoolexecutor runworker unknown source at java base java util concurrent threadpoolexecutor worker run unknown source at java base java lang thread run unknown source caused by com google devtools build lib remote executionstatusexception unknown findmissingblobs is too large digests are required in one request at com google devtools build lib remote grpcremoteexecutor handlestatus grpcremoteexecutor java at com google devtools build lib remote grpcremoteexecutor getoperationresponse grpcremoteexecutor java at com google devtools build lib remote grpcremoteexecutor lambda executeremotely grpcremoteexecutor java at com google devtools build lib remote retrier execute retrier java at com google devtools build lib remote remoteretrier execute remoteretrier java at com google devtools build lib remote remoteretrier execute remoteretrier java at com google devtools build lib remote grpcremoteexecutor lambda executeremotely grpcremoteexecutor java at com google devtools build lib remote util utils refreshifunauthenticated utils java at com google devtools build lib remote grpcremoteexecutor executeremotely grpcremoteexecutor java more
0
4,308
7,200,409,698
IssuesEvent
2018-02-05 18:57:54
zcash/zips
https://api.github.com/repos/zcash/zips
opened
Meta-Proposal: Move all ZIP drafts into master branch.
ZIP process
I'd personally prefer to have all ZIP drafts typically in the master branch, with relatively small iterations on any given ZIP going through PRs. The reason is that I want to do a `git clone` and browse all ZIPs, both mature/finalized ZIPs as well as drafts and proposals. This would help particularly where multiple ZIP proposals are interacting. In this approach, the scope of changes and discussion for a PR may be smaller and simpler, because acceptance of the PR would *not be* acceptance of a ZIP proposal. It would only be acceptance of the contents of a proposal. Thoughts?
1.0
Meta-Proposal: Move all ZIP drafts into master branch. - I'd personally prefer to have all ZIP drafts typically in the master branch, with relatively small iterations on any given ZIP going through PRs. The reason is that I want to do a `git clone` and browse all ZIPs, both mature/finalized ZIPs as well as drafts and proposals. This would help particularly where multiple ZIP proposals are interacting. In this approach, the scope of changes and discussion for a PR may be smaller and simpler, because acceptance of the PR would *not be* acceptance of a ZIP proposal. It would only be acceptance of the contents of a proposal. Thoughts?
process
meta proposal move all zip drafts into master branch i d personally prefer to have all zip drafts typically in the master branch with relatively small iterations on any given zip going through prs the reason is that i want to do a git clone and browse all zips both mature finalized zips as well as drafts and proposals this would help particularly where multiple zip proposals are interacting in this approach the scope of changes and discussion for a pr may be smaller and simpler because acceptance of the pr would not be acceptance of a zip proposal it would only be acceptance of the contents of a proposal thoughts
1
275,804
30,306,722,840
IssuesEvent
2023-07-10 09:58:23
GeoNode/geonode
https://api.github.com/repos/GeoNode/geonode
closed
GNIP 94: Generic and pluggable OIDC SocialAccount Provider for GeoNode
gnip security
# GNIP 94: Generic and pluggable OIDC SocialAccount Provider for GeoNode ## Overview Currently GeoNode provides by default 2 quite old and outdated `SOCIALACCOUNT` providers, `LinkedIn` and `Facebook`, which are based on `OAuth2` and make use of very old plugins. Today the reference protocol is `OIDC`, and GeoNode users increasingly need to be able to link their accounts to providers that support and implement this technology. Two of the most commonly used providers are Google and Microsoft Azure, for example. The purpose of this proposal is to revise the `SocialAccount` and `SocialProvider` classes in GeoNode in order to not only make them capable of handling the more modern OIDC protocol but also to take advantage of some of its features, such as extracting user information from the `id_token` after validating its origin, among others. Furthermore, a significant benefit of this technology is that the protocol is now a widely respected standard among all providers. This allows us to create dynamic, modular, and scalable structures that enable GeoNode to handle almost all use cases through a few configuration parameters. ### Proposed By @afabiani ### Assigned to Release This proposal is for GeoNode 4.1.2+. ### State * [x] Under Discussion * [x] In Progress * [x] Completed * [ ] Rejected * [ ] Deferred ### Proposal ## The `geonode_openid_connect` social account provider This will be the generic class which will allow us handling the OIDC social login ```python class GenericOpenIDConnectProvider(OAuth2Provider): id = "geonode_openid_connect" name = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("NAME", "GeoNode OpenIDConnect") account_class = import_class_module( getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}) .get(PROVIDER_ID, {}) .get( "ACCOUNT_CLASS", "geonode.people.socialaccount.providers.geonode_openid_connect.provider.GenericOpenIDConnectProviderAccount", ) ) def get_default_scope(self): scope = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("SCOPE", "") return scope def get_auth_params(self, request, action): ret = super(GenericOpenIDConnectProvider, self).get_auth_params(request, action) if action == AuthAction.REAUTHENTICATE: ret["prompt"] = ( getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}) .get(PROVIDER_ID, {}) .get("AUTH_PARAMS", {}) .get("prompt", "") ) return ret def extract_uid(self, data): return data.get("sub", data.get("id")) def extract_common_fields(self, data): _common_fields = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("COMMON_FIELDS", {}) return _common_fields def extract_email_addresses(self, data): addresses = [] email = data.get("email") if email: addresses.append( EmailAddress( email=email, verified=data.get("email_verified", False), primary=True, ) ) return addresses ``` Notice that the class is fully customizable, from the `name`, the `account class` to the `common fields` that we would like to extract from the `JSON` user-info response. All those parameters can be driver by the GeoNode `settings` ## A fully pluggable Provider configuration From the `settings` point of view, configuring the provider will simply require to fill the common customizable properties of the OIDC authority. As an instance, the following ones are two sample configurations for `Google` and `Microsofr Azure` ```python INSTALLED_APPS += ("geonode.people.socialaccount.providers.geonode_openid_connect",) _AZURE_TENANT_ID = os.getenv("MICROSOFT_TENANT_ID", "") _AZURE_SOCIALACCOUNT_PROVIDER = { "NAME": "Microsoft Azure", "SCOPE": [ "User.Read", "openid", ], "AUTH_PARAMS": { "access_type": "online", "prompt": "select_account", }, "COMMON_FIELDS": {"email": "mail", "last_name": "surname", "first_name": "givenName"}, "ACCOUNT_CLASS": "allauth.socialaccount.providers.azure.provider.AzureAccount", "ACCESS_TOKEN_URL": f"https://login.microsoftonline.com/{_AZURE_TENANT_ID}/oauth2/v2.0/token", "AUTHORIZE_URL": f"https://login.microsoftonline.com/{_AZURE_TENANT_ID}/oauth2/v2.0/authorize", "PROFILE_URL": "https://graph.microsoft.com/v1.0/me", } _GOOGLE_SOCIALACCOUNT_PROVIDER = { "NAME": "Google", "SCOPE": [ "profile", "email", ], "AUTH_PARAMS": { "access_type": "online", "prompt": "select_account consent", }, "COMMON_FIELDS": {"email": "email", "last_name": "family_name", "first_name": "given_name"}, "ACCOUNT_CLASS": "allauth.socialaccount.providers.google.provider.GoogleAccount", "ACCESS_TOKEN_URL": "https://oauth2.googleapis.com/token", "AUTHORIZE_URL": "https://accounts.google.com/o/oauth2/v2/auth", "ID_TOKEN_ISSUER": "https://accounts.google.com", "OAUTH_PKCE_ENABLED": True, } SOCIALACCOUNT_PROVIDERS = { "geonode_openid_connect": _AZURE_SOCIALACCOUNT_PROVIDER, } ``` By updating the `SOCIALACCOUNT_PROVIDERS` dictionary you can easily switch from a provider to another. Notice how their configuration is mostly the same. The only values the user should provide are the correct OIDC endpoints specific to the authority we would like to use. ## The default `AccountAdapter` Within the GeoNode classes will be available also a generic account provider which is defined like this ```python SOCIALACCOUNT_ADAPTER = os.environ.get("SOCIALACCOUNT_ADAPTER", "geonode.people.adapters.GenericOpenIDConnectAdapter") SOCIALACCOUNT_PROFILE_EXTRACTORS = { "geonode_openid_connect": "geonode.people.profileextractors.OpenIDExtractor", } ``` The `GenericOpenIDConnectAdapter` extends few common `login` methods which will allow us to seemlessly extract the user information both from the `UserInfoURI` and the `IDToken` ```python PROVIDER_ID = getattr(settings, "SOCIALACCOUNT_OIDC_PROVIDER", "geonode_openid_connect") ACCESS_TOKEN_URL = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("ACCESS_TOKEN_URL", "") AUTHORIZE_URL = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("AUTHORIZE_URL", "") PROFILE_URL = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("PROFILE_URL", "") ID_TOKEN_ISSUER = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("ID_TOKEN_ISSUER", "") class GenericOpenIDConnectAdapter(OAuth2Adapter, SocialAccountAdapter): provider_id = PROVIDER_ID access_token_url = ACCESS_TOKEN_URL authorize_url = AUTHORIZE_URL profile_url = PROFILE_URL id_token_issuer = ID_TOKEN_ISSUER def complete_login(self, request, app, token, response, **kwargs): extra_data = {} if self.profile_url: headers = {"Authorization": "Bearer {0}".format(token.token)} resp = requests.get(self.profile_url, headers=headers) profile_data = resp.json() extra_data.update(profile_data) elif "id_token" in response: try: extra_data = jwt.decode( response["id_token"], # Since the token was received by direct communication # protected by TLS between this library and Google, we # are allowed to skip checking the token signature # according to the OpenID Connect Core 1.0 # specification. # https://openid.net/specs/openid-connect-core-1_0.html#IDTokenValidation options={ "verify_signature": False, "verify_iss": True, "verify_aud": True, "verify_exp": True, }, issuer=self.id_token_issuer, audience=app.client_id, ) except jwt.PyJWTError as e: raise OAuth2Error("Invalid id_token") from e login = self.get_provider().sociallogin_from_response(request, extra_data) return login def save_user(self, request, sociallogin, form=None): user = super(SocialAccountAdapter, self).save_user(request, sociallogin, form=form) extractor = get_data_extractor(sociallogin.account.provider) try: groups = extractor.extract_groups(sociallogin.account.extra_data) or extractor.extract_roles( sociallogin.account.extra_data ) # check here if user is member already of other groups and remove it form the ones that are not declared here... for groupprofile in user.group_list_all(): groupprofile.leave(user) for group_name in groups: groupprofile = GroupProfile.objects.filter(slug=group_name).first() if groupprofile: groupprofile.join(user) except (AttributeError, NotImplementedError): pass # extractor doesn't define a method for extracting field return user ``` The `complete_login` method will check whether we want to extract the user `extra-info` Json data from the `id_token` or from the `user-info` endpoint. The `save_user` method will extract the commond fields from the Json data by making a mapping between the Json properties and the GeoNode `UserProfile` model. Moreover it will check if the returned data declares some `Groups` or `Roles` the user belongs to. In that case it will check if the corresponding `GroupProfile` exists already on the GeoNode database and automatically assign the user to it. ### Backwards Compatibility No backwards compatibility. ## Future evolution Add more sample configurations for the most popular OIDC providers. ## Feedback - None yet. ## Voting Project Steering Committee: * Alessio Fabiani: * Francesco Bartoli: * Giovanni Allegri: +1 * Toni Schoenbuchner: +1 * Florian Hoedt: +1 ## Links Remove unused links below. * [Email Discussion]() * [Pull Request]() * [Mail Discussion]() * [Linked Issue]()
True
GNIP 94: Generic and pluggable OIDC SocialAccount Provider for GeoNode - # GNIP 94: Generic and pluggable OIDC SocialAccount Provider for GeoNode ## Overview Currently GeoNode provides by default 2 quite old and outdated `SOCIALACCOUNT` providers, `LinkedIn` and `Facebook`, which are based on `OAuth2` and make use of very old plugins. Today the reference protocol is `OIDC`, and GeoNode users increasingly need to be able to link their accounts to providers that support and implement this technology. Two of the most commonly used providers are Google and Microsoft Azure, for example. The purpose of this proposal is to revise the `SocialAccount` and `SocialProvider` classes in GeoNode in order to not only make them capable of handling the more modern OIDC protocol but also to take advantage of some of its features, such as extracting user information from the `id_token` after validating its origin, among others. Furthermore, a significant benefit of this technology is that the protocol is now a widely respected standard among all providers. This allows us to create dynamic, modular, and scalable structures that enable GeoNode to handle almost all use cases through a few configuration parameters. ### Proposed By @afabiani ### Assigned to Release This proposal is for GeoNode 4.1.2+. ### State * [x] Under Discussion * [x] In Progress * [x] Completed * [ ] Rejected * [ ] Deferred ### Proposal ## The `geonode_openid_connect` social account provider This will be the generic class which will allow us handling the OIDC social login ```python class GenericOpenIDConnectProvider(OAuth2Provider): id = "geonode_openid_connect" name = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("NAME", "GeoNode OpenIDConnect") account_class = import_class_module( getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}) .get(PROVIDER_ID, {}) .get( "ACCOUNT_CLASS", "geonode.people.socialaccount.providers.geonode_openid_connect.provider.GenericOpenIDConnectProviderAccount", ) ) def get_default_scope(self): scope = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("SCOPE", "") return scope def get_auth_params(self, request, action): ret = super(GenericOpenIDConnectProvider, self).get_auth_params(request, action) if action == AuthAction.REAUTHENTICATE: ret["prompt"] = ( getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}) .get(PROVIDER_ID, {}) .get("AUTH_PARAMS", {}) .get("prompt", "") ) return ret def extract_uid(self, data): return data.get("sub", data.get("id")) def extract_common_fields(self, data): _common_fields = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("COMMON_FIELDS", {}) return _common_fields def extract_email_addresses(self, data): addresses = [] email = data.get("email") if email: addresses.append( EmailAddress( email=email, verified=data.get("email_verified", False), primary=True, ) ) return addresses ``` Notice that the class is fully customizable, from the `name`, the `account class` to the `common fields` that we would like to extract from the `JSON` user-info response. All those parameters can be driver by the GeoNode `settings` ## A fully pluggable Provider configuration From the `settings` point of view, configuring the provider will simply require to fill the common customizable properties of the OIDC authority. As an instance, the following ones are two sample configurations for `Google` and `Microsofr Azure` ```python INSTALLED_APPS += ("geonode.people.socialaccount.providers.geonode_openid_connect",) _AZURE_TENANT_ID = os.getenv("MICROSOFT_TENANT_ID", "") _AZURE_SOCIALACCOUNT_PROVIDER = { "NAME": "Microsoft Azure", "SCOPE": [ "User.Read", "openid", ], "AUTH_PARAMS": { "access_type": "online", "prompt": "select_account", }, "COMMON_FIELDS": {"email": "mail", "last_name": "surname", "first_name": "givenName"}, "ACCOUNT_CLASS": "allauth.socialaccount.providers.azure.provider.AzureAccount", "ACCESS_TOKEN_URL": f"https://login.microsoftonline.com/{_AZURE_TENANT_ID}/oauth2/v2.0/token", "AUTHORIZE_URL": f"https://login.microsoftonline.com/{_AZURE_TENANT_ID}/oauth2/v2.0/authorize", "PROFILE_URL": "https://graph.microsoft.com/v1.0/me", } _GOOGLE_SOCIALACCOUNT_PROVIDER = { "NAME": "Google", "SCOPE": [ "profile", "email", ], "AUTH_PARAMS": { "access_type": "online", "prompt": "select_account consent", }, "COMMON_FIELDS": {"email": "email", "last_name": "family_name", "first_name": "given_name"}, "ACCOUNT_CLASS": "allauth.socialaccount.providers.google.provider.GoogleAccount", "ACCESS_TOKEN_URL": "https://oauth2.googleapis.com/token", "AUTHORIZE_URL": "https://accounts.google.com/o/oauth2/v2/auth", "ID_TOKEN_ISSUER": "https://accounts.google.com", "OAUTH_PKCE_ENABLED": True, } SOCIALACCOUNT_PROVIDERS = { "geonode_openid_connect": _AZURE_SOCIALACCOUNT_PROVIDER, } ``` By updating the `SOCIALACCOUNT_PROVIDERS` dictionary you can easily switch from a provider to another. Notice how their configuration is mostly the same. The only values the user should provide are the correct OIDC endpoints specific to the authority we would like to use. ## The default `AccountAdapter` Within the GeoNode classes will be available also a generic account provider which is defined like this ```python SOCIALACCOUNT_ADAPTER = os.environ.get("SOCIALACCOUNT_ADAPTER", "geonode.people.adapters.GenericOpenIDConnectAdapter") SOCIALACCOUNT_PROFILE_EXTRACTORS = { "geonode_openid_connect": "geonode.people.profileextractors.OpenIDExtractor", } ``` The `GenericOpenIDConnectAdapter` extends few common `login` methods which will allow us to seemlessly extract the user information both from the `UserInfoURI` and the `IDToken` ```python PROVIDER_ID = getattr(settings, "SOCIALACCOUNT_OIDC_PROVIDER", "geonode_openid_connect") ACCESS_TOKEN_URL = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("ACCESS_TOKEN_URL", "") AUTHORIZE_URL = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("AUTHORIZE_URL", "") PROFILE_URL = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("PROFILE_URL", "") ID_TOKEN_ISSUER = getattr(settings, "SOCIALACCOUNT_PROVIDERS", {}).get(PROVIDER_ID, {}).get("ID_TOKEN_ISSUER", "") class GenericOpenIDConnectAdapter(OAuth2Adapter, SocialAccountAdapter): provider_id = PROVIDER_ID access_token_url = ACCESS_TOKEN_URL authorize_url = AUTHORIZE_URL profile_url = PROFILE_URL id_token_issuer = ID_TOKEN_ISSUER def complete_login(self, request, app, token, response, **kwargs): extra_data = {} if self.profile_url: headers = {"Authorization": "Bearer {0}".format(token.token)} resp = requests.get(self.profile_url, headers=headers) profile_data = resp.json() extra_data.update(profile_data) elif "id_token" in response: try: extra_data = jwt.decode( response["id_token"], # Since the token was received by direct communication # protected by TLS between this library and Google, we # are allowed to skip checking the token signature # according to the OpenID Connect Core 1.0 # specification. # https://openid.net/specs/openid-connect-core-1_0.html#IDTokenValidation options={ "verify_signature": False, "verify_iss": True, "verify_aud": True, "verify_exp": True, }, issuer=self.id_token_issuer, audience=app.client_id, ) except jwt.PyJWTError as e: raise OAuth2Error("Invalid id_token") from e login = self.get_provider().sociallogin_from_response(request, extra_data) return login def save_user(self, request, sociallogin, form=None): user = super(SocialAccountAdapter, self).save_user(request, sociallogin, form=form) extractor = get_data_extractor(sociallogin.account.provider) try: groups = extractor.extract_groups(sociallogin.account.extra_data) or extractor.extract_roles( sociallogin.account.extra_data ) # check here if user is member already of other groups and remove it form the ones that are not declared here... for groupprofile in user.group_list_all(): groupprofile.leave(user) for group_name in groups: groupprofile = GroupProfile.objects.filter(slug=group_name).first() if groupprofile: groupprofile.join(user) except (AttributeError, NotImplementedError): pass # extractor doesn't define a method for extracting field return user ``` The `complete_login` method will check whether we want to extract the user `extra-info` Json data from the `id_token` or from the `user-info` endpoint. The `save_user` method will extract the commond fields from the Json data by making a mapping between the Json properties and the GeoNode `UserProfile` model. Moreover it will check if the returned data declares some `Groups` or `Roles` the user belongs to. In that case it will check if the corresponding `GroupProfile` exists already on the GeoNode database and automatically assign the user to it. ### Backwards Compatibility No backwards compatibility. ## Future evolution Add more sample configurations for the most popular OIDC providers. ## Feedback - None yet. ## Voting Project Steering Committee: * Alessio Fabiani: * Francesco Bartoli: * Giovanni Allegri: +1 * Toni Schoenbuchner: +1 * Florian Hoedt: +1 ## Links Remove unused links below. * [Email Discussion]() * [Pull Request]() * [Mail Discussion]() * [Linked Issue]()
non_process
gnip generic and pluggable oidc socialaccount provider for geonode gnip generic and pluggable oidc socialaccount provider for geonode overview currently geonode provides by default quite old and outdated socialaccount providers linkedin and facebook which are based on and make use of very old plugins today the reference protocol is oidc and geonode users increasingly need to be able to link their accounts to providers that support and implement this technology two of the most commonly used providers are google and microsoft azure for example the purpose of this proposal is to revise the socialaccount and socialprovider classes in geonode in order to not only make them capable of handling the more modern oidc protocol but also to take advantage of some of its features such as extracting user information from the id token after validating its origin among others furthermore a significant benefit of this technology is that the protocol is now a widely respected standard among all providers this allows us to create dynamic modular and scalable structures that enable geonode to handle almost all use cases through a few configuration parameters proposed by afabiani assigned to release this proposal is for geonode state under discussion in progress completed rejected deferred proposal the geonode openid connect social account provider this will be the generic class which will allow us handling the oidc social login python class genericopenidconnectprovider id geonode openid connect name getattr settings socialaccount providers get provider id get name geonode openidconnect account class import class module getattr settings socialaccount providers get provider id get account class geonode people socialaccount providers geonode openid connect provider genericopenidconnectprovideraccount def get default scope self scope getattr settings socialaccount providers get provider id get scope return scope def get auth params self request action ret super genericopenidconnectprovider self get auth params request action if action authaction reauthenticate ret getattr settings socialaccount providers get provider id get auth params get prompt return ret def extract uid self data return data get sub data get id def extract common fields self data common fields getattr settings socialaccount providers get provider id get common fields return common fields def extract email addresses self data addresses email data get email if email addresses append emailaddress email email verified data get email verified false primary true return addresses notice that the class is fully customizable from the name the account class to the common fields that we would like to extract from the json user info response all those parameters can be driver by the geonode settings a fully pluggable provider configuration from the settings point of view configuring the provider will simply require to fill the common customizable properties of the oidc authority as an instance the following ones are two sample configurations for google and microsofr azure python installed apps geonode people socialaccount providers geonode openid connect azure tenant id os getenv microsoft tenant id azure socialaccount provider name microsoft azure scope user read openid auth params access type online prompt select account common fields email mail last name surname first name givenname account class allauth socialaccount providers azure provider azureaccount access token url f authorize url f profile url google socialaccount provider name google scope profile email auth params access type online prompt select account consent common fields email email last name family name first name given name account class allauth socialaccount providers google provider googleaccount access token url authorize url id token issuer oauth pkce enabled true socialaccount providers geonode openid connect azure socialaccount provider by updating the socialaccount providers dictionary you can easily switch from a provider to another notice how their configuration is mostly the same the only values the user should provide are the correct oidc endpoints specific to the authority we would like to use the default accountadapter within the geonode classes will be available also a generic account provider which is defined like this python socialaccount adapter os environ get socialaccount adapter geonode people adapters genericopenidconnectadapter socialaccount profile extractors geonode openid connect geonode people profileextractors openidextractor the genericopenidconnectadapter extends few common login methods which will allow us to seemlessly extract the user information both from the userinfouri and the idtoken python provider id getattr settings socialaccount oidc provider geonode openid connect access token url getattr settings socialaccount providers get provider id get access token url authorize url getattr settings socialaccount providers get provider id get authorize url profile url getattr settings socialaccount providers get provider id get profile url id token issuer getattr settings socialaccount providers get provider id get id token issuer class genericopenidconnectadapter socialaccountadapter provider id provider id access token url access token url authorize url authorize url profile url profile url id token issuer id token issuer def complete login self request app token response kwargs extra data if self profile url headers authorization bearer format token token resp requests get self profile url headers headers profile data resp json extra data update profile data elif id token in response try extra data jwt decode response since the token was received by direct communication protected by tls between this library and google we are allowed to skip checking the token signature according to the openid connect core specification options verify signature false verify iss true verify aud true verify exp true issuer self id token issuer audience app client id except jwt pyjwterror as e raise invalid id token from e login self get provider sociallogin from response request extra data return login def save user self request sociallogin form none user super socialaccountadapter self save user request sociallogin form form extractor get data extractor sociallogin account provider try groups extractor extract groups sociallogin account extra data or extractor extract roles sociallogin account extra data check here if user is member already of other groups and remove it form the ones that are not declared here for groupprofile in user group list all groupprofile leave user for group name in groups groupprofile groupprofile objects filter slug group name first if groupprofile groupprofile join user except attributeerror notimplementederror pass extractor doesn t define a method for extracting field return user the complete login method will check whether we want to extract the user extra info json data from the id token or from the user info endpoint the save user method will extract the commond fields from the json data by making a mapping between the json properties and the geonode userprofile model moreover it will check if the returned data declares some groups or roles the user belongs to in that case it will check if the corresponding groupprofile exists already on the geonode database and automatically assign the user to it backwards compatibility no backwards compatibility future evolution add more sample configurations for the most popular oidc providers feedback none yet voting project steering committee alessio fabiani francesco bartoli giovanni allegri toni schoenbuchner florian hoedt links remove unused links below
0
66,773
27,578,582,198
IssuesEvent
2023-03-08 14:44:26
openstreetmap/operations
https://api.github.com/repos/openstreetmap/operations
closed
forum content migration to discourse
service:discourse
Steps identified: - [x] dump or access to fluxBB mysql - [x] initial import test -> https://forum-import-test.openstreetmap.org/ - [x] retrieve osm_id and generate OAuth2 linkage - [x] avoid duplicate accounts for existing Discourse accounts (dedup based on username) - [x] ~~retrieve sticky topics flag ?~~ - [x] fix quotes in posts - [x] fix external links in posts - [x] fix internal forum links in posts, recreate Discourse internal links ? - [x] recreate mapping between old topic/post id to new ones - [x] link redirect to new URLs - [x] automatic language detection on posts to add language tag to them - [x] ~~retrieve avatars ?~~ --- In order to minimize work time on your side, I propose to test the content migration on a fresh Discourse setup I can install temporarily on OSM-FR servers (I can give you access to it if you want). For this Iโ€™ll need a read only access to the fluxBB mysql (or do a mysql_dump of it). Regarding user private data, the script only needs these fields in the โ€œusersโ€ table: id, username, realname name, url website, email email, registered created_at, registration_ip registration_ip_address, last_visit last_visit_time, last_email_sent last_emailed_at, location, group_id If you prefer, you can limit the read only access to these fields on that table and, of course, I guarantee not to use personal data in any way outside of this migration process. Script improvements - Iโ€™ve seen that sticky topics are not handled by the script, Iโ€™ll check if it is possible to improve that as well as getting the avatar for the users and some additional user preferences. - Iโ€™ll check also how to keep the mapping between id of old topics and posts and new ones, in order to have create the URL redirections. Once a first migration test looks fine, we can share the URL of the temporary discourse to have more eyes looking at the result to fix residual problems and iterate if required. Finally, if the migration looks ok and is globally approved, youโ€™ll simply have to run the updated migration script on the real instance. Is this process ok for you ? Regarding timing, I can spend my next week-end on that. --- Regarding login management, I didn't know the OSM account was mandatory for the fluxBB forum and also for Discourse login. As @tomhughes mentionned on Discourse, this will require additional access to login details. Was fluxBB modified to deal with that ?
1.0
forum content migration to discourse - Steps identified: - [x] dump or access to fluxBB mysql - [x] initial import test -> https://forum-import-test.openstreetmap.org/ - [x] retrieve osm_id and generate OAuth2 linkage - [x] avoid duplicate accounts for existing Discourse accounts (dedup based on username) - [x] ~~retrieve sticky topics flag ?~~ - [x] fix quotes in posts - [x] fix external links in posts - [x] fix internal forum links in posts, recreate Discourse internal links ? - [x] recreate mapping between old topic/post id to new ones - [x] link redirect to new URLs - [x] automatic language detection on posts to add language tag to them - [x] ~~retrieve avatars ?~~ --- In order to minimize work time on your side, I propose to test the content migration on a fresh Discourse setup I can install temporarily on OSM-FR servers (I can give you access to it if you want). For this Iโ€™ll need a read only access to the fluxBB mysql (or do a mysql_dump of it). Regarding user private data, the script only needs these fields in the โ€œusersโ€ table: id, username, realname name, url website, email email, registered created_at, registration_ip registration_ip_address, last_visit last_visit_time, last_email_sent last_emailed_at, location, group_id If you prefer, you can limit the read only access to these fields on that table and, of course, I guarantee not to use personal data in any way outside of this migration process. Script improvements - Iโ€™ve seen that sticky topics are not handled by the script, Iโ€™ll check if it is possible to improve that as well as getting the avatar for the users and some additional user preferences. - Iโ€™ll check also how to keep the mapping between id of old topics and posts and new ones, in order to have create the URL redirections. Once a first migration test looks fine, we can share the URL of the temporary discourse to have more eyes looking at the result to fix residual problems and iterate if required. Finally, if the migration looks ok and is globally approved, youโ€™ll simply have to run the updated migration script on the real instance. Is this process ok for you ? Regarding timing, I can spend my next week-end on that. --- Regarding login management, I didn't know the OSM account was mandatory for the fluxBB forum and also for Discourse login. As @tomhughes mentionned on Discourse, this will require additional access to login details. Was fluxBB modified to deal with that ?
non_process
forum content migration to discourse steps identified dump or access to fluxbb mysql initial import test retrieve osm id and generate linkage avoid duplicate accounts for existing discourse accounts dedup based on username retrieve sticky topics flag fix quotes in posts fix external links in posts fix internal forum links in posts recreate discourse internal links recreate mapping between old topic post id to new ones link redirect to new urls automatic language detection on posts to add language tag to them retrieve avatars in order to minimize work time on your side i propose to test the content migration on a fresh discourse setup i can install temporarily on osm fr servers i can give you access to it if you want for this iโ€™ll need a read only access to the fluxbb mysql or do a mysql dump of it regarding user private data the script only needs these fields in the โ€œusersโ€ table id username realname name url website email email registered created at registration ip registration ip address last visit last visit time last email sent last emailed at location group id if you prefer you can limit the read only access to these fields on that table and of course i guarantee not to use personal data in any way outside of this migration process script improvements iโ€™ve seen that sticky topics are not handled by the script iโ€™ll check if it is possible to improve that as well as getting the avatar for the users and some additional user preferences iโ€™ll check also how to keep the mapping between id of old topics and posts and new ones in order to have create the url redirections once a first migration test looks fine we can share the url of the temporary discourse to have more eyes looking at the result to fix residual problems and iterate if required finally if the migration looks ok and is globally approved youโ€™ll simply have to run the updated migration script on the real instance is this process ok for you regarding timing i can spend my next week end on that regarding login management i didn t know the osm account was mandatory for the fluxbb forum and also for discourse login as tomhughes mentionned on discourse this will require additional access to login details was fluxbb modified to deal with that
0
10,353
13,178,625,254
IssuesEvent
2020-08-12 09:27:45
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
Add tests for System.Diagnostics.Process.Responding
area-System.Diagnostics.Process test enhancement
https://github.com/dotnet/runtime/pull/38385 added a test for `Process.Responding`, which however was disabled since it depends on the current implementation characteristics of Wordpad. We should replace it with a test that consistently reproduces the behaviour.
1.0
Add tests for System.Diagnostics.Process.Responding - https://github.com/dotnet/runtime/pull/38385 added a test for `Process.Responding`, which however was disabled since it depends on the current implementation characteristics of Wordpad. We should replace it with a test that consistently reproduces the behaviour.
process
add tests for system diagnostics process responding added a test for process responding which however was disabled since it depends on the current implementation characteristics of wordpad we should replace it with a test that consistently reproduces the behaviour
1
5,098
7,879,188,901
IssuesEvent
2018-06-26 12:43:30
amarbajric/EBUSA-AIM17
https://api.github.com/repos/amarbajric/EBUSA-AIM17
closed
Business Plan Document - LaTeX
BP business plan business processes in progress
Combine all individual Business Plan and Process files into one large file.
1.0
Business Plan Document - LaTeX - Combine all individual Business Plan and Process files into one large file.
process
business plan document latex combine all individual business plan and process files into one large file
1
680,395
23,268,834,101
IssuesEvent
2022-08-04 20:19:43
metabase/metabase
https://api.github.com/repos/metabase/metabase
opened
Admin > Troubleshooting > Logs page shows spinner if no logs
Type:Bug Priority:P3
This bug kinda requires another bug to surface, but still hints at error handling and robustness of the page could be improved. The `Admin > Troubleshooting > Logs` page continuously fires requests to `api/util/logs`. If there is a problem with our logger and the response is `[]` each time, the spinner continues to show rather than some empty collection. Makes us think if the api returns a 500 for some reason the page would not have a good error message. Low priority since it only surfaces when other issues compound with it. #### UI when logs return empty <img width="1363" alt="image" src="https://user-images.githubusercontent.com/6377293/182944722-99d06d22-4274-446d-ae87-de6941aa09e3.png"> #### UI when api throws an error <img width="1168" alt="image" src="https://user-images.githubusercontent.com/6377293/182944904-f0ba305b-f630-45aa-a264-b42cd817fa9b.png">
1.0
Admin > Troubleshooting > Logs page shows spinner if no logs - This bug kinda requires another bug to surface, but still hints at error handling and robustness of the page could be improved. The `Admin > Troubleshooting > Logs` page continuously fires requests to `api/util/logs`. If there is a problem with our logger and the response is `[]` each time, the spinner continues to show rather than some empty collection. Makes us think if the api returns a 500 for some reason the page would not have a good error message. Low priority since it only surfaces when other issues compound with it. #### UI when logs return empty <img width="1363" alt="image" src="https://user-images.githubusercontent.com/6377293/182944722-99d06d22-4274-446d-ae87-de6941aa09e3.png"> #### UI when api throws an error <img width="1168" alt="image" src="https://user-images.githubusercontent.com/6377293/182944904-f0ba305b-f630-45aa-a264-b42cd817fa9b.png">
non_process
admin troubleshooting logs page shows spinner if no logs this bug kinda requires another bug to surface but still hints at error handling and robustness of the page could be improved the admin troubleshooting logs page continuously fires requests to api util logs if there is a problem with our logger and the response is each time the spinner continues to show rather than some empty collection makes us think if the api returns a for some reason the page would not have a good error message low priority since it only surfaces when other issues compound with it ui when logs return empty img width alt image src ui when api throws an error img width alt image src
0
2,316
5,139,967,750
IssuesEvent
2017-01-12 02:16:18
vuejs/vue-loader
https://api.github.com/repos/vuejs/vue-loader
closed
Pug/Jade included file changes does not fires webpack
pre-processor
Pug/Jade includes changes does not fires webpack, i believe it's a problem on [template-loader.js](https://github.com/vuejs/vue-loader/blob/master/lib/template-loader.js).
1.0
Pug/Jade included file changes does not fires webpack - Pug/Jade includes changes does not fires webpack, i believe it's a problem on [template-loader.js](https://github.com/vuejs/vue-loader/blob/master/lib/template-loader.js).
process
pug jade included file changes does not fires webpack pug jade includes changes does not fires webpack i believe it s a problem on
1
17,286
23,094,748,300
IssuesEvent
2022-07-26 18:22:29
GSA/EDX
https://api.github.com/repos/GSA/EDX
closed
Schedule team rituals in 2 week cadences
process
* [ ] planning - day 1 * [ ] review - last day * [ ] retro - last day * [ ] grooming (recurring and ad-hoc)
1.0
Schedule team rituals in 2 week cadences - * [ ] planning - day 1 * [ ] review - last day * [ ] retro - last day * [ ] grooming (recurring and ad-hoc)
process
schedule team rituals in week cadences planning day review last day retro last day grooming recurring and ad hoc
1
3,737
6,733,146,711
IssuesEvent
2017-10-18 13:59:23
york-region-tpss/stp
https://api.github.com/repos/york-region-tpss/stp
closed
Extra Work Report - Improve Detail Rows Drop Down
enhancement process workflow report ui ux
Add animation for the detail rows when sliding down and slide up.
1.0
Extra Work Report - Improve Detail Rows Drop Down - Add animation for the detail rows when sliding down and slide up.
process
extra work report improve detail rows drop down add animation for the detail rows when sliding down and slide up
1
606,161
18,756,125,741
IssuesEvent
2021-11-05 10:59:17
betagouv/service-national-universel
https://api.github.com/repos/betagouv/service-national-universel
closed
fix: mail - notif de confirmation de creation de compte
enhancement priority-HIGH
### Fonctionnalitรฉ liรฉe ร  un problรจme ? _No response_ ### Fonctionnalitรฉ envoyer le template de mail 219 a la creation du compte ### Commentaires _No response_
1.0
fix: mail - notif de confirmation de creation de compte - ### Fonctionnalitรฉ liรฉe ร  un problรจme ? _No response_ ### Fonctionnalitรฉ envoyer le template de mail 219 a la creation du compte ### Commentaires _No response_
non_process
fix mail notif de confirmation de creation de compte fonctionnalitรฉ liรฉe ร  un problรจme no response fonctionnalitรฉ envoyer le template de mail a la creation du compte commentaires no response
0
253,299
8,053,846,126
IssuesEvent
2018-08-02 01:32:09
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Server still experience crash loop with "this"
High Priority
This is still happening pretty much on a daily or twice daily basis. When it occurs the server stays in a crash loop over and over on this same error until stopped and a backup is employed. I know this was supposed to be resolved in 7.5 but it does not seem to have helped. --BEGIN DUMP-- Dump Time 07/02/2018 02:01:42 Exception Exception: ArgumentException Message:Delegate to an instance method cannot have null 'this'. Source:mscorlib System.ArgumentException: Delegate to an instance method cannot have null 'this'. at System.MulticastDelegate.ThrowNullThisInDelegateToInstance() at Eco.Core.Plugins.StorageManager..ctor() --END DUMP-- https://www.dropbox.com/s/0qo0ido04ts1syz/BigWorld_July1_Dump.zip?dl=0 https://www.dropbox.com/s/s3y8riiw643p6jf/BigWorld_Monday_Dump.zip?dl=0 https://www.dropbox.com/s/4d62tlcpt34gmy2/Game.db?dl=0 https://www.dropbox.com/s/48vhbzdf4mf5zvi/Game.eco?dl=0
1.0
Server still experience crash loop with "this" - This is still happening pretty much on a daily or twice daily basis. When it occurs the server stays in a crash loop over and over on this same error until stopped and a backup is employed. I know this was supposed to be resolved in 7.5 but it does not seem to have helped. --BEGIN DUMP-- Dump Time 07/02/2018 02:01:42 Exception Exception: ArgumentException Message:Delegate to an instance method cannot have null 'this'. Source:mscorlib System.ArgumentException: Delegate to an instance method cannot have null 'this'. at System.MulticastDelegate.ThrowNullThisInDelegateToInstance() at Eco.Core.Plugins.StorageManager..ctor() --END DUMP-- https://www.dropbox.com/s/0qo0ido04ts1syz/BigWorld_July1_Dump.zip?dl=0 https://www.dropbox.com/s/s3y8riiw643p6jf/BigWorld_Monday_Dump.zip?dl=0 https://www.dropbox.com/s/4d62tlcpt34gmy2/Game.db?dl=0 https://www.dropbox.com/s/48vhbzdf4mf5zvi/Game.eco?dl=0
non_process
server still experience crash loop with this this is still happening pretty much on a daily or twice daily basis when it occurs the server stays in a crash loop over and over on this same error until stopped and a backup is employed i know this was supposed to be resolved in but it does not seem to have helped begin dump dump time exception exception argumentexception message delegate to an instance method cannot have null this source mscorlib system argumentexception delegate to an instance method cannot have null this at system multicastdelegate thrownullthisindelegatetoinstance at eco core plugins storagemanager ctor end dump
0
3,408
6,522,201,324
IssuesEvent
2017-08-29 00:47:05
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
Various fixes for grabABI
apps-grabABI status-inprocess type-bug
- [x] Either hide verbose or give it a purpose. - [x] The word 'indexed' should appear in the event signatures where there are indexed params ~~- [ ] grabABI used to report on the constructor, now it ignores the constructor.~~ ~~- [ ] default function is not handled properly at all~~
1.0
Various fixes for grabABI - - [x] Either hide verbose or give it a purpose. - [x] The word 'indexed' should appear in the event signatures where there are indexed params ~~- [ ] grabABI used to report on the constructor, now it ignores the constructor.~~ ~~- [ ] default function is not handled properly at all~~
process
various fixes for grababi either hide verbose or give it a purpose the word indexed should appear in the event signatures where there are indexed params grababi used to report on the constructor now it ignores the constructor default function is not handled properly at all
1
15,472
3,466,468,287
IssuesEvent
2015-12-22 03:51:03
infiniteautomation/ma-core-public
https://api.github.com/repos/infiniteautomation/ma-core-public
closed
Watchlist - SQL Error on watchlist delete
Bug Ready for Testing
Steve alerted us to [modules public #13](https://github.com/infiniteautomation/ma-modules-public/issues/13) but I'm moving it here since we're churning through these and I have replication instructions. The error is: ------------ PreparedStatementCallback; SQL [INSERT INTO selectedWatchList (userId, watchListId) VALUES (?,?)]; Unique index or primary key violation: "PRIMARY KEY ON PUBLIC.SELECTEDWATCHLIST(USERID)"; SQL statement: INSERT INTO selectedWatchList (userId, watchListId) VALUES (?,?) [23505-181]; nested exception is org.h2.jdbc.JdbcSQLException: Unique index or primary key violation: "PRIMARY KEY ON PUBLIC.SELECTEDWATCHLIST(USERID)"; SQL statement: INSERT INTO selectedWatchList (userId, watchListId) VALUES (?,?) [23505-181] ------------ To replicate, make a few random watchlists with admin1. Now, switch users to admin2 and delete some watchlists' points and then the watchlist.
1.0
Watchlist - SQL Error on watchlist delete - Steve alerted us to [modules public #13](https://github.com/infiniteautomation/ma-modules-public/issues/13) but I'm moving it here since we're churning through these and I have replication instructions. The error is: ------------ PreparedStatementCallback; SQL [INSERT INTO selectedWatchList (userId, watchListId) VALUES (?,?)]; Unique index or primary key violation: "PRIMARY KEY ON PUBLIC.SELECTEDWATCHLIST(USERID)"; SQL statement: INSERT INTO selectedWatchList (userId, watchListId) VALUES (?,?) [23505-181]; nested exception is org.h2.jdbc.JdbcSQLException: Unique index or primary key violation: "PRIMARY KEY ON PUBLIC.SELECTEDWATCHLIST(USERID)"; SQL statement: INSERT INTO selectedWatchList (userId, watchListId) VALUES (?,?) [23505-181] ------------ To replicate, make a few random watchlists with admin1. Now, switch users to admin2 and delete some watchlists' points and then the watchlist.
non_process
watchlist sql error on watchlist delete steve alerted us to but i m moving it here since we re churning through these and i have replication instructions the error is preparedstatementcallback sql unique index or primary key violation primary key on public selectedwatchlist userid sql statement insert into selectedwatchlist userid watchlistid values nested exception is org jdbc jdbcsqlexception unique index or primary key violation primary key on public selectedwatchlist userid sql statement insert into selectedwatchlist userid watchlistid values to replicate make a few random watchlists with now switch users to and delete some watchlists points and then the watchlist
0
157,076
19,914,000,952
IssuesEvent
2022-01-25 20:18:16
dmyers87/BoomTown.FuzzySharp
https://api.github.com/repos/dmyers87/BoomTown.FuzzySharp
opened
CVE-2019-0820 (High) detected in system.text.regularexpressions.4.3.0.nupkg
security vulnerability
## CVE-2019-0820 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.text.regularexpressions.4.3.0.nupkg</b></p></summary> <p>Provides the System.Text.RegularExpressions.Regex class, an implementation of a regular expression e...</p> <p>Library home page: <a href="https://api.nuget.org/packages/system.text.regularexpressions.4.3.0.nupkg">https://api.nuget.org/packages/system.text.regularexpressions.4.3.0.nupkg</a></p> <p>Path to dependency file: /BoomTown.FuzzySharp.csproj</p> <p>Path to vulnerable library: /usr/share/dotnet/sdk/NuGetFallbackFolder/system.text.regularexpressions/4.3.0/system.text.regularexpressions.4.3.0.nupkg</p> <p> Dependency Hierarchy: - netstandard.library.1.6.1.nupkg (Root Library) - :x: **system.text.regularexpressions.4.3.0.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dmyers87/BoomTown.FuzzySharp/commit/1aeb2dc78e06dab00f85d3488bae294664a92a22">1aeb2dc78e06dab00f85d3488bae294664a92a22</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A denial of service vulnerability exists when .NET Framework and .NET Core improperly process RegEx strings, aka '.NET Framework and .NET Core Denial of Service Vulnerability'. This CVE ID is unique from CVE-2019-0980, CVE-2019-0981. <p>Publish Date: 2019-05-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820>CVE-2019-0820</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cmhx-cq75-c4mj">https://github.com/advisories/GHSA-cmhx-cq75-c4mj</a></p> <p>Release Date: 2019-05-16</p> <p>Fix Resolution: System.Text.RegularExpressions - 4.3.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Nuget","packageName":"System.Text.RegularExpressions","packageVersion":"4.3.0","packageFilePaths":["/BoomTown.FuzzySharp.csproj"],"isTransitiveDependency":true,"dependencyTree":"NETStandard.Library:1.6.1;System.Text.RegularExpressions:4.3.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"System.Text.RegularExpressions - 4.3.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-0820","vulnerabilityDetails":"A denial of service vulnerability exists when .NET Framework and .NET Core improperly process RegEx strings, aka \u0027.NET Framework and .NET Core Denial of Service Vulnerability\u0027. This CVE ID is unique from CVE-2019-0980, CVE-2019-0981.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-0820 (High) detected in system.text.regularexpressions.4.3.0.nupkg - ## CVE-2019-0820 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.text.regularexpressions.4.3.0.nupkg</b></p></summary> <p>Provides the System.Text.RegularExpressions.Regex class, an implementation of a regular expression e...</p> <p>Library home page: <a href="https://api.nuget.org/packages/system.text.regularexpressions.4.3.0.nupkg">https://api.nuget.org/packages/system.text.regularexpressions.4.3.0.nupkg</a></p> <p>Path to dependency file: /BoomTown.FuzzySharp.csproj</p> <p>Path to vulnerable library: /usr/share/dotnet/sdk/NuGetFallbackFolder/system.text.regularexpressions/4.3.0/system.text.regularexpressions.4.3.0.nupkg</p> <p> Dependency Hierarchy: - netstandard.library.1.6.1.nupkg (Root Library) - :x: **system.text.regularexpressions.4.3.0.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dmyers87/BoomTown.FuzzySharp/commit/1aeb2dc78e06dab00f85d3488bae294664a92a22">1aeb2dc78e06dab00f85d3488bae294664a92a22</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A denial of service vulnerability exists when .NET Framework and .NET Core improperly process RegEx strings, aka '.NET Framework and .NET Core Denial of Service Vulnerability'. This CVE ID is unique from CVE-2019-0980, CVE-2019-0981. <p>Publish Date: 2019-05-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820>CVE-2019-0820</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cmhx-cq75-c4mj">https://github.com/advisories/GHSA-cmhx-cq75-c4mj</a></p> <p>Release Date: 2019-05-16</p> <p>Fix Resolution: System.Text.RegularExpressions - 4.3.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Nuget","packageName":"System.Text.RegularExpressions","packageVersion":"4.3.0","packageFilePaths":["/BoomTown.FuzzySharp.csproj"],"isTransitiveDependency":true,"dependencyTree":"NETStandard.Library:1.6.1;System.Text.RegularExpressions:4.3.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"System.Text.RegularExpressions - 4.3.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-0820","vulnerabilityDetails":"A denial of service vulnerability exists when .NET Framework and .NET Core improperly process RegEx strings, aka \u0027.NET Framework and .NET Core Denial of Service Vulnerability\u0027. This CVE ID is unique from CVE-2019-0980, CVE-2019-0981.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in system text regularexpressions nupkg cve high severity vulnerability vulnerable library system text regularexpressions nupkg provides the system text regularexpressions regex class an implementation of a regular expression e library home page a href path to dependency file boomtown fuzzysharp csproj path to vulnerable library usr share dotnet sdk nugetfallbackfolder system text regularexpressions system text regularexpressions nupkg dependency hierarchy netstandard library nupkg root library x system text regularexpressions nupkg vulnerable library found in head commit a href found in base branch master vulnerability details a denial of service vulnerability exists when net framework and net core improperly process regex strings aka net framework and net core denial of service vulnerability this cve id is unique from cve cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution system text regularexpressions isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree netstandard library system text regularexpressions isminimumfixversionavailable true minimumfixversion system text regularexpressions isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a denial of service vulnerability exists when net framework and net core improperly process regex strings aka net framework and net core denial of service vulnerability this cve id is unique from cve cve vulnerabilityurl
0
10,465
13,242,097,204
IssuesEvent
2020-08-19 09:13:55
googleapis/python-spanner-django
https://api.github.com/repos/googleapis/python-spanner-django
opened
Use nox instead of tox for testing
api: spanner type: process
I've noticed that Django Spanner uses `tox` as a test base. Other Google Python packages are using `nox` with generated noxfiles (with lint, docs, coverage, system and unit testing sessions) - see [python-spanner noxfile.py](https://github.com/googleapis/python-spanner/blob/master/noxfile.py) as an example. If we're going to move this API to the original Python Spanner client, we probably should use `nox` here to stay in touch with the original test configurations.
1.0
Use nox instead of tox for testing - I've noticed that Django Spanner uses `tox` as a test base. Other Google Python packages are using `nox` with generated noxfiles (with lint, docs, coverage, system and unit testing sessions) - see [python-spanner noxfile.py](https://github.com/googleapis/python-spanner/blob/master/noxfile.py) as an example. If we're going to move this API to the original Python Spanner client, we probably should use `nox` here to stay in touch with the original test configurations.
process
use nox instead of tox for testing i ve noticed that django spanner uses tox as a test base other google python packages are using nox with generated noxfiles with lint docs coverage system and unit testing sessions see as an example if we re going to move this api to the original python spanner client we probably should use nox here to stay in touch with the original test configurations
1
166,724
20,721,161,875
IssuesEvent
2022-03-13 12:15:19
tabacws-remediation-demos/Java-Demo-Rem
https://api.github.com/repos/tabacws-remediation-demos/Java-Demo-Rem
reopened
CVE-2018-1313 (Medium) detected in derby-10.8.3.0.jar
security vulnerability
## CVE-2018-1313 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>derby-10.8.3.0.jar</b></p></summary> <p>Contains the core Apache Derby database engine, which also includes the embedded JDBC driver.</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /ository/org/apache/derby/derby/10.8.3.0/derby-10.8.3.0.jar</p> <p> Dependency Hierarchy: - :x: **derby-10.8.3.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tabacws-remediation-demos/Java-Demo-Rem/commit/018eccd7980798c5f39292e9c7133e829e9ed7a3">018eccd7980798c5f39292e9c7133e829e9ed7a3</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Derby 10.3.1.4 to 10.14.1.0, a specially-crafted network packet can be used to request the Derby Network Server to boot a database whose location and contents are under the user's control. If the Derby Network Server is not running with a Java Security Manager policy file, the attack is successful. If the server is using a policy file, the policy file must permit the database location to be read for the attack to work. The default Derby Network Server policy file distributed with the affected releases includes a permissive policy as the default Network Server policy, which allows the attack to work. <p>Publish Date: 2018-05-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1313>CVE-2018-1313</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1313">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1313</a></p> <p>Release Date: 2018-05-07</p> <p>Fix Resolution: org.apache.derby:derbynet:10.14.2.0;org.apache.derby:derby:10.14.2.0;org.apache.derby:derbytools:10.14.2.0;org.apache.derby:derbyclient:10.14.2.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.derby","packageName":"derby","packageVersion":"10.8.3.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.derby:derby:10.8.3.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.derby:derbynet:10.14.2.0;org.apache.derby:derby:10.14.2.0;org.apache.derby:derbytools:10.14.2.0;org.apache.derby:derbyclient:10.14.2.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-1313","vulnerabilityDetails":"In Apache Derby 10.3.1.4 to 10.14.1.0, a specially-crafted network packet can be used to request the Derby Network Server to boot a database whose location and contents are under the user\u0027s control. If the Derby Network Server is not running with a Java Security Manager policy file, the attack is successful. If the server is using a policy file, the policy file must permit the database location to be read for the attack to work. The default Derby Network Server policy file distributed with the affected releases includes a permissive policy as the default Network Server policy, which allows the attack to work.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1313","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-1313 (Medium) detected in derby-10.8.3.0.jar - ## CVE-2018-1313 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>derby-10.8.3.0.jar</b></p></summary> <p>Contains the core Apache Derby database engine, which also includes the embedded JDBC driver.</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /ository/org/apache/derby/derby/10.8.3.0/derby-10.8.3.0.jar</p> <p> Dependency Hierarchy: - :x: **derby-10.8.3.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tabacws-remediation-demos/Java-Demo-Rem/commit/018eccd7980798c5f39292e9c7133e829e9ed7a3">018eccd7980798c5f39292e9c7133e829e9ed7a3</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Derby 10.3.1.4 to 10.14.1.0, a specially-crafted network packet can be used to request the Derby Network Server to boot a database whose location and contents are under the user's control. If the Derby Network Server is not running with a Java Security Manager policy file, the attack is successful. If the server is using a policy file, the policy file must permit the database location to be read for the attack to work. The default Derby Network Server policy file distributed with the affected releases includes a permissive policy as the default Network Server policy, which allows the attack to work. <p>Publish Date: 2018-05-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1313>CVE-2018-1313</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1313">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1313</a></p> <p>Release Date: 2018-05-07</p> <p>Fix Resolution: org.apache.derby:derbynet:10.14.2.0;org.apache.derby:derby:10.14.2.0;org.apache.derby:derbytools:10.14.2.0;org.apache.derby:derbyclient:10.14.2.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.derby","packageName":"derby","packageVersion":"10.8.3.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.derby:derby:10.8.3.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.derby:derbynet:10.14.2.0;org.apache.derby:derby:10.14.2.0;org.apache.derby:derbytools:10.14.2.0;org.apache.derby:derbyclient:10.14.2.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-1313","vulnerabilityDetails":"In Apache Derby 10.3.1.4 to 10.14.1.0, a specially-crafted network packet can be used to request the Derby Network Server to boot a database whose location and contents are under the user\u0027s control. If the Derby Network Server is not running with a Java Security Manager policy file, the attack is successful. If the server is using a policy file, the policy file must permit the database location to be read for the attack to work. The default Derby Network Server policy file distributed with the affected releases includes a permissive policy as the default Network Server policy, which allows the attack to work.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1313","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in derby jar cve medium severity vulnerability vulnerable library derby jar contains the core apache derby database engine which also includes the embedded jdbc driver path to dependency file pom xml path to vulnerable library ository org apache derby derby derby jar dependency hierarchy x derby jar vulnerable library found in head commit a href found in base branch main vulnerability details in apache derby to a specially crafted network packet can be used to request the derby network server to boot a database whose location and contents are under the user s control if the derby network server is not running with a java security manager policy file the attack is successful if the server is using a policy file the policy file must permit the database location to be read for the attack to work the default derby network server policy file distributed with the affected releases includes a permissive policy as the default network server policy which allows the attack to work publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache derby derbynet org apache derby derby org apache derby derbytools org apache derby derbyclient isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache derby derby isminimumfixversionavailable true minimumfixversion org apache derby derbynet org apache derby derby org apache derby derbytools org apache derby derbyclient isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in apache derby to a specially crafted network packet can be used to request the derby network server to boot a database whose location and contents are under the user control if the derby network server is not running with a java security manager policy file the attack is successful if the server is using a policy file the policy file must permit the database location to be read for the attack to work the default derby network server policy file distributed with the affected releases includes a permissive policy as the default network server policy which allows the attack to work vulnerabilityurl
0
20,998
27,865,180,891
IssuesEvent
2023-03-21 09:46:36
UnitTestBot/UTBotJava
https://api.github.com/repos/UnitTestBot/UTBotJava
closed
`IllegalArgumentException`s in Instrumented process for `MockReturnObjectExample`
ctg-bug comp-instrumented-process
**Description** There are IllegalArgumentException(s) that arguments provided for of `calculate` and `calculateFromArray` methods are incorrect. **To Reproduce** 1. Run the 'utbot' project in IntelliJ Idea 2022.2.4 Ultimate 2. [Install plugin built from unit-test-bot/rc3102023 branch](https://github.com/UnitTestBot/UTBotJava/actions/runs/4448325159) 3. Generate tests for `utbot-sample/src/test/java/org/utbot/mock/MockReturnObjectExample` with default settings: Symbolic + Fuzzing **Expected behavior** Instrumented process should be executed correctly. Test cases generated by Fuzzing are expected. **Actual behavior** There are Errors reports for `calculate` and `calculateFromArray`. There are no test methods generated by Fuzzing. **Visual proofs (screenshots, logs, images)** ~~~java 16:55:11.653 | ERROR | ConcreteExecutor | executeAsync, response(ERROR) org.utbot.instrumentation.util.InstrumentedProcessError: Error in the instrumented process |> com.jetbrains.rd.util.reactive.RdFault: InvocationPhase, reason: org.utbot.instrumentation.instrumentation.execution.phases.ExecutionPhaseError: InvocationPhase at org.utbot.instrumentation.instrumentation.execution.phases.InvocationPhase.wrapError(InvocationPhase.kt:22) at org.utbot.instrumentation.instrumentation.execution.phases.ExecutionPhaseKt.start(ExecutionPhase.kt:30) at org.utbot.instrumentation.instrumentation.execution.phases.PhasesController.executePhaseInTimeout(PhasesController.kt:55) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:108) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:48) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:152) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:149) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1$2$1.invoke(ClientProcessUtil.kt:113) at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:86) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1.invoke(ClientProcessUtil.kt:112) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.lang.IllegalArgumentException: signature=calculateFromArray()I expecting this, but provided argument list is empty at org.utbot.instrumentation.instrumentation.InvokeInstrumentation.invoke-BWLJW6A(InvokeInstrumentation.kt:49) at org.utbot.instrumentation.instrumentation.InvokeInstrumentation.invoke(InvokeInstrumentation.kt:21) at org.utbot.instrumentation.instrumentation.Instrumentation$DefaultImpls.invoke$default(Instrumentation.kt:21) at org.utbot.instrumentation.instrumentation.execution.phases.InvocationPhase.invoke-0E7RQCE(InvocationPhase.kt:31) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation$invoke$1$concreteResult$1.invoke-IoAF18A(UtExecutionInstrumentation.kt:109) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation$invoke$1$concreteResult$1.invoke(UtExecutionInstrumentation.kt:108) at org.utbot.instrumentation.instrumentation.execution.phases.PhasesController$executePhaseInTimeout$1$result$1.invoke(PhasesController.kt:61) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$2.invoke(ThreadUtil.kt:56) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$1.invoke(ThreadUtil.kt:47) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$1.invoke(ThreadUtil.kt:43) at kotlin.concurrent.ThreadsKt$thread$thread$1.run(Thread.kt:30) at com.jetbrains.rd.framework.RdTaskResult$Companion.read(TaskInterfaces.kt:30) at com.jetbrains.rd.framework.impl.CallSiteWiredRdTask.onWireReceived(RdTask.kt:104) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$1.invoke(RdTask.kt:278) at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$2.invokeSuspend(RdTask.kt:287) at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) at kotlinx.coroutines.EventLoopImplBase.processNextEvent(EventLoop.common.kt:284) at kotlinx.coroutines.BlockingCoroutine.joinBlocking(Builders.kt:85) at kotlinx.coroutines.BuildersKt__BuildersKt.runBlocking(Builders.kt:59) at kotlinx.coroutines.BuildersKt.runBlocking(Unknown Source) at kotlinx.coroutines.BuildersKt__BuildersKt.runBlocking$default(Builders.kt:38) at kotlinx.coroutines.BuildersKt.runBlocking$default(Unknown Source) at org.utbot.common.ConcurrencyKt.runBlockingWithCancellationPredicate(Concurrency.kt:38) at org.utbot.framework.plugin.api.TestCaseGenerator$generate$3.invoke(TestCaseGenerator.kt:156) at org.utbot.framework.plugin.api.TestCaseGenerator$generate$3.invoke(TestCaseGenerator.kt:155) at org.utbot.common.ConcurrencyKt.runIgnoringCancellationException(Concurrency.kt:47) at org.utbot.framework.plugin.api.TestCaseGenerator.generate(TestCaseGenerator.kt:155) at org.utbot.framework.process.EngineProcessMainKt$setup$3.invoke(EngineProcessMain.kt:111) at org.utbot.framework.process.EngineProcessMainKt$setup$3.invoke(EngineProcessMain.kt:97) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1$2$1.invoke(ClientProcessUtil.kt:113) at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:86) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1.invoke(ClientProcessUtil.kt:112) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) at org.utbot.instrumentation.ConcreteExecutor.withProcess(ConcreteExecutor.kt:218) ~[utbot-instrumentation-2023.3.jar:?] at org.utbot.instrumentation.ConcreteExecutor$withProcess$1.invokeSuspend(ConcreteExecutor.kt) ~[utbot-instrumentation-2023.3.jar:?] at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) [kotlin-stdlib-1.8.0.jar:1.8.0-release-345(1.8.0)] at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:104) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.EventLoopImplBase.processNextEvent(EventLoop.common.kt:284) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.BlockingCoroutine.joinBlocking(Builders.kt:85) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.BuildersKt__BuildersKt.runBlocking(Builders.kt:59) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.BuildersKt.runBlocking(Unknown Source) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.BuildersKt__BuildersKt.runBlocking$default(Builders.kt:38) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.BuildersKt.runBlocking$default(Unknown Source) [utbot-instrumentation-2023.3.jar:?] at org.utbot.common.ConcurrencyKt.runBlockingWithCancellationPredicate(Concurrency.kt:38) [utbot-core-2023.3.jar:?] at org.utbot.framework.plugin.api.TestCaseGenerator$generate$3.invoke(TestCaseGenerator.kt:156) [utbot-framework-2023.3.jar:?] at org.utbot.framework.plugin.api.TestCaseGenerator$generate$3.invoke(TestCaseGenerator.kt:155) [utbot-framework-2023.3.jar:?] at org.utbot.common.ConcurrencyKt.runIgnoringCancellationException(Concurrency.kt:47) [utbot-core-2023.3.jar:?] at org.utbot.framework.plugin.api.TestCaseGenerator.generate(TestCaseGenerator.kt:155) [utbot-framework-2023.3.jar:?] at org.utbot.framework.process.EngineProcessMainKt$setup$3.invoke(EngineProcessMain.kt:111) [utbot-framework-2023.3.jar:?] at org.utbot.framework.process.EngineProcessMainKt$setup$3.invoke(EngineProcessMain.kt:97) [utbot-framework-2023.3.jar:?] at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1$2$1.invoke(ClientProcessUtil.kt:113) [utbot-rd-2023.3.jar:?] at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:86) [utbot-rd-2023.3.jar:?] at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1.invoke(ClientProcessUtil.kt:112) [utbot-rd-2023.3.jar:?] at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) [rd-core-2022.2.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:829) [?:?] Caused by: com.jetbrains.rd.util.reactive.RdFault: InvocationPhase, reason: org.utbot.instrumentation.instrumentation.execution.phases.ExecutionPhaseError: InvocationPhase at org.utbot.instrumentation.instrumentation.execution.phases.InvocationPhase.wrapError(InvocationPhase.kt:22) at org.utbot.instrumentation.instrumentation.execution.phases.ExecutionPhaseKt.start(ExecutionPhase.kt:30) at org.utbot.instrumentation.instrumentation.execution.phases.PhasesController.executePhaseInTimeout(PhasesController.kt:55) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:108) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:48) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:152) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:149) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1$2$1.invoke(ClientProcessUtil.kt:113) at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:86) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1.invoke(ClientProcessUtil.kt:112) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.lang.IllegalArgumentException: signature=calculateFromArray()I expecting this, but provided argument list is empty at org.utbot.instrumentation.instrumentation.InvokeInstrumentation.invoke-BWLJW6A(InvokeInstrumentation.kt:49) at org.utbot.instrumentation.instrumentation.InvokeInstrumentation.invoke(InvokeInstrumentation.kt:21) at org.utbot.instrumentation.instrumentation.Instrumentation$DefaultImpls.invoke$default(Instrumentation.kt:21) at org.utbot.instrumentation.instrumentation.execution.phases.InvocationPhase.invoke-0E7RQCE(InvocationPhase.kt:31) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation$invoke$1$concreteResult$1.invoke-IoAF18A(UtExecutionInstrumentation.kt:109) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation$invoke$1$concreteResult$1.invoke(UtExecutionInstrumentation.kt:108) at org.utbot.instrumentation.instrumentation.execution.phases.PhasesController$executePhaseInTimeout$1$result$1.invoke(PhasesController.kt:61) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$2.invoke(ThreadUtil.kt:56) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$1.invoke(ThreadUtil.kt:47) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$1.invoke(ThreadUtil.kt:43) at kotlin.concurrent.ThreadsKt$thread$thread$1.run(Thread.kt:30) at com.jetbrains.rd.framework.RdTaskResult$Companion.read(TaskInterfaces.kt:30) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.CallSiteWiredRdTask.onWireReceived(RdTask.kt:104) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$1.invoke(RdTask.kt:278) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$2.invokeSuspend(RdTask.kt:287) ~[rd-framework-2022.2.1.jar:?] at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) ~[kotlin-stdlib-1.8.0.jar:1.8.0-release-345(1.8.0)] at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) ~[utbot-instrumentation-2023.3.jar:?] ... 28 more ~~~ **Environment** Windows 10 Pro IntelliJ IDEA 2022.2.24 Ultimate **Additional context** When running this code in a separate project - there were 1000+ exceptions of this kind. [utbot-engine-current - MockReturnObjectExample.log](https://github.com/UnitTestBot/UTBotJava/files/11019142/utbot-engine-current.-.MockReturnObjectExample.log)
1.0
`IllegalArgumentException`s in Instrumented process for `MockReturnObjectExample` - **Description** There are IllegalArgumentException(s) that arguments provided for of `calculate` and `calculateFromArray` methods are incorrect. **To Reproduce** 1. Run the 'utbot' project in IntelliJ Idea 2022.2.4 Ultimate 2. [Install plugin built from unit-test-bot/rc3102023 branch](https://github.com/UnitTestBot/UTBotJava/actions/runs/4448325159) 3. Generate tests for `utbot-sample/src/test/java/org/utbot/mock/MockReturnObjectExample` with default settings: Symbolic + Fuzzing **Expected behavior** Instrumented process should be executed correctly. Test cases generated by Fuzzing are expected. **Actual behavior** There are Errors reports for `calculate` and `calculateFromArray`. There are no test methods generated by Fuzzing. **Visual proofs (screenshots, logs, images)** ~~~java 16:55:11.653 | ERROR | ConcreteExecutor | executeAsync, response(ERROR) org.utbot.instrumentation.util.InstrumentedProcessError: Error in the instrumented process |> com.jetbrains.rd.util.reactive.RdFault: InvocationPhase, reason: org.utbot.instrumentation.instrumentation.execution.phases.ExecutionPhaseError: InvocationPhase at org.utbot.instrumentation.instrumentation.execution.phases.InvocationPhase.wrapError(InvocationPhase.kt:22) at org.utbot.instrumentation.instrumentation.execution.phases.ExecutionPhaseKt.start(ExecutionPhase.kt:30) at org.utbot.instrumentation.instrumentation.execution.phases.PhasesController.executePhaseInTimeout(PhasesController.kt:55) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:108) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:48) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:152) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:149) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1$2$1.invoke(ClientProcessUtil.kt:113) at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:86) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1.invoke(ClientProcessUtil.kt:112) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.lang.IllegalArgumentException: signature=calculateFromArray()I expecting this, but provided argument list is empty at org.utbot.instrumentation.instrumentation.InvokeInstrumentation.invoke-BWLJW6A(InvokeInstrumentation.kt:49) at org.utbot.instrumentation.instrumentation.InvokeInstrumentation.invoke(InvokeInstrumentation.kt:21) at org.utbot.instrumentation.instrumentation.Instrumentation$DefaultImpls.invoke$default(Instrumentation.kt:21) at org.utbot.instrumentation.instrumentation.execution.phases.InvocationPhase.invoke-0E7RQCE(InvocationPhase.kt:31) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation$invoke$1$concreteResult$1.invoke-IoAF18A(UtExecutionInstrumentation.kt:109) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation$invoke$1$concreteResult$1.invoke(UtExecutionInstrumentation.kt:108) at org.utbot.instrumentation.instrumentation.execution.phases.PhasesController$executePhaseInTimeout$1$result$1.invoke(PhasesController.kt:61) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$2.invoke(ThreadUtil.kt:56) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$1.invoke(ThreadUtil.kt:47) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$1.invoke(ThreadUtil.kt:43) at kotlin.concurrent.ThreadsKt$thread$thread$1.run(Thread.kt:30) at com.jetbrains.rd.framework.RdTaskResult$Companion.read(TaskInterfaces.kt:30) at com.jetbrains.rd.framework.impl.CallSiteWiredRdTask.onWireReceived(RdTask.kt:104) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$1.invoke(RdTask.kt:278) at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$2.invokeSuspend(RdTask.kt:287) at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) at kotlinx.coroutines.EventLoopImplBase.processNextEvent(EventLoop.common.kt:284) at kotlinx.coroutines.BlockingCoroutine.joinBlocking(Builders.kt:85) at kotlinx.coroutines.BuildersKt__BuildersKt.runBlocking(Builders.kt:59) at kotlinx.coroutines.BuildersKt.runBlocking(Unknown Source) at kotlinx.coroutines.BuildersKt__BuildersKt.runBlocking$default(Builders.kt:38) at kotlinx.coroutines.BuildersKt.runBlocking$default(Unknown Source) at org.utbot.common.ConcurrencyKt.runBlockingWithCancellationPredicate(Concurrency.kt:38) at org.utbot.framework.plugin.api.TestCaseGenerator$generate$3.invoke(TestCaseGenerator.kt:156) at org.utbot.framework.plugin.api.TestCaseGenerator$generate$3.invoke(TestCaseGenerator.kt:155) at org.utbot.common.ConcurrencyKt.runIgnoringCancellationException(Concurrency.kt:47) at org.utbot.framework.plugin.api.TestCaseGenerator.generate(TestCaseGenerator.kt:155) at org.utbot.framework.process.EngineProcessMainKt$setup$3.invoke(EngineProcessMain.kt:111) at org.utbot.framework.process.EngineProcessMainKt$setup$3.invoke(EngineProcessMain.kt:97) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1$2$1.invoke(ClientProcessUtil.kt:113) at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:86) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1.invoke(ClientProcessUtil.kt:112) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) at org.utbot.instrumentation.ConcreteExecutor.withProcess(ConcreteExecutor.kt:218) ~[utbot-instrumentation-2023.3.jar:?] at org.utbot.instrumentation.ConcreteExecutor$withProcess$1.invokeSuspend(ConcreteExecutor.kt) ~[utbot-instrumentation-2023.3.jar:?] at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) [kotlin-stdlib-1.8.0.jar:1.8.0-release-345(1.8.0)] at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:104) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.EventLoopImplBase.processNextEvent(EventLoop.common.kt:284) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.BlockingCoroutine.joinBlocking(Builders.kt:85) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.BuildersKt__BuildersKt.runBlocking(Builders.kt:59) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.BuildersKt.runBlocking(Unknown Source) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.BuildersKt__BuildersKt.runBlocking$default(Builders.kt:38) [utbot-instrumentation-2023.3.jar:?] at kotlinx.coroutines.BuildersKt.runBlocking$default(Unknown Source) [utbot-instrumentation-2023.3.jar:?] at org.utbot.common.ConcurrencyKt.runBlockingWithCancellationPredicate(Concurrency.kt:38) [utbot-core-2023.3.jar:?] at org.utbot.framework.plugin.api.TestCaseGenerator$generate$3.invoke(TestCaseGenerator.kt:156) [utbot-framework-2023.3.jar:?] at org.utbot.framework.plugin.api.TestCaseGenerator$generate$3.invoke(TestCaseGenerator.kt:155) [utbot-framework-2023.3.jar:?] at org.utbot.common.ConcurrencyKt.runIgnoringCancellationException(Concurrency.kt:47) [utbot-core-2023.3.jar:?] at org.utbot.framework.plugin.api.TestCaseGenerator.generate(TestCaseGenerator.kt:155) [utbot-framework-2023.3.jar:?] at org.utbot.framework.process.EngineProcessMainKt$setup$3.invoke(EngineProcessMain.kt:111) [utbot-framework-2023.3.jar:?] at org.utbot.framework.process.EngineProcessMainKt$setup$3.invoke(EngineProcessMain.kt:97) [utbot-framework-2023.3.jar:?] at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1$2$1.invoke(ClientProcessUtil.kt:113) [utbot-rd-2023.3.jar:?] at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:86) [utbot-rd-2023.3.jar:?] at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1.invoke(ClientProcessUtil.kt:112) [utbot-rd-2023.3.jar:?] at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) [rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) [rd-core-2022.2.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] at java.lang.Thread.run(Thread.java:829) [?:?] Caused by: com.jetbrains.rd.util.reactive.RdFault: InvocationPhase, reason: org.utbot.instrumentation.instrumentation.execution.phases.ExecutionPhaseError: InvocationPhase at org.utbot.instrumentation.instrumentation.execution.phases.InvocationPhase.wrapError(InvocationPhase.kt:22) at org.utbot.instrumentation.instrumentation.execution.phases.ExecutionPhaseKt.start(ExecutionPhase.kt:30) at org.utbot.instrumentation.instrumentation.execution.phases.PhasesController.executePhaseInTimeout(PhasesController.kt:55) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:108) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation.invoke(UtExecutionInstrumentation.kt:48) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:152) at org.utbot.instrumentation.process.InstrumentedProcessMainKt$setup$2.invoke(InstrumentedProcessMain.kt:149) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1$2$1.invoke(ClientProcessUtil.kt:113) at org.utbot.rd.IdleWatchdog.wrapActive(ClientProcessUtil.kt:86) at org.utbot.rd.IdleWatchdog$measureTimeForActiveCall$1.invoke(ClientProcessUtil.kt:112) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:182) at com.jetbrains.rd.framework.IRdEndpoint$set$1.invoke(TaskInterfaces.kt:173) at com.jetbrains.rd.framework.impl.RdCall.onWireReceived(RdTask.kt:360) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) at com.jetbrains.rd.util.threading.SingleThreadSchedulerBase$queue$1.run(SingleThreadScheduler.kt:41) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.lang.IllegalArgumentException: signature=calculateFromArray()I expecting this, but provided argument list is empty at org.utbot.instrumentation.instrumentation.InvokeInstrumentation.invoke-BWLJW6A(InvokeInstrumentation.kt:49) at org.utbot.instrumentation.instrumentation.InvokeInstrumentation.invoke(InvokeInstrumentation.kt:21) at org.utbot.instrumentation.instrumentation.Instrumentation$DefaultImpls.invoke$default(Instrumentation.kt:21) at org.utbot.instrumentation.instrumentation.execution.phases.InvocationPhase.invoke-0E7RQCE(InvocationPhase.kt:31) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation$invoke$1$concreteResult$1.invoke-IoAF18A(UtExecutionInstrumentation.kt:109) at org.utbot.instrumentation.instrumentation.execution.UtExecutionInstrumentation$invoke$1$concreteResult$1.invoke(UtExecutionInstrumentation.kt:108) at org.utbot.instrumentation.instrumentation.execution.phases.PhasesController$executePhaseInTimeout$1$result$1.invoke(PhasesController.kt:61) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$2.invoke(ThreadUtil.kt:56) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$1.invoke(ThreadUtil.kt:47) at org.utbot.common.ThreadBasedExecutor$invokeWithTimeout$1.invoke(ThreadUtil.kt:43) at kotlin.concurrent.ThreadsKt$thread$thread$1.run(Thread.kt:30) at com.jetbrains.rd.framework.RdTaskResult$Companion.read(TaskInterfaces.kt:30) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.CallSiteWiredRdTask.onWireReceived(RdTask.kt:104) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:57) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2$2.invoke(MessageBroker.kt:12) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.ProtocolContexts.readMessageContextAndInvoke(ProtocolContexts.kt:151) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:56) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.MessageBroker$invoke$2.invoke(MessageBroker.kt:12) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$1.invoke(RdTask.kt:278) ~[rd-framework-2022.2.1.jar:?] at com.jetbrains.rd.framework.impl.RdCall$createResponseScheduler$1$queue$2.invokeSuspend(RdTask.kt:287) ~[rd-framework-2022.2.1.jar:?] at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) ~[kotlin-stdlib-1.8.0.jar:1.8.0-release-345(1.8.0)] at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) ~[utbot-instrumentation-2023.3.jar:?] ... 28 more ~~~ **Environment** Windows 10 Pro IntelliJ IDEA 2022.2.24 Ultimate **Additional context** When running this code in a separate project - there were 1000+ exceptions of this kind. [utbot-engine-current - MockReturnObjectExample.log](https://github.com/UnitTestBot/UTBotJava/files/11019142/utbot-engine-current.-.MockReturnObjectExample.log)
process
illegalargumentexception s in instrumented process for mockreturnobjectexample description there are illegalargumentexception s that arguments provided for of calculate and calculatefromarray methods are incorrect to reproduce run the utbot project in intellij idea ultimate generate tests for utbot sample src test java org utbot mock mockreturnobjectexample with default settings symbolic fuzzing expected behavior instrumented process should be executed correctly test cases generated by fuzzing are expected actual behavior there are errors reports for calculate and calculatefromarray there are no test methods generated by fuzzing visual proofs screenshots logs images java error concreteexecutor executeasync response error org utbot instrumentation util instrumentedprocesserror error in the instrumented process com jetbrains rd util reactive rdfault invocationphase reason org utbot instrumentation instrumentation execution phases executionphaseerror invocationphase at org utbot instrumentation instrumentation execution phases invocationphase wraperror invocationphase kt at org utbot instrumentation instrumentation execution phases executionphasekt start executionphase kt at org utbot instrumentation instrumentation execution phases phasescontroller executephaseintimeout phasescontroller kt at org utbot instrumentation instrumentation execution utexecutioninstrumentation invoke utexecutioninstrumentation kt at org utbot instrumentation instrumentation execution utexecutioninstrumentation invoke utexecutioninstrumentation kt at org utbot instrumentation process instrumentedprocessmainkt setup invoke instrumentedprocessmain kt at org utbot instrumentation process instrumentedprocessmainkt setup invoke instrumentedprocessmain kt at org utbot rd idlewatchdog measuretimeforactivecall invoke clientprocessutil kt at org utbot rd idlewatchdog wrapactive clientprocessutil kt at org utbot rd idlewatchdog measuretimeforactivecall invoke clientprocessutil kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework impl rdcall onwirereceived rdtask kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl protocolcontexts readmessagecontextandinvoke protocolcontexts kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd util threading singlethreadschedulerbase queue run singlethreadscheduler kt at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java caused by java lang illegalargumentexception signature calculatefromarray i expecting this but provided argument list is empty at org utbot instrumentation instrumentation invokeinstrumentation invoke invokeinstrumentation kt at org utbot instrumentation instrumentation invokeinstrumentation invoke invokeinstrumentation kt at org utbot instrumentation instrumentation instrumentation defaultimpls invoke default instrumentation kt at org utbot instrumentation instrumentation execution phases invocationphase invoke invocationphase kt at org utbot instrumentation instrumentation execution utexecutioninstrumentation invoke concreteresult invoke utexecutioninstrumentation kt at org utbot instrumentation instrumentation execution utexecutioninstrumentation invoke concreteresult invoke utexecutioninstrumentation kt at org utbot instrumentation instrumentation execution phases phasescontroller executephaseintimeout result invoke phasescontroller kt at org utbot common threadbasedexecutor invokewithtimeout invoke threadutil kt at org utbot common threadbasedexecutor invokewithtimeout invoke threadutil kt at org utbot common threadbasedexecutor invokewithtimeout invoke threadutil kt at kotlin concurrent threadskt thread thread run thread kt at com jetbrains rd framework rdtaskresult companion read taskinterfaces kt at com jetbrains rd framework impl callsitewiredrdtask onwirereceived rdtask kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl protocolcontexts readmessagecontextandinvoke protocolcontexts kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl rdcall createresponsescheduler queue invoke rdtask kt at com jetbrains rd framework impl rdcall createresponsescheduler queue invokesuspend rdtask kt at kotlin coroutines jvm internal basecontinuationimpl resumewith continuationimpl kt at kotlinx coroutines dispatchedtask run dispatchedtask kt at kotlinx coroutines eventloopimplbase processnextevent eventloop common kt at kotlinx coroutines blockingcoroutine joinblocking builders kt at kotlinx coroutines builderskt builderskt runblocking builders kt at kotlinx coroutines builderskt runblocking unknown source at kotlinx coroutines builderskt builderskt runblocking default builders kt at kotlinx coroutines builderskt runblocking default unknown source at org utbot common concurrencykt runblockingwithcancellationpredicate concurrency kt at org utbot framework plugin api testcasegenerator generate invoke testcasegenerator kt at org utbot framework plugin api testcasegenerator generate invoke testcasegenerator kt at org utbot common concurrencykt runignoringcancellationexception concurrency kt at org utbot framework plugin api testcasegenerator generate testcasegenerator kt at org utbot framework process engineprocessmainkt setup invoke engineprocessmain kt at org utbot framework process engineprocessmainkt setup invoke engineprocessmain kt at org utbot rd idlewatchdog measuretimeforactivecall invoke clientprocessutil kt at org utbot rd idlewatchdog wrapactive clientprocessutil kt at org utbot rd idlewatchdog measuretimeforactivecall invoke clientprocessutil kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework impl rdcall onwirereceived rdtask kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl protocolcontexts readmessagecontextandinvoke protocolcontexts kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd util threading singlethreadschedulerbase queue run singlethreadscheduler kt at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java at org utbot instrumentation concreteexecutor withprocess concreteexecutor kt at org utbot instrumentation concreteexecutor withprocess invokesuspend concreteexecutor kt at kotlin coroutines jvm internal basecontinuationimpl resumewith continuationimpl kt at kotlinx coroutines dispatchedtask run dispatchedtask kt at kotlinx coroutines eventloopimplbase processnextevent eventloop common kt at kotlinx coroutines blockingcoroutine joinblocking builders kt at kotlinx coroutines builderskt builderskt runblocking builders kt at kotlinx coroutines builderskt runblocking unknown source at kotlinx coroutines builderskt builderskt runblocking default builders kt at kotlinx coroutines builderskt runblocking default unknown source at org utbot common concurrencykt runblockingwithcancellationpredicate concurrency kt at org utbot framework plugin api testcasegenerator generate invoke testcasegenerator kt at org utbot framework plugin api testcasegenerator generate invoke testcasegenerator kt at org utbot common concurrencykt runignoringcancellationexception concurrency kt at org utbot framework plugin api testcasegenerator generate testcasegenerator kt at org utbot framework process engineprocessmainkt setup invoke engineprocessmain kt at org utbot framework process engineprocessmainkt setup invoke engineprocessmain kt at org utbot rd idlewatchdog measuretimeforactivecall invoke clientprocessutil kt at org utbot rd idlewatchdog wrapactive clientprocessutil kt at org utbot rd idlewatchdog measuretimeforactivecall invoke clientprocessutil kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework impl rdcall onwirereceived rdtask kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl protocolcontexts readmessagecontextandinvoke protocolcontexts kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd util threading singlethreadschedulerbase queue run singlethreadscheduler kt at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by com jetbrains rd util reactive rdfault invocationphase reason org utbot instrumentation instrumentation execution phases executionphaseerror invocationphase at org utbot instrumentation instrumentation execution phases invocationphase wraperror invocationphase kt at org utbot instrumentation instrumentation execution phases executionphasekt start executionphase kt at org utbot instrumentation instrumentation execution phases phasescontroller executephaseintimeout phasescontroller kt at org utbot instrumentation instrumentation execution utexecutioninstrumentation invoke utexecutioninstrumentation kt at org utbot instrumentation instrumentation execution utexecutioninstrumentation invoke utexecutioninstrumentation kt at org utbot instrumentation process instrumentedprocessmainkt setup invoke instrumentedprocessmain kt at org utbot instrumentation process instrumentedprocessmainkt setup invoke instrumentedprocessmain kt at org utbot rd idlewatchdog measuretimeforactivecall invoke clientprocessutil kt at org utbot rd idlewatchdog wrapactive clientprocessutil kt at org utbot rd idlewatchdog measuretimeforactivecall invoke clientprocessutil kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework irdendpoint set invoke taskinterfaces kt at com jetbrains rd framework impl rdcall onwirereceived rdtask kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl protocolcontexts readmessagecontextandinvoke protocolcontexts kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd util threading singlethreadschedulerbase queue run singlethreadscheduler kt at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java caused by java lang illegalargumentexception signature calculatefromarray i expecting this but provided argument list is empty at org utbot instrumentation instrumentation invokeinstrumentation invoke invokeinstrumentation kt at org utbot instrumentation instrumentation invokeinstrumentation invoke invokeinstrumentation kt at org utbot instrumentation instrumentation instrumentation defaultimpls invoke default instrumentation kt at org utbot instrumentation instrumentation execution phases invocationphase invoke invocationphase kt at org utbot instrumentation instrumentation execution utexecutioninstrumentation invoke concreteresult invoke utexecutioninstrumentation kt at org utbot instrumentation instrumentation execution utexecutioninstrumentation invoke concreteresult invoke utexecutioninstrumentation kt at org utbot instrumentation instrumentation execution phases phasescontroller executephaseintimeout result invoke phasescontroller kt at org utbot common threadbasedexecutor invokewithtimeout invoke threadutil kt at org utbot common threadbasedexecutor invokewithtimeout invoke threadutil kt at org utbot common threadbasedexecutor invokewithtimeout invoke threadutil kt at kotlin concurrent threadskt thread thread run thread kt at com jetbrains rd framework rdtaskresult companion read taskinterfaces kt at com jetbrains rd framework impl callsitewiredrdtask onwirereceived rdtask kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl protocolcontexts readmessagecontextandinvoke protocolcontexts kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework messagebroker invoke invoke messagebroker kt at com jetbrains rd framework impl rdcall createresponsescheduler queue invoke rdtask kt at com jetbrains rd framework impl rdcall createresponsescheduler queue invokesuspend rdtask kt at kotlin coroutines jvm internal basecontinuationimpl resumewith continuationimpl kt at kotlinx coroutines dispatchedtask run dispatchedtask kt more environment windows pro intellij idea ultimate additional context when running this code in a separate project there were exceptions of this kind
1
15,709
19,848,719,239
IssuesEvent
2022-01-21 09:51:21
ooi-data/CE02SHSM-RID26-08-SPKIRB000-recovered_host-spkir_abj_dcl_instrument_recovered
https://api.github.com/repos/ooi-data/CE02SHSM-RID26-08-SPKIRB000-recovered_host-spkir_abj_dcl_instrument_recovered
opened
๐Ÿ›‘ Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:51:21.302493. ## Details Flow name: `CE02SHSM-RID26-08-SPKIRB000-recovered_host-spkir_abj_dcl_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
๐Ÿ›‘ Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:51:21.302493. ## Details Flow name: `CE02SHSM-RID26-08-SPKIRB000-recovered_host-spkir_abj_dcl_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
๐Ÿ›‘ processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host spkir abj dcl instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
538,919
15,780,898,657
IssuesEvent
2021-04-01 10:34:34
robotframework/robotframework
https://api.github.com/repos/robotframework/robotframework
closed
Using `BuiltIn().run_keyword` in listener `start/end_suite` method creates invalid output.xml
bug priority: critical
In our system we use a listener class, in which we use several BuiltIn().run_keyword() calls. In versions prior to 4.0, ie 3.2.2, this worked nicely and gave no error. In version 4.0 however, even though the run_keyword() calls succeed and relevant test gets a PASS status we get an error at the bottom of console report: `[ ERROR ] Reading XML source '/folder/output.xml' failed: AttributeError: 'TestSuite' object has no attribute 'body'` I feel this is related to the recent <body> tag change of output.xml file. Here is a quick demonstration: test.robot ``` *** Test Cases *** NOOP Log Do Nothing. ``` ListenerClass.py ``` from robot.libraries.BuiltIn import BuiltIn class ListenerClass(object): ROBOT_LISTENER_API_VERSION = 3 def start_suite(self, suite, result): BuiltIn().run_keyword("Log", "this is a test", "console=True") ``` When this is run with the command: `robot --listener ListenerClass.py test.robot` The output will be: ``` ============================================================================== Test ============================================================================== this is a test NOOP | PASS | ------------------------------------------------------------------------------ Test | PASS | 1 test, 1 passed, 0 failed ============================================================================== Output: /home/tuncay/bugdemo/output.xml [ ERROR ] Reading XML source '/home/tuncay/bugdemo/output.xml' failed: AttributeError: 'TestSuite' object has no attribute 'body' Try --help for usage information. ``` Note the ERROR on the bottom.
1.0
Using `BuiltIn().run_keyword` in listener `start/end_suite` method creates invalid output.xml - In our system we use a listener class, in which we use several BuiltIn().run_keyword() calls. In versions prior to 4.0, ie 3.2.2, this worked nicely and gave no error. In version 4.0 however, even though the run_keyword() calls succeed and relevant test gets a PASS status we get an error at the bottom of console report: `[ ERROR ] Reading XML source '/folder/output.xml' failed: AttributeError: 'TestSuite' object has no attribute 'body'` I feel this is related to the recent <body> tag change of output.xml file. Here is a quick demonstration: test.robot ``` *** Test Cases *** NOOP Log Do Nothing. ``` ListenerClass.py ``` from robot.libraries.BuiltIn import BuiltIn class ListenerClass(object): ROBOT_LISTENER_API_VERSION = 3 def start_suite(self, suite, result): BuiltIn().run_keyword("Log", "this is a test", "console=True") ``` When this is run with the command: `robot --listener ListenerClass.py test.robot` The output will be: ``` ============================================================================== Test ============================================================================== this is a test NOOP | PASS | ------------------------------------------------------------------------------ Test | PASS | 1 test, 1 passed, 0 failed ============================================================================== Output: /home/tuncay/bugdemo/output.xml [ ERROR ] Reading XML source '/home/tuncay/bugdemo/output.xml' failed: AttributeError: 'TestSuite' object has no attribute 'body' Try --help for usage information. ``` Note the ERROR on the bottom.
non_process
using builtin run keyword in listener start end suite method creates invalid output xml in our system we use a listener class in which we use several builtin run keyword calls in versions prior to ie this worked nicely and gave no error in version however even though the run keyword calls succeed and relevant test gets a pass status we get an error at the bottom of console report reading xml source folder output xml failed attributeerror testsuite object has no attribute body i feel this is related to the recent tag change of output xml file here is a quick demonstration test robot test cases noop log do nothing listenerclass py from robot libraries builtin import builtin class listenerclass object robot listener api version def start suite self suite result builtin run keyword log this is a test console true when this is run with the command robot listener listenerclass py test robot the output will be test this is a test noop pass test pass test passed failed output home tuncay bugdemo output xml reading xml source home tuncay bugdemo output xml failed attributeerror testsuite object has no attribute body try help for usage information note the error on the bottom
0
2,011
4,835,845,757
IssuesEvent
2016-11-08 17:53:55
LazyTroll/WikiCode
https://api.github.com/repos/LazyTroll/WikiCode
closed
ะกะพะทะดะฐั‚ัŒ ะฝะพะฒั‹ะน ะผะพะดัƒะปัŒ WikiVersions
introduction process task
ะกะฒะพะตะพะฑั€ะฐะทะฝะฐั ัะธัั‚ะตะผะฐ ะบะพะฝั‚ั€ะพะปั ะฒะตั€ัะธะน ะดะปั ะบะพะฝัะฟะตะบั‚ะพะฒ. ะœะพะดัƒะปัŽ ะผะพะถะฝะพ ัƒะบะฐะทะฐั‚ัŒ, ะบะฐะบะธะต ะฟะฐั€ะฐะณั€ะฐั„ั‹ ะฑั‹ะปะธ ะดะพะฑะฐะฒะปะตะฝั‹, ะบะฐะบะธะต ะธะทะผะตะฝะตะฝั‹, ะบะฐะบะธะต ัƒะดะฐะปะตะฝั‹ ะธ ะฟั€ะพั‡ะตะต. ะะฐ ะพัะฝะพะฒะต ัั‚ะพะณะพ, WikiVersions ะดะพะปะถะตะฝ ะฟะพั€ะพะถะดะฐั‚ัŒ ะฝะพะฒัƒัŽ ะฒะตั€ัะธัŽ. ะ”ะพะปะถะตะฝ ั…ั€ะฐะฝะธั‚ัŒ ะฒ ัะตะฑะต ะปะธัˆัŒ ั‚ะพะปัŒะบะพ ะธะทะผะตะฝะตะฝะธั ะธ ั‚ะตะบัƒั‰ัƒัŽ ะฐะบั‚ัƒะฐะปัŒะฝัƒัŽ ะฒะตั€ัะธัŽ ะบะพะฝัะฟะตะบั‚ะฐ. ะขะฐะบะถะต, ะพะฝ ะดะพะปะถะตะฝ ัƒะผะตั‚ัŒ ะปะตะณะบะพ ะฟะตั€ะตะบะปัŽั‡ะฐั‚ัŒ ะฒะตั€ัะธะธ, ะดะพะฑะฐะฒะปัั/ัƒะฑะธั€ะฐั ะธะทะผะตะฝะตะฝะธั.
1.0
ะกะพะทะดะฐั‚ัŒ ะฝะพะฒั‹ะน ะผะพะดัƒะปัŒ WikiVersions - ะกะฒะพะตะพะฑั€ะฐะทะฝะฐั ัะธัั‚ะตะผะฐ ะบะพะฝั‚ั€ะพะปั ะฒะตั€ัะธะน ะดะปั ะบะพะฝัะฟะตะบั‚ะพะฒ. ะœะพะดัƒะปัŽ ะผะพะถะฝะพ ัƒะบะฐะทะฐั‚ัŒ, ะบะฐะบะธะต ะฟะฐั€ะฐะณั€ะฐั„ั‹ ะฑั‹ะปะธ ะดะพะฑะฐะฒะปะตะฝั‹, ะบะฐะบะธะต ะธะทะผะตะฝะตะฝั‹, ะบะฐะบะธะต ัƒะดะฐะปะตะฝั‹ ะธ ะฟั€ะพั‡ะตะต. ะะฐ ะพัะฝะพะฒะต ัั‚ะพะณะพ, WikiVersions ะดะพะปะถะตะฝ ะฟะพั€ะพะถะดะฐั‚ัŒ ะฝะพะฒัƒัŽ ะฒะตั€ัะธัŽ. ะ”ะพะปะถะตะฝ ั…ั€ะฐะฝะธั‚ัŒ ะฒ ัะตะฑะต ะปะธัˆัŒ ั‚ะพะปัŒะบะพ ะธะทะผะตะฝะตะฝะธั ะธ ั‚ะตะบัƒั‰ัƒัŽ ะฐะบั‚ัƒะฐะปัŒะฝัƒัŽ ะฒะตั€ัะธัŽ ะบะพะฝัะฟะตะบั‚ะฐ. ะขะฐะบะถะต, ะพะฝ ะดะพะปะถะตะฝ ัƒะผะตั‚ัŒ ะปะตะณะบะพ ะฟะตั€ะตะบะปัŽั‡ะฐั‚ัŒ ะฒะตั€ัะธะธ, ะดะพะฑะฐะฒะปัั/ัƒะฑะธั€ะฐั ะธะทะผะตะฝะตะฝะธั.
process
ัะพะทะดะฐั‚ัŒ ะฝะพะฒั‹ะน ะผะพะดัƒะปัŒ wikiversions ัะฒะพะตะพะฑั€ะฐะทะฝะฐั ัะธัั‚ะตะผะฐ ะบะพะฝั‚ั€ะพะปั ะฒะตั€ัะธะน ะดะปั ะบะพะฝัะฟะตะบั‚ะพะฒ ะผะพะดัƒะปัŽ ะผะพะถะฝะพ ัƒะบะฐะทะฐั‚ัŒ ะบะฐะบะธะต ะฟะฐั€ะฐะณั€ะฐั„ั‹ ะฑั‹ะปะธ ะดะพะฑะฐะฒะปะตะฝั‹ ะบะฐะบะธะต ะธะทะผะตะฝะตะฝั‹ ะบะฐะบะธะต ัƒะดะฐะปะตะฝั‹ ะธ ะฟั€ะพั‡ะตะต ะฝะฐ ะพัะฝะพะฒะต ัั‚ะพะณะพ wikiversions ะดะพะปะถะตะฝ ะฟะพั€ะพะถะดะฐั‚ัŒ ะฝะพะฒัƒัŽ ะฒะตั€ัะธัŽ ะดะพะปะถะตะฝ ั…ั€ะฐะฝะธั‚ัŒ ะฒ ัะตะฑะต ะปะธัˆัŒ ั‚ะพะปัŒะบะพ ะธะทะผะตะฝะตะฝะธั ะธ ั‚ะตะบัƒั‰ัƒัŽ ะฐะบั‚ัƒะฐะปัŒะฝัƒัŽ ะฒะตั€ัะธัŽ ะบะพะฝัะฟะตะบั‚ะฐ ั‚ะฐะบะถะต ะพะฝ ะดะพะปะถะตะฝ ัƒะผะตั‚ัŒ ะปะตะณะบะพ ะฟะตั€ะตะบะปัŽั‡ะฐั‚ัŒ ะฒะตั€ัะธะธ ะดะพะฑะฐะฒะปัั ัƒะฑะธั€ะฐั ะธะทะผะตะฝะตะฝะธั
1
15,803
11,712,656,863
IssuesEvent
2020-03-09 08:49:48
DivanteLtd/shopware-pwa
https://api.github.com/repos/DivanteLtd/shopware-pwa
closed
Change deployment from Zeit to Storefrontcloud
Core Team Infrastructure
Acceptance criteria - [x] PR/branch is able to deploy to shopware-pwa instance on cloud
1.0
Change deployment from Zeit to Storefrontcloud - Acceptance criteria - [x] PR/branch is able to deploy to shopware-pwa instance on cloud
non_process
change deployment from zeit to storefrontcloud acceptance criteria pr branch is able to deploy to shopware pwa instance on cloud
0
6,881
10,023,501,238
IssuesEvent
2019-07-16 19:21:28
shirou/gopsutil
https://api.github.com/repos/shirou/gopsutil
closed
Find top processes by their resource consumption
package:process
How to get top processes utilization CPU, Mem, Disk? Currently via process.Pids() we can fetch all processes. Is it another way? Additionally how to calculate CPU percentage for process? For example memory: MemoryPercent()
1.0
Find top processes by their resource consumption - How to get top processes utilization CPU, Mem, Disk? Currently via process.Pids() we can fetch all processes. Is it another way? Additionally how to calculate CPU percentage for process? For example memory: MemoryPercent()
process
find top processes by their resource consumption how to get top processes utilization cpu mem disk currently via process pids we can fetch all processes is it another way additionally how to calculate cpu percentage for process for example memory memorypercent
1
18,035
24,045,065,967
IssuesEvent
2022-09-16 07:33:07
OI-wiki/OI-wiki
https://api.github.com/repos/OI-wiki/OI-wiki
closed
SA้ƒจๅˆ†ไธŠไธ‹่ฎฐๅทไธ็ปŸไธ€
Help needed for format / ้œ€่ฆๅธฎๅŠฉไฟฎๆญฃๆ ผๅผ Need Processing / ้œ€่ฆๅค„็†
็ŽฐๅœจSA้ƒจๅˆ†ไธŠไธ‹่ฎฐๅทไธ็ปŸไธ€๏ผŒๆ˜ฏไธๆ˜ฏๅบ”่ฏฅ็ปŸไธ€ไธ€ไธ‹๏ผŸ _Originally posted by @partychicken in https://github.com/24OI/OI-wiki/pull/1031#issuecomment-471888174_ ๆˆ‘่ง‰ๅพ—ไธ€ไธช pr ๅšๅฅฝไธ€ไปถไบ‹ๅฐฑๅฏไปฅๅ•ฆ๏ผˆ ไธ้œ€่ฆไธ€ๆญฅๅˆฐไฝ๏ผˆ๏ผˆ
1.0
SA้ƒจๅˆ†ไธŠไธ‹่ฎฐๅทไธ็ปŸไธ€ - ็ŽฐๅœจSA้ƒจๅˆ†ไธŠไธ‹่ฎฐๅทไธ็ปŸไธ€๏ผŒๆ˜ฏไธๆ˜ฏๅบ”่ฏฅ็ปŸไธ€ไธ€ไธ‹๏ผŸ _Originally posted by @partychicken in https://github.com/24OI/OI-wiki/pull/1031#issuecomment-471888174_ ๆˆ‘่ง‰ๅพ—ไธ€ไธช pr ๅšๅฅฝไธ€ไปถไบ‹ๅฐฑๅฏไปฅๅ•ฆ๏ผˆ ไธ้œ€่ฆไธ€ๆญฅๅˆฐไฝ๏ผˆ๏ผˆ
process
sa้ƒจๅˆ†ไธŠไธ‹่ฎฐๅทไธ็ปŸไธ€ ็Žฐๅœจsa้ƒจๅˆ†ไธŠไธ‹่ฎฐๅทไธ็ปŸไธ€๏ผŒๆ˜ฏไธๆ˜ฏๅบ”่ฏฅ็ปŸไธ€ไธ€ไธ‹๏ผŸ originally posted by partychicken in ๆˆ‘่ง‰ๅพ—ไธ€ไธช pr ๅšๅฅฝไธ€ไปถไบ‹ๅฐฑๅฏไปฅๅ•ฆ๏ผˆ ไธ้œ€่ฆไธ€ๆญฅๅˆฐไฝ๏ผˆ๏ผˆ
1
336,276
24,491,569,930
IssuesEvent
2022-10-10 02:57:23
udoprog/audio
https://api.github.com/repos/udoprog/audio
opened
Unify documentation under a few common concepts
documentation
This library uses the following concepts: * Buffer refers to an object holding an audio buffer. * A sample is a single value from with in a buffer from a single channel. * Channel refers to a single channel inside of an audio buffer. * A frame refers to the group of channel at a given offset inside of an audio buffer. The primary abstraction are *buffers* that contain 0 or more *channels*. Each buffer can also be viewed as a sequence of frames, where each frame contains the corresponding sample from every channel in that buffer.
1.0
Unify documentation under a few common concepts - This library uses the following concepts: * Buffer refers to an object holding an audio buffer. * A sample is a single value from with in a buffer from a single channel. * Channel refers to a single channel inside of an audio buffer. * A frame refers to the group of channel at a given offset inside of an audio buffer. The primary abstraction are *buffers* that contain 0 or more *channels*. Each buffer can also be viewed as a sequence of frames, where each frame contains the corresponding sample from every channel in that buffer.
non_process
unify documentation under a few common concepts this library uses the following concepts buffer refers to an object holding an audio buffer a sample is a single value from with in a buffer from a single channel channel refers to a single channel inside of an audio buffer a frame refers to the group of channel at a given offset inside of an audio buffer the primary abstraction are buffers that contain or more channels each buffer can also be viewed as a sequence of frames where each frame contains the corresponding sample from every channel in that buffer
0
49,217
12,299,571,807
IssuesEvent
2020-05-11 12:37:36
doxygen/doxygen
https://api.github.com/repos/doxygen/doxygen
closed
error when use make and make install commands line
install/build invalid needinfo
Hello, error when use make and make install commands line : ![HereMyErrorAboutMakeCommandLineAndMakeInstallCommandLine](https://user-images.githubusercontent.com/30596513/81493215-f1f3db80-929e-11ea-97a3-8e3db762f528.jpg) Thank you in advance to repair the problems, Regards. Azaretdodo.
1.0
error when use make and make install commands line - Hello, error when use make and make install commands line : ![HereMyErrorAboutMakeCommandLineAndMakeInstallCommandLine](https://user-images.githubusercontent.com/30596513/81493215-f1f3db80-929e-11ea-97a3-8e3db762f528.jpg) Thank you in advance to repair the problems, Regards. Azaretdodo.
non_process
error when use make and make install commands line hello error when use make and make install commands line thank you in advance to repair the problems regards azaretdodo
0
9,159
12,217,599,744
IssuesEvent
2020-05-01 17:32:37
kubeflow/testing
https://api.github.com/repos/kubeflow/testing
closed
Replace kubeflow-bot with a GitHub App and GitHub Org
area/engprod kind/process lifecycle/stale priority/p2
Right now we use a user account https://github.com/kubeflow-bot as a robot account. This is really a pain to manage because there is no easy way to make multiple people administrators of this account. Using a GitHub App is also much better for programmatic modification. We can easily create GitHub Apps within our existing Kubeflow GitHub org. GitHub Apps however can't "own" repositories. So when our CI/CD bot opens up PRs to update kubeflow/manifests where should it create those new branches? Possible options 1. Use a branch on kubeflow/manifests 1. Create a new GitHub org "kubeflow-bots" and put forks of any repos needed by bots there. I think I prefer the second option. * Using a fork seems cleaner then polluting kubeflow repos with a bunch of auto created PRs * Using separate forks seems like it would require giving the bot less privileges
1.0
Replace kubeflow-bot with a GitHub App and GitHub Org - Right now we use a user account https://github.com/kubeflow-bot as a robot account. This is really a pain to manage because there is no easy way to make multiple people administrators of this account. Using a GitHub App is also much better for programmatic modification. We can easily create GitHub Apps within our existing Kubeflow GitHub org. GitHub Apps however can't "own" repositories. So when our CI/CD bot opens up PRs to update kubeflow/manifests where should it create those new branches? Possible options 1. Use a branch on kubeflow/manifests 1. Create a new GitHub org "kubeflow-bots" and put forks of any repos needed by bots there. I think I prefer the second option. * Using a fork seems cleaner then polluting kubeflow repos with a bunch of auto created PRs * Using separate forks seems like it would require giving the bot less privileges
process
replace kubeflow bot with a github app and github org right now we use a user account as a robot account this is really a pain to manage because there is no easy way to make multiple people administrators of this account using a github app is also much better for programmatic modification we can easily create github apps within our existing kubeflow github org github apps however can t own repositories so when our ci cd bot opens up prs to update kubeflow manifests where should it create those new branches possible options use a branch on kubeflow manifests create a new github org kubeflow bots and put forks of any repos needed by bots there i think i prefer the second option using a fork seems cleaner then polluting kubeflow repos with a bunch of auto created prs using separate forks seems like it would require giving the bot less privileges
1
174,024
14,444,487,927
IssuesEvent
2020-12-07 21:19:42
WilliamBronzo/OpenCV_Python_UFRN_DCA
https://api.github.com/repos/WilliamBronzo/OpenCV_Python_UFRN_DCA
opened
Unidade 2
documentation enhancement
- [ ] Questรฃo 7: - [x] Fazer o programa. - [ ] Apresentar o problema da lista. - [ ] Explicar o programa. - [ ] Melhorar a apresentaรงรฃo do GitHub. - [ ] Questรฃo 8: - [x] Fazer o programa. - [ ] Apresentar o problema da lista. - [ ] Explicar o programa. - [ ] Melhorar a apresentaรงรฃo do GitHub. - [ ] Questรฃo 9: - [x] Fazer o programa. - [ ] Apresentar o problema da lista. - [ ] Explicar o programa. - [ ] Melhorar a apresentaรงรฃo do GitHub.
1.0
Unidade 2 - - [ ] Questรฃo 7: - [x] Fazer o programa. - [ ] Apresentar o problema da lista. - [ ] Explicar o programa. - [ ] Melhorar a apresentaรงรฃo do GitHub. - [ ] Questรฃo 8: - [x] Fazer o programa. - [ ] Apresentar o problema da lista. - [ ] Explicar o programa. - [ ] Melhorar a apresentaรงรฃo do GitHub. - [ ] Questรฃo 9: - [x] Fazer o programa. - [ ] Apresentar o problema da lista. - [ ] Explicar o programa. - [ ] Melhorar a apresentaรงรฃo do GitHub.
non_process
unidade questรฃo fazer o programa apresentar o problema da lista explicar o programa melhorar a apresentaรงรฃo do github questรฃo fazer o programa apresentar o problema da lista explicar o programa melhorar a apresentaรงรฃo do github questรฃo fazer o programa apresentar o problema da lista explicar o programa melhorar a apresentaรงรฃo do github
0
159,649
6,050,184,859
IssuesEvent
2017-06-12 20:29:46
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Add support for minions to import and run docker images from tarball
priority/awaiting-more-evidence sig/cluster-lifecycle sig/node triaged
The docker private registry is still not fully flushed out and many organizations do not want to upload their images to the global docker hub. For these scenarios, finding a solution to push down a tarball file and letting the minions import it into their local repository and running it is a more optimal choice. Can we add this support?
1.0
Add support for minions to import and run docker images from tarball - The docker private registry is still not fully flushed out and many organizations do not want to upload their images to the global docker hub. For these scenarios, finding a solution to push down a tarball file and letting the minions import it into their local repository and running it is a more optimal choice. Can we add this support?
non_process
add support for minions to import and run docker images from tarball the docker private registry is still not fully flushed out and many organizations do not want to upload their images to the global docker hub for these scenarios finding a solution to push down a tarball file and letting the minions import it into their local repository and running it is a more optimal choice can we add this support
0
18,859
24,781,207,664
IssuesEvent
2022-10-24 05:16:16
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Error in migration engine: entered unreachable code
kind/bug process/candidate tech/engines/datamodel topic: error reporting topic: mysql team/schema topic: prisma db push
<!-- If required, please update the title to be clear and descriptive --> Command: `prisma db push` Version: `4.4.0` Binary Version: `f352a33b70356f46311da8b00d83386dd9f145d6` Report: https://prisma-errors.netlify.app/report/14375 OS: `x64 win32 10.0.22000` Rust Stacktrace: ``` Starting migration engine RPC server Aborting migration because at least one unexecutable step was detected. Attempting best_effort_reset [migration-engine\connectors\sql-migration-connector\src\sql_renderer\mysql_renderer.rs:522:97] internal error: entered unreachable code ``` ## Context We're hitting this line: https://github.com/prisma/prisma-engines/blob/4.4.0/migration-engine/connectors/sql-migration-connector/src/sql_renderer/mysql_renderer.rs#L522 ```rust fn render_default<'a>(column: ColumnWalker<'a>, default: &'a DefaultValue) -> Cow<'a, str> { match default.kind() { // ... DefaultKind::DbGenerated(None) | DefaultKind::Sequence(_) | DefaultKind::UniqueRowid => unreachable!(), } } ```
1.0
Error in migration engine: entered unreachable code - <!-- If required, please update the title to be clear and descriptive --> Command: `prisma db push` Version: `4.4.0` Binary Version: `f352a33b70356f46311da8b00d83386dd9f145d6` Report: https://prisma-errors.netlify.app/report/14375 OS: `x64 win32 10.0.22000` Rust Stacktrace: ``` Starting migration engine RPC server Aborting migration because at least one unexecutable step was detected. Attempting best_effort_reset [migration-engine\connectors\sql-migration-connector\src\sql_renderer\mysql_renderer.rs:522:97] internal error: entered unreachable code ``` ## Context We're hitting this line: https://github.com/prisma/prisma-engines/blob/4.4.0/migration-engine/connectors/sql-migration-connector/src/sql_renderer/mysql_renderer.rs#L522 ```rust fn render_default<'a>(column: ColumnWalker<'a>, default: &'a DefaultValue) -> Cow<'a, str> { match default.kind() { // ... DefaultKind::DbGenerated(None) | DefaultKind::Sequence(_) | DefaultKind::UniqueRowid => unreachable!(), } } ```
process
error in migration engine entered unreachable code command prisma db push version binary version report os rust stacktrace starting migration engine rpc server aborting migration because at least one unexecutable step was detected attempting best effort reset internal error entered unreachable code context we re hitting this line rust fn render default column columnwalker default a defaultvalue cow match default kind defaultkind dbgenerated none defaultkind sequence defaultkind uniquerowid unreachable
1
2,341
5,145,230,815
IssuesEvent
2017-01-12 20:58:21
glos/myglos
https://api.github.com/repos/glos/myglos
reopened
GLOS By the Numbers (extremes)
enhancement - future/proposed process
Becky would like to programmatically capture various GLOS by the numbers (extremes like 'highest waves'). - Short term have them emailed to GLOS to use in marketing. - Longer term would like it to be something that displays on the portal itself. Kelly confirmed capturing โ€˜extremesโ€™ data is feasible but would like to wait until after annual meeting. Tad will put on the next fiscal year planning. See also : https://drive.google.com/open?id=0B7pJZVmOuWvzV0hkdVdWOW9XRGs
1.0
GLOS By the Numbers (extremes) - Becky would like to programmatically capture various GLOS by the numbers (extremes like 'highest waves'). - Short term have them emailed to GLOS to use in marketing. - Longer term would like it to be something that displays on the portal itself. Kelly confirmed capturing โ€˜extremesโ€™ data is feasible but would like to wait until after annual meeting. Tad will put on the next fiscal year planning. See also : https://drive.google.com/open?id=0B7pJZVmOuWvzV0hkdVdWOW9XRGs
process
glos by the numbers extremes becky would like to programmatically capture various glos by the numbers extremes like highest waves short term have them emailed to glos to use in marketing longer term would like it to be something that displays on the portal itself kelly confirmed capturing โ€˜extremesโ€™ data is feasible but would like to wait until after annual meeting tad will put on the next fiscal year planning see also
1
16,885
22,163,378,496
IssuesEvent
2022-06-04 21:11:06
New-Time-Development/.github
https://api.github.com/repos/New-Time-Development/.github
closed
Create new Readme Design
Work in process enhancement priority: high
Hes working on a new design of the file and the general Organisation profile
1.0
Create new Readme Design - Hes working on a new design of the file and the general Organisation profile
process
create new readme design hes working on a new design of the file and the general organisation profile
1
87,438
17,267,508,845
IssuesEvent
2021-07-22 15:22:41
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
closed
Fix decoders number repetition
threatintel threatintel/decoders
|Wazuh version|Component|Install type|Install method|Platform| |---|---|---|---|---| | 4.3| Decoder | - | - | - | <!-- Whenever possible, issues should be created for bug reporting and feature requests. For questions related to the user experience, please refer: - Wazuh mailing list: https://groups.google.com/forum/#!forum/wazuh - Join Wazuh on Slack: https://wazuh.com/community/join-us-on-slack Please fill the table above. Feel free to extend it at your convenience. --> Hello team, The decoders [0555-aws-eks-authenticator_decoders.xml](https://github.com/wazuh/wazuh/blob/master/ruleset/decoders/0555-aws-eks-authenticator_decoders.xml) and [0555-fireeye_decoders.xml](https://github.com/wazuh/wazuh/blob/master/ruleset/decoders/0555-fireeye_decoders.xml) has the same number. Also, the decoders [0550-oracledb_decoders.xml](https://github.com/wazuh/wazuh/blob/master/ruleset/decoders/0550-oracledb_decoders.xml) and [0550-arbor_decoders.xml](https://github.com/wazuh/wazuh/blob/master/ruleset/decoders/0550-arbor_decoders.xml) also has the same number. We should change this to unique numbers. <!-- You may want to set debug options `<component>.debug=2` (see https://documentation.wazuh.com/current/user-manual/reference/internal-options.html) to get verbose logs. This may help investigate the issue. -->
1.0
Fix decoders number repetition - |Wazuh version|Component|Install type|Install method|Platform| |---|---|---|---|---| | 4.3| Decoder | - | - | - | <!-- Whenever possible, issues should be created for bug reporting and feature requests. For questions related to the user experience, please refer: - Wazuh mailing list: https://groups.google.com/forum/#!forum/wazuh - Join Wazuh on Slack: https://wazuh.com/community/join-us-on-slack Please fill the table above. Feel free to extend it at your convenience. --> Hello team, The decoders [0555-aws-eks-authenticator_decoders.xml](https://github.com/wazuh/wazuh/blob/master/ruleset/decoders/0555-aws-eks-authenticator_decoders.xml) and [0555-fireeye_decoders.xml](https://github.com/wazuh/wazuh/blob/master/ruleset/decoders/0555-fireeye_decoders.xml) has the same number. Also, the decoders [0550-oracledb_decoders.xml](https://github.com/wazuh/wazuh/blob/master/ruleset/decoders/0550-oracledb_decoders.xml) and [0550-arbor_decoders.xml](https://github.com/wazuh/wazuh/blob/master/ruleset/decoders/0550-arbor_decoders.xml) also has the same number. We should change this to unique numbers. <!-- You may want to set debug options `<component>.debug=2` (see https://documentation.wazuh.com/current/user-manual/reference/internal-options.html) to get verbose logs. This may help investigate the issue. -->
non_process
fix decoders number repetition wazuh version component install type install method platform decoder whenever possible issues should be created for bug reporting and feature requests for questions related to the user experience please refer wazuh mailing list join wazuh on slack please fill the table above feel free to extend it at your convenience hello team the decoders and has the same number also the decoders and also has the same number we should change this to unique numbers you may want to set debug options debug see to get verbose logs this may help investigate the issue
0
20,575
27,236,637,365
IssuesEvent
2023-02-21 16:47:24
alphagov/govuk-design-system
https://api.github.com/repos/alphagov/govuk-design-system
opened
Build a publishing plan for Exit this Page
process
## What Scope remaining stories for mvp and build a publishing plan. ## Why So the team and stakeholders know what's left to do and we can give ourselves healthy targets to work towards. ## Who needs to work on this Kelly ##ย Who needs to review this Steve, Katrina, Ciandelle, David, Owen, Beeps, Calvin ##ย Done when - [ ] Run a workshop to explore remaining actions - [ ] Estimate timescales - [ ] Document decisions
1.0
Build a publishing plan for Exit this Page - ## What Scope remaining stories for mvp and build a publishing plan. ## Why So the team and stakeholders know what's left to do and we can give ourselves healthy targets to work towards. ## Who needs to work on this Kelly ##ย Who needs to review this Steve, Katrina, Ciandelle, David, Owen, Beeps, Calvin ##ย Done when - [ ] Run a workshop to explore remaining actions - [ ] Estimate timescales - [ ] Document decisions
process
build a publishing plan for exit this page what scope remaining stories for mvp and build a publishing plan why so the team and stakeholders know what s left to do and we can give ourselves healthy targets to work towards who needs to work on this kelly ย who needs to review this steve katrina ciandelle david owen beeps calvin ย done when run a workshop to explore remaining actions estimate timescales document decisions
1
313,821
23,493,171,324
IssuesEvent
2022-08-17 20:59:17
mattermost/focalboard
https://api.github.com/repos/mattermost/focalboard
closed
Doc: Add Help Doc note to inform users โ€œEnable Publicly-Shared Boardsโ€ will be on by default for Cloud
Documentation
## Summary Add note in Help Docs that the System Console option to โ€œEnable Publicly-Shared Boardsโ€ will be on by default for Cloud. ## (Optional) Additional context and/or screenshot https://docs.google.com/document/d/1VxIcDnI5jc68mQSNKSICv-0npkN8nGl2cnkWIMQUiEk/edit?disco=AAAAXtYBIkI
1.0
Doc: Add Help Doc note to inform users โ€œEnable Publicly-Shared Boardsโ€ will be on by default for Cloud - ## Summary Add note in Help Docs that the System Console option to โ€œEnable Publicly-Shared Boardsโ€ will be on by default for Cloud. ## (Optional) Additional context and/or screenshot https://docs.google.com/document/d/1VxIcDnI5jc68mQSNKSICv-0npkN8nGl2cnkWIMQUiEk/edit?disco=AAAAXtYBIkI
non_process
doc add help doc note to inform users โ€œenable publicly shared boardsโ€ will be on by default for cloud summary add note in help docs that the system console option to โ€œenable publicly shared boardsโ€ will be on by default for cloud optional additional context and or screenshot
0
16,912
22,240,016,438
IssuesEvent
2022-06-09 03:33:23
streamnative/pulsar-spark
https://api.github.com/repos/streamnative/pulsar-spark
closed
[FEATURE] Allow configure predefined subscription for metadata reader
type/feature compute/data-processing
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when \[...] **Describe the solution you'd like** Allow users to provide a predefined pulsar subscription to manage the cursor. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
1.0
[FEATURE] Allow configure predefined subscription for metadata reader - **Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when \[...] **Describe the solution you'd like** Allow users to provide a predefined pulsar subscription to manage the cursor. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
process
allow configure predefined subscription for metadata reader is your feature request related to a problem please describe a clear and concise description of what the problem is ex i m always frustrated when describe the solution you d like allow users to provide a predefined pulsar subscription to manage the cursor describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
1
71,537
9,526,949,717
IssuesEvent
2019-04-29 00:29:50
KaTeX/KaTeX
https://api.github.com/repos/KaTeX/KaTeX
closed
Mhchem not loaded on katex.org
contrib documentation good first issue
Hi! I tried to enable mhchem extension that shipped in the new 0.10.1 version on my site. It seems to work although some rendering seems a bit odd (will make another issue for that) so I wanted compare with katex.org test page. It seems that mhchem is not loaded there. I realize that might be by design since mhchem is an extension, but maybe in this case it is worth it to add?
1.0
Mhchem not loaded on katex.org - Hi! I tried to enable mhchem extension that shipped in the new 0.10.1 version on my site. It seems to work although some rendering seems a bit odd (will make another issue for that) so I wanted compare with katex.org test page. It seems that mhchem is not loaded there. I realize that might be by design since mhchem is an extension, but maybe in this case it is worth it to add?
non_process
mhchem not loaded on katex org hi i tried to enable mhchem extension that shipped in the new version on my site it seems to work although some rendering seems a bit odd will make another issue for that so i wanted compare with katex org test page it seems that mhchem is not loaded there i realize that might be by design since mhchem is an extension but maybe in this case it is worth it to add
0
789,934
27,810,132,321
IssuesEvent
2023-03-18 02:40:08
antoinecarme/pyaf
https://api.github.com/repos/antoinecarme/pyaf
opened
Final Touch 3 : report plot filenames in the logs
class:enhancement class:Doc priority:high topic:plotting status:in_progress
The call cForecastEngine.standardPlots produces a lot of png files by default. Report these file names in the model logs. some thing like : ``` INFO:pyaf.timing:('OPERATION_START', ('PLOTTING', {'Signals': ['Ozone']})) INFO:pyaf.std:SAVING_PLOT ('Trend', 'outputs/my_ozone_Ozone_Trend_decomp_output.png') INFO:pyaf.std:SAVING_PLOT ('Cycle', 'outputs/my_ozone_Ozone_Cycle_decomp_output.png') INFO:pyaf.std:SAVING_PLOT ('AR', 'outputs/my_ozone_Ozone_AR_decomp_output.png') INFO:pyaf.std:SAVING_PLOT ('TransformedForecast', 'outputs/my_ozone_Ozone_TransformedForecast_decomp_output.png') INFO:pyaf.std:SAVING_PLOT ('Forecast', 'outputs/my_ozone_Ozone_Forecast_decomp_output.png') INFO:pyaf.std:SAVING_PLOT ('PredictionIntervals', 'outputs/my_ozone_Ozone_prediction_intervals_output.png') INFO:pyaf.std:SAVING_PLOT ('Quantiles', 'outputs/my_ozone_Ozone_quantiles_output.png') INFO:pyaf.timing:('OPERATION_END_ELAPSED', 5.11, ('PLOTTING', {'Signals': ['Ozone']})) ```
1.0
Final Touch 3 : report plot filenames in the logs - The call cForecastEngine.standardPlots produces a lot of png files by default. Report these file names in the model logs. some thing like : ``` INFO:pyaf.timing:('OPERATION_START', ('PLOTTING', {'Signals': ['Ozone']})) INFO:pyaf.std:SAVING_PLOT ('Trend', 'outputs/my_ozone_Ozone_Trend_decomp_output.png') INFO:pyaf.std:SAVING_PLOT ('Cycle', 'outputs/my_ozone_Ozone_Cycle_decomp_output.png') INFO:pyaf.std:SAVING_PLOT ('AR', 'outputs/my_ozone_Ozone_AR_decomp_output.png') INFO:pyaf.std:SAVING_PLOT ('TransformedForecast', 'outputs/my_ozone_Ozone_TransformedForecast_decomp_output.png') INFO:pyaf.std:SAVING_PLOT ('Forecast', 'outputs/my_ozone_Ozone_Forecast_decomp_output.png') INFO:pyaf.std:SAVING_PLOT ('PredictionIntervals', 'outputs/my_ozone_Ozone_prediction_intervals_output.png') INFO:pyaf.std:SAVING_PLOT ('Quantiles', 'outputs/my_ozone_Ozone_quantiles_output.png') INFO:pyaf.timing:('OPERATION_END_ELAPSED', 5.11, ('PLOTTING', {'Signals': ['Ozone']})) ```
non_process
final touch report plot filenames in the logs the call cforecastengine standardplots produces a lot of png files by default report these file names in the model logs some thing like info pyaf timing operation start plotting signals info pyaf std saving plot trend outputs my ozone ozone trend decomp output png info pyaf std saving plot cycle outputs my ozone ozone cycle decomp output png info pyaf std saving plot ar outputs my ozone ozone ar decomp output png info pyaf std saving plot transformedforecast outputs my ozone ozone transformedforecast decomp output png info pyaf std saving plot forecast outputs my ozone ozone forecast decomp output png info pyaf std saving plot predictionintervals outputs my ozone ozone prediction intervals output png info pyaf std saving plot quantiles outputs my ozone ozone quantiles output png info pyaf timing operation end elapsed plotting signals
0
281,851
30,888,974,379
IssuesEvent
2023-08-04 02:04:46
hshivhare67/kernel_v4.1.15_CVE-2019-10220
https://api.github.com/repos/hshivhare67/kernel_v4.1.15_CVE-2019-10220
reopened
CVE-2020-15436 (Medium) detected in linuxlinux-4.4.302
Mend: dependency security vulnerability
## CVE-2020-15436 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/block_dev.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Use-after-free vulnerability in fs/block_dev.c in the Linux kernel before 5.8 allows local users to gain privileges or cause a denial of service by leveraging improper access to a certain error field. Mend Note: After conducting further research, Mend has determined that versions v2.6.12-rc2 through v4.4.228, v4.5-rc1 through v4.9.228, v4.10-rc1 through v4.14.185, v4.15-rc1 through v4.19.129, v5.0-rc1 through v5.4.48, v5.5-rc1 through v5.7.5 and v5.8-rc1 of Linux Kernel are vulnerable to CVE-2020-15436. <p>Publish Date: 2020-11-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-15436>CVE-2020-15436</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-15436">https://www.linuxkernelcves.com/cves/CVE-2020-15436</a></p> <p>Release Date: 2020-11-23</p> <p>Fix Resolution: v4.4.229, v4.9.229, v4.14.186, v4.19.130, v5.4.49, v5.7.6, v5.8-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15436 (Medium) detected in linuxlinux-4.4.302 - ## CVE-2020-15436 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/block_dev.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Use-after-free vulnerability in fs/block_dev.c in the Linux kernel before 5.8 allows local users to gain privileges or cause a denial of service by leveraging improper access to a certain error field. Mend Note: After conducting further research, Mend has determined that versions v2.6.12-rc2 through v4.4.228, v4.5-rc1 through v4.9.228, v4.10-rc1 through v4.14.185, v4.15-rc1 through v4.19.129, v5.0-rc1 through v5.4.48, v5.5-rc1 through v5.7.5 and v5.8-rc1 of Linux Kernel are vulnerable to CVE-2020-15436. <p>Publish Date: 2020-11-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-15436>CVE-2020-15436</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-15436">https://www.linuxkernelcves.com/cves/CVE-2020-15436</a></p> <p>Release Date: 2020-11-23</p> <p>Fix Resolution: v4.4.229, v4.9.229, v4.14.186, v4.19.130, v5.4.49, v5.7.6, v5.8-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files fs block dev c vulnerability details use after free vulnerability in fs block dev c in the linux kernel before allows local users to gain privileges or cause a denial of service by leveraging improper access to a certain error field mend note after conducting further research mend has determined that versions through through through through through through and of linux kernel are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
447,815
12,893,530,278
IssuesEvent
2020-07-13 21:50:51
DSpace/dspace-angular
https://api.github.com/repos/DSpace/dspace-angular
opened
DSpace (Solr) Statistics - Workflow Reports (same as v6.x)
Difficulty: High component: statistics medium priority
From release plan spreadsheet Estimate from release plan: 22 hours Expressing interest: none No additional notes
1.0
DSpace (Solr) Statistics - Workflow Reports (same as v6.x) - From release plan spreadsheet Estimate from release plan: 22 hours Expressing interest: none No additional notes
non_process
dspace solr statistics workflow reports same as x from release plan spreadsheet estimate from release plan hours expressing interest none no additional notes
0
11,204
13,957,703,691
IssuesEvent
2020-10-24 08:14:00
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
NL: link between dataset and Atom based service
Geoportal Harvesting process NL - The Netherlands
Dear Angelo, In this issue tracker i found following information: The link between a Dataset/Series and a View Service is established via the MetadataURL element inside the Layer element of the capabilities document of the OGC WMS service. The link between a Dataset/Series and a Download Service is established: for a WFS, via the SpatialDataSetIdentifier element inside Extended Capabilities element of the capabilities document of the OGC WFS service. for an Atom based service, via the link (rel=&quot;describedby&quot;) element inside the entry inside top feed Can you explain why the used elements for WFS are not the same as the used elements for Atom? The WFS also contains the link to the metadata of the dataset which is used for the Atom based service. The Atom also contains the SpatialDataSetIdentifier elements which is used for the WFS. Best regards Ine
1.0
NL: link between dataset and Atom based service - Dear Angelo, In this issue tracker i found following information: The link between a Dataset/Series and a View Service is established via the MetadataURL element inside the Layer element of the capabilities document of the OGC WMS service. The link between a Dataset/Series and a Download Service is established: for a WFS, via the SpatialDataSetIdentifier element inside Extended Capabilities element of the capabilities document of the OGC WFS service. for an Atom based service, via the link (rel=&quot;describedby&quot;) element inside the entry inside top feed Can you explain why the used elements for WFS are not the same as the used elements for Atom? The WFS also contains the link to the metadata of the dataset which is used for the Atom based service. The Atom also contains the SpatialDataSetIdentifier elements which is used for the WFS. Best regards Ine
process
nl link between dataset and atom based service dear angelo in this issue tracker i found following information the link between a dataset series and a view service is established via the metadataurl element inside the layer element of the capabilities document of the ogc wms service the link between a dataset series and a download service is established for a wfs via the spatialdatasetidentifier element inside extended capabilities element of the capabilities document of the ogc wfs service for an atom based service via the link rel quot describedby quot element inside the entry inside top feed can you explain why the used elements for wfs are not the same as the used elements for atom the wfs also contains the link to the metadata of the dataset which is used for the atom based service the atom also contains the spatialdatasetidentifier elements which is used for the wfs best regards ine
1
11,137
13,957,691,993
IssuesEvent
2020-10-24 08:10:34
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
DE - harvesting timeframe & filter
DE - Germany Geoportal Harvesting process
Dear Angelo, Could you please tell me, What is the current timeframe honoured by the INSPIRE Geoportal for Germany&#39;s Discovery Service? Which OGC Filter is used by the INSPIRE Geoportal for Germany&#39;s Discovery Service? Related to my first question: on the INSPIRE Service Register page (http://inspire-geoportal.ec.europa.eu/INSPIRERegistry/) I can see, that the Pull Operation Frequency is set to &quot;monthly&quot;. So the following timeframe isn&#39;t longer valid, right? &lt;ns2:SchedulerDetails&gt; &lt;ns2:RecachingFrequency&gt;weekly&lt;/ns2:RecachingFrequency&gt; &lt;ns2:RecachingStartTimeFrame&gt; &lt;ns2:WeekDay&gt;monday&lt;/ns2:WeekDay&gt; &lt;ns2:WeekDay&gt;tuesday&lt;/ns2:WeekDay&gt; &lt;ns2:WeekDay&gt;thursday&lt;/ns2:WeekDay&gt; &lt;ns2:WeekDay&gt;friday&lt;/ns2:WeekDay&gt; &lt;ns2:WeekDay&gt;saturday&lt;/ns2:WeekDay&gt; &lt;ns2:TimeWindow&gt; &lt;ns2:AfterSpecificTime&gt;03:00:00&lt;/ns2:AfterSpecificTime&gt; &lt;ns2:BeforeSpecificTime&gt;22:00:00&lt;/ns2:BeforeSpecificTime&gt; &lt;/ns2:TimeWindow&gt; &lt;/ns2:RecachingStartTimeFrame&gt; &lt;ns2:MaxHoursToCompletePullOperation&gt;72&lt;/ns2:MaxHoursToCompletePullOperation&gt; &lt;/ns2:SchedulerDetails&gt; Related to my second question: I updated the Discovery Filter Text on the INSPIRE Service Register page (http://inspire-geoportal.ec.europa.eu/INSPIRERegistry/) to &lt;ogc:Filter xmlns:ogc=&quot;http://www.opengis.net/ogc&quot;&gt; &lt;ogc:PropertyIsEqualTo&gt; &lt;ogc:PropertyName&gt;apiso:subject&lt;/ogc:PropertyName&gt; &lt;ogc:Literal&gt;inspireidentifiziert&lt;/ogc:Literal&gt; &lt;/ogc:PropertyIsEqualTo&gt; &lt;/ogc:Filter&gt; But on this page (http://inspire-geoportal.ec.europa.eu/resources/INSPIREResourcesReports/resourcesReport_2019-03-20/) there is still the &quot;old version&quot; standing: Thanks in advance and best regards, Anja
1.0
DE - harvesting timeframe & filter - Dear Angelo, Could you please tell me, What is the current timeframe honoured by the INSPIRE Geoportal for Germany&#39;s Discovery Service? Which OGC Filter is used by the INSPIRE Geoportal for Germany&#39;s Discovery Service? Related to my first question: on the INSPIRE Service Register page (http://inspire-geoportal.ec.europa.eu/INSPIRERegistry/) I can see, that the Pull Operation Frequency is set to &quot;monthly&quot;. So the following timeframe isn&#39;t longer valid, right? &lt;ns2:SchedulerDetails&gt; &lt;ns2:RecachingFrequency&gt;weekly&lt;/ns2:RecachingFrequency&gt; &lt;ns2:RecachingStartTimeFrame&gt; &lt;ns2:WeekDay&gt;monday&lt;/ns2:WeekDay&gt; &lt;ns2:WeekDay&gt;tuesday&lt;/ns2:WeekDay&gt; &lt;ns2:WeekDay&gt;thursday&lt;/ns2:WeekDay&gt; &lt;ns2:WeekDay&gt;friday&lt;/ns2:WeekDay&gt; &lt;ns2:WeekDay&gt;saturday&lt;/ns2:WeekDay&gt; &lt;ns2:TimeWindow&gt; &lt;ns2:AfterSpecificTime&gt;03:00:00&lt;/ns2:AfterSpecificTime&gt; &lt;ns2:BeforeSpecificTime&gt;22:00:00&lt;/ns2:BeforeSpecificTime&gt; &lt;/ns2:TimeWindow&gt; &lt;/ns2:RecachingStartTimeFrame&gt; &lt;ns2:MaxHoursToCompletePullOperation&gt;72&lt;/ns2:MaxHoursToCompletePullOperation&gt; &lt;/ns2:SchedulerDetails&gt; Related to my second question: I updated the Discovery Filter Text on the INSPIRE Service Register page (http://inspire-geoportal.ec.europa.eu/INSPIRERegistry/) to &lt;ogc:Filter xmlns:ogc=&quot;http://www.opengis.net/ogc&quot;&gt; &lt;ogc:PropertyIsEqualTo&gt; &lt;ogc:PropertyName&gt;apiso:subject&lt;/ogc:PropertyName&gt; &lt;ogc:Literal&gt;inspireidentifiziert&lt;/ogc:Literal&gt; &lt;/ogc:PropertyIsEqualTo&gt; &lt;/ogc:Filter&gt; But on this page (http://inspire-geoportal.ec.europa.eu/resources/INSPIREResourcesReports/resourcesReport_2019-03-20/) there is still the &quot;old version&quot; standing: Thanks in advance and best regards, Anja
process
de harvesting timeframe filter dear angelo could you please tell me what is the current timeframe honoured by the inspire geoportal for germany s discovery service which ogc filter is used by the inspire geoportal for germany s discovery service related to my first question on the inspire service register page i can see that the pull operation frequency is set to quot monthly quot so the following timeframe isn t longer valid right lt schedulerdetails gt lt recachingfrequency gt weekly lt recachingfrequency gt lt recachingstarttimeframe gt lt weekday gt monday lt weekday gt lt weekday gt tuesday lt weekday gt lt weekday gt thursday lt weekday gt lt weekday gt friday lt weekday gt lt weekday gt saturday lt weekday gt lt timewindow gt lt afterspecifictime gt lt afterspecifictime gt lt beforespecifictime gt lt beforespecifictime gt lt timewindow gt lt recachingstarttimeframe gt lt maxhourstocompletepulloperation gt lt maxhourstocompletepulloperation gt lt schedulerdetails gt related to my second question i updated the discovery filter text on the inspire service register page to lt ogc filter xmlns ogc quot lt ogc propertyisequalto gt lt ogc propertyname gt apiso subject lt ogc propertyname gt lt ogc literal gt inspireidentifiziert lt ogc literal gt lt ogc propertyisequalto gt lt ogc filter gt but on this page there is still the quot old version quot standing thanks in advance and best regards anja
1
158,093
24,786,910,468
IssuesEvent
2022-10-24 10:33:48
baloise-incubator/design-system
https://api.github.com/repos/baloise-incubator/design-system
closed
Initial value does not get set on remote typeahead when multiple remote typeaheads are displayed
๐Ÿ› Bug ๐Ÿง‘โ€๐Ÿ’ป Technical Design System
The setting of a initial value of the remote typeaheads do not work when multiple remote typeaheads are displayed on one page. I provided an example branch with the name (feat/remote_typeahead_initialization_bug) in this [repository](https://github.com/baloise/design-system-examples). On initial run of the application sometimes it works. But when you do a reload of the page it does not work anymore. Hint: I did a small analysis and could be that this function is the reason for the issue: ``` updateInputValue(value) { if (this.updateInputValueTimer) { clearTimeout(this.updateInputValueTimer); } this.updateInputValueTimer = setTimeout(() => { if (!lodash_isnil(this.inputElement)) { this.inputElement.value = value; this.inputValue = value; } }, 0); } ``` since it uses timeoutes maybe it is a timing issue.
1.0
Initial value does not get set on remote typeahead when multiple remote typeaheads are displayed - The setting of a initial value of the remote typeaheads do not work when multiple remote typeaheads are displayed on one page. I provided an example branch with the name (feat/remote_typeahead_initialization_bug) in this [repository](https://github.com/baloise/design-system-examples). On initial run of the application sometimes it works. But when you do a reload of the page it does not work anymore. Hint: I did a small analysis and could be that this function is the reason for the issue: ``` updateInputValue(value) { if (this.updateInputValueTimer) { clearTimeout(this.updateInputValueTimer); } this.updateInputValueTimer = setTimeout(() => { if (!lodash_isnil(this.inputElement)) { this.inputElement.value = value; this.inputValue = value; } }, 0); } ``` since it uses timeoutes maybe it is a timing issue.
non_process
initial value does not get set on remote typeahead when multiple remote typeaheads are displayed the setting of a initial value of the remote typeaheads do not work when multiple remote typeaheads are displayed on one page i provided an example branch with the name feat remote typeahead initialization bug in this on initial run of the application sometimes it works but when you do a reload of the page it does not work anymore hint i did a small analysis and could be that this function is the reason for the issue updateinputvalue value if this updateinputvaluetimer cleartimeout this updateinputvaluetimer this updateinputvaluetimer settimeout if lodash isnil this inputelement this inputelement value value this inputvalue value since it uses timeoutes maybe it is a timing issue
0
11,793
14,619,868,790
IssuesEvent
2020-12-22 18:38:32
ewen-lbh/portfolio
https://api.github.com/repos/ewen-lbh/portfolio
closed
Implement `hide title`
processing styling
Use this to hide the <h1> (with sth that makes it so screen readers stil see it) when ie the first media is a wordmark and putting the h1 _then_ the logo is repetitve and ugly (see humanr)
1.0
Implement `hide title` - Use this to hide the <h1> (with sth that makes it so screen readers stil see it) when ie the first media is a wordmark and putting the h1 _then_ the logo is repetitve and ugly (see humanr)
process
implement hide title use this to hide the with sth that makes it so screen readers stil see it when ie the first media is a wordmark and putting the then the logo is repetitve and ugly see humanr
1
464,267
13,309,237,811
IssuesEvent
2020-08-26 03:23:26
creativecommons/vocabulary
https://api.github.com/repos/creativecommons/vocabulary
closed
Implement Input variations from Design Language
enhancement good first issue help wanted improvement:enhancement priority:moderate project:cccatalog-frontend
The Design Language specifies a number of input variations that we are missing: - [x] Inputs with labels: https://www.figma.com/file/l4Mt3dn3Ndtrvrb4aLcwXI/Design-Library?node-id=923%3A0 (fixed via: #532) - [x] Inputs with names and descriptions: https://www.figma.com/file/l4Mt3dn3Ndtrvrb4aLcwXI/Design-Library?node-id=7177%3A29 (fixed via: #522) - [ ] Inputs with icons: https://www.figma.com/file/l4Mt3dn3Ndtrvrb4aLcwXI/Design-Library?node-id=924%3A1626 - This issue has a [related bug](https://github.com/creativecommons/fonts/issues/110) These should be implemented in individual PRs referencing this issue.
1.0
Implement Input variations from Design Language - The Design Language specifies a number of input variations that we are missing: - [x] Inputs with labels: https://www.figma.com/file/l4Mt3dn3Ndtrvrb4aLcwXI/Design-Library?node-id=923%3A0 (fixed via: #532) - [x] Inputs with names and descriptions: https://www.figma.com/file/l4Mt3dn3Ndtrvrb4aLcwXI/Design-Library?node-id=7177%3A29 (fixed via: #522) - [ ] Inputs with icons: https://www.figma.com/file/l4Mt3dn3Ndtrvrb4aLcwXI/Design-Library?node-id=924%3A1626 - This issue has a [related bug](https://github.com/creativecommons/fonts/issues/110) These should be implemented in individual PRs referencing this issue.
non_process
implement input variations from design language the design language specifies a number of input variations that we are missing inputs with labels fixed via inputs with names and descriptions fixed via inputs with icons this issue has a these should be implemented in individual prs referencing this issue
0
19,746
26,107,933,355
IssuesEvent
2022-12-27 15:32:48
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
logstransformprocessor deadlocks under load
bug processor/logstransform
### Component(s) processor/logstransform ### What happened? ## Description There is a deadlock scenario caused in the logs transform processor under load. The primary issue is that every invocation of `logstransformprocessor.processLogs` shares the same stanza pipeline and a common instance of the [emitter](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/main/pkg/stanza/adapter/emitter.go). Each invocation pushes log messages into the processing channel pipeline (blocking). ```go // Add the logs to the chain err := ltp.fromConverter.Batch(ld) if err != nil { return ld, err } func (c *FromPdataConverter) Batch(pLogs plog.Logs) error { for i := 0; i < pLogs.ResourceLogs().Len(); i++ { rls := pLogs.ResourceLogs().At(i) for j := 0; j < rls.ScopeLogs().Len(); j++ { scope := rls.ScopeLogs().At(j) item := fromConverterWorkerItem{ Resource: rls.Resource(), Scope: scope, LogRecordSlice: scope.LogRecords(), } select { case c.workerChan <- item: continue case <-c.stopChan: return nil } } } return nil } ``` Then waits on an output channel that is at the tail end of that pipeline. ```go for { select { case <-doneChan: ltp.logger.Debug("loop stopped") return ld, errors.New("processor interrupted") case output, ok := <-ltp.outputChannel: if !ok { return ld, errors.New("processor encountered an issue receiving logs from stanza operators pipeline") } if output.err != nil { return ld, err } return output.logs, nil } } ``` However, because of the implementation of the shared emitter, each invocation to processLogs isn't necessarily going to receive the logs that came through it's pipeline. As each log message being sent through the stanza pipeline is processed, it gets pushed into a batch in LogEmitter. Logs pushed to the outputChannel to be processed only occur if at least 100 (max batch size) messages were pushed in or the timeout occurs. ```go // emitter.go // flusher flushes the current batch every flush interval. Intended to be run as a goroutine func (e *LogEmitter) flusher(ctx context.Context) { defer e.wg.Done() ticker := time.NewTicker(e.flushInterval) defer ticker.Stop() for { select { case <-ticker.C: if oldBatch := e.makeNewBatch(); len(oldBatch) > 0 { e.flush(ctx, oldBatch) } case <-ctx.Done(): return } } } ``` It can cause a situation where multiple concurrent invocations push messages to be processed, but only 1 invocation receives the batch to select from. Each additional invocation is blocked waiting to select from the outputChannel that has nothing in it. ``` // All concurrent processLogs pushes 25 messages to stanza pipeline -> emitter puts into batch -> batch has 100 items -> ltp selects from output and returns processLogs pushes 25 messages to stanza pipeline -> emitter puts into batch -> batch has 0 items -> no op -> ltp blocks on select from output processLogs pushes 25 messages to stanza pipeline -> emitter puts into batch -> batch has 0 items -> no op -> ltp blocks on select from output processLogs pushes 25 messages to stanza pipeline -> emitter puts into batch -> batch has 0 items -> no op -> ltp blocks on select from output ``` The bottom 3 invocations are blocked because there are no more logs being emitted. If a new invocation comes in, only one of the blocked routines will become unblocked. If the upstream receivers of those pipelines do not timeout to cancel their context, it effectively creates a deadlock. ``` โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ processLogs() โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ” โ”‚ โ”‚ โ”‚ processLogs() โ”œโ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ” โ”‚ โ””โ”€โ”‚ โ”‚ processLogs() โ”‚โ”€โ”€โ”€โ” โ””โ”€โ”ค Pushes messages to be processed on a stanza โ”‚ โ”‚ โ”‚ pipeline where emitter is the last stanza in โ”‚ โ”‚ โ”‚ the pipeline. Then waits to receive batched โ”‚ โ”‚ โ”‚ logs on a output chan that receives the โ”‚ โ”‚ โ”‚ batches from the emitter. โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚ emitter.flush() โ—„โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ emits at most 100 messages per โ”‚ โ”‚ process logs invocation. โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ ``` ## Steps to Reproduce I have example test cases here where you can see the effect of this. https://github.com/open-telemetry/opentelemetry-collector-contrib/commit/7fb27bba2a73974b55c82d4bd3140d3be33c7187 ## Expected Result Each batch of log messages pushed into logstransform should be fully processed by processLogs. ## Actual Result The logs output by each call to processLogs are somewhat arbitrary because emitted batches could contain a combination of logs from other asynchronous invocations. ### Collector version v0.66.0 ### Environment information ## Environment OS: (e.g., "Ubuntu 20.04") Compiler(if manually compiled): (e.g., "go 14.2") ### OpenTelemetry Collector configuration _No response_ ### Log output ```shell 3 @ 0x43c9b6 0x44c83c 0xc8a5d4 0xc42df8 0x9f0139 0xc8e504 0x46e021 # 0xc8a5d3 github.com/open-telemetry/opentelemetry-collector-contrib/processor/logstransformprocessor.(*logsTransformProcessor).processLogs+0xd3 /obfuscated/opentelemetry-collector-contrib/processor/logstransformprocessor/processor.go:129 # 0xc42df7 go.opentelemetry.io/collector/processor/processorhelper.NewLogsProcessor.func1+0xf7 /obfuscated/go/pkg/mod/go.opentelemetry.io/collector@v0.66.1-0.20221202005155-1c54042beb70/processor/processorhelper/logs.go:62 # 0x9f0138 go.opentelemetry.io/collector/consumer.ConsumeLogsFunc.ConsumeLogs+0x38 /obfuscated/go/pkg/mod/go.opentelemetry.io/collector/consumer@v0.66.1-0.20221202005155-1c54042beb70/logs.go:36 # 0xc8e503 github.com/open-telemetry/opentelemetry-collector-contrib/processor/logstransformprocessor.TestLogsTransformProcessor_Deadlock.func1+0x83 /obfuscated/source/oss/opentelemetry-collector-contrib/processor/logstransformprocessor/processor_test.go:185 ``` ### Additional context _No response_
1.0
logstransformprocessor deadlocks under load - ### Component(s) processor/logstransform ### What happened? ## Description There is a deadlock scenario caused in the logs transform processor under load. The primary issue is that every invocation of `logstransformprocessor.processLogs` shares the same stanza pipeline and a common instance of the [emitter](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/main/pkg/stanza/adapter/emitter.go). Each invocation pushes log messages into the processing channel pipeline (blocking). ```go // Add the logs to the chain err := ltp.fromConverter.Batch(ld) if err != nil { return ld, err } func (c *FromPdataConverter) Batch(pLogs plog.Logs) error { for i := 0; i < pLogs.ResourceLogs().Len(); i++ { rls := pLogs.ResourceLogs().At(i) for j := 0; j < rls.ScopeLogs().Len(); j++ { scope := rls.ScopeLogs().At(j) item := fromConverterWorkerItem{ Resource: rls.Resource(), Scope: scope, LogRecordSlice: scope.LogRecords(), } select { case c.workerChan <- item: continue case <-c.stopChan: return nil } } } return nil } ``` Then waits on an output channel that is at the tail end of that pipeline. ```go for { select { case <-doneChan: ltp.logger.Debug("loop stopped") return ld, errors.New("processor interrupted") case output, ok := <-ltp.outputChannel: if !ok { return ld, errors.New("processor encountered an issue receiving logs from stanza operators pipeline") } if output.err != nil { return ld, err } return output.logs, nil } } ``` However, because of the implementation of the shared emitter, each invocation to processLogs isn't necessarily going to receive the logs that came through it's pipeline. As each log message being sent through the stanza pipeline is processed, it gets pushed into a batch in LogEmitter. Logs pushed to the outputChannel to be processed only occur if at least 100 (max batch size) messages were pushed in or the timeout occurs. ```go // emitter.go // flusher flushes the current batch every flush interval. Intended to be run as a goroutine func (e *LogEmitter) flusher(ctx context.Context) { defer e.wg.Done() ticker := time.NewTicker(e.flushInterval) defer ticker.Stop() for { select { case <-ticker.C: if oldBatch := e.makeNewBatch(); len(oldBatch) > 0 { e.flush(ctx, oldBatch) } case <-ctx.Done(): return } } } ``` It can cause a situation where multiple concurrent invocations push messages to be processed, but only 1 invocation receives the batch to select from. Each additional invocation is blocked waiting to select from the outputChannel that has nothing in it. ``` // All concurrent processLogs pushes 25 messages to stanza pipeline -> emitter puts into batch -> batch has 100 items -> ltp selects from output and returns processLogs pushes 25 messages to stanza pipeline -> emitter puts into batch -> batch has 0 items -> no op -> ltp blocks on select from output processLogs pushes 25 messages to stanza pipeline -> emitter puts into batch -> batch has 0 items -> no op -> ltp blocks on select from output processLogs pushes 25 messages to stanza pipeline -> emitter puts into batch -> batch has 0 items -> no op -> ltp blocks on select from output ``` The bottom 3 invocations are blocked because there are no more logs being emitted. If a new invocation comes in, only one of the blocked routines will become unblocked. If the upstream receivers of those pipelines do not timeout to cancel their context, it effectively creates a deadlock. ``` โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ processLogs() โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ” โ”‚ โ”‚ โ”‚ processLogs() โ”œโ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ” โ”‚ โ””โ”€โ”‚ โ”‚ processLogs() โ”‚โ”€โ”€โ”€โ” โ””โ”€โ”ค Pushes messages to be processed on a stanza โ”‚ โ”‚ โ”‚ pipeline where emitter is the last stanza in โ”‚ โ”‚ โ”‚ the pipeline. Then waits to receive batched โ”‚ โ”‚ โ”‚ logs on a output chan that receives the โ”‚ โ”‚ โ”‚ batches from the emitter. โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚ emitter.flush() โ—„โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ emits at most 100 messages per โ”‚ โ”‚ process logs invocation. โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ ``` ## Steps to Reproduce I have example test cases here where you can see the effect of this. https://github.com/open-telemetry/opentelemetry-collector-contrib/commit/7fb27bba2a73974b55c82d4bd3140d3be33c7187 ## Expected Result Each batch of log messages pushed into logstransform should be fully processed by processLogs. ## Actual Result The logs output by each call to processLogs are somewhat arbitrary because emitted batches could contain a combination of logs from other asynchronous invocations. ### Collector version v0.66.0 ### Environment information ## Environment OS: (e.g., "Ubuntu 20.04") Compiler(if manually compiled): (e.g., "go 14.2") ### OpenTelemetry Collector configuration _No response_ ### Log output ```shell 3 @ 0x43c9b6 0x44c83c 0xc8a5d4 0xc42df8 0x9f0139 0xc8e504 0x46e021 # 0xc8a5d3 github.com/open-telemetry/opentelemetry-collector-contrib/processor/logstransformprocessor.(*logsTransformProcessor).processLogs+0xd3 /obfuscated/opentelemetry-collector-contrib/processor/logstransformprocessor/processor.go:129 # 0xc42df7 go.opentelemetry.io/collector/processor/processorhelper.NewLogsProcessor.func1+0xf7 /obfuscated/go/pkg/mod/go.opentelemetry.io/collector@v0.66.1-0.20221202005155-1c54042beb70/processor/processorhelper/logs.go:62 # 0x9f0138 go.opentelemetry.io/collector/consumer.ConsumeLogsFunc.ConsumeLogs+0x38 /obfuscated/go/pkg/mod/go.opentelemetry.io/collector/consumer@v0.66.1-0.20221202005155-1c54042beb70/logs.go:36 # 0xc8e503 github.com/open-telemetry/opentelemetry-collector-contrib/processor/logstransformprocessor.TestLogsTransformProcessor_Deadlock.func1+0x83 /obfuscated/source/oss/opentelemetry-collector-contrib/processor/logstransformprocessor/processor_test.go:185 ``` ### Additional context _No response_
process
logstransformprocessor deadlocks under load component s processor logstransform what happened description there is a deadlock scenario caused in the logs transform processor under load the primary issue is that every invocation of logstransformprocessor processlogs shares the same stanza pipeline and a common instance of the each invocation pushes log messages into the processing channel pipeline blocking go add the logs to the chain err ltp fromconverter batch ld if err nil return ld err func c frompdataconverter batch plogs plog logs error for i i plogs resourcelogs len i rls plogs resourcelogs at i for j j rls scopelogs len j scope rls scopelogs at j item fromconverterworkeritem resource rls resource scope scope logrecordslice scope logrecords select case c workerchan item continue case c stopchan return nil return nil then waits on an output channel that is at the tail end of that pipeline go for select case donechan ltp logger debug loop stopped return ld errors new processor interrupted case output ok ltp outputchannel if ok return ld errors new processor encountered an issue receiving logs from stanza operators pipeline if output err nil return ld err return output logs nil however because of the implementation of the shared emitter each invocation to processlogs isn t necessarily going to receive the logs that came through it s pipeline as each log message being sent through the stanza pipeline is processed it gets pushed into a batch in logemitter logs pushed to the outputchannel to be processed only occur if at least max batch size messages were pushed in or the timeout occurs go emitter go flusher flushes the current batch every flush interval intended to be run as a goroutine func e logemitter flusher ctx context context defer e wg done ticker time newticker e flushinterval defer ticker stop for select case ticker c if oldbatch e makenewbatch len oldbatch e flush ctx oldbatch case ctx done return it can cause a situation where multiple concurrent invocations push messages to be processed but only invocation receives the batch to select from each additional invocation is blocked waiting to select from the outputchannel that has nothing in it all concurrent processlogs pushes messages to stanza pipeline emitter puts into batch batch has items ltp selects from output and returns processlogs pushes messages to stanza pipeline emitter puts into batch batch has items no op ltp blocks on select from output processlogs pushes messages to stanza pipeline emitter puts into batch batch has items no op ltp blocks on select from output processlogs pushes messages to stanza pipeline emitter puts into batch batch has items no op ltp blocks on select from output the bottom invocations are blocked because there are no more logs being emitted if a new invocation comes in only one of the blocked routines will become unblocked if the upstream receivers of those pipelines do not timeout to cancel their context it effectively creates a deadlock โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ processlogs โ”œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ” โ”‚ โ”‚ โ”‚ processlogs โ”œโ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”ดโ”€โ” โ”‚ โ””โ”€โ”‚ โ”‚ processlogs โ”‚โ”€โ”€โ”€โ” โ””โ”€โ”ค pushes messages to be processed on a stanza โ”‚ โ”‚ โ”‚ pipeline where emitter is the last stanza in โ”‚ โ”‚ โ”‚ the pipeline then waits to receive batched โ”‚ โ”‚ โ”‚ logs on a output chan that receives the โ”‚ โ”‚ โ”‚ batches from the emitter โ”‚ โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ” โ”‚ โ”‚ emitter flush โ—„โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ โ”‚ emits at most messages per โ”‚ โ”‚ process logs invocation โ”‚ โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜ steps to reproduce i have example test cases here where you can see the effect of this expected result each batch of log messages pushed into logstransform should be fully processed by processlogs actual result the logs output by each call to processlogs are somewhat arbitrary because emitted batches could contain a combination of logs from other asynchronous invocations collector version environment information environment os e g ubuntu compiler if manually compiled e g go opentelemetry collector configuration no response log output shell github com open telemetry opentelemetry collector contrib processor logstransformprocessor logstransformprocessor processlogs obfuscated opentelemetry collector contrib processor logstransformprocessor processor go go opentelemetry io collector processor processorhelper newlogsprocessor obfuscated go pkg mod go opentelemetry io collector processor processorhelper logs go go opentelemetry io collector consumer consumelogsfunc consumelogs obfuscated go pkg mod go opentelemetry io collector consumer logs go github com open telemetry opentelemetry collector contrib processor logstransformprocessor testlogstransformprocessor deadlock obfuscated source oss opentelemetry collector contrib processor logstransformprocessor processor test go additional context no response
1
114,541
4,635,799,496
IssuesEvent
2016-09-29 08:35:49
dhis2/maintenance-app
https://api.github.com/repos/dhis2/maintenance-app
closed
Add aggregate export fields to data dimensional objects
enhancement priority:high
Add the following fields: Fields --- JSON field | UI label | UI type --- | --- | --- aggregateExportCategoryOptionCombo | Category option combination for aggregate data export | Input field aggregateExportAttributeOptionCombo | Attribute option combination for aggregate data export | Input field The fields should be added to the add / edit object screens for the following objects: Objects --- - Data element - Indicator - Data set
1.0
Add aggregate export fields to data dimensional objects - Add the following fields: Fields --- JSON field | UI label | UI type --- | --- | --- aggregateExportCategoryOptionCombo | Category option combination for aggregate data export | Input field aggregateExportAttributeOptionCombo | Attribute option combination for aggregate data export | Input field The fields should be added to the add / edit object screens for the following objects: Objects --- - Data element - Indicator - Data set
non_process
add aggregate export fields to data dimensional objects add the following fields fields json field ui label ui type aggregateexportcategoryoptioncombo category option combination for aggregate data export input field aggregateexportattributeoptioncombo attribute option combination for aggregate data export input field the fields should be added to the add edit object screens for the following objects objects data element indicator data set
0
22,141
30,684,380,957
IssuesEvent
2023-07-26 11:19:32
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Number of Source Control integrations supported
automation/svc triaged assigned-to-author doc-enhancement process-automation/subsvc Pri2
Hi, I didnยดt find any information about the number of source controls which can be integrated for a a single automation account. Can this be added please? --- #### Document Details โš  *Do not edit this section. It is required for docs.microsoft.com โžŸ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration) * Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SGSneha * Microsoft Alias: **v-ssudhir**
1.0
Number of Source Control integrations supported - Hi, I didnยดt find any information about the number of source controls which can be integrated for a a single automation account. Can this be added please? --- #### Document Details โš  *Do not edit this section. It is required for docs.microsoft.com โžŸ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration) * Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SGSneha * Microsoft Alias: **v-ssudhir**
process
number of source control integrations supported hi i didnยดt find any information about the number of source controls which can be integrated for a a single automation account can this be added please document details โš  do not edit this section it is required for docs microsoft com โžŸ github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias v ssudhir
1
9,096
12,166,992,136
IssuesEvent
2020-04-27 10:08:45
pwittchen/ReactiveSensors
https://api.github.com/repos/pwittchen/ReactiveSensors
closed
Release 0.4.2 [rx2, rx3]
release process
Release notes: - made getters of ReactiveSensorEvent consistent: `getAccuracy()` -> `accuracy()`
1.0
Release 0.4.2 [rx2, rx3] - Release notes: - made getters of ReactiveSensorEvent consistent: `getAccuracy()` -> `accuracy()`
process
release release notes made getters of reactivesensorevent consistent getaccuracy accuracy
1
18,592
24,569,637,902
IssuesEvent
2022-10-13 07:36:36
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Edit admin details screen > Deactivated admins > UI issue
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
Edit admin details screen > Deactivated admins > UI issue 1. there should be a little gap between the title and placeholder 2. Place holder text should be center aligned to the grayed out area 3. Disabled icon should get displayed when the admin mouse hovered on the dropdown ![image](https://user-images.githubusercontent.com/71445210/191196037-44fdaf44-c376-4125-8f14-61b91913a3ac.png)
3.0
[PM] Edit admin details screen > Deactivated admins > UI issue - Edit admin details screen > Deactivated admins > UI issue 1. there should be a little gap between the title and placeholder 2. Place holder text should be center aligned to the grayed out area 3. Disabled icon should get displayed when the admin mouse hovered on the dropdown ![image](https://user-images.githubusercontent.com/71445210/191196037-44fdaf44-c376-4125-8f14-61b91913a3ac.png)
process
edit admin details screen deactivated admins ui issue edit admin details screen deactivated admins ui issue there should be a little gap between the title and placeholder place holder text should be center aligned to the grayed out area disabled icon should get displayed when the admin mouse hovered on the dropdown
1
2,974
5,963,412,005
IssuesEvent
2017-05-30 04:51:49
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
generated links to topics under a chunked to-content topic point to wrong id
bug preprocess/chunking
This issue is related to the HTML5 transform. When linking to a child of a chunked="to-content" topic the resulting href does not correspond to the `@id ` of the target in the output: `<xref href="child-topic-2.dita#child-topic-2"/>` where _child-topic-2.dita_ is under a chunked topic, outputs as `<a class="xref" href="head-topic-chunked.html#child-topic-2">`. Whereas _child-topic-2.dita_ outputs as `<article class="topic nested1" aria-labelledby="ariaid-title3" id="head-topic-chunked__child-topic-2">`. Note the original `@id` concatenated with the chunked topic `@id`. The same issue is true with a keyref instead of an href. See samples for more info. I used DITA-OT 2.4.6 and develop produce the output. [DITA.zip](https://github.com/dita-ot/dita-ot/files/1024668/DITA.zip) [HTML5.zip](https://github.com/dita-ot/dita-ot/files/1024665/HTML5.zip) [HTML5-develop.zip](https://github.com/dita-ot/dita-ot/files/1024686/HTML5-develop.zip)
1.0
generated links to topics under a chunked to-content topic point to wrong id - This issue is related to the HTML5 transform. When linking to a child of a chunked="to-content" topic the resulting href does not correspond to the `@id ` of the target in the output: `<xref href="child-topic-2.dita#child-topic-2"/>` where _child-topic-2.dita_ is under a chunked topic, outputs as `<a class="xref" href="head-topic-chunked.html#child-topic-2">`. Whereas _child-topic-2.dita_ outputs as `<article class="topic nested1" aria-labelledby="ariaid-title3" id="head-topic-chunked__child-topic-2">`. Note the original `@id` concatenated with the chunked topic `@id`. The same issue is true with a keyref instead of an href. See samples for more info. I used DITA-OT 2.4.6 and develop produce the output. [DITA.zip](https://github.com/dita-ot/dita-ot/files/1024668/DITA.zip) [HTML5.zip](https://github.com/dita-ot/dita-ot/files/1024665/HTML5.zip) [HTML5-develop.zip](https://github.com/dita-ot/dita-ot/files/1024686/HTML5-develop.zip)
process
generated links to topics under a chunked to content topic point to wrong id this issue is related to the transform when linking to a child of a chunked to content topic the resulting href does not correspond to the id of the target in the output where child topic dita is under a chunked topic outputs as whereas child topic dita outputs as note the original id concatenated with the chunked topic id the same issue is true with a keyref instead of an href see samples for more info i used dita ot and develop produce the output
1
161,721
20,155,308,355
IssuesEvent
2022-02-09 15:57:21
kapseliboi/Node-Data
https://api.github.com/repos/kapseliboi/Node-Data
opened
CVE-2019-20920 (High) detected in handlebars-4.0.12.tgz
security vulnerability
## CVE-2019-20920 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.12.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.12.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.12.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - istanbul-0.4.2.tgz (Root Library) - :x: **handlebars-4.0.12.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/Node-Data/commit/289c77565fc637d4c0e4bf4a9a1e81df96cd190a">289c77565fc637d4c0e4bf4a9a1e81df96cd190a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Handlebars before 3.0.8 and 4.x before 4.5.3 is vulnerable to Arbitrary Code Execution. The lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript. This can be used to run arbitrary code on a server processing Handlebars templates or in a victim's browser (effectively serving as XSS). <p>Publish Date: 2020-09-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20920>CVE-2019-20920</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p> <p>Release Date: 2020-10-15</p> <p>Fix Resolution (handlebars): 4.5.3</p> <p>Direct dependency fix Resolution (istanbul): 0.4.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-20920 (High) detected in handlebars-4.0.12.tgz - ## CVE-2019-20920 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.12.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.12.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.12.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - istanbul-0.4.2.tgz (Root Library) - :x: **handlebars-4.0.12.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/Node-Data/commit/289c77565fc637d4c0e4bf4a9a1e81df96cd190a">289c77565fc637d4c0e4bf4a9a1e81df96cd190a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Handlebars before 3.0.8 and 4.x before 4.5.3 is vulnerable to Arbitrary Code Execution. The lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript. This can be used to run arbitrary code on a server processing Handlebars templates or in a victim's browser (effectively serving as XSS). <p>Publish Date: 2020-09-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20920>CVE-2019-20920</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p> <p>Release Date: 2020-10-15</p> <p>Fix Resolution (handlebars): 4.5.3</p> <p>Direct dependency fix Resolution (istanbul): 0.4.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in handlebars tgz cve high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file package json path to vulnerable library node modules handlebars package json dependency hierarchy istanbul tgz root library x handlebars tgz vulnerable library found in head commit a href found in base branch master vulnerability details handlebars before and x before is vulnerable to arbitrary code execution the lookup helper fails to properly validate templates allowing attackers to submit templates that execute arbitrary javascript this can be used to run arbitrary code on a server processing handlebars templates or in a victim s browser effectively serving as xss publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars direct dependency fix resolution istanbul step up your open source security game with whitesource
0
21,951
30,452,306,170
IssuesEvent
2023-07-16 13:00:18
tc3oliver/wiki
https://api.github.com/repos/tc3oliver/wiki
opened
โฌ…๏ธŽ
Gitalk [02]Natural Language Processing
https://wiki.meowcoder.com/#/wiki/[03]%E4%BA%BA%E5%B7%A5%E6%99%BA%E6%85%A7/[02]Natural%20Language%20Processing/ ้€™่ฃกๆ˜ฏๆˆ‘่จ˜้Œ„่‡ชๅทฑๅญธ็ฟ’็š„ๅœฐๆ–น๏ผŒไนŸๆ˜ฏ่ˆ‡ๅคงๅฎถๅˆ†ไบซๅฟƒๅพ—็š„ๅนณๅฐใ€‚ๅœจ้€™่ฃก๏ผŒๆ‚จๅฏไปฅๆ‰พๅˆฐๆˆ‘ๅฐๅ„็จฎไธป้กŒ็š„ๅญธ็ฟ’็ญ†่จ˜๏ผŒๅŒ…ๆ‹ฌๆŠ€่ก“ใ€่จญ่จˆใ€ๅ‰ตๆ„ๅ’Œ็”Ÿๆดป็ญ‰ๆ–น้ขใ€‚
1.0
โฌ…๏ธŽ - https://wiki.meowcoder.com/#/wiki/[03]%E4%BA%BA%E5%B7%A5%E6%99%BA%E6%85%A7/[02]Natural%20Language%20Processing/ ้€™่ฃกๆ˜ฏๆˆ‘่จ˜้Œ„่‡ชๅทฑๅญธ็ฟ’็š„ๅœฐๆ–น๏ผŒไนŸๆ˜ฏ่ˆ‡ๅคงๅฎถๅˆ†ไบซๅฟƒๅพ—็š„ๅนณๅฐใ€‚ๅœจ้€™่ฃก๏ผŒๆ‚จๅฏไปฅๆ‰พๅˆฐๆˆ‘ๅฐๅ„็จฎไธป้กŒ็š„ๅญธ็ฟ’็ญ†่จ˜๏ผŒๅŒ…ๆ‹ฌๆŠ€่ก“ใ€่จญ่จˆใ€ๅ‰ตๆ„ๅ’Œ็”Ÿๆดป็ญ‰ๆ–น้ขใ€‚
process
โฌ…๏ธŽ ba ba ba natural ้€™่ฃกๆ˜ฏๆˆ‘่จ˜้Œ„่‡ชๅทฑๅญธ็ฟ’็š„ๅœฐๆ–น๏ผŒไนŸๆ˜ฏ่ˆ‡ๅคงๅฎถๅˆ†ไบซๅฟƒๅพ—็š„ๅนณๅฐใ€‚ๅœจ้€™่ฃก๏ผŒๆ‚จๅฏไปฅๆ‰พๅˆฐๆˆ‘ๅฐๅ„็จฎไธป้กŒ็š„ๅญธ็ฟ’็ญ†่จ˜๏ผŒๅŒ…ๆ‹ฌๆŠ€่ก“ใ€่จญ่จˆใ€ๅ‰ตๆ„ๅ’Œ็”Ÿๆดป็ญ‰ๆ–น้ขใ€‚
1
21,661
30,107,980,761
IssuesEvent
2023-06-30 04:21:33
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
[Mirror] zig 0.11.0-dev.3886+0c1bfe271
P2 type: process team-OSS mirror request
### Please list the URLs of the archives you'd like to mirror: https://ziglang.org/builds/zig-windows-x86_64-0.11.0-dev.3886+0c1bfe271.zip https://ziglang.org/builds/zig-windows-x86-0.11.0-dev.3886+0c1bfe271.zip https://ziglang.org/builds/zig-windows-aarch64-0.11.0-dev.3886+0c1bfe271.zip https://ziglang.org/builds/zig-macos-aarch64-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-macos-x86_64-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-x86_64-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-x86-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-aarch64-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-riscv64-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-powerpc64le-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-powerpc-0.11.0-dev.3886+0c1bfe271.tar.xz
1.0
[Mirror] zig 0.11.0-dev.3886+0c1bfe271 - ### Please list the URLs of the archives you'd like to mirror: https://ziglang.org/builds/zig-windows-x86_64-0.11.0-dev.3886+0c1bfe271.zip https://ziglang.org/builds/zig-windows-x86-0.11.0-dev.3886+0c1bfe271.zip https://ziglang.org/builds/zig-windows-aarch64-0.11.0-dev.3886+0c1bfe271.zip https://ziglang.org/builds/zig-macos-aarch64-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-macos-x86_64-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-x86_64-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-x86-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-aarch64-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-riscv64-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-powerpc64le-0.11.0-dev.3886+0c1bfe271.tar.xz https://ziglang.org/builds/zig-linux-powerpc-0.11.0-dev.3886+0c1bfe271.tar.xz
process
zig dev please list the urls of the archives you d like to mirror
1
67,412
27,835,073,512
IssuesEvent
2023-03-20 08:55:11
hashicorp/terraform-provider-azurerm
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
closed
Application Gateway WAF policy doesn't support operator "any"
enhancement service/application-gateway
### Is there an existing issue for this? - [X] I have searched the existing issues ### Community Note Hi Team, I am trying to create the custom rule in application Gateway WAF policy in the azure portal operator "any " is exist but terraform doesn't support and there is no possible value to support "any" operator * Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community [--->](url) ### Terraform Version 0.13 ### AzureRM Provider Version 3.9.0 ### Affected Resource(s)/Data Source(s) azurerm_web_application_firewall_policy ### Terraform Configuration Files ```hcl waf_custom_rules = [ { name = XXXXXX priority = 1 rule_type = "MatchRule" action = "Allow" match_conditions = [ { match_variables = [ { match_variable = "RequestHeaders" selector = "Host" } ] match_values = XXXXXX operator = "Equal" negation_condition = false transforms = ["Lowercase"] }, { match_variables = [ { match_variable = "RequestHeaders" selector = XXXXXX } ] match_values = XXXXXX operator = "BeginsWith" negation_condition = false transforms = [] } ] ``` ### Debug Output/Panic Output ```shell we should required to use operator "any" for custon rules ``` ### Expected Behaviour _No response_ ### Actual Behaviour _No response_ ### Steps to Reproduce _No response_ ### Important Factoids _No response_ ### References _No response_
1.0
Application Gateway WAF policy doesn't support operator "any" - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Community Note Hi Team, I am trying to create the custom rule in application Gateway WAF policy in the azure portal operator "any " is exist but terraform doesn't support and there is no possible value to support "any" operator * Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community [--->](url) ### Terraform Version 0.13 ### AzureRM Provider Version 3.9.0 ### Affected Resource(s)/Data Source(s) azurerm_web_application_firewall_policy ### Terraform Configuration Files ```hcl waf_custom_rules = [ { name = XXXXXX priority = 1 rule_type = "MatchRule" action = "Allow" match_conditions = [ { match_variables = [ { match_variable = "RequestHeaders" selector = "Host" } ] match_values = XXXXXX operator = "Equal" negation_condition = false transforms = ["Lowercase"] }, { match_variables = [ { match_variable = "RequestHeaders" selector = XXXXXX } ] match_values = XXXXXX operator = "BeginsWith" negation_condition = false transforms = [] } ] ``` ### Debug Output/Panic Output ```shell we should required to use operator "any" for custon rules ``` ### Expected Behaviour _No response_ ### Actual Behaviour _No response_ ### Steps to Reproduce _No response_ ### Important Factoids _No response_ ### References _No response_
non_process
application gateway waf policy doesn t support operator any is there an existing issue for this i have searched the existing issues community note hi team i am trying to create the custom rule in application gateway waf policy in the azure portal operator any is exist but terraform doesn t support and there is no possible value to support any operator please vote on this issue by adding a thumbsup to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment thank you for keeping this note for the community url terraform version azurerm provider version affected resource s data source s azurerm web application firewall policy terraform configuration files hcl waf custom rules name xxxxxx priority rule type matchrule action allow match conditions match variables match variable requestheaders selector host match values xxxxxx operator equal negation condition false transforms match variables match variable requestheaders selector xxxxxx match values xxxxxx operator beginswith negation condition false transforms debug output panic output shell we should required to use operator any for custon rules expected behaviour no response actual behaviour no response steps to reproduce no response important factoids no response references no response
0
11,156
13,957,693,663
IssuesEvent
2020-10-24 08:11:02
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
FR: Missing resources in Geoportal
FR - France Geoportal Harvesting process
Collected from the Geoportal Workshop online survey answers: EXAMPLE OF IGN INSPIRE WFS metadata: http://inspire-geoportal.ec.europa.eu/proxybrowser/#q=%22INSPIRE%20IGN%20WFS%22 has coupled resources : &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_GeographicalName_1-0.xml#MD_DataIdentification&quot;/&gt; &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_Administrative_Unit_1-0.xml#MD_DataIdentification&quot;/&gt; &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_Adresses_1-0.xml#MD_DataIdentification&quot;/&gt; &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_Hydrography_1-0.xml#MD_DataIdentification&quot;/&gt; &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_TransportNetworks_1-0.xml#MD_DataIdentification&quot;/&gt; &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_Buildings_1-0.xml#MD_DataIdentification&quot;/&gt; That are designated as non-downloadable in the thematic viewer. For example it is coupled with: http://inspire-geoportal.ec.europa.eu/proxybrowser/#q=IGNF_ELF_Buildings_1-0
1.0
FR: Missing resources in Geoportal - Collected from the Geoportal Workshop online survey answers: EXAMPLE OF IGN INSPIRE WFS metadata: http://inspire-geoportal.ec.europa.eu/proxybrowser/#q=%22INSPIRE%20IGN%20WFS%22 has coupled resources : &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_GeographicalName_1-0.xml#MD_DataIdentification&quot;/&gt; &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_Administrative_Unit_1-0.xml#MD_DataIdentification&quot;/&gt; &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_Adresses_1-0.xml#MD_DataIdentification&quot;/&gt; &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_Hydrography_1-0.xml#MD_DataIdentification&quot;/&gt; &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_TransportNetworks_1-0.xml#MD_DataIdentification&quot;/&gt; &lt;srv:operatesOn xlink:type=&quot;simple&quot; xlink:href=&quot;http://wxs.ign.fr/inspire/csw?service=CSW&amp;request=GetRecordbYiD&amp;version=2.0.2&amp;typeNames=csw:Record&amp;resultType=results&amp;ID=IGNF_ELF_Buildings_1-0.xml#MD_DataIdentification&quot;/&gt; That are designated as non-downloadable in the thematic viewer. For example it is coupled with: http://inspire-geoportal.ec.europa.eu/proxybrowser/#q=IGNF_ELF_Buildings_1-0
process
fr missing resources in geoportal collected from the geoportal workshop online survey answers example of ign inspire wfs metadata has coupled resources lt srv operateson xlink type quot simple quot xlink href quot lt srv operateson xlink type quot simple quot xlink href quot lt srv operateson xlink type quot simple quot xlink href quot lt srv operateson xlink type quot simple quot xlink href quot lt srv operateson xlink type quot simple quot xlink href quot lt srv operateson xlink type quot simple quot xlink href quot that are designated as non downloadable in the thematic viewer for example it is coupled with
1
19,564
25,885,352,397
IssuesEvent
2022-12-14 14:12:17
python/cpython
https://api.github.com/repos/python/cpython
closed
Cannot use 'multiprocessing'
type-bug OS-mac expert-multiprocessing
Python 3.10.9 MacOS M2 Ventura@13.1 Cannot use 'multiprocessing'. Prompt: **ValueError:<multiprocessing. pool. ApplyResult object at 0x1273d5ff0>not ready** Code reference: [https://docs.python.org/zh-cn/3.10/library/multiprocessing.html?highlight=multiprocessing#module -multiprocessing.pool](url) ``` from multiprocessing import Pool import time def f(x): return x*x if __name__ == '__main__': with Pool(processes=4) as pool: # start 4 worker processes result = pool.apply_async(f, (10,)) # evaluate "f(10)" asynchronously in a single process print(result.get(timeout=1)) ```
1.0
Cannot use 'multiprocessing' - Python 3.10.9 MacOS M2 Ventura@13.1 Cannot use 'multiprocessing'. Prompt: **ValueError:<multiprocessing. pool. ApplyResult object at 0x1273d5ff0>not ready** Code reference: [https://docs.python.org/zh-cn/3.10/library/multiprocessing.html?highlight=multiprocessing#module -multiprocessing.pool](url) ``` from multiprocessing import Pool import time def f(x): return x*x if __name__ == '__main__': with Pool(processes=4) as pool: # start 4 worker processes result = pool.apply_async(f, (10,)) # evaluate "f(10)" asynchronously in a single process print(result.get(timeout=1)) ```
process
cannot use multiprocessing python macos ventura cannot use multiprocessing prompt valueerror not ready code reference url from multiprocessing import pool import time def f x return x x if name main with pool processes as pool start worker processes result pool apply async f evaluate f asynchronously in a single process print result get timeout
1
113,913
11,826,451,036
IssuesEvent
2020-03-21 17:56:36
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
closed
Description of bootstrapping for RandomForest estimators
Documentation good first issue help wanted
#### Describe the issue linked to the documentation The docstrings for [sklearn.ensemble.RandomForestClassifier](https://scikit-learn.org/stable/modules/generated/sklearn.ensemble.RandomForestClassifier.html) and [sklearn.ensemble.RandomForestClassifier](https://scikit-learn.org/stable/modules/generated/sklearn.ensemble.RandomForestRegressor.html) states that: > The sub-sample size is always the same as the original input sample size but the samples are drawn with replacement if `bootstrap=True` (default). However, the new `max_samples` parameter has been added in version 0.22, so it needs updating (or removing). #### Suggest a potential alternative/fix A simple change would be to mention this new parameter and to borrow the documentation for the `bootstrap` parameter: > The sub-sample size is controlled with the `max_samples` parameter if `bootstrap=True` (default), otherwise the whole dataset is used to build each tree.
1.0
Description of bootstrapping for RandomForest estimators - #### Describe the issue linked to the documentation The docstrings for [sklearn.ensemble.RandomForestClassifier](https://scikit-learn.org/stable/modules/generated/sklearn.ensemble.RandomForestClassifier.html) and [sklearn.ensemble.RandomForestClassifier](https://scikit-learn.org/stable/modules/generated/sklearn.ensemble.RandomForestRegressor.html) states that: > The sub-sample size is always the same as the original input sample size but the samples are drawn with replacement if `bootstrap=True` (default). However, the new `max_samples` parameter has been added in version 0.22, so it needs updating (or removing). #### Suggest a potential alternative/fix A simple change would be to mention this new parameter and to borrow the documentation for the `bootstrap` parameter: > The sub-sample size is controlled with the `max_samples` parameter if `bootstrap=True` (default), otherwise the whole dataset is used to build each tree.
non_process
description of bootstrapping for randomforest estimators describe the issue linked to the documentation the docstrings for and states that the sub sample size is always the same as the original input sample size but the samples are drawn with replacement if bootstrap true default however the new max samples parameter has been added in version so it needs updating or removing suggest a potential alternative fix a simple change would be to mention this new parameter and to borrow the documentation for the bootstrap parameter the sub sample size is controlled with the max samples parameter if bootstrap true default otherwise the whole dataset is used to build each tree
0
92,718
18,940,563,449
IssuesEvent
2021-11-18 02:01:01
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Performance Improvement: Value Range Test
tenet-performance area-CodeGen-coreclr duplicate
### Description JIT cannot at this time perform conditional move or set operations. For example with ternary operators. I'd like to bring up different ideas to reduce branching and am asking if such morphing can be achieved now with the current JIT. _I'm also trying to get more familiar with the JIT optimizer, so I maybe can integrate such ideas myself. I have everything I need to do so, I just have to give myself patience and time._ Most often, there is code to check for a range of values... for example ```csharp bool Condition1 = 'A' <= chr && chr <= 'Z'; bool Condition2 = chr is >= 'A' and <= 'Z'; ``` This can be morphed into: ```csharp bool Condition1 = ((unint) chr - 'A') <= (unint)('Z' - 'A'); ``` Both must be constants, the comparison needs to be unsigned. Benefit is the reduction of 2x branching into 1x branching. **Question: Is such morphing possible now? Without major code rewrite?** Morphing would be: ```csharp bool Condition1a = C1 <= val && val <= C2; bool Condition1b = C1-C1 <= val-C1 && val-C1 <= C2-C1; bool Condition1c = 0 <= (nuint)(val-C1) && (nuint)(val-C1) <= (nuint)(C2-C1); bool Condition1d = true && val-C1 <= C2-C1; bool Condition1e = val-C1 <= C2-C1; ``` * it will work both signed/unsigned **when C1 <= C2** and at stage 1c both values are cast to unint. (native int to have val-C1 overflow) * it will also work for "<" comparison when C1/C2 is adjusted (shifted by 1) ### Analysis I wrote some helper functions that will achieve this morphing. And it works. It will be faster, but I cannot add any useful benchmark, as the CPU branch predictor will counter any benefit. _Maybe by unrolling a large loop and feeding it with data that will always mispredict the branch..._ ```csharp [MethodImpl(MethodImplOptions.AggressiveInlining | MethodImplOptions.AggressiveOptimization)] public static bool CondA(int value, int low, int high) => low <= value && value <= high; [MethodImpl(MethodImplOptions.AggressiveInlining | MethodImplOptions.AggressiveOptimization)] public static bool CondB(int value, int low, int high) => (nuint)(value - low) <= (nuint)(high - low); [MethodImpl(MethodImplOptions.AggressiveInlining | MethodImplOptions.AggressiveOptimization)] public static int TestA(int value) => CondA(value, 'A', 'Z') ? 1 : 2; [MethodImpl(MethodImplOptions.AggressiveInlining | MethodImplOptions.AggressiveOptimization)] public static int TestB(int value) => CondB(value, 'A', 'Z') ? 1 : 2; ``` ----- TestA disassembly: ```assembly G_M52463_IG01: ;; bbWeight=1 PerfScore 0.00 G_M52463_IG02: cmp ecx, 65 jl SHORT G_M52463_IG05 ;; bbWeight=1 PerfScore 1.25 G_M52463_IG03: cmp ecx, 90 jg SHORT G_M52463_IG05 mov eax, 1 ;; bbWeight=0.50 PerfScore 0.75 G_M52463_IG04: ret ;; bbWeight=0.50 PerfScore 0.50 G_M52463_IG05: mov eax, 2 ;; bbWeight=0.50 PerfScore 0.12 G_M52463_IG06: ret ;; bbWeight=0.50 PerfScore 0.50 ; Total bytes of code 22, prolog size 0, PerfScore 5.33, instruction count 8, allocated bytes for code 22 (MethodHash=bdf53310) for method SharperDisasm.RegisterExtensions:TestA(int):int ; ============================================================ ``` ----- TestB disassembly: ```assembly G_M7244_IG01: ;; bbWeight=1 PerfScore 0.00 G_M7244_IG02: add ecx, -65 movsxd rax, ecx cmp rax, 25 jbe SHORT G_M7244_IG05 ;; bbWeight=1 PerfScore 1.75 G_M7244_IG03: mov eax, 2 ;; bbWeight=0.50 PerfScore 0.12 G_M7244_IG04: ret ;; bbWeight=0.50 PerfScore 0.50 G_M7244_IG05: mov eax, 1 ;; bbWeight=0.50 PerfScore 0.12 G_M7244_IG06: ret ;; bbWeight=0.50 PerfScore 0.50 ; Total bytes of code 24, prolog size 0, PerfScore 5.40, instruction count 8, allocated bytes for code 24 (MethodHash=e215e3b3) for method SharperDisasm.RegisterExtensions:TestB(int):int ; ============================================================ ```
1.0
Performance Improvement: Value Range Test - ### Description JIT cannot at this time perform conditional move or set operations. For example with ternary operators. I'd like to bring up different ideas to reduce branching and am asking if such morphing can be achieved now with the current JIT. _I'm also trying to get more familiar with the JIT optimizer, so I maybe can integrate such ideas myself. I have everything I need to do so, I just have to give myself patience and time._ Most often, there is code to check for a range of values... for example ```csharp bool Condition1 = 'A' <= chr && chr <= 'Z'; bool Condition2 = chr is >= 'A' and <= 'Z'; ``` This can be morphed into: ```csharp bool Condition1 = ((unint) chr - 'A') <= (unint)('Z' - 'A'); ``` Both must be constants, the comparison needs to be unsigned. Benefit is the reduction of 2x branching into 1x branching. **Question: Is such morphing possible now? Without major code rewrite?** Morphing would be: ```csharp bool Condition1a = C1 <= val && val <= C2; bool Condition1b = C1-C1 <= val-C1 && val-C1 <= C2-C1; bool Condition1c = 0 <= (nuint)(val-C1) && (nuint)(val-C1) <= (nuint)(C2-C1); bool Condition1d = true && val-C1 <= C2-C1; bool Condition1e = val-C1 <= C2-C1; ``` * it will work both signed/unsigned **when C1 <= C2** and at stage 1c both values are cast to unint. (native int to have val-C1 overflow) * it will also work for "<" comparison when C1/C2 is adjusted (shifted by 1) ### Analysis I wrote some helper functions that will achieve this morphing. And it works. It will be faster, but I cannot add any useful benchmark, as the CPU branch predictor will counter any benefit. _Maybe by unrolling a large loop and feeding it with data that will always mispredict the branch..._ ```csharp [MethodImpl(MethodImplOptions.AggressiveInlining | MethodImplOptions.AggressiveOptimization)] public static bool CondA(int value, int low, int high) => low <= value && value <= high; [MethodImpl(MethodImplOptions.AggressiveInlining | MethodImplOptions.AggressiveOptimization)] public static bool CondB(int value, int low, int high) => (nuint)(value - low) <= (nuint)(high - low); [MethodImpl(MethodImplOptions.AggressiveInlining | MethodImplOptions.AggressiveOptimization)] public static int TestA(int value) => CondA(value, 'A', 'Z') ? 1 : 2; [MethodImpl(MethodImplOptions.AggressiveInlining | MethodImplOptions.AggressiveOptimization)] public static int TestB(int value) => CondB(value, 'A', 'Z') ? 1 : 2; ``` ----- TestA disassembly: ```assembly G_M52463_IG01: ;; bbWeight=1 PerfScore 0.00 G_M52463_IG02: cmp ecx, 65 jl SHORT G_M52463_IG05 ;; bbWeight=1 PerfScore 1.25 G_M52463_IG03: cmp ecx, 90 jg SHORT G_M52463_IG05 mov eax, 1 ;; bbWeight=0.50 PerfScore 0.75 G_M52463_IG04: ret ;; bbWeight=0.50 PerfScore 0.50 G_M52463_IG05: mov eax, 2 ;; bbWeight=0.50 PerfScore 0.12 G_M52463_IG06: ret ;; bbWeight=0.50 PerfScore 0.50 ; Total bytes of code 22, prolog size 0, PerfScore 5.33, instruction count 8, allocated bytes for code 22 (MethodHash=bdf53310) for method SharperDisasm.RegisterExtensions:TestA(int):int ; ============================================================ ``` ----- TestB disassembly: ```assembly G_M7244_IG01: ;; bbWeight=1 PerfScore 0.00 G_M7244_IG02: add ecx, -65 movsxd rax, ecx cmp rax, 25 jbe SHORT G_M7244_IG05 ;; bbWeight=1 PerfScore 1.75 G_M7244_IG03: mov eax, 2 ;; bbWeight=0.50 PerfScore 0.12 G_M7244_IG04: ret ;; bbWeight=0.50 PerfScore 0.50 G_M7244_IG05: mov eax, 1 ;; bbWeight=0.50 PerfScore 0.12 G_M7244_IG06: ret ;; bbWeight=0.50 PerfScore 0.50 ; Total bytes of code 24, prolog size 0, PerfScore 5.40, instruction count 8, allocated bytes for code 24 (MethodHash=e215e3b3) for method SharperDisasm.RegisterExtensions:TestB(int):int ; ============================================================ ```
non_process
performance improvement value range test description jit cannot at this time perform conditional move or set operations for example with ternary operators i d like to bring up different ideas to reduce branching and am asking if such morphing can be achieved now with the current jit i m also trying to get more familiar with the jit optimizer so i maybe can integrate such ideas myself i have everything i need to do so i just have to give myself patience and time most often there is code to check for a range of values for example csharp bool a chr chr z bool chr is a and z this can be morphed into csharp bool unint chr a unint z a both must be constants the comparison needs to be unsigned benefit is the reduction of branching into branching question is such morphing possible now without major code rewrite morphing would be csharp bool val val bool val val bool nuint val nuint val nuint bool true val bool val it will work both signed unsigned when and at stage both values are cast to unint native int to have val overflow it will also work for comparison when is adjusted shifted by analysis i wrote some helper functions that will achieve this morphing and it works it will be faster but i cannot add any useful benchmark as the cpu branch predictor will counter any benefit maybe by unrolling a large loop and feeding it with data that will always mispredict the branch csharp public static bool conda int value int low int high low value value high public static bool condb int value int low int high nuint value low nuint high low public static int testa int value conda value a z public static int testb int value condb value a z testa disassembly assembly g bbweight perfscore g cmp ecx jl short g bbweight perfscore g cmp ecx jg short g mov eax bbweight perfscore g ret bbweight perfscore g mov eax bbweight perfscore g ret bbweight perfscore total bytes of code prolog size perfscore instruction count allocated bytes for code methodhash for method sharperdisasm registerextensions testa int int testb disassembly assembly g bbweight perfscore g add ecx movsxd rax ecx cmp rax jbe short g bbweight perfscore g mov eax bbweight perfscore g ret bbweight perfscore g mov eax bbweight perfscore g ret bbweight perfscore total bytes of code prolog size perfscore instruction count allocated bytes for code methodhash for method sharperdisasm registerextensions testb int int
0
14,569
17,692,054,971
IssuesEvent
2021-08-24 11:14:11
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Unable to migration autoincrementing BigInt to autoincrementing Int
bug/2-confirmed kind/bug process/candidate team/migrations topic: db push
### Bug description I tried to create a prisma model with connects to a Postgres database. For the ids for the table, I want to use the BigInt datatype, as it says it's supported. However, when I try to push the new model I get `Error querying the database: db error: ERROR: type "bigserial" does not exist` and I'm unable to push the model ### How to reproduce 1. Create a new prisma instance as per documentation, connecting to a postgres database 2. Create the model as described below 3. Push the model using `prisma db push` 4. Very the error returned is `Error: Database error Error querying the database: db error: ERROR: type "bigserial" does not exist 0: sql_migration_connector::sql_database_step_applier::apply_migration at migration-engine/connectors/sql-migration-connector/src/sql_database_step_applier.rs:11 1: migration_core::api::SchemaPush at migration-engine/core/src/api.rs:161` ### Expected behavior `push` command should happen as expected as BigInt is supported ### Prisma information ```prisma enum EntityType { club individual } enum Status { active dormant } generator client { provider = "prisma-client-js" } datasource db { provider = "postgresql" url = env("DATABASE_URL") } model club_entity { id BigInt @db.BigInt @id @default(autoincrement()) name String @unique email String type EntityType @default(club) status Status points Int @default(0) } model club_transactions { id BigInt @db.BigInt @id @default(autoincrement()) description String pointChange Int clubId Int } ``` ### Environment & setup - Client OS is MacOS - Database: PostsgreSQL - Node: 14.15.4 ### Prisma Version ``` prisma : 2.29.1 @prisma/client : 2.29.1 Current platform : darwin Query Engine : query-engine 1be4cd60b89afa04b192acb1ef47758a39810f3a (at node_modules/@prisma/engines/query-engine-darwin) Migration Engine : migration-engine-cli 1be4cd60b89afa04b192acb1ef47758a39810f3a (at node_modules/@prisma/engines/migration-engine-darwin) Introspection Engine : introspection-core 1be4cd60b89afa04b192acb1ef47758a39810f3a (at node_modules/@prisma/engines/introspection-engine-darwin) Format Binary : prisma-fmt 1be4cd60b89afa04b192acb1ef47758a39810f3a (at node_modules/@prisma/engines/prisma-fmt-darwin) Default Engines Hash : 1be4cd60b89afa04b192acb1ef47758a39810f3a Studio : 0.419.0 ```
1.0
Unable to migration autoincrementing BigInt to autoincrementing Int - ### Bug description I tried to create a prisma model with connects to a Postgres database. For the ids for the table, I want to use the BigInt datatype, as it says it's supported. However, when I try to push the new model I get `Error querying the database: db error: ERROR: type "bigserial" does not exist` and I'm unable to push the model ### How to reproduce 1. Create a new prisma instance as per documentation, connecting to a postgres database 2. Create the model as described below 3. Push the model using `prisma db push` 4. Very the error returned is `Error: Database error Error querying the database: db error: ERROR: type "bigserial" does not exist 0: sql_migration_connector::sql_database_step_applier::apply_migration at migration-engine/connectors/sql-migration-connector/src/sql_database_step_applier.rs:11 1: migration_core::api::SchemaPush at migration-engine/core/src/api.rs:161` ### Expected behavior `push` command should happen as expected as BigInt is supported ### Prisma information ```prisma enum EntityType { club individual } enum Status { active dormant } generator client { provider = "prisma-client-js" } datasource db { provider = "postgresql" url = env("DATABASE_URL") } model club_entity { id BigInt @db.BigInt @id @default(autoincrement()) name String @unique email String type EntityType @default(club) status Status points Int @default(0) } model club_transactions { id BigInt @db.BigInt @id @default(autoincrement()) description String pointChange Int clubId Int } ``` ### Environment & setup - Client OS is MacOS - Database: PostsgreSQL - Node: 14.15.4 ### Prisma Version ``` prisma : 2.29.1 @prisma/client : 2.29.1 Current platform : darwin Query Engine : query-engine 1be4cd60b89afa04b192acb1ef47758a39810f3a (at node_modules/@prisma/engines/query-engine-darwin) Migration Engine : migration-engine-cli 1be4cd60b89afa04b192acb1ef47758a39810f3a (at node_modules/@prisma/engines/migration-engine-darwin) Introspection Engine : introspection-core 1be4cd60b89afa04b192acb1ef47758a39810f3a (at node_modules/@prisma/engines/introspection-engine-darwin) Format Binary : prisma-fmt 1be4cd60b89afa04b192acb1ef47758a39810f3a (at node_modules/@prisma/engines/prisma-fmt-darwin) Default Engines Hash : 1be4cd60b89afa04b192acb1ef47758a39810f3a Studio : 0.419.0 ```
process
unable to migration autoincrementing bigint to autoincrementing int bug description i tried to create a prisma model with connects to a postgres database for the ids for the table i want to use the bigint datatype as it says it s supported however when i try to push the new model i get error querying the database db error error type bigserial does not exist and i m unable to push the model how to reproduce create a new prisma instance as per documentation connecting to a postgres database create the model as described below push the model using prisma db push very the error returned is error database error error querying the database db error error type bigserial does not exist sql migration connector sql database step applier apply migration at migration engine connectors sql migration connector src sql database step applier rs migration core api schemapush at migration engine core src api rs expected behavior push command should happen as expected as bigint is supported prisma information prisma enum entitytype club individual enum status active dormant generator client provider prisma client js datasource db provider postgresql url env database url model club entity id bigint db bigint id default autoincrement name string unique email string type entitytype default club status status points int default model club transactions id bigint db bigint id default autoincrement description string pointchange int clubid int environment setup client os is macos database postsgresql node prisma version prisma prisma client current platform darwin query engine query engine at node modules prisma engines query engine darwin migration engine migration engine cli at node modules prisma engines migration engine darwin introspection engine introspection core at node modules prisma engines introspection engine darwin format binary prisma fmt at node modules prisma engines prisma fmt darwin default engines hash studio
1
187,741
22,045,882,701
IssuesEvent
2022-05-30 01:36:26
DavidSpek/pipelines
https://api.github.com/repos/DavidSpek/pipelines
opened
CVE-2022-29216 (High) detected in tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl
security vulnerability
## CVE-2022-29216 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /contrib/components/openvino/ovms-deployer/containers/requirements.txt</p> <p>Path to vulnerable library: /contrib/components/openvino/ovms-deployer/containers/requirements.txt,/samples/core/ai_platform/training</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DavidSpek/pipelines/commit/6f7433f006e282c4f25441e7502b80d73751e38f">6f7433f006e282c4f25441e7502b80d73751e38f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. Prior to versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4, TensorFlow's `saved_model_cli` tool is vulnerable to a code injection. This can be used to open a reverse shell. This code path was maintained for compatibility reasons as the maintainers had several test cases where numpy expressions were used as arguments. However, given that the tool is always run manually, the impact of this is still not severe. The maintainers have now removed the `safe=False` argument, so all parsing is done without calling `eval`. The patch is available in versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4. <p>Publish Date: 2022-05-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29216>CVE-2022-29216</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29216">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29216</a></p> <p>Release Date: 2022-05-21</p> <p>Fix Resolution: tensorflow - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-cpu - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-gpu - 2.6.4,2.7.2,2.8.1,2.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-29216 (High) detected in tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl - ## CVE-2022-29216 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /contrib/components/openvino/ovms-deployer/containers/requirements.txt</p> <p>Path to vulnerable library: /contrib/components/openvino/ovms-deployer/containers/requirements.txt,/samples/core/ai_platform/training</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DavidSpek/pipelines/commit/6f7433f006e282c4f25441e7502b80d73751e38f">6f7433f006e282c4f25441e7502b80d73751e38f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. Prior to versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4, TensorFlow's `saved_model_cli` tool is vulnerable to a code injection. This can be used to open a reverse shell. This code path was maintained for compatibility reasons as the maintainers had several test cases where numpy expressions were used as arguments. However, given that the tool is always run manually, the impact of this is still not severe. The maintainers have now removed the `safe=False` argument, so all parsing is done without calling `eval`. The patch is available in versions 2.9.0, 2.8.1, 2.7.2, and 2.6.4. <p>Publish Date: 2022-05-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29216>CVE-2022-29216</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29216">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29216</a></p> <p>Release Date: 2022-05-21</p> <p>Fix Resolution: tensorflow - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-cpu - 2.6.4,2.7.2,2.8.1,2.9.0;tensorflow-gpu - 2.6.4,2.7.2,2.8.1,2.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file contrib components openvino ovms deployer containers requirements txt path to vulnerable library contrib components openvino ovms deployer containers requirements txt samples core ai platform training dependency hierarchy x tensorflow whl vulnerable library found in head commit a href found in base branch master vulnerability details tensorflow is an open source platform for machine learning prior to versions and tensorflow s saved model cli tool is vulnerable to a code injection this can be used to open a reverse shell this code path was maintained for compatibility reasons as the maintainers had several test cases where numpy expressions were used as arguments however given that the tool is always run manually the impact of this is still not severe the maintainers have now removed the safe false argument so all parsing is done without calling eval the patch is available in versions and publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with mend
0
450,659
31,934,476,970
IssuesEvent
2023-09-19 09:33:38
mariuzka/popy
https://api.github.com/repos/mariuzka/popy
closed
examples in Example Gallery
documentation
Wenn die Examples, die aktuell in `examples` sind, in `docs/examples/` wรคren, kรถnnte man sie in der Dokumentation direkt anzeigen lassen.
1.0
examples in Example Gallery - Wenn die Examples, die aktuell in `examples` sind, in `docs/examples/` wรคren, kรถnnte man sie in der Dokumentation direkt anzeigen lassen.
non_process
examples in example gallery wenn die examples die aktuell in examples sind in docs examples wรคren kรถnnte man sie in der dokumentation direkt anzeigen lassen
0
278,543
24,160,372,828
IssuesEvent
2022-09-22 11:07:46
wpeventmanager/wp-event-manager
https://api.github.com/repos/wpeventmanager/wp-event-manager
closed
Event Dashboard - Filter is not working properly
In Testing
Go to Event Dashboard > Click on the filter icon button. Using pagination option move second or thrid page. Now type event name which is availabel on the first page. Observed that events are availabel but it is not searching recored and display the message "You do not have any active events."
1.0
Event Dashboard - Filter is not working properly - Go to Event Dashboard > Click on the filter icon button. Using pagination option move second or thrid page. Now type event name which is availabel on the first page. Observed that events are availabel but it is not searching recored and display the message "You do not have any active events."
non_process
event dashboard filter is not working properly go to event dashboard click on the filter icon button using pagination option move second or thrid page now type event name which is availabel on the first page observed that events are availabel but it is not searching recored and display the message you do not have any active events
0
13,837
16,600,330,053
IssuesEvent
2021-06-01 18:29:52
icra/ecam
https://api.github.com/repos/icra/ecam
closed
group sludge management equations for onsite treatment
in process
wwo_KPI_GHG_landapp(), wwo_KPI_GHG_landfil(), wwo_KPI_GHG_dumping(), wwo_KPI_GHG_trck(), wwo_KPI_GHG_urine(),
1.0
group sludge management equations for onsite treatment - wwo_KPI_GHG_landapp(), wwo_KPI_GHG_landfil(), wwo_KPI_GHG_dumping(), wwo_KPI_GHG_trck(), wwo_KPI_GHG_urine(),
process
group sludge management equations for onsite treatment wwo kpi ghg landapp wwo kpi ghg landfil wwo kpi ghg dumping wwo kpi ghg trck wwo kpi ghg urine
1
350,824
31,932,353,976
IssuesEvent
2023-09-19 08:17:48
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix array.test_jax___rlshift__
JAX Frontend Sub Task Failing Test
| | | |---|---| |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6202565153/job/16841552478"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/6202565153/job/16841552478"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6202565153/job/16841552478"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/6202565153/job/16841552478"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6202565153/job/16841552478"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix array.test_jax___rlshift__ - | | | |---|---| |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6202565153/job/16841552478"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/6202565153/job/16841552478"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6202565153/job/16841552478"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/6202565153/job/16841552478"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6202565153/job/16841552478"><img src=https://img.shields.io/badge/-failure-red></a>
non_process
fix array test jax rlshift numpy a href src jax a href src tensorflow a href src torch a href src paddle a href src
0
294,923
25,414,934,468
IssuesEvent
2022-11-22 22:45:02
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
closed
[chip-test] chip_sw_flash_rma_unlocked
Component:ChipLevelTest
### Test point name [chip_sw_flash_rma_unlocked](https://github.com/lowRISC/opentitan/blob/f3e95c6d0bc66483c94bc07fed6aab8b9d38b25c/hw/top_earlgrey/data/chip_testplan.hjson#L2703) ### Host side component SystemVerilog ### OpenTitanTool infrastructure implemented _No response_ ### Contact person @tjaychen, @cindychip ### Checklist Please fill out this checklist as items are completed. Link to PRs and issues as appropriate. - [x] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes) - [x] Device-side (C) component developed - [x] Bazel build rules developed - [x] Host-side component developed - [x] HJSON test plan updated with test name (so it shows up in the dashboard) - [x] Test added to dvsim nightly regression (and passing at time of checking) This test requires a minor update based on the manufacturing testplan sync. The test needs to initiate an ID check from the host side.
1.0
[chip-test] chip_sw_flash_rma_unlocked - ### Test point name [chip_sw_flash_rma_unlocked](https://github.com/lowRISC/opentitan/blob/f3e95c6d0bc66483c94bc07fed6aab8b9d38b25c/hw/top_earlgrey/data/chip_testplan.hjson#L2703) ### Host side component SystemVerilog ### OpenTitanTool infrastructure implemented _No response_ ### Contact person @tjaychen, @cindychip ### Checklist Please fill out this checklist as items are completed. Link to PRs and issues as appropriate. - [x] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes) - [x] Device-side (C) component developed - [x] Bazel build rules developed - [x] Host-side component developed - [x] HJSON test plan updated with test name (so it shows up in the dashboard) - [x] Test added to dvsim nightly regression (and passing at time of checking) This test requires a minor update based on the manufacturing testplan sync. The test needs to initiate an ID check from the host side.
non_process
chip sw flash rma unlocked test point name host side component systemverilog opentitantool infrastructure implemented no response contact person tjaychen cindychip checklist please fill out this checklist as items are completed link to prs and issues as appropriate check if existing test covers most or all of this testpoint if so either extend said test to cover all points or skip the next checkboxes device side c component developed bazel build rules developed host side component developed hjson test plan updated with test name so it shows up in the dashboard test added to dvsim nightly regression and passing at time of checking this test requires a minor update based on the manufacturing testplan sync the test needs to initiate an id check from the host side
0
11,986
14,737,132,566
IssuesEvent
2021-01-07 00:57:52
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Error "We're sorry, but something went wrong"
anc-ops anc-process anp-1.5 ant-bug ant-support
In GitLab by @kdjstudios on Apr 18, 2018, 10:38 **Submitted by:** "Kimberly Gagner" <kimberly.gagner@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-18-64776/conversation **Server:** Internal **Client/Site:** Billerica **Account:** Multiple **Issue:** I have noticed this morning you cannot bring up any account by searching in the search bar to the right. You get the error โ€œWeโ€™re sorry, but something went wrongโ€ The only way you can bring up an account is on the left hand side by choosing the site, account name, etc.
1.0
Error "We're sorry, but something went wrong" - In GitLab by @kdjstudios on Apr 18, 2018, 10:38 **Submitted by:** "Kimberly Gagner" <kimberly.gagner@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-18-64776/conversation **Server:** Internal **Client/Site:** Billerica **Account:** Multiple **Issue:** I have noticed this morning you cannot bring up any account by searching in the search bar to the right. You get the error โ€œWeโ€™re sorry, but something went wrongโ€ The only way you can bring up an account is on the left hand side by choosing the site, account name, etc.
process
error we re sorry but something went wrong in gitlab by kdjstudios on apr submitted by kimberly gagner helpdesk server internal client site billerica account multiple issue i have noticed this morning you cannot bring up any account by searching in the search bar to the right you get the error โ€œweโ€™re sorry but something went wrongโ€ the only way you can bring up an account is on the left hand side by choosing the site account name etc
1
20,382
27,038,217,252
IssuesEvent
2023-02-13 01:08:32
TeamAidemy/ds-paper-summaries
https://api.github.com/repos/TeamAidemy/ds-paper-summaries
opened
When Attention Meets Fast Recurrence: Training Language Models with Reduced Compute
Natural language processing
Lei, Tao. 2021. โ€œWhen Attention Meets Fast Recurrence: Training Language Models with Reduced Compute.โ€ In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 7633โ€“48. Online and Punta Cana, Dominican Republic: Association for Computational Linguistics. https://aclanthology.org/2021.emnlp-main.602/ - ไฝŽ่จˆ็ฎ—ใ‚ณใ‚นใƒˆใง่จ€่ชžใƒขใƒ‡ใƒชใƒณใ‚ฐใŒๅฏ่ƒฝใชSRU++ใ‚’ๆๆกˆ - Attentionใจ้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏใซ็€็›ฎ - Transformer็ณปใฎ3ๅ€ใ‹ใ‚‰25ๅ€ๅฐ‘ใชใ„่จˆ็ฎ—้‡ใงใ€ๅŒ็ญ‰ใฎๆ€ง่ƒฝใ‚’็คบใ—ใŸ Introductionใง **"Is attention all we need for modeling?"** ใจใ€ใ‚ใฎๆœ‰ๅใช่จ€่‘‰ใ‚’ใ‚‚ใ˜ใ‚Šใ€ๅฎŸ้š›ใซAttentionใ‚’ๆŒŸใ‚€ใ ใ‘ใงๅคงๅน…ใชๆ”นๅ–„ใ‚’่ฆ‹ใ›ใ€ใใฎๅจๅŠ›ใ‚’่ฆ‹ใ›ใฆใใ‚Œใ‚‹้ข็™ฝใ„ไป•ไบ‹ใ€‚ ## Abstract >Large language models have become increasingly difficult to train because of the growing computation time and cost. In this work, we present SRU++, a highly-efficient architecture that combines fast recurrence and attention for sequence modeling. SRU++ exhibits strong modeling capacity and training efficiency. On standard language modeling tasks such as Enwik8, Wiki-103 and Billion Word datasets, our model obtains better bits-per-character and perplexity while using 3x-10x less training cost compared to top-performing Transformer models. For instance, our model achieves a state-of-the-art result on the Enwik8 dataset using 1.6 days of training on an 8-GPU machine. We further demonstrate that SRU++ requires minimal attention for near state-of-the-art performance. Our results suggest jointly leveraging fast recurrence with little attention as a promising direction for accelerating model training and inference. ๅคง่ฆๆจกใช่จ€่ชžใƒขใƒ‡ใƒซใฏใ€่จˆ็ฎ—ๆ™‚้–“ใ‚„ใ‚ณใ‚นใƒˆใŒๅข—ๅคงใ™ใ‚‹ใŸใ‚ใ€ๅญฆ็ฟ’ใŒๅ›ฐ้›ฃใซใชใฃใฆใใฆใ„ใ‚‹ใ€‚ๆœฌ็ ”็ฉถใงใฏใ€ใ‚ทใƒผใ‚ฑใƒณใ‚นใƒขใƒ‡ใƒชใƒณใ‚ฐใฎใŸใ‚ใซ้ซ˜้€Ÿใชๅ†ๅธฐๆ€งใจๆณจๆ„ๅŠ›ใ‚’็ต„ใฟๅˆใ‚ใ›ใŸ้ซ˜ๅŠน็އใชใ‚ขใƒผใ‚ญใƒ†ใ‚ฏใƒใƒฃใงใ‚ใ‚‹SRU++ใ‚’็™บ่กจใ™ใ‚‹ใ€‚SRU++ใฏใ€ๅผทๅŠ›ใชใƒขใƒ‡ใƒชใƒณใ‚ฐ่ƒฝๅŠ›ใจๅญฆ็ฟ’ๅŠน็އใ‚’็™บๆฎใ™ใ‚‹ใ€‚Enwik8ใ€Wiki-103ใ€Billion Wordใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใชใฉใฎๆจ™ๆบ–็š„ใช่จ€่ชžใƒขใƒ‡ใƒชใƒณใ‚ฐใ‚ฟใ‚นใ‚ฏใซใŠใ„ใฆใ€ๆˆ‘ใ€…ใฎใƒขใƒ‡ใƒซใฏใ€ไธŠไฝใฎTransformerใƒขใƒ‡ใƒซใจๆฏ”่ผƒใ—ใฆใ€3ๅ€ใ‹ใ‚‰10ๅ€ๅฐ‘ใชใ„ๅญฆ็ฟ’ใ‚ณใ‚นใƒˆใงใ€ๅ„ชใ‚ŒใŸๆ–‡ๅญ—ใ‚ใŸใ‚Šใฎใƒ“ใƒƒใƒˆๆ•ฐใจใƒ‘ใƒผใƒ—ใƒฌใ‚ญใ‚ทใƒ†ใ‚ฃใ‚’็ฒๅพ—ใ™ใ‚‹ใ“ใจใŒใงใใ‚‹ใ€‚ไพ‹ใˆใฐใ€Enwik8ใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใงใฏใ€8GPUใฎใƒžใ‚ทใƒณใง1.6ๆ—ฅ้–“ใฎๅญฆ็ฟ’ใ‚’่กŒใ„ใ€ๆœ€ๅ…ˆ็ซฏใฎ็ตๆžœใ‚’้”ๆˆใ—ใพใ—ใŸใ€‚ใ•ใ‚‰ใซใ€SRU++ใฏๆœ€ๅฐ้™ใฎๆณจๆ„ใงๆœ€ๆ–ฐ้‹ญใซ่ฟ‘ใ„ๆ€ง่ƒฝใ‚’็™บๆฎใ™ใ‚‹ใ“ใจใ‚’ๅฎŸ่จผใ—ใฆใ„ใพใ™ใ€‚ใ“ใฎ็ตๆžœใฏใ€ใƒขใƒ‡ใƒซใฎๅญฆ็ฟ’ใจๆŽจ่ซ–ใ‚’้ซ˜้€ŸๅŒ–ใ™ใ‚‹ใŸใ‚ใฎๆœ‰ๆœ›ใชๆ–นๅ‘ๆ€งใจใ—ใฆใ€ๅฐ‘ใชใ„ๆณจๆ„ใง้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใ‚นใ‚’ๅ…ฑๅŒใงๆดป็”จใ™ใ‚‹ใ“ใจใ‚’็คบๅ”†ใ—ใฆใ„ใพใ™ใ€‚ ## ใ‚ณใƒผใƒ‰ https://github.com/asappresearch/sru ## ่งฃๆฑบใ—ใŸ่ชฒ้กŒ/ๅ…ˆ่กŒ็ ”็ฉถใจใฎๆฏ”่ผƒ - ่ฟ‘ๅนดใฎ่จ€่ชžใƒขใƒ‡ใƒซใฏใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใฎใ‚ตใ‚คใ‚บใ€ใ‚ขใƒผใ‚ญใƒ†ใ‚ฏใƒใƒฃใฎใ‚ตใ‚คใ‚บใจใ‚‚ใซๅคง่ฆๆจกๅŒ–ใ—ใฆใใฆใŠใ‚Šใ€่จˆ็ฎ—ใ‚ณใ‚นใƒˆใŒใฉใ‚“ใฉใ‚“้ซ˜ใใชใฃใฆใใฆใ„ใ‚‹ใ€‚ - ่ฟ‘ๅนดใฎ็ ”็ฉถใ‹ใ‚‰Attentionใจ้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏใ‚’็ต„ใฟๅˆใ‚ใ›ใ‚‹ใ“ใจใงใ€ใ‚ˆใ‚ŠๅŠน็އ็š„ใชใƒขใƒ‡ใƒซใ‚’ๆง‹็ฏ‰ใงใใ‚‹ๅฏ่ƒฝๆ€งใŒ็คบๅ”†ใ•ใ‚Œใฆใ„ใŸ (Bradbury et al., 2017; Zhang and Sennrich, 2019) - ๆœฌ่ซ–ๆ–‡ใงใฏใใฎใ‚ขใ‚คใƒ‡ใ‚ขใ‚’ๆคœ่จผใ—ใŸใ€‚ - 2018ๅนดใซๅ ฑๅ‘Šใ—ใŸ[SRU](https://arxiv.org/abs/1709.02755) (Lei et al., 2018) ใ‚’ใƒ™ใƒผใ‚นใซใ—ใฆใ„ใ‚‹ใ€‚ ## ๆŠ€่ก“ใƒปๆ‰‹ๆณ•ใฎใƒใ‚คใƒณใƒˆ [![Image from Gyazo](https://i.gyazo.com/96ed23604cb1fa034305cae81ec835c9.png)](https://gyazo.com/96ed23604cb1fa034305cae81ec835c9) - ใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏ (SRU) ใฎๅ…ฅๅŠ›ใฎ็ทšๅฝขๅค‰ๆ›ใ‚’ใ€Attentionใ‚’็”จใ„ใŸๅค‰ๆ›ใซ็ฝฎๆ› - ๅ…ทไฝ“็š„ใซใฏใ€ใƒ™ใƒผใ‚นใจใชใฃใŸSRUใงใฏ $\mathbf{U}$ ใ‚’ไปฅไธ‹ใฎใ‚ˆใ†ใซ่จˆ็ฎ—ใ—ใฆใ„ใ‚‹ใ€‚ $$ \displaystyle\mathbf{U}^{\top}\ \displaystyle=\ \left(\begin{array}[]{l}\mathbf{W}\\ \mathbf{W}^{\prime}\\ \mathbf{W}^{\prime\prime}\end{array}\right)\mathbf{X}^{\top} $$ - ไธ€ๆ–นใงใ€SRU++ ใงใฏๅ†…้ƒจใงAttentionใ‚’็”จใ„ใฆไปฅไธ‹ใฎ่จˆ็ฎ—ใ‚’่กŒใฃใฆใ„ใ‚‹ใ€‚ $$ \begin{array}{l} \mathbf{Q} \displaystyle=\mathbf{W}^{q}\mathbf{X}^{\top} \\ \mathbf{K} \displaystyle=\mathbf{W}^{k}\mathbf{Q}^{\top} \\ \mathbf{V} \displaystyle=\mathbf{W}^{v}\mathbf{Q}^{\top} \end{array} $$ $$ \displaystyle\mathbf{A}^{\top}=\text{softmax}\left(\frac{\mathbf{Q}^{\top}\mathbf{K}}{\sqrt{d^{\prime}}}\right)\mathbf{V}^{\top}. $$ $$ \displaystyle\mathbf{U}^{\top}=\mathbf{W}^{o}\left(\mathbf{Q}+\alpha\cdot\mathbf{A}\right). $$ - ็ทšๅฝขๅค‰ๆ›ใ‹ใ‚‰ใ‚ˆใ‚Š่ค‡้›‘ๆ€งใฎ้ซ˜ใ„ๆผ”็ฎ—ใซๅค‰ๆ›ดใ™ใ‚‹ใ“ใจใง่กจ็พๅŠ›ใŒๅ‘ไธŠ - ๅพŒๆฎตใฎใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏ็ฎ‡ๆ‰€ใฏSRUใจๅŒๆง˜ - ใ“ใ‚ŒใŒไฝ•ๆ•…ๅŠนๆžœ็š„ใซใฏใŸใ‚‰ใใฎ๏ผŸใฎ่ญฐ่ซ–ใฏ **Fig.5** ๅ‘จ่พบใ‚’ๅ‚็…งใ€‚ ## ่ฉ•ไพกๆŒ‡ๆจ™ - ENWIK8 (**Table 3**), WIKI103 (**Table 5**), BILLION WORD (**Table 6**) ใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใ‚’็”จใ„ใ€Transformer็ณปใฎใƒขใƒ‡ใƒซใจๆฏ”่ผƒใ€‚ - ่ฉ•ไพกๆŒ‡ๆจ™ใฏPerplexityใจGPU days. (ENWIK8ใฏBPC) - ใ™ในใฆใฎใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใซใŠใ„ใฆใ€Transformer็ณปใฎ3ๅ€ใ‹ใ‚‰25ๅ€ๅฐ‘ใชใ„่จˆ็ฎ—้‡ใงใ€ๅŒ็ญ‰ใฎ็ตๆžœใ‚’้”ๆˆใ€‚ - ใพใŸใ€WIKI-103ใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใฎๆŽจ่ซ–ใ‚นใƒ”ใƒผใƒ‰ใŒๆ•ฐๅ€้€Ÿใ„ใ€‚ (**Table 7**) - IWSLT'14ใฎDeโ†’En็ฟป่จณใ‚ฟใ‚นใ‚ฏใงใ‚‚TransfomrerใจๅŒ็จ‹ๅบฆใฎๆ€ง่ƒฝใ€‚ (**Table 8**) - Attentionใ‚’ๆŒฟๅ…ฅใ™ใ‚‹ๅ ดๆ‰€ใซใคใ„ใฆใ‚‚ๆคœ่จŽใ—ใŸใจใ“ใ‚ใ€ๅ…ฅๅŠ›ๅดใฎ1ๅฑค็›ฎใŒๆœ€ใ‚‚ๅŠนๆžœ็š„ใงใ‚ใฃใŸใ€‚ (**Fig.4**) ## ๆฎ‹ใ•ใ‚ŒใŸ่ชฒ้กŒใƒป่ญฐ่ซ– - Attention่‡ชไฝ“ใฎ้ซ˜้€ŸๅŒ–ใจใฏ็•ฐใชใ‚‹ใ‚ขใ‚คใƒ‡ใ‚ขใงใ‚ใ‚‹ใŸใ‚ใ€Attention่‡ชไฝ“ใฎ้ซ˜้€ŸๅŒ–ใจๅˆใ‚ใ›ใ‚Œใฐใ€ใ‚ˆใ‚Šๅ…จไฝ“ใฎ้ซ˜้€ŸๅŒ–ใŒ้€ฒใ‚ใ‚‰ใ‚Œใ‚‹ใ‹ใ‚‚ใ—ใ‚Œใชใ„ใ€‚ - ๅŒๆง˜ใซใ€ๆœฌๆๆกˆๆ‰‹ๆณ•ใฏใƒชใ‚ซใƒฌใƒณใƒˆๅฎŸ่ฃ…ใ‚„ๆญฃ่ฆๅŒ–ใ€ๆœ€้ฉๅŒ–ๆŠ€่ก“ใฎ้€ฒๆญฉใจ็ต„ใฟๅˆใ‚ใ›ใ‚‰ใ‚Œใ‚‹ใจ่€ƒใˆใ‚‰ใ‚Œใ‚‹ใ€‚ - Introductionใงใฏใƒขใƒ‡ใƒซใฎใ‚ตใ‚คใ‚บใซใ‚‚่งฆใ‚Œใ‚‰ใ‚Œใฆใ„ใŸใŒใ€ใƒ‘ใƒฉใƒกใƒผใ‚ฟๆ•ฐใŒๅคšใ„ใ“ใจใซๅค‰ใ‚ใ‚Šใฏใชใ—ใ€‚ใƒ‘ใƒฉใƒกใƒผใ‚ฟๆ•ฐใŒใƒใƒƒใ‚ฏใซใชใ‚‹ๅ ดๅˆใซใฏใ“ใฎๆ‰‹ๆณ•ใฏไฝฟใˆใชใ„ใ€‚(**Table 3**) - ใ€Œ่ปฝ้‡ๅŒ–ใ•ใ‚ŒใŸใŸใ‚ใ‚จใƒƒใ‚ธใƒ‡ใƒใ‚คใ‚นใซใ‚‚่ผ‰ใ›ใ‚„ใ™ใ„๏ผใ€ใจใ„ใ†ๆ–นๅ‘ใ‚’็›ฎๆŒ‡ใ—ใŸไป•ไบ‹ใงใฏใชใ„ใ€‚ ## ้‡่ฆใชๅผ•็”จ - Transformer - Vaswani, Ashish, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. โ€œAttention Is All You Need.โ€ arXiv [cs.CL]. arXiv. http://arxiv.org/abs/1706.03762. - Attentionใจ้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏใ‚’็ต„ใฟๅˆใ‚ใ›ใŸๅŠน็އ็š„ใชใƒขใƒ‡ใƒซๆง‹็ฏ‰ใฎๅฏ่ƒฝๆ€งใ‚’็คบๅ”† - Bradbury, James, Stephen Merity, Caiming Xiong, and Richard Socher. 2016. โ€œQuasi-Recurrent Neural Networks.โ€ arXiv [cs.NE]. arXiv. http://arxiv.org/abs/1611.01576. - Zhang, Biao, and Rico Sennrich. 2019. โ€œA Lightweight Recurrent Network for Sequence Modeling.โ€ In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 1538โ€“48. Florence, Italy: Association for Computational Linguistics. - Tramsformer-XL. ๆฏ”่ผƒๅฏพ่ฑกใจใชใฃใฆใ„ใ‚‹ใ€่ซ–ๆ–‡ๆŠ•็จฟๆ™‚็‚นใงใฎๆœ€ๅผทใƒขใƒ‡ใƒซ - Dai, Zihang, Zhilin Yang, Yiming Yang, Jaime Carbonell, Quoc V. Le, and Ruslan Salakhutdinov. 2019. โ€œTransformer-XL: Attentive Language Models Beyond a Fixed-Length Context.โ€ arXiv [cs.LG]. arXiv. http://arxiv.org/abs/1901.02860. - SRU. ๅŒ่‘—่€…ใฎๅ…ˆ่กŒ็ ”็ฉถ - Lei, Tao, Yu Zhang, Sida I. Wang, Hui Dai, and Yoav Artzi. 2017. โ€œSimple Recurrent Units for Highly Parallelizable Recurrence.โ€ arXiv [cs.CL]. arXiv. http://arxiv.org/abs/1709.02755. ## ๅ‚่€ƒๆƒ…ๅ ฑ - [Transformerๆ—ฅๆœฌ่ชž่ฆ็ด„](https://github.com/TeamAidemy/ds-paper-summaries/issues/1)
1.0
When Attention Meets Fast Recurrence: Training Language Models with Reduced Compute - Lei, Tao. 2021. โ€œWhen Attention Meets Fast Recurrence: Training Language Models with Reduced Compute.โ€ In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 7633โ€“48. Online and Punta Cana, Dominican Republic: Association for Computational Linguistics. https://aclanthology.org/2021.emnlp-main.602/ - ไฝŽ่จˆ็ฎ—ใ‚ณใ‚นใƒˆใง่จ€่ชžใƒขใƒ‡ใƒชใƒณใ‚ฐใŒๅฏ่ƒฝใชSRU++ใ‚’ๆๆกˆ - Attentionใจ้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏใซ็€็›ฎ - Transformer็ณปใฎ3ๅ€ใ‹ใ‚‰25ๅ€ๅฐ‘ใชใ„่จˆ็ฎ—้‡ใงใ€ๅŒ็ญ‰ใฎๆ€ง่ƒฝใ‚’็คบใ—ใŸ Introductionใง **"Is attention all we need for modeling?"** ใจใ€ใ‚ใฎๆœ‰ๅใช่จ€่‘‰ใ‚’ใ‚‚ใ˜ใ‚Šใ€ๅฎŸ้š›ใซAttentionใ‚’ๆŒŸใ‚€ใ ใ‘ใงๅคงๅน…ใชๆ”นๅ–„ใ‚’่ฆ‹ใ›ใ€ใใฎๅจๅŠ›ใ‚’่ฆ‹ใ›ใฆใใ‚Œใ‚‹้ข็™ฝใ„ไป•ไบ‹ใ€‚ ## Abstract >Large language models have become increasingly difficult to train because of the growing computation time and cost. In this work, we present SRU++, a highly-efficient architecture that combines fast recurrence and attention for sequence modeling. SRU++ exhibits strong modeling capacity and training efficiency. On standard language modeling tasks such as Enwik8, Wiki-103 and Billion Word datasets, our model obtains better bits-per-character and perplexity while using 3x-10x less training cost compared to top-performing Transformer models. For instance, our model achieves a state-of-the-art result on the Enwik8 dataset using 1.6 days of training on an 8-GPU machine. We further demonstrate that SRU++ requires minimal attention for near state-of-the-art performance. Our results suggest jointly leveraging fast recurrence with little attention as a promising direction for accelerating model training and inference. ๅคง่ฆๆจกใช่จ€่ชžใƒขใƒ‡ใƒซใฏใ€่จˆ็ฎ—ๆ™‚้–“ใ‚„ใ‚ณใ‚นใƒˆใŒๅข—ๅคงใ™ใ‚‹ใŸใ‚ใ€ๅญฆ็ฟ’ใŒๅ›ฐ้›ฃใซใชใฃใฆใใฆใ„ใ‚‹ใ€‚ๆœฌ็ ”็ฉถใงใฏใ€ใ‚ทใƒผใ‚ฑใƒณใ‚นใƒขใƒ‡ใƒชใƒณใ‚ฐใฎใŸใ‚ใซ้ซ˜้€Ÿใชๅ†ๅธฐๆ€งใจๆณจๆ„ๅŠ›ใ‚’็ต„ใฟๅˆใ‚ใ›ใŸ้ซ˜ๅŠน็އใชใ‚ขใƒผใ‚ญใƒ†ใ‚ฏใƒใƒฃใงใ‚ใ‚‹SRU++ใ‚’็™บ่กจใ™ใ‚‹ใ€‚SRU++ใฏใ€ๅผทๅŠ›ใชใƒขใƒ‡ใƒชใƒณใ‚ฐ่ƒฝๅŠ›ใจๅญฆ็ฟ’ๅŠน็އใ‚’็™บๆฎใ™ใ‚‹ใ€‚Enwik8ใ€Wiki-103ใ€Billion Wordใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใชใฉใฎๆจ™ๆบ–็š„ใช่จ€่ชžใƒขใƒ‡ใƒชใƒณใ‚ฐใ‚ฟใ‚นใ‚ฏใซใŠใ„ใฆใ€ๆˆ‘ใ€…ใฎใƒขใƒ‡ใƒซใฏใ€ไธŠไฝใฎTransformerใƒขใƒ‡ใƒซใจๆฏ”่ผƒใ—ใฆใ€3ๅ€ใ‹ใ‚‰10ๅ€ๅฐ‘ใชใ„ๅญฆ็ฟ’ใ‚ณใ‚นใƒˆใงใ€ๅ„ชใ‚ŒใŸๆ–‡ๅญ—ใ‚ใŸใ‚Šใฎใƒ“ใƒƒใƒˆๆ•ฐใจใƒ‘ใƒผใƒ—ใƒฌใ‚ญใ‚ทใƒ†ใ‚ฃใ‚’็ฒๅพ—ใ™ใ‚‹ใ“ใจใŒใงใใ‚‹ใ€‚ไพ‹ใˆใฐใ€Enwik8ใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใงใฏใ€8GPUใฎใƒžใ‚ทใƒณใง1.6ๆ—ฅ้–“ใฎๅญฆ็ฟ’ใ‚’่กŒใ„ใ€ๆœ€ๅ…ˆ็ซฏใฎ็ตๆžœใ‚’้”ๆˆใ—ใพใ—ใŸใ€‚ใ•ใ‚‰ใซใ€SRU++ใฏๆœ€ๅฐ้™ใฎๆณจๆ„ใงๆœ€ๆ–ฐ้‹ญใซ่ฟ‘ใ„ๆ€ง่ƒฝใ‚’็™บๆฎใ™ใ‚‹ใ“ใจใ‚’ๅฎŸ่จผใ—ใฆใ„ใพใ™ใ€‚ใ“ใฎ็ตๆžœใฏใ€ใƒขใƒ‡ใƒซใฎๅญฆ็ฟ’ใจๆŽจ่ซ–ใ‚’้ซ˜้€ŸๅŒ–ใ™ใ‚‹ใŸใ‚ใฎๆœ‰ๆœ›ใชๆ–นๅ‘ๆ€งใจใ—ใฆใ€ๅฐ‘ใชใ„ๆณจๆ„ใง้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใ‚นใ‚’ๅ…ฑๅŒใงๆดป็”จใ™ใ‚‹ใ“ใจใ‚’็คบๅ”†ใ—ใฆใ„ใพใ™ใ€‚ ## ใ‚ณใƒผใƒ‰ https://github.com/asappresearch/sru ## ่งฃๆฑบใ—ใŸ่ชฒ้กŒ/ๅ…ˆ่กŒ็ ”็ฉถใจใฎๆฏ”่ผƒ - ่ฟ‘ๅนดใฎ่จ€่ชžใƒขใƒ‡ใƒซใฏใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใฎใ‚ตใ‚คใ‚บใ€ใ‚ขใƒผใ‚ญใƒ†ใ‚ฏใƒใƒฃใฎใ‚ตใ‚คใ‚บใจใ‚‚ใซๅคง่ฆๆจกๅŒ–ใ—ใฆใใฆใŠใ‚Šใ€่จˆ็ฎ—ใ‚ณใ‚นใƒˆใŒใฉใ‚“ใฉใ‚“้ซ˜ใใชใฃใฆใใฆใ„ใ‚‹ใ€‚ - ่ฟ‘ๅนดใฎ็ ”็ฉถใ‹ใ‚‰Attentionใจ้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏใ‚’็ต„ใฟๅˆใ‚ใ›ใ‚‹ใ“ใจใงใ€ใ‚ˆใ‚ŠๅŠน็އ็š„ใชใƒขใƒ‡ใƒซใ‚’ๆง‹็ฏ‰ใงใใ‚‹ๅฏ่ƒฝๆ€งใŒ็คบๅ”†ใ•ใ‚Œใฆใ„ใŸ (Bradbury et al., 2017; Zhang and Sennrich, 2019) - ๆœฌ่ซ–ๆ–‡ใงใฏใใฎใ‚ขใ‚คใƒ‡ใ‚ขใ‚’ๆคœ่จผใ—ใŸใ€‚ - 2018ๅนดใซๅ ฑๅ‘Šใ—ใŸ[SRU](https://arxiv.org/abs/1709.02755) (Lei et al., 2018) ใ‚’ใƒ™ใƒผใ‚นใซใ—ใฆใ„ใ‚‹ใ€‚ ## ๆŠ€่ก“ใƒปๆ‰‹ๆณ•ใฎใƒใ‚คใƒณใƒˆ [![Image from Gyazo](https://i.gyazo.com/96ed23604cb1fa034305cae81ec835c9.png)](https://gyazo.com/96ed23604cb1fa034305cae81ec835c9) - ใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏ (SRU) ใฎๅ…ฅๅŠ›ใฎ็ทšๅฝขๅค‰ๆ›ใ‚’ใ€Attentionใ‚’็”จใ„ใŸๅค‰ๆ›ใซ็ฝฎๆ› - ๅ…ทไฝ“็š„ใซใฏใ€ใƒ™ใƒผใ‚นใจใชใฃใŸSRUใงใฏ $\mathbf{U}$ ใ‚’ไปฅไธ‹ใฎใ‚ˆใ†ใซ่จˆ็ฎ—ใ—ใฆใ„ใ‚‹ใ€‚ $$ \displaystyle\mathbf{U}^{\top}\ \displaystyle=\ \left(\begin{array}[]{l}\mathbf{W}\\ \mathbf{W}^{\prime}\\ \mathbf{W}^{\prime\prime}\end{array}\right)\mathbf{X}^{\top} $$ - ไธ€ๆ–นใงใ€SRU++ ใงใฏๅ†…้ƒจใงAttentionใ‚’็”จใ„ใฆไปฅไธ‹ใฎ่จˆ็ฎ—ใ‚’่กŒใฃใฆใ„ใ‚‹ใ€‚ $$ \begin{array}{l} \mathbf{Q} \displaystyle=\mathbf{W}^{q}\mathbf{X}^{\top} \\ \mathbf{K} \displaystyle=\mathbf{W}^{k}\mathbf{Q}^{\top} \\ \mathbf{V} \displaystyle=\mathbf{W}^{v}\mathbf{Q}^{\top} \end{array} $$ $$ \displaystyle\mathbf{A}^{\top}=\text{softmax}\left(\frac{\mathbf{Q}^{\top}\mathbf{K}}{\sqrt{d^{\prime}}}\right)\mathbf{V}^{\top}. $$ $$ \displaystyle\mathbf{U}^{\top}=\mathbf{W}^{o}\left(\mathbf{Q}+\alpha\cdot\mathbf{A}\right). $$ - ็ทšๅฝขๅค‰ๆ›ใ‹ใ‚‰ใ‚ˆใ‚Š่ค‡้›‘ๆ€งใฎ้ซ˜ใ„ๆผ”็ฎ—ใซๅค‰ๆ›ดใ™ใ‚‹ใ“ใจใง่กจ็พๅŠ›ใŒๅ‘ไธŠ - ๅพŒๆฎตใฎใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏ็ฎ‡ๆ‰€ใฏSRUใจๅŒๆง˜ - ใ“ใ‚ŒใŒไฝ•ๆ•…ๅŠนๆžœ็š„ใซใฏใŸใ‚‰ใใฎ๏ผŸใฎ่ญฐ่ซ–ใฏ **Fig.5** ๅ‘จ่พบใ‚’ๅ‚็…งใ€‚ ## ่ฉ•ไพกๆŒ‡ๆจ™ - ENWIK8 (**Table 3**), WIKI103 (**Table 5**), BILLION WORD (**Table 6**) ใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใ‚’็”จใ„ใ€Transformer็ณปใฎใƒขใƒ‡ใƒซใจๆฏ”่ผƒใ€‚ - ่ฉ•ไพกๆŒ‡ๆจ™ใฏPerplexityใจGPU days. (ENWIK8ใฏBPC) - ใ™ในใฆใฎใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใซใŠใ„ใฆใ€Transformer็ณปใฎ3ๅ€ใ‹ใ‚‰25ๅ€ๅฐ‘ใชใ„่จˆ็ฎ—้‡ใงใ€ๅŒ็ญ‰ใฎ็ตๆžœใ‚’้”ๆˆใ€‚ - ใพใŸใ€WIKI-103ใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใฎๆŽจ่ซ–ใ‚นใƒ”ใƒผใƒ‰ใŒๆ•ฐๅ€้€Ÿใ„ใ€‚ (**Table 7**) - IWSLT'14ใฎDeโ†’En็ฟป่จณใ‚ฟใ‚นใ‚ฏใงใ‚‚TransfomrerใจๅŒ็จ‹ๅบฆใฎๆ€ง่ƒฝใ€‚ (**Table 8**) - Attentionใ‚’ๆŒฟๅ…ฅใ™ใ‚‹ๅ ดๆ‰€ใซใคใ„ใฆใ‚‚ๆคœ่จŽใ—ใŸใจใ“ใ‚ใ€ๅ…ฅๅŠ›ๅดใฎ1ๅฑค็›ฎใŒๆœ€ใ‚‚ๅŠนๆžœ็š„ใงใ‚ใฃใŸใ€‚ (**Fig.4**) ## ๆฎ‹ใ•ใ‚ŒใŸ่ชฒ้กŒใƒป่ญฐ่ซ– - Attention่‡ชไฝ“ใฎ้ซ˜้€ŸๅŒ–ใจใฏ็•ฐใชใ‚‹ใ‚ขใ‚คใƒ‡ใ‚ขใงใ‚ใ‚‹ใŸใ‚ใ€Attention่‡ชไฝ“ใฎ้ซ˜้€ŸๅŒ–ใจๅˆใ‚ใ›ใ‚Œใฐใ€ใ‚ˆใ‚Šๅ…จไฝ“ใฎ้ซ˜้€ŸๅŒ–ใŒ้€ฒใ‚ใ‚‰ใ‚Œใ‚‹ใ‹ใ‚‚ใ—ใ‚Œใชใ„ใ€‚ - ๅŒๆง˜ใซใ€ๆœฌๆๆกˆๆ‰‹ๆณ•ใฏใƒชใ‚ซใƒฌใƒณใƒˆๅฎŸ่ฃ…ใ‚„ๆญฃ่ฆๅŒ–ใ€ๆœ€้ฉๅŒ–ๆŠ€่ก“ใฎ้€ฒๆญฉใจ็ต„ใฟๅˆใ‚ใ›ใ‚‰ใ‚Œใ‚‹ใจ่€ƒใˆใ‚‰ใ‚Œใ‚‹ใ€‚ - Introductionใงใฏใƒขใƒ‡ใƒซใฎใ‚ตใ‚คใ‚บใซใ‚‚่งฆใ‚Œใ‚‰ใ‚Œใฆใ„ใŸใŒใ€ใƒ‘ใƒฉใƒกใƒผใ‚ฟๆ•ฐใŒๅคšใ„ใ“ใจใซๅค‰ใ‚ใ‚Šใฏใชใ—ใ€‚ใƒ‘ใƒฉใƒกใƒผใ‚ฟๆ•ฐใŒใƒใƒƒใ‚ฏใซใชใ‚‹ๅ ดๅˆใซใฏใ“ใฎๆ‰‹ๆณ•ใฏไฝฟใˆใชใ„ใ€‚(**Table 3**) - ใ€Œ่ปฝ้‡ๅŒ–ใ•ใ‚ŒใŸใŸใ‚ใ‚จใƒƒใ‚ธใƒ‡ใƒใ‚คใ‚นใซใ‚‚่ผ‰ใ›ใ‚„ใ™ใ„๏ผใ€ใจใ„ใ†ๆ–นๅ‘ใ‚’็›ฎๆŒ‡ใ—ใŸไป•ไบ‹ใงใฏใชใ„ใ€‚ ## ้‡่ฆใชๅผ•็”จ - Transformer - Vaswani, Ashish, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. โ€œAttention Is All You Need.โ€ arXiv [cs.CL]. arXiv. http://arxiv.org/abs/1706.03762. - Attentionใจ้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏใ‚’็ต„ใฟๅˆใ‚ใ›ใŸๅŠน็އ็š„ใชใƒขใƒ‡ใƒซๆง‹็ฏ‰ใฎๅฏ่ƒฝๆ€งใ‚’็คบๅ”† - Bradbury, James, Stephen Merity, Caiming Xiong, and Richard Socher. 2016. โ€œQuasi-Recurrent Neural Networks.โ€ arXiv [cs.NE]. arXiv. http://arxiv.org/abs/1611.01576. - Zhang, Biao, and Rico Sennrich. 2019. โ€œA Lightweight Recurrent Network for Sequence Modeling.โ€ In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 1538โ€“48. Florence, Italy: Association for Computational Linguistics. - Tramsformer-XL. ๆฏ”่ผƒๅฏพ่ฑกใจใชใฃใฆใ„ใ‚‹ใ€่ซ–ๆ–‡ๆŠ•็จฟๆ™‚็‚นใงใฎๆœ€ๅผทใƒขใƒ‡ใƒซ - Dai, Zihang, Zhilin Yang, Yiming Yang, Jaime Carbonell, Quoc V. Le, and Ruslan Salakhutdinov. 2019. โ€œTransformer-XL: Attentive Language Models Beyond a Fixed-Length Context.โ€ arXiv [cs.LG]. arXiv. http://arxiv.org/abs/1901.02860. - SRU. ๅŒ่‘—่€…ใฎๅ…ˆ่กŒ็ ”็ฉถ - Lei, Tao, Yu Zhang, Sida I. Wang, Hui Dai, and Yoav Artzi. 2017. โ€œSimple Recurrent Units for Highly Parallelizable Recurrence.โ€ arXiv [cs.CL]. arXiv. http://arxiv.org/abs/1709.02755. ## ๅ‚่€ƒๆƒ…ๅ ฑ - [Transformerๆ—ฅๆœฌ่ชž่ฆ็ด„](https://github.com/TeamAidemy/ds-paper-summaries/issues/1)
process
when attention meets fast recurrence training language models with reduced compute lei tao โ€œwhen attention meets fast recurrence training language models with reduced compute โ€ in proceedings of the conference on empirical methods in natural language processing โ€“ online and punta cana dominican republic association for computational linguistics ไฝŽ่จˆ็ฎ—ใ‚ณใ‚นใƒˆใง่จ€่ชžใƒขใƒ‡ใƒชใƒณใ‚ฐใŒๅฏ่ƒฝใชsru ใ‚’ๆๆกˆ attentionใจ้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏใซ็€็›ฎ ใ€ๅŒ็ญ‰ใฎๆ€ง่ƒฝใ‚’็คบใ—ใŸ introductionใง is attention all we need for modeling ใจใ€ใ‚ใฎๆœ‰ๅใช่จ€่‘‰ใ‚’ใ‚‚ใ˜ใ‚Šใ€ๅฎŸ้š›ใซattentionใ‚’ๆŒŸใ‚€ใ ใ‘ใงๅคงๅน…ใชๆ”นๅ–„ใ‚’่ฆ‹ใ›ใ€ใใฎๅจๅŠ›ใ‚’่ฆ‹ใ›ใฆใใ‚Œใ‚‹้ข็™ฝใ„ไป•ไบ‹ใ€‚ abstract large language models have become increasingly difficult to train because of the growing computation time and cost in this work we present sru a highly efficient architecture that combines fast recurrence and attention for sequence modeling sru exhibits strong modeling capacity and training efficiency on standard language modeling tasks such as wiki and billion word datasets our model obtains better bits per character and perplexity while using less training cost compared to top performing transformer models for instance our model achieves a state of the art result on the dataset using days of training on an gpu machine we further demonstrate that sru requires minimal attention for near state of the art performance our results suggest jointly leveraging fast recurrence with little attention as a promising direction for accelerating model training and inference ๅคง่ฆๆจกใช่จ€่ชžใƒขใƒ‡ใƒซใฏใ€่จˆ็ฎ—ๆ™‚้–“ใ‚„ใ‚ณใ‚นใƒˆใŒๅข—ๅคงใ™ใ‚‹ใŸใ‚ใ€ๅญฆ็ฟ’ใŒๅ›ฐ้›ฃใซใชใฃใฆใใฆใ„ใ‚‹ใ€‚ๆœฌ็ ”็ฉถใงใฏใ€ใ‚ทใƒผใ‚ฑใƒณใ‚นใƒขใƒ‡ใƒชใƒณใ‚ฐใฎใŸใ‚ใซ้ซ˜้€Ÿใชๅ†ๅธฐๆ€งใจๆณจๆ„ๅŠ›ใ‚’็ต„ใฟๅˆใ‚ใ›ใŸ้ซ˜ๅŠน็އใชใ‚ขใƒผใ‚ญใƒ†ใ‚ฏใƒใƒฃใงใ‚ใ‚‹sru ใ‚’็™บ่กจใ™ใ‚‹ใ€‚sru ใฏใ€ๅผทๅŠ›ใชใƒขใƒ‡ใƒชใƒณใ‚ฐ่ƒฝๅŠ›ใจๅญฆ็ฟ’ๅŠน็އใ‚’็™บๆฎใ™ใ‚‹ใ€‚ ใ€wiki ใ€billion wordใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใชใฉใฎๆจ™ๆบ–็š„ใช่จ€่ชžใƒขใƒ‡ใƒชใƒณใ‚ฐใ‚ฟใ‚นใ‚ฏใซใŠใ„ใฆใ€ๆˆ‘ใ€…ใฎใƒขใƒ‡ใƒซใฏใ€ไธŠไฝใฎtransformerใƒขใƒ‡ใƒซใจๆฏ”่ผƒใ—ใฆใ€ ใ€ๅ„ชใ‚ŒใŸๆ–‡ๅญ—ใ‚ใŸใ‚Šใฎใƒ“ใƒƒใƒˆๆ•ฐใจใƒ‘ใƒผใƒ—ใƒฌใ‚ญใ‚ทใƒ†ใ‚ฃใ‚’็ฒๅพ—ใ™ใ‚‹ใ“ใจใŒใงใใ‚‹ใ€‚ไพ‹ใˆใฐใ€ ใ€ ใ€ๆœ€ๅ…ˆ็ซฏใฎ็ตๆžœใ‚’้”ๆˆใ—ใพใ—ใŸใ€‚ใ•ใ‚‰ใซใ€sru ใฏๆœ€ๅฐ้™ใฎๆณจๆ„ใงๆœ€ๆ–ฐ้‹ญใซ่ฟ‘ใ„ๆ€ง่ƒฝใ‚’็™บๆฎใ™ใ‚‹ใ“ใจใ‚’ๅฎŸ่จผใ—ใฆใ„ใพใ™ใ€‚ใ“ใฎ็ตๆžœใฏใ€ใƒขใƒ‡ใƒซใฎๅญฆ็ฟ’ใจๆŽจ่ซ–ใ‚’้ซ˜้€ŸๅŒ–ใ™ใ‚‹ใŸใ‚ใฎๆœ‰ๆœ›ใชๆ–นๅ‘ๆ€งใจใ—ใฆใ€ๅฐ‘ใชใ„ๆณจๆ„ใง้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใ‚นใ‚’ๅ…ฑๅŒใงๆดป็”จใ™ใ‚‹ใ“ใจใ‚’็คบๅ”†ใ—ใฆใ„ใพใ™ใ€‚ ใ‚ณใƒผใƒ‰ ่งฃๆฑบใ—ใŸ่ชฒ้กŒ ๅ…ˆ่กŒ็ ”็ฉถใจใฎๆฏ”่ผƒ ่ฟ‘ๅนดใฎ่จ€่ชžใƒขใƒ‡ใƒซใฏใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใฎใ‚ตใ‚คใ‚บใ€ใ‚ขใƒผใ‚ญใƒ†ใ‚ฏใƒใƒฃใฎใ‚ตใ‚คใ‚บใจใ‚‚ใซๅคง่ฆๆจกๅŒ–ใ—ใฆใใฆใŠใ‚Šใ€่จˆ็ฎ—ใ‚ณใ‚นใƒˆใŒใฉใ‚“ใฉใ‚“้ซ˜ใใชใฃใฆใใฆใ„ใ‚‹ใ€‚ ่ฟ‘ๅนดใฎ็ ”็ฉถใ‹ใ‚‰attentionใจ้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏใ‚’็ต„ใฟๅˆใ‚ใ›ใ‚‹ใ“ใจใงใ€ใ‚ˆใ‚ŠๅŠน็އ็š„ใชใƒขใƒ‡ใƒซใ‚’ๆง‹็ฏ‰ใงใใ‚‹ๅฏ่ƒฝๆ€งใŒ็คบๅ”†ใ•ใ‚Œใฆใ„ใŸ bradbury et al zhang and sennrich ๆœฌ่ซ–ๆ–‡ใงใฏใใฎใ‚ขใ‚คใƒ‡ใ‚ขใ‚’ๆคœ่จผใ—ใŸใ€‚ lei et al ใ‚’ใƒ™ใƒผใ‚นใซใ—ใฆใ„ใ‚‹ใ€‚ ๆŠ€่ก“ใƒปๆ‰‹ๆณ•ใฎใƒใ‚คใƒณใƒˆ ใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏ sru ใฎๅ…ฅๅŠ›ใฎ็ทšๅฝขๅค‰ๆ›ใ‚’ใ€attentionใ‚’็”จใ„ใŸๅค‰ๆ›ใซ็ฝฎๆ› ๅ…ทไฝ“็š„ใซใฏใ€ใƒ™ใƒผใ‚นใจใชใฃใŸsruใงใฏ mathbf u ใ‚’ไปฅไธ‹ใฎใ‚ˆใ†ใซ่จˆ็ฎ—ใ—ใฆใ„ใ‚‹ใ€‚ displaystyle mathbf u top displaystyle left begin array l mathbf w mathbf w prime mathbf w prime prime end array right mathbf x top ไธ€ๆ–นใงใ€sru ใงใฏๅ†…้ƒจใงattentionใ‚’็”จใ„ใฆไปฅไธ‹ใฎ่จˆ็ฎ—ใ‚’่กŒใฃใฆใ„ใ‚‹ใ€‚ begin array l mathbf q displaystyle mathbf w q mathbf x top mathbf k displaystyle mathbf w k mathbf q top mathbf v displaystyle mathbf w v mathbf q top end array displaystyle mathbf a top text softmax left frac mathbf q top mathbf k sqrt d prime right mathbf v top displaystyle mathbf u top mathbf w o left mathbf q alpha cdot mathbf a right ็ทšๅฝขๅค‰ๆ›ใ‹ใ‚‰ใ‚ˆใ‚Š่ค‡้›‘ๆ€งใฎ้ซ˜ใ„ๆผ”็ฎ—ใซๅค‰ๆ›ดใ™ใ‚‹ใ“ใจใง่กจ็พๅŠ›ใŒๅ‘ไธŠ ๅพŒๆฎตใฎใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏ็ฎ‡ๆ‰€ใฏsruใจๅŒๆง˜ ใ“ใ‚ŒใŒไฝ•ๆ•…ๅŠนๆžœ็š„ใซใฏใŸใ‚‰ใใฎ๏ผŸใฎ่ญฐ่ซ–ใฏ fig ๅ‘จ่พบใ‚’ๅ‚็…งใ€‚ ่ฉ•ไพกๆŒ‡ๆจ™ table table billion word table ใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใ‚’็”จใ„ใ€transformer็ณปใฎใƒขใƒ‡ใƒซใจๆฏ”่ผƒใ€‚ ่ฉ•ไพกๆŒ‡ๆจ™ใฏperplexityใจgpu days ใ™ในใฆใฎใƒ‡ใƒผใ‚ฟใ‚ปใƒƒใƒˆใซใŠใ„ใฆใ€ ใ€ๅŒ็ญ‰ใฎ็ตๆžœใ‚’้”ๆˆใ€‚ ใพใŸใ€wiki ใ€‚ table iwslt โ†’en็ฟป่จณใ‚ฟใ‚นใ‚ฏใงใ‚‚transfomrerใจๅŒ็จ‹ๅบฆใฎๆ€ง่ƒฝใ€‚ table attentionใ‚’ๆŒฟๅ…ฅใ™ใ‚‹ๅ ดๆ‰€ใซใคใ„ใฆใ‚‚ๆคœ่จŽใ—ใŸใจใ“ใ‚ใ€ ใ€‚ fig ๆฎ‹ใ•ใ‚ŒใŸ่ชฒ้กŒใƒป่ญฐ่ซ– attention่‡ชไฝ“ใฎ้ซ˜้€ŸๅŒ–ใจใฏ็•ฐใชใ‚‹ใ‚ขใ‚คใƒ‡ใ‚ขใงใ‚ใ‚‹ใŸใ‚ใ€attention่‡ชไฝ“ใฎ้ซ˜้€ŸๅŒ–ใจๅˆใ‚ใ›ใ‚Œใฐใ€ใ‚ˆใ‚Šๅ…จไฝ“ใฎ้ซ˜้€ŸๅŒ–ใŒ้€ฒใ‚ใ‚‰ใ‚Œใ‚‹ใ‹ใ‚‚ใ—ใ‚Œใชใ„ใ€‚ ๅŒๆง˜ใซใ€ๆœฌๆๆกˆๆ‰‹ๆณ•ใฏใƒชใ‚ซใƒฌใƒณใƒˆๅฎŸ่ฃ…ใ‚„ๆญฃ่ฆๅŒ–ใ€ๆœ€้ฉๅŒ–ๆŠ€่ก“ใฎ้€ฒๆญฉใจ็ต„ใฟๅˆใ‚ใ›ใ‚‰ใ‚Œใ‚‹ใจ่€ƒใˆใ‚‰ใ‚Œใ‚‹ใ€‚ introductionใงใฏใƒขใƒ‡ใƒซใฎใ‚ตใ‚คใ‚บใซใ‚‚่งฆใ‚Œใ‚‰ใ‚Œใฆใ„ใŸใŒใ€ใƒ‘ใƒฉใƒกใƒผใ‚ฟๆ•ฐใŒๅคšใ„ใ“ใจใซๅค‰ใ‚ใ‚Šใฏใชใ—ใ€‚ใƒ‘ใƒฉใƒกใƒผใ‚ฟๆ•ฐใŒใƒใƒƒใ‚ฏใซใชใ‚‹ๅ ดๅˆใซใฏใ“ใฎๆ‰‹ๆณ•ใฏไฝฟใˆใชใ„ใ€‚ table ใ€Œ่ปฝ้‡ๅŒ–ใ•ใ‚ŒใŸใŸใ‚ใ‚จใƒƒใ‚ธใƒ‡ใƒใ‚คใ‚นใซใ‚‚่ผ‰ใ›ใ‚„ใ™ใ„๏ผใ€ใจใ„ใ†ๆ–นๅ‘ใ‚’็›ฎๆŒ‡ใ—ใŸไป•ไบ‹ใงใฏใชใ„ใ€‚ ้‡่ฆใชๅผ•็”จ transformer vaswani ashish noam shazeer niki parmar jakob uszkoreit llion jones aidan n gomez lukasz kaiser and illia polosukhin โ€œattention is all you need โ€ arxiv arxiv attentionใจ้ซ˜้€Ÿใƒชใ‚ซใƒฌใƒณใƒˆใƒใƒƒใƒˆใƒฏใƒผใ‚ฏใ‚’็ต„ใฟๅˆใ‚ใ›ใŸๅŠน็އ็š„ใชใƒขใƒ‡ใƒซๆง‹็ฏ‰ใฎๅฏ่ƒฝๆ€งใ‚’็คบๅ”† bradbury james stephen merity caiming xiong and richard socher โ€œquasi recurrent neural networks โ€ arxiv arxiv zhang biao and rico sennrich โ€œa lightweight recurrent network for sequence modeling โ€ in proceedings of the annual meeting of the association for computational linguistics โ€“ florence italy association for computational linguistics tramsformer xl ๆฏ”่ผƒๅฏพ่ฑกใจใชใฃใฆใ„ใ‚‹ใ€่ซ–ๆ–‡ๆŠ•็จฟๆ™‚็‚นใงใฎๆœ€ๅผทใƒขใƒ‡ใƒซ dai zihang zhilin yang yiming yang jaime carbonell quoc v le and ruslan salakhutdinov โ€œtransformer xl attentive language models beyond a fixed length context โ€ arxiv arxiv sru ๅŒ่‘—่€…ใฎๅ…ˆ่กŒ็ ”็ฉถ lei tao yu zhang sida i wang hui dai and yoav artzi โ€œsimple recurrent units for highly parallelizable recurrence โ€ arxiv arxiv ๅ‚่€ƒๆƒ…ๅ ฑ
1
21,276
28,442,550,284
IssuesEvent
2023-04-16 04:02:59
cse442-at-ub/project_s23-team-infinity
https://api.github.com/repos/cse442-at-ub/project_s23-team-infinity
closed
Create back-end PHP for "create an event" in the database
Processing Task Sprint 3
**Task tests** *Test 1* 1) Verify that when event is created for user, correct user ID associated with event (foreign key with Users table) 2) Verify that there is some form of secondary authentication when creating an event for a user (token) 3) Verify that even if provided with empty necessary fields (title, time, date), the event does not populate the Events table 4) Verify that new events are placed into Events table with correct fields
1.0
Create back-end PHP for "create an event" in the database - **Task tests** *Test 1* 1) Verify that when event is created for user, correct user ID associated with event (foreign key with Users table) 2) Verify that there is some form of secondary authentication when creating an event for a user (token) 3) Verify that even if provided with empty necessary fields (title, time, date), the event does not populate the Events table 4) Verify that new events are placed into Events table with correct fields
process
create back end php for create an event in the database task tests test verify that when event is created for user correct user id associated with event foreign key with users table verify that there is some form of secondary authentication when creating an event for a user token verify that even if provided with empty necessary fields title time date the event does not populate the events table verify that new events are placed into events table with correct fields
1
20,989
27,853,672,687
IssuesEvent
2023-03-20 20:48:19
keras-team/keras-cv
https://api.github.com/repos/keras-team/keras-cv
opened
Support segmentation masks in `CutMix` layer
contribution-welcome preprocessing augmentation
This should follow the same structure as segmentation mask augmentation in our other preprocessing layers. Treating masks like 1-channel images for the purpose of applying CutMix should be a simple and effective approach for this.
1.0
Support segmentation masks in `CutMix` layer - This should follow the same structure as segmentation mask augmentation in our other preprocessing layers. Treating masks like 1-channel images for the purpose of applying CutMix should be a simple and effective approach for this.
process
support segmentation masks in cutmix layer this should follow the same structure as segmentation mask augmentation in our other preprocessing layers treating masks like channel images for the purpose of applying cutmix should be a simple and effective approach for this
1
23,892
16,677,373,519
IssuesEvent
2021-06-07 17:59:55
aws-amplify/amplify-android
https://api.github.com/repos/aws-amplify/amplify-android
closed
Facing gradle build failed in Windows 10 even though working fine on Mac
Developer Tools Feature Request Improvement Infrastructure
When I try to run ```gradlew build``` facing build Failed in windows 10 and working fine in Mac this is the logs I get after running ```gradlew build``` ```Microsoft Windows [Version 10.0.19041.508] (c) 2020 Microsoft Corporation. All rights reserved. C:\MyData\Amplify original\amplify-android>gradlew build --stacktrace Welcome to Gradle 6.6.1! Here are the highlights of this release: - Experimental build configuration caching - Built-in conventions for handling credentials - Java compilation supports --release flag For more details see https://docs.gradle.org/6.6.1/release-notes.html Starting a Gradle Daemon, 1 incompatible Daemon could not be reused, use --status for details > Task :aws-analytics-pinpoint:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:50: Using a static member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:51: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:52: Using a static member import should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:53: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:54: Using a static member import should be avoided - org.junit.Assert.fail. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:139:57: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:140:58: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:157:22: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:158:22: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:273:27: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:274:28: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:299:31: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:300:32: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :30: Using a static member import should be avoided - org.mockito.Mockito.inOrder. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :31: Using a static member import should be avoided - org.mockito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :32: Using a static member import should be avoided - org.mockito.Mockito.never. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :33: Using a static member import should be avoided - org.mockito.Mockito.verify. [AvoidStaticImport] > Task :aws-analytics-pinpoint:checkstyle FAILED FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':aws-analytics-pinpoint:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-analytics-pinpoint/build/reports/checkstyle/checkstyle. html Checkstyle files with violations: 2 Checkstyle violations by severity: [error:17] * Try: Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Exception is: org.gradle.api.tasks.TaskExecutionException: Execution failed for task ':aws-analytics-pinpoint:checkstyle'. at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.lambda$executeIfValid$1(ExecuteActionsTaskExecuter.java:208) at org.gradle.internal.Try$Failure.ifSuccessfulOrElse(Try.java:263) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeIfValid(ExecuteActionsTaskExecuter.java:206) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.execute(ExecuteActionsTaskExecuter.java:187) at org.gradle.api.internal.tasks.execution.CleanupStaleOutputsExecuter.execute(CleanupStaleOutputsExecuter.java:114) at org.gradle.api.internal.tasks.execution.FinalizePropertiesTaskExecuter.execute(FinalizePropertiesTaskExecuter.java:46) at org.gradle.api.internal.tasks.execution.ResolveTaskExecutionModeExecuter.execute(ResolveTaskExecutionModeExecuter.java:62) at org.gradle.api.internal.tasks.execution.SkipTaskWithNoActionsExecuter.execute(SkipTaskWithNoActionsExecuter.java:57) at org.gradle.api.internal.tasks.execution.SkipOnlyIfTaskExecuter.execute(SkipOnlyIfTaskExecuter.java:56) at org.gradle.api.internal.tasks.execution.CatchExceptionTaskExecuter.execute(CatchExceptionTaskExecuter.java:36) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.executeTask(EventFiringTaskExecuter.java:77) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:55) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:52) at org.gradle.internal.operations.DefaultBuildOperationExecutor$CallableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:409) at org.gradle.internal.operations.DefaultBuildOperationExecutor$CallableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:399) at org.gradle.internal.operations.DefaultBuildOperationExecutor$1.execute(DefaultBuildOperationExecutor.java:157) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:242) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:150) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:94) at org.gradle.internal.operations.DelegatingBuildOperationExecutor.call(DelegatingBuildOperationExecutor.java:36) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter.execute(EventFiringTaskExecuter.java:52) at org.gradle.execution.plan.LocalTaskNodeExecutor.execute(LocalTaskNodeExecutor.java:41) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:372) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:359) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:352) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:338) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.lambda$run$0(DefaultPlanExecutor.java:127) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.execute(DefaultPlanExecutor.java:191) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.executeNextNode(DefaultPlanExecutor.java:182) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.run(DefaultPlanExecutor.java:124) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) Caused by: org.gradle.api.GradleException: Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-analytics-pinp oint/build/reports/checkstyle/checkstyle.html Checkstyle files with violations: 2 Checkstyle violations by severity: [error:17] at org.gradle.api.plugins.quality.internal.CheckstyleInvoker$_invoke_closure1.doCall(CheckstyleInvoker.groovy:106) at org.gradle.util.ClosureBackedAction.execute(ClosureBackedAction.java:71) at org.gradle.util.ClosureBackedAction.execute(ClosureBackedAction.java:52) at org.gradle.api.internal.project.antbuilder.DefaultIsolatedAntBuilder$2.execute(DefaultIsolatedAntBuilder.java:154) at org.gradle.api.internal.project.antbuilder.DefaultIsolatedAntBuilder$2.execute(DefaultIsolatedAntBuilder.java:136) at org.gradle.api.internal.project.antbuilder.ClassPathToClassLoaderCache.withCachedClassLoader(ClassPathToClassLoaderCache.java:135) at org.gradle.api.internal.project.antbuilder.DefaultIsolatedAntBuilder.execute(DefaultIsolatedAntBuilder.java:130) at org.gradle.api.internal.project.IsolatedAntBuilder$execute$0.call(Unknown Source) at org.gradle.api.plugins.quality.internal.CheckstyleInvoker.invoke(CheckstyleInvoker.groovy:51) at org.gradle.api.plugins.quality.Checkstyle.run(Checkstyle.java:151) at org.gradle.internal.reflect.JavaMethod.invoke(JavaMethod.java:104) at org.gradle.api.internal.project.taskfactory.StandardTaskAction.doExecute(StandardTaskAction.java:58) at org.gradle.api.internal.project.taskfactory.StandardTaskAction.execute(StandardTaskAction.java:51) at org.gradle.api.internal.project.taskfactory.StandardTaskAction.execute(StandardTaskAction.java:29) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$3.run(ExecuteActionsTaskExecuter.java:570) at org.gradle.internal.operations.DefaultBuildOperationExecutor$RunnableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:395) at org.gradle.internal.operations.DefaultBuildOperationExecutor$RunnableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:387) at org.gradle.internal.operations.DefaultBuildOperationExecutor$1.execute(DefaultBuildOperationExecutor.java:157) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:242) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:150) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:84) at org.gradle.internal.operations.DelegatingBuildOperationExecutor.run(DelegatingBuildOperationExecutor.java:31) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeAction(ExecuteActionsTaskExecuter.java:555) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeActions(ExecuteActionsTaskExecuter.java:538) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.access$300(ExecuteActionsTaskExecuter.java:109) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$TaskExecution.executeWithPreviousOutputFiles(ExecuteActionsTaskExecuter.java:279) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$TaskExecution.execute(ExecuteActionsTaskExecuter.java:268) at org.gradle.internal.execution.steps.ExecuteStep.lambda$execute$1(ExecuteStep.java:33) at org.gradle.internal.execution.steps.ExecuteStep.execute(ExecuteStep.java:33) at org.gradle.internal.execution.steps.ExecuteStep.execute(ExecuteStep.java:26) at org.gradle.internal.execution.steps.CleanupOutputsStep.execute(CleanupOutputsStep.java:67) at org.gradle.internal.execution.steps.CleanupOutputsStep.execute(CleanupOutputsStep.java:36) at org.gradle.internal.execution.steps.ResolveInputChangesStep.execute(ResolveInputChangesStep.java:49) at org.gradle.internal.execution.steps.ResolveInputChangesStep.execute(ResolveInputChangesStep.java:34) at org.gradle.internal.execution.steps.CancelExecutionStep.execute(CancelExecutionStep.java:43) at org.gradle.internal.execution.steps.TimeoutStep.executeWithoutTimeout(TimeoutStep.java:73) at org.gradle.internal.execution.steps.TimeoutStep.execute(TimeoutStep.java:54) at org.gradle.internal.execution.steps.CatchExceptionStep.execute(CatchExceptionStep.java:34) at org.gradle.internal.execution.steps.CreateOutputsStep.execute(CreateOutputsStep.java:44) at org.gradle.internal.execution.steps.SnapshotOutputsStep.execute(SnapshotOutputsStep.java:54) at org.gradle.internal.execution.steps.SnapshotOutputsStep.execute(SnapshotOutputsStep.java:38) at org.gradle.internal.execution.steps.BroadcastChangingOutputsStep.execute(BroadcastChangingOutputsStep.java:49) at org.gradle.internal.execution.steps.CacheStep.executeWithoutCache(CacheStep.java:159) at org.gradle.internal.execution.steps.CacheStep.execute(CacheStep.java:72) at org.gradle.internal.execution.steps.CacheStep.execute(CacheStep.java:43) at org.gradle.internal.execution.steps.StoreExecutionStateStep.execute(StoreExecutionStateStep.java:44) at org.gradle.internal.execution.steps.StoreExecutionStateStep.execute(StoreExecutionStateStep.java:33) at org.gradle.internal.execution.steps.RecordOutputsStep.execute(RecordOutputsStep.java:38) at org.gradle.internal.execution.steps.RecordOutputsStep.execute(RecordOutputsStep.java:24) at org.gradle.internal.execution.steps.SkipUpToDateStep.executeBecause(SkipUpToDateStep.java:92) at org.gradle.internal.execution.steps.SkipUpToDateStep.lambda$execute$0(SkipUpToDateStep.java:85) at org.gradle.internal.execution.steps.SkipUpToDateStep.execute(SkipUpToDateStep.java:55) at org.gradle.internal.execution.steps.SkipUpToDateStep.execute(SkipUpToDateStep.java:39) at org.gradle.internal.execution.steps.ResolveChangesStep.execute(ResolveChangesStep.java:76) at org.gradle.internal.execution.steps.ResolveChangesStep.execute(ResolveChangesStep.java:37) at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsFinishedStep.execute(MarkSnapshottingInputsFinishedStep.java:36) at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsFinishedStep.execute(MarkSnapshottingInputsFinishedStep.java:26) at org.gradle.internal.execution.steps.ResolveCachingStateStep.execute(ResolveCachingStateStep.java:94) at org.gradle.internal.execution.steps.ResolveCachingStateStep.execute(ResolveCachingStateStep.java:49) at org.gradle.internal.execution.steps.CaptureStateBeforeExecutionStep.execute(CaptureStateBeforeExecutionStep.java:79) at org.gradle.internal.execution.steps.CaptureStateBeforeExecutionStep.execute(CaptureStateBeforeExecutionStep.java:53) at org.gradle.internal.execution.steps.ValidateStep.execute(ValidateStep.java:74) at org.gradle.internal.execution.steps.SkipEmptyWorkStep.lambda$execute$2(SkipEmptyWorkStep.java:78) at org.gradle.internal.execution.steps.SkipEmptyWorkStep.execute(SkipEmptyWorkStep.java:78) at org.gradle.internal.execution.steps.SkipEmptyWorkStep.execute(SkipEmptyWorkStep.java:34) at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsStartedStep.execute(MarkSnapshottingInputsStartedStep.java:39) at org.gradle.internal.execution.steps.LoadExecutionStateStep.execute(LoadExecutionStateStep.java:40) at org.gradle.internal.execution.steps.LoadExecutionStateStep.execute(LoadExecutionStateStep.java:28) at org.gradle.internal.execution.impl.DefaultWorkExecutor.execute(DefaultWorkExecutor.java:33) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeIfValid(ExecuteActionsTaskExecuter.java:195) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.execute(ExecuteActionsTaskExecuter.java:187) at org.gradle.api.internal.tasks.execution.CleanupStaleOutputsExecuter.execute(CleanupStaleOutputsExecuter.java:114) at org.gradle.api.internal.tasks.execution.FinalizePropertiesTaskExecuter.execute(FinalizePropertiesTaskExecuter.java:46) at org.gradle.api.internal.tasks.execution.ResolveTaskExecutionModeExecuter.execute(ResolveTaskExecutionModeExecuter.java:62) at org.gradle.api.internal.tasks.execution.SkipTaskWithNoActionsExecuter.execute(SkipTaskWithNoActionsExecuter.java:57) at org.gradle.api.internal.tasks.execution.SkipOnlyIfTaskExecuter.execute(SkipOnlyIfTaskExecuter.java:56) at org.gradle.api.internal.tasks.execution.CatchExceptionTaskExecuter.execute(CatchExceptionTaskExecuter.java:36) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.executeTask(EventFiringTaskExecuter.java:77) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:55) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:52) at org.gradle.internal.operations.DefaultBuildOperationExecutor$CallableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:409) at org.gradle.internal.operations.DefaultBuildOperationExecutor$CallableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:399) at org.gradle.internal.operations.DefaultBuildOperationExecutor$1.execute(DefaultBuildOperationExecutor.java:157) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:242) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:150) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:94) at org.gradle.internal.operations.DelegatingBuildOperationExecutor.call(DelegatingBuildOperationExecutor.java:36) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter.execute(EventFiringTaskExecuter.java:52) at org.gradle.execution.plan.LocalTaskNodeExecutor.execute(LocalTaskNodeExecutor.java:41) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:372) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:359) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:352) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:338) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.lambda$run$0(DefaultPlanExecutor.java:127) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.execute(DefaultPlanExecutor.java:191) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.executeNextNode(DefaultPlanExecutor.java:182) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.run(DefaultPlanExecutor.java:124) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 37s 43 actionable tasks: 43 executed C:\MyData\Amplify original\amplify-android>gradlew build > Task :aws-analytics-pinpoint:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:50: Using a static member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:51: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:52: Using a static member import should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:53: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:54: Using a static member import should be avoided - org.junit.Assert.fail. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:139:57: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:140:58: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:157:22: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:158:22: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:273:27: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:274:28: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:299:31: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:300:32: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :30: Using a static member import should be avoided - org.mockito.Mockito.inOrder. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :31: Using a static member import should be avoided - org.mockito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :32: Using a static member import should be avoided - org.mockito.Mockito.never. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :33: Using a static member import should be avoided - org.mockito.Mockito.verify. [AvoidStaticImport] > Task :aws-analytics-pinpoint:checkstyle FAILED FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':aws-analytics-pinpoint:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-analytics-pinpoint/build/reports/checkstyle/checkstyle. html Checkstyle files with violations: 2 Checkstyle violations by severity: [error:17] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 13s 64 actionable tasks: 22 executed, 42 up-to-date C:\MyData\Amplify original\amplify-android>gradlew build --scan > Task :aws-analytics-pinpoint:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:50: Using a static member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:51: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:52: Using a static member import should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:53: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:54: Using a static member import should be avoided - org.junit.Assert.fail. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:139:57: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:140:58: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:157:22: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:158:22: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:273:27: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:274:28: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:299:31: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:300:32: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :30: Using a static member import should be avoided - org.mockito.Mockito.inOrder. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :31: Using a static member import should be avoided - org.mockito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :32: Using a static member import should be avoided - org.mockito.Mockito.never. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :33: Using a static member import should be avoided - org.mockito.Mockito.verify. [AvoidStaticImport] > Task :aws-analytics-pinpoint:checkstyle FAILED > Task :aws-api:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:47: Usi ng a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:48: Usi ng a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:84:18: '29' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:124:42: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:184:18: '29' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:192:22: '30' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:302:20: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\GraphQLInstrumentationTest.java:42: Using a st atic member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\GraphQLInstrumentationTest.java:43: Using a st atic member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:35: Using a st atic member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:36: Using a st atic member import should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:37: Using a st atic member import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:38: Using a st atic member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:41: Using a stat ic member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:42: Using a stat ic member import should be avoided - org.junit.Assert.assertNotEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:43: Using a stat ic member import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\TestApiCategory.java:32: Using a static member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:27: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:28: Using a static member import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:29: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:74: Using a static member impor t should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:75: Using a static member impor t should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:76: Using a static member impor t should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:77: Using a static member impor t should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:78: Using a static member impor t should be avoided - org.mockito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:79: Using a static member impor t should be avoided - org.mockito.Mockito.when. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:100:25: '8080' is a magic numbe r. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:39: Using a static mem ber import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:40: Using a static mem ber import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:41: Using a static mem ber import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:42: Using a static mem ber import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:100:19: '16' is a magi c number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:121:19: '28' is a magi c number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\ApiSelectorTest.java:26: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:103:18: '19' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:104:27: '2000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:104:36: '15' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:155:51: '123456 7890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncSigV4SignerInterceptorFactoryTest.java:31: Usi ng a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:47: Using a stati c member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:48: Using a stati c member import should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:49: Using a stati c member import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:143:29: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:146:41: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:146:44: '7' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:197:29: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:200:41: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:200:44: '7' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:253:37: '11' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:253:41: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:382:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:390:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:398:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:406:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:414:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:422:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:29: Using a static member import sh ould be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:47:17: '3.5f' is a magic number. [M agicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:62:35: '4' is a magic number. [Magi cNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:63:33: '5.5f' is a magic number. [M agicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\SubscriptionAuthorizerTest.java:32: Using a static me mber import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:28: Using a static member im port should be avoided - org.mockito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:29: Using a static member im port should be avoided - org.mockito.Mockito.verify. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:30: Using a static member im port should be avoided - org.mockito.Mockito.verifyNoInteractions. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TypeMakerTest.java:26: Using a static member import s hould be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TypeMakerTest.java:27: Using a static member import s hould be avoided - org.junit.Assert.assertThrows. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\utils\RestRequestFactoryTest.java:30: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\utils\RestRequestFactoryTest.java:31: Using a static member import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] > Task :aws-api:checkstyle FAILED FAILURE: Build completed with 2 failures. 1: Task failed with an exception. ----------- * What went wrong: Execution failed for task ':aws-analytics-pinpoint:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-analytics-pinpoint/build/reports/checkstyle/checkstyle. html Checkstyle files with violations: 2 Checkstyle violations by severity: [error:17] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== 2: Task failed with an exception. ----------- * What went wrong: Execution failed for task ':aws-api:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-api/build/reports/checkstyle/checkstyle.html Checkstyle files with violations: 17 Checkstyle violations by severity: [error:68] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings Publishing a build scan to scans.gradle.com requires accepting the Gradle Terms of Service defined at https://gradle.com/terms-of-service. Do you accept these terms? [yes, no] yGradle Terms of Service accepted. Publishing build scan... https://gradle.com/s/mbezreddtb23y C:\MyData\Amplify original\amplify-android>gradlew build --status PID STATUS INFO 15212 IDLE 6.6.1 14376 IDLE 6.6.1 Only Daemons for the current Gradle version are displayed. See https://docs.gradle.org/6.6.1/userguide/gradle_daemon.html#sec:status C:\MyData\Amplify original\amplify-android>clear 'clear' is not recognized as an internal or external command, operable program or batch file. C:\MyData\Amplify original\amplify-android>cr 'cr' is not recognized as an internal or external command, operable program or batch file. C:\MyData\Amplify original\amplify-android>cls C:\MyData\Amplify original\amplify-android>gradlew build --scan > Task :aws-analytics-pinpoint:checkstyle FAILED [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:50: Using a static member import should b e avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:51: Using a static member import should b e avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:52: Using a static member import should b e avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:53: Using a static member import should b e avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:54: Using a static member import should b e avoided - org.junit.Assert.fail. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:139:57: '3.14' is a magic number. [MagicN umber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:140:58: '42' is a magic number. [MagicNum ber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:157:22: '3.14' is a magic number. [MagicN umber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:158:22: '42' is a magic number. [MagicNum ber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:273:27: '47.6154086' is a magic number. [ MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:274:28: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:299:31: '47.6154086' is a magic number. [ MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:300:32: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java:30: Using a static member import should be avoided - org.mo ckito.Mockito.inOrder. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java:31: Using a static member import should be avoided - org.mo ckito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java:32: Using a static member import should be avoided - org.mo ckito.Mockito.never. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java:33: Using a static member import should be avoided - org.mo ckito.Mockito.verify. [AvoidStaticImport] > Task :aws-api-appsync:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQlRequestTest.java:26: Using a static member import should be avoided - org.junit.Assert.asse rtEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQlRequestTest.java:28: Missing a Javadoc comment. [MissingJavadocType] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\api\aws\DefaultGraphQLRequestOptions.java:21: Missing a Javadoc comment. [MissingJavadocType] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\api\aws\SelectionSetTest.java:28: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\api\aws\SelectionSetTest.java:30: Missing a Javadoc comment. [MissingJavadocType] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:28: Using a static member import should be avoided - org.junit.Assert.a ssertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:29: Using a static member import should be avoided - org.junit.Assert.a ssertThrows. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:68:32: '2001' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:70:40: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:90:32: '2001' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:92:40: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:27: Using a static member import should be avoided - org.junit.Asse rt.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:68:46: '2001' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:68:55: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:68:58: '4' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:68:61: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:68:64: '6' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:73:48: '1000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:28: Using a static member import should be avoided - org.junit.Assert.a ssertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:29: Using a static member import should be avoided - org.junit.Assert.a ssertThrows. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:73:32: '1970' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:78:34: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:79:39: '4' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:97:32: '1970' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:102:34: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:103:39: '4' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:26: Using a static member import should be avoided - org.junit.Ass ert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:27: Using a static member import should be avoided - org.junit.Ass ert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:78:59: '5_000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:79:60: '5_000_000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:92:59: '5_000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:93:60: '6_000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:94:59: '7_000' is a magic number. [MagicNumber] > Task :aws-api-appsync:checkstyle FAILED > Task :aws-api:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:47: Using a static member import should be avoided - org.junit.Asse rt.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:48: Using a static member import should be avoided - org.junit.Asse rt.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:84:18: '29' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:124:42: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:184:18: '29' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:192:22: '30' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:302:20: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\GraphQLInstrumentationTest.java:42: Using a static member import should be avoided - androidx.test.core.ap p.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\GraphQLInstrumentationTest.java:43: Using a static member import should be avoided - org.junit.Assert.asse rtEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:35: Using a static member import should be avoided - org.junit.Assert.asse rtEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:36: Using a static member import should be avoided - org.junit.Assert.asse rtFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:37: Using a static member import should be avoided - org.junit.Assert.asse rtNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:38: Using a static member import should be avoided - org.junit.Assert.asse rtTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:41: Using a static member import should be avoided - androidx.test.core.app. ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:42: Using a static member import should be avoided - org.junit.Assert.assert NotEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:43: Using a static member import should be avoided - org.junit.Assert.assert NotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\TestApiCategory.java:32: Using a static member import should be avoided - androidx.test.core.app.Applicati onProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:27: Using a static member import should be avoided - org.junit.Assert.as sertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:28: Using a static member import should be avoided - org.junit.Assert.as sertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:29: Using a static member import should be avoided - org.junit.Assert.as sertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:74: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidS taticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:75: Using a static member import should be avoided - org.junit.Assert.assertFalse. [AvoidSt aticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:76: Using a static member import should be avoided - org.junit.Assert.assertNotNull. [Avoid StaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:77: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidSta ticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:78: Using a static member import should be avoided - org.mockito.Mockito.mock. [AvoidStatic Import] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:79: Using a static member import should be avoided - org.mockito.Mockito.when. [AvoidStatic Import] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:100:25: '8080' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:39: Using a static member import should be avoided - androidx.test.core.app.Applic ationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:40: Using a static member import should be avoided - org.junit.Assert.assertEquals . [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:41: Using a static member import should be avoided - org.junit.Assert.assertNotNul l. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:42: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:100:19: '16' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:121:19: '28' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\ApiSelectorTest.java:26: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidSt aticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:103:18: '19' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:104:27: '2000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:104:36: '15' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:155:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncSigV4SignerInterceptorFactoryTest.java:31: Using a static member import should be avoided - org.junit.Asse rt.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:47: Using a static member import should be avoided - org.junit.Assert.assertE quals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:48: Using a static member import should be avoided - org.junit.Assert.assertF alse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:49: Using a static member import should be avoided - org.junit.Assert.assertN otNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:143:29: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:146:41: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:146:44: '7' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:197:29: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:200:41: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:200:44: '7' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:253:37: '11' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:253:41: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:382:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:390:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:398:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:406:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:414:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:422:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:29: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStati cImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:47:17: '3.5f' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:62:35: '4' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:63:33: '5.5f' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\SubscriptionAuthorizerTest.java:32: Using a static member import should be avoided - org.junit.Assert.assertEqual s. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:28: Using a static member import should be avoided - org.mockito.Mockito.mock. [AvoidSta ticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:29: Using a static member import should be avoided - org.mockito.Mockito.verify. [AvoidS taticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:30: Using a static member import should be avoided - org.mockito.Mockito.verifyNoInterac tions. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TypeMakerTest.java:26: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStat icImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TypeMakerTest.java:27: Using a static member import should be avoided - org.junit.Assert.assertThrows. [AvoidStat icImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\utils\RestRequestFactoryTest.java:30: Using a static member import should be avoided - org.junit.Assert.assertEqu als. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\utils\RestRequestFactoryTest.java:31: Using a static member import should be avoided - org.junit.Assert.assertNot Null. [AvoidStaticImport] > Task :aws-api:checkstyle FAILED FAILURE: Build completed with 3 failures. 1: Task failed with an exception. ----------- * What went wrong: Execution failed for task ':aws-analytics-pinpoint:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-analytics-pinpoint/build/reports/checkstyle/checkstyle.html Checkstyle files with violations: 2 Checkstyle violations by severity: [error:17] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== 2: Task failed with an exception. ----------- * What went wrong: Execution failed for task ':aws-api-appsync:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-api-appsync/build/reports/checkstyle/checkstyle.html Checkstyle files with violations: 7 Checkstyle violations by severity: [error:33] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== 3: Task failed with an exception. ----------- * What went wrong: Execution failed for task ':aws-api:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-api/build/reports/checkstyle/checkstyle.html Checkstyle files with violations: 17 Checkstyle violations by severity: [error:68] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 12s 236 actionable tasks: 17 executed, 219 up-to-date Publishing a build scan to scans.gradle.com requires accepting the Gradle Terms of Service defined at https://gradle.com/terms-of-service. Do you accept these terms? [yes, no] yes Gradle Terms of Service accepted. Publishing build scan... https://gradle.com/s/n5yh2atjmu274 <-------------> 0% WAITING > IDLE > IDLE > IDLE > IDLE C:\MyData\Amplify original\amplify-android>```
1.0
Facing gradle build failed in Windows 10 even though working fine on Mac - When I try to run ```gradlew build``` facing build Failed in windows 10 and working fine in Mac this is the logs I get after running ```gradlew build``` ```Microsoft Windows [Version 10.0.19041.508] (c) 2020 Microsoft Corporation. All rights reserved. C:\MyData\Amplify original\amplify-android>gradlew build --stacktrace Welcome to Gradle 6.6.1! Here are the highlights of this release: - Experimental build configuration caching - Built-in conventions for handling credentials - Java compilation supports --release flag For more details see https://docs.gradle.org/6.6.1/release-notes.html Starting a Gradle Daemon, 1 incompatible Daemon could not be reused, use --status for details > Task :aws-analytics-pinpoint:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:50: Using a static member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:51: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:52: Using a static member import should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:53: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:54: Using a static member import should be avoided - org.junit.Assert.fail. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:139:57: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:140:58: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:157:22: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:158:22: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:273:27: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:274:28: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:299:31: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:300:32: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :30: Using a static member import should be avoided - org.mockito.Mockito.inOrder. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :31: Using a static member import should be avoided - org.mockito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :32: Using a static member import should be avoided - org.mockito.Mockito.never. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :33: Using a static member import should be avoided - org.mockito.Mockito.verify. [AvoidStaticImport] > Task :aws-analytics-pinpoint:checkstyle FAILED FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':aws-analytics-pinpoint:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-analytics-pinpoint/build/reports/checkstyle/checkstyle. html Checkstyle files with violations: 2 Checkstyle violations by severity: [error:17] * Try: Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Exception is: org.gradle.api.tasks.TaskExecutionException: Execution failed for task ':aws-analytics-pinpoint:checkstyle'. at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.lambda$executeIfValid$1(ExecuteActionsTaskExecuter.java:208) at org.gradle.internal.Try$Failure.ifSuccessfulOrElse(Try.java:263) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeIfValid(ExecuteActionsTaskExecuter.java:206) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.execute(ExecuteActionsTaskExecuter.java:187) at org.gradle.api.internal.tasks.execution.CleanupStaleOutputsExecuter.execute(CleanupStaleOutputsExecuter.java:114) at org.gradle.api.internal.tasks.execution.FinalizePropertiesTaskExecuter.execute(FinalizePropertiesTaskExecuter.java:46) at org.gradle.api.internal.tasks.execution.ResolveTaskExecutionModeExecuter.execute(ResolveTaskExecutionModeExecuter.java:62) at org.gradle.api.internal.tasks.execution.SkipTaskWithNoActionsExecuter.execute(SkipTaskWithNoActionsExecuter.java:57) at org.gradle.api.internal.tasks.execution.SkipOnlyIfTaskExecuter.execute(SkipOnlyIfTaskExecuter.java:56) at org.gradle.api.internal.tasks.execution.CatchExceptionTaskExecuter.execute(CatchExceptionTaskExecuter.java:36) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.executeTask(EventFiringTaskExecuter.java:77) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:55) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:52) at org.gradle.internal.operations.DefaultBuildOperationExecutor$CallableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:409) at org.gradle.internal.operations.DefaultBuildOperationExecutor$CallableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:399) at org.gradle.internal.operations.DefaultBuildOperationExecutor$1.execute(DefaultBuildOperationExecutor.java:157) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:242) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:150) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:94) at org.gradle.internal.operations.DelegatingBuildOperationExecutor.call(DelegatingBuildOperationExecutor.java:36) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter.execute(EventFiringTaskExecuter.java:52) at org.gradle.execution.plan.LocalTaskNodeExecutor.execute(LocalTaskNodeExecutor.java:41) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:372) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:359) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:352) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:338) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.lambda$run$0(DefaultPlanExecutor.java:127) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.execute(DefaultPlanExecutor.java:191) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.executeNextNode(DefaultPlanExecutor.java:182) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.run(DefaultPlanExecutor.java:124) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) Caused by: org.gradle.api.GradleException: Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-analytics-pinp oint/build/reports/checkstyle/checkstyle.html Checkstyle files with violations: 2 Checkstyle violations by severity: [error:17] at org.gradle.api.plugins.quality.internal.CheckstyleInvoker$_invoke_closure1.doCall(CheckstyleInvoker.groovy:106) at org.gradle.util.ClosureBackedAction.execute(ClosureBackedAction.java:71) at org.gradle.util.ClosureBackedAction.execute(ClosureBackedAction.java:52) at org.gradle.api.internal.project.antbuilder.DefaultIsolatedAntBuilder$2.execute(DefaultIsolatedAntBuilder.java:154) at org.gradle.api.internal.project.antbuilder.DefaultIsolatedAntBuilder$2.execute(DefaultIsolatedAntBuilder.java:136) at org.gradle.api.internal.project.antbuilder.ClassPathToClassLoaderCache.withCachedClassLoader(ClassPathToClassLoaderCache.java:135) at org.gradle.api.internal.project.antbuilder.DefaultIsolatedAntBuilder.execute(DefaultIsolatedAntBuilder.java:130) at org.gradle.api.internal.project.IsolatedAntBuilder$execute$0.call(Unknown Source) at org.gradle.api.plugins.quality.internal.CheckstyleInvoker.invoke(CheckstyleInvoker.groovy:51) at org.gradle.api.plugins.quality.Checkstyle.run(Checkstyle.java:151) at org.gradle.internal.reflect.JavaMethod.invoke(JavaMethod.java:104) at org.gradle.api.internal.project.taskfactory.StandardTaskAction.doExecute(StandardTaskAction.java:58) at org.gradle.api.internal.project.taskfactory.StandardTaskAction.execute(StandardTaskAction.java:51) at org.gradle.api.internal.project.taskfactory.StandardTaskAction.execute(StandardTaskAction.java:29) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$3.run(ExecuteActionsTaskExecuter.java:570) at org.gradle.internal.operations.DefaultBuildOperationExecutor$RunnableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:395) at org.gradle.internal.operations.DefaultBuildOperationExecutor$RunnableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:387) at org.gradle.internal.operations.DefaultBuildOperationExecutor$1.execute(DefaultBuildOperationExecutor.java:157) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:242) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:150) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:84) at org.gradle.internal.operations.DelegatingBuildOperationExecutor.run(DelegatingBuildOperationExecutor.java:31) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeAction(ExecuteActionsTaskExecuter.java:555) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeActions(ExecuteActionsTaskExecuter.java:538) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.access$300(ExecuteActionsTaskExecuter.java:109) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$TaskExecution.executeWithPreviousOutputFiles(ExecuteActionsTaskExecuter.java:279) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter$TaskExecution.execute(ExecuteActionsTaskExecuter.java:268) at org.gradle.internal.execution.steps.ExecuteStep.lambda$execute$1(ExecuteStep.java:33) at org.gradle.internal.execution.steps.ExecuteStep.execute(ExecuteStep.java:33) at org.gradle.internal.execution.steps.ExecuteStep.execute(ExecuteStep.java:26) at org.gradle.internal.execution.steps.CleanupOutputsStep.execute(CleanupOutputsStep.java:67) at org.gradle.internal.execution.steps.CleanupOutputsStep.execute(CleanupOutputsStep.java:36) at org.gradle.internal.execution.steps.ResolveInputChangesStep.execute(ResolveInputChangesStep.java:49) at org.gradle.internal.execution.steps.ResolveInputChangesStep.execute(ResolveInputChangesStep.java:34) at org.gradle.internal.execution.steps.CancelExecutionStep.execute(CancelExecutionStep.java:43) at org.gradle.internal.execution.steps.TimeoutStep.executeWithoutTimeout(TimeoutStep.java:73) at org.gradle.internal.execution.steps.TimeoutStep.execute(TimeoutStep.java:54) at org.gradle.internal.execution.steps.CatchExceptionStep.execute(CatchExceptionStep.java:34) at org.gradle.internal.execution.steps.CreateOutputsStep.execute(CreateOutputsStep.java:44) at org.gradle.internal.execution.steps.SnapshotOutputsStep.execute(SnapshotOutputsStep.java:54) at org.gradle.internal.execution.steps.SnapshotOutputsStep.execute(SnapshotOutputsStep.java:38) at org.gradle.internal.execution.steps.BroadcastChangingOutputsStep.execute(BroadcastChangingOutputsStep.java:49) at org.gradle.internal.execution.steps.CacheStep.executeWithoutCache(CacheStep.java:159) at org.gradle.internal.execution.steps.CacheStep.execute(CacheStep.java:72) at org.gradle.internal.execution.steps.CacheStep.execute(CacheStep.java:43) at org.gradle.internal.execution.steps.StoreExecutionStateStep.execute(StoreExecutionStateStep.java:44) at org.gradle.internal.execution.steps.StoreExecutionStateStep.execute(StoreExecutionStateStep.java:33) at org.gradle.internal.execution.steps.RecordOutputsStep.execute(RecordOutputsStep.java:38) at org.gradle.internal.execution.steps.RecordOutputsStep.execute(RecordOutputsStep.java:24) at org.gradle.internal.execution.steps.SkipUpToDateStep.executeBecause(SkipUpToDateStep.java:92) at org.gradle.internal.execution.steps.SkipUpToDateStep.lambda$execute$0(SkipUpToDateStep.java:85) at org.gradle.internal.execution.steps.SkipUpToDateStep.execute(SkipUpToDateStep.java:55) at org.gradle.internal.execution.steps.SkipUpToDateStep.execute(SkipUpToDateStep.java:39) at org.gradle.internal.execution.steps.ResolveChangesStep.execute(ResolveChangesStep.java:76) at org.gradle.internal.execution.steps.ResolveChangesStep.execute(ResolveChangesStep.java:37) at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsFinishedStep.execute(MarkSnapshottingInputsFinishedStep.java:36) at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsFinishedStep.execute(MarkSnapshottingInputsFinishedStep.java:26) at org.gradle.internal.execution.steps.ResolveCachingStateStep.execute(ResolveCachingStateStep.java:94) at org.gradle.internal.execution.steps.ResolveCachingStateStep.execute(ResolveCachingStateStep.java:49) at org.gradle.internal.execution.steps.CaptureStateBeforeExecutionStep.execute(CaptureStateBeforeExecutionStep.java:79) at org.gradle.internal.execution.steps.CaptureStateBeforeExecutionStep.execute(CaptureStateBeforeExecutionStep.java:53) at org.gradle.internal.execution.steps.ValidateStep.execute(ValidateStep.java:74) at org.gradle.internal.execution.steps.SkipEmptyWorkStep.lambda$execute$2(SkipEmptyWorkStep.java:78) at org.gradle.internal.execution.steps.SkipEmptyWorkStep.execute(SkipEmptyWorkStep.java:78) at org.gradle.internal.execution.steps.SkipEmptyWorkStep.execute(SkipEmptyWorkStep.java:34) at org.gradle.internal.execution.steps.legacy.MarkSnapshottingInputsStartedStep.execute(MarkSnapshottingInputsStartedStep.java:39) at org.gradle.internal.execution.steps.LoadExecutionStateStep.execute(LoadExecutionStateStep.java:40) at org.gradle.internal.execution.steps.LoadExecutionStateStep.execute(LoadExecutionStateStep.java:28) at org.gradle.internal.execution.impl.DefaultWorkExecutor.execute(DefaultWorkExecutor.java:33) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeIfValid(ExecuteActionsTaskExecuter.java:195) at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.execute(ExecuteActionsTaskExecuter.java:187) at org.gradle.api.internal.tasks.execution.CleanupStaleOutputsExecuter.execute(CleanupStaleOutputsExecuter.java:114) at org.gradle.api.internal.tasks.execution.FinalizePropertiesTaskExecuter.execute(FinalizePropertiesTaskExecuter.java:46) at org.gradle.api.internal.tasks.execution.ResolveTaskExecutionModeExecuter.execute(ResolveTaskExecutionModeExecuter.java:62) at org.gradle.api.internal.tasks.execution.SkipTaskWithNoActionsExecuter.execute(SkipTaskWithNoActionsExecuter.java:57) at org.gradle.api.internal.tasks.execution.SkipOnlyIfTaskExecuter.execute(SkipOnlyIfTaskExecuter.java:56) at org.gradle.api.internal.tasks.execution.CatchExceptionTaskExecuter.execute(CatchExceptionTaskExecuter.java:36) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.executeTask(EventFiringTaskExecuter.java:77) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:55) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter$1.call(EventFiringTaskExecuter.java:52) at org.gradle.internal.operations.DefaultBuildOperationExecutor$CallableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:409) at org.gradle.internal.operations.DefaultBuildOperationExecutor$CallableBuildOperationWorker.execute(DefaultBuildOperationExecutor.java:399) at org.gradle.internal.operations.DefaultBuildOperationExecutor$1.execute(DefaultBuildOperationExecutor.java:157) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:242) at org.gradle.internal.operations.DefaultBuildOperationExecutor.execute(DefaultBuildOperationExecutor.java:150) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:94) at org.gradle.internal.operations.DelegatingBuildOperationExecutor.call(DelegatingBuildOperationExecutor.java:36) at org.gradle.api.internal.tasks.execution.EventFiringTaskExecuter.execute(EventFiringTaskExecuter.java:52) at org.gradle.execution.plan.LocalTaskNodeExecutor.execute(LocalTaskNodeExecutor.java:41) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:372) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$InvokeNodeExecutorsAction.execute(DefaultTaskExecutionGraph.java:359) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:352) at org.gradle.execution.taskgraph.DefaultTaskExecutionGraph$BuildOperationAwareExecutionAction.execute(DefaultTaskExecutionGraph.java:338) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.lambda$run$0(DefaultPlanExecutor.java:127) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.execute(DefaultPlanExecutor.java:191) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.executeNextNode(DefaultPlanExecutor.java:182) at org.gradle.execution.plan.DefaultPlanExecutor$ExecutorWorker.run(DefaultPlanExecutor.java:124) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 37s 43 actionable tasks: 43 executed C:\MyData\Amplify original\amplify-android>gradlew build > Task :aws-analytics-pinpoint:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:50: Using a static member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:51: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:52: Using a static member import should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:53: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:54: Using a static member import should be avoided - org.junit.Assert.fail. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:139:57: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:140:58: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:157:22: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:158:22: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:273:27: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:274:28: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:299:31: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:300:32: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :30: Using a static member import should be avoided - org.mockito.Mockito.inOrder. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :31: Using a static member import should be avoided - org.mockito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :32: Using a static member import should be avoided - org.mockito.Mockito.never. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :33: Using a static member import should be avoided - org.mockito.Mockito.verify. [AvoidStaticImport] > Task :aws-analytics-pinpoint:checkstyle FAILED FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':aws-analytics-pinpoint:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-analytics-pinpoint/build/reports/checkstyle/checkstyle. html Checkstyle files with violations: 2 Checkstyle violations by severity: [error:17] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 13s 64 actionable tasks: 22 executed, 42 up-to-date C:\MyData\Amplify original\amplify-android>gradlew build --scan > Task :aws-analytics-pinpoint:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:50: Using a static member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:51: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:52: Using a static member import should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:53: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:54: Using a static member import should be avoided - org.junit.Assert.fail. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:139:57: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:140:58: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:157:22: '3.14' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:158:22: '42' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:273:27: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:274:28: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:299:31: '47.6154086' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointIns trumentedTest.java:300:32: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :30: Using a static member import should be avoided - org.mockito.Mockito.inOrder. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :31: Using a static member import should be avoided - org.mockito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :32: Using a static member import should be avoided - org.mockito.Mockito.never. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java :33: Using a static member import should be avoided - org.mockito.Mockito.verify. [AvoidStaticImport] > Task :aws-analytics-pinpoint:checkstyle FAILED > Task :aws-api:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:47: Usi ng a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:48: Usi ng a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:84:18: '29' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:124:42: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:184:18: '29' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:192:22: '30' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:302:20: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\GraphQLInstrumentationTest.java:42: Using a st atic member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\GraphQLInstrumentationTest.java:43: Using a st atic member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:35: Using a st atic member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:36: Using a st atic member import should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:37: Using a st atic member import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:38: Using a st atic member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:41: Using a stat ic member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:42: Using a stat ic member import should be avoided - org.junit.Assert.assertNotEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:43: Using a stat ic member import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\TestApiCategory.java:32: Using a static member import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:27: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:28: Using a static member import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:29: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:74: Using a static member impor t should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:75: Using a static member impor t should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:76: Using a static member impor t should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:77: Using a static member impor t should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:78: Using a static member impor t should be avoided - org.mockito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:79: Using a static member impor t should be avoided - org.mockito.Mockito.when. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:100:25: '8080' is a magic numbe r. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:39: Using a static mem ber import should be avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:40: Using a static mem ber import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:41: Using a static mem ber import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:42: Using a static mem ber import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:100:19: '16' is a magi c number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:121:19: '28' is a magi c number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\ApiSelectorTest.java:26: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:103:18: '19' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:104:27: '2000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:104:36: '15' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:155:51: '123456 7890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncSigV4SignerInterceptorFactoryTest.java:31: Usi ng a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:47: Using a stati c member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:48: Using a stati c member import should be avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:49: Using a stati c member import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:143:29: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:146:41: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:146:44: '7' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:197:29: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:200:41: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:200:44: '7' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:253:37: '11' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:253:41: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:382:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:390:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:398:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:406:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:414:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:422:51: '12345678 90000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:29: Using a static member import sh ould be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:47:17: '3.5f' is a magic number. [M agicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:62:35: '4' is a magic number. [Magi cNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:63:33: '5.5f' is a magic number. [M agicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\SubscriptionAuthorizerTest.java:32: Using a static me mber import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:28: Using a static member im port should be avoided - org.mockito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:29: Using a static member im port should be avoided - org.mockito.Mockito.verify. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:30: Using a static member im port should be avoided - org.mockito.Mockito.verifyNoInteractions. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TypeMakerTest.java:26: Using a static member import s hould be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TypeMakerTest.java:27: Using a static member import s hould be avoided - org.junit.Assert.assertThrows. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\utils\RestRequestFactoryTest.java:30: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\utils\RestRequestFactoryTest.java:31: Using a static member import should be avoided - org.junit.Assert.assertNotNull. [AvoidStaticImport] > Task :aws-api:checkstyle FAILED FAILURE: Build completed with 2 failures. 1: Task failed with an exception. ----------- * What went wrong: Execution failed for task ':aws-analytics-pinpoint:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-analytics-pinpoint/build/reports/checkstyle/checkstyle. html Checkstyle files with violations: 2 Checkstyle violations by severity: [error:17] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== 2: Task failed with an exception. ----------- * What went wrong: Execution failed for task ':aws-api:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-api/build/reports/checkstyle/checkstyle.html Checkstyle files with violations: 17 Checkstyle violations by severity: [error:68] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings Publishing a build scan to scans.gradle.com requires accepting the Gradle Terms of Service defined at https://gradle.com/terms-of-service. Do you accept these terms? [yes, no] yGradle Terms of Service accepted. Publishing build scan... https://gradle.com/s/mbezreddtb23y C:\MyData\Amplify original\amplify-android>gradlew build --status PID STATUS INFO 15212 IDLE 6.6.1 14376 IDLE 6.6.1 Only Daemons for the current Gradle version are displayed. See https://docs.gradle.org/6.6.1/userguide/gradle_daemon.html#sec:status C:\MyData\Amplify original\amplify-android>clear 'clear' is not recognized as an internal or external command, operable program or batch file. C:\MyData\Amplify original\amplify-android>cr 'cr' is not recognized as an internal or external command, operable program or batch file. C:\MyData\Amplify original\amplify-android>cls C:\MyData\Amplify original\amplify-android>gradlew build --scan > Task :aws-analytics-pinpoint:checkstyle FAILED [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:50: Using a static member import should b e avoided - androidx.test.core.app.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:51: Using a static member import should b e avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:52: Using a static member import should b e avoided - org.junit.Assert.assertFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:53: Using a static member import should b e avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:54: Using a static member import should b e avoided - org.junit.Assert.fail. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:139:57: '3.14' is a magic number. [MagicN umber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:140:58: '42' is a magic number. [MagicNum ber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:157:22: '3.14' is a magic number. [MagicN umber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:158:22: '42' is a magic number. [MagicNum ber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:273:27: '47.6154086' is a magic number. [ MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:274:28: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:299:31: '47.6154086' is a magic number. [ MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\androidTest\java\com\amplifyframework\analytics\pinpoint\AnalyticsPinpointInstrumentedTest.java:300:32: '-122.3349685' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java:30: Using a static member import should be avoided - org.mo ckito.Mockito.inOrder. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java:31: Using a static member import should be avoided - org.mo ckito.Mockito.mock. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java:32: Using a static member import should be avoided - org.mo ckito.Mockito.never. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-analytics-pinpoint\src\test\java\com\amplifyframework\analytics\pinpoint\AutoSessionTrackerTest.java:33: Using a static member import should be avoided - org.mo ckito.Mockito.verify. [AvoidStaticImport] > Task :aws-api-appsync:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQlRequestTest.java:26: Using a static member import should be avoided - org.junit.Assert.asse rtEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQlRequestTest.java:28: Missing a Javadoc comment. [MissingJavadocType] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\api\aws\DefaultGraphQLRequestOptions.java:21: Missing a Javadoc comment. [MissingJavadocType] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\api\aws\SelectionSetTest.java:28: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\api\aws\SelectionSetTest.java:30: Missing a Javadoc comment. [MissingJavadocType] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:28: Using a static member import should be avoided - org.junit.Assert.a ssertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:29: Using a static member import should be avoided - org.junit.Assert.a ssertThrows. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:68:32: '2001' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:70:40: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:90:32: '2001' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTest.java:92:40: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:27: Using a static member import should be avoided - org.junit.Asse rt.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:68:46: '2001' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:68:55: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:68:58: '4' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:68:61: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:68:64: '6' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalDateTimeTest.java:73:48: '1000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:28: Using a static member import should be avoided - org.junit.Assert.a ssertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:29: Using a static member import should be avoided - org.junit.Assert.a ssertThrows. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:73:32: '1970' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:78:34: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:79:39: '4' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:97:32: '1970' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:102:34: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimeTest.java:103:39: '4' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:26: Using a static member import should be avoided - org.junit.Ass ert.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:27: Using a static member import should be avoided - org.junit.Ass ert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:78:59: '5_000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:79:60: '5_000_000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:92:59: '5_000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:93:60: '6_000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api-appsync\src\test\java\com\amplifyframework\core\model\temporal\TemporalTimestampTest.java:94:59: '7_000' is a magic number. [MagicNumber] > Task :aws-api-appsync:checkstyle FAILED > Task :aws-api:checkstyle [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:47: Using a static member import should be avoided - org.junit.Asse rt.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:48: Using a static member import should be avoided - org.junit.Asse rt.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:84:18: '29' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:124:42: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:184:18: '29' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:192:22: '30' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\CodeGenerationInstrumentationTest.java:302:20: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\GraphQLInstrumentationTest.java:42: Using a static member import should be avoided - androidx.test.core.ap p.ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\GraphQLInstrumentationTest.java:43: Using a static member import should be avoided - org.junit.Assert.asse rtEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:35: Using a static member import should be avoided - org.junit.Assert.asse rtEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:36: Using a static member import should be avoided - org.junit.Assert.asse rtFalse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:37: Using a static member import should be avoided - org.junit.Assert.asse rtNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\RestApiInstrumentationTest.java:38: Using a static member import should be avoided - org.junit.Assert.asse rtTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:41: Using a static member import should be avoided - androidx.test.core.app. ApplicationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:42: Using a static member import should be avoided - org.junit.Assert.assert NotEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\SubscriptionEndpointTest.java:43: Using a static member import should be avoided - org.junit.Assert.assert NotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\androidTest\java\com\amplifyframework\api\aws\TestApiCategory.java:32: Using a static member import should be avoided - androidx.test.core.app.Applicati onProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:27: Using a static member import should be avoided - org.junit.Assert.as sertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:28: Using a static member import should be avoided - org.junit.Assert.as sertNotNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginConfigurationReaderTest.java:29: Using a static member import should be avoided - org.junit.Assert.as sertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:74: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidS taticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:75: Using a static member import should be avoided - org.junit.Assert.assertFalse. [AvoidSt aticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:76: Using a static member import should be avoided - org.junit.Assert.assertNotNull. [Avoid StaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:77: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidSta ticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:78: Using a static member import should be avoided - org.mockito.Mockito.mock. [AvoidStatic Import] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:79: Using a static member import should be avoided - org.mockito.Mockito.when. [AvoidStatic Import] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginTest.java:100:25: '8080' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:39: Using a static member import should be avoided - androidx.test.core.app.Applic ationProvider.getApplicationContext. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:40: Using a static member import should be avoided - org.junit.Assert.assertEquals . [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:41: Using a static member import should be avoided - org.junit.Assert.assertNotNul l. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:42: Using a static member import should be avoided - org.junit.Assert.assertTrue. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:100:19: '16' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AWSApiPluginUserAgentTest.java:121:19: '28' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\ApiSelectorTest.java:26: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidSt aticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:103:18: '19' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:104:27: '2000' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:104:36: '15' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncGraphQLRequestFactoryTest.java:155:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\AppSyncSigV4SignerInterceptorFactoryTest.java:31: Using a static member import should be avoided - org.junit.Asse rt.assertEquals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:47: Using a static member import should be avoided - org.junit.Assert.assertE quals. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:48: Using a static member import should be avoided - org.junit.Assert.assertF alse. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:49: Using a static member import should be avoided - org.junit.Assert.assertN otNull. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:143:29: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:146:41: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:146:44: '7' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:197:29: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:200:41: '5' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:200:44: '7' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:253:37: '11' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:253:41: '3' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:382:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:390:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:398:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:406:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:414:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonGraphQLResponseFactoryTest.java:422:51: '1234567890000L' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:29: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStati cImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:47:17: '3.5f' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:62:35: '4' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\GsonUtilTest.java:63:33: '5.5f' is a magic number. [MagicNumber] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\SubscriptionAuthorizerTest.java:32: Using a static member import should be avoided - org.junit.Assert.assertEqual s. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:28: Using a static member import should be avoided - org.mockito.Mockito.mock. [AvoidSta ticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:29: Using a static member import should be avoided - org.mockito.Mockito.verify. [AvoidS taticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TimeoutWatchdogTest.java:30: Using a static member import should be avoided - org.mockito.Mockito.verifyNoInterac tions. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TypeMakerTest.java:26: Using a static member import should be avoided - org.junit.Assert.assertEquals. [AvoidStat icImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\TypeMakerTest.java:27: Using a static member import should be avoided - org.junit.Assert.assertThrows. [AvoidStat icImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\utils\RestRequestFactoryTest.java:30: Using a static member import should be avoided - org.junit.Assert.assertEqu als. [AvoidStaticImport] [ant:checkstyle] [ERROR] C:\MyData\Amplify original\amplify-android\aws-api\src\test\java\com\amplifyframework\api\aws\utils\RestRequestFactoryTest.java:31: Using a static member import should be avoided - org.junit.Assert.assertNot Null. [AvoidStaticImport] > Task :aws-api:checkstyle FAILED FAILURE: Build completed with 3 failures. 1: Task failed with an exception. ----------- * What went wrong: Execution failed for task ':aws-analytics-pinpoint:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-analytics-pinpoint/build/reports/checkstyle/checkstyle.html Checkstyle files with violations: 2 Checkstyle violations by severity: [error:17] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== 2: Task failed with an exception. ----------- * What went wrong: Execution failed for task ':aws-api-appsync:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-api-appsync/build/reports/checkstyle/checkstyle.html Checkstyle files with violations: 7 Checkstyle violations by severity: [error:33] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== 3: Task failed with an exception. ----------- * What went wrong: Execution failed for task ':aws-api:checkstyle'. > Checkstyle rule violations were found. See the report at: file:///C:/MyData/Amplify%20original/amplify-android/aws-api/build/reports/checkstyle/checkstyle.html Checkstyle files with violations: 17 Checkstyle violations by severity: [error:68] * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. ============================================================================== * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.6.1/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 12s 236 actionable tasks: 17 executed, 219 up-to-date Publishing a build scan to scans.gradle.com requires accepting the Gradle Terms of Service defined at https://gradle.com/terms-of-service. Do you accept these terms? [yes, no] yes Gradle Terms of Service accepted. Publishing build scan... https://gradle.com/s/n5yh2atjmu274 <-------------> 0% WAITING > IDLE > IDLE > IDLE > IDLE C:\MyData\Amplify original\amplify-android>```
non_process
facing gradle build failed in windows even though working fine on mac when i try to run gradlew build facing build failed in windows and working fine in mac this is the logs i get after running gradlew build microsoft windows c microsoft corporation all rights reserved c mydata amplify original amplify android gradlew build stacktrace welcome to gradle here are the highlights of this release experimental build configuration caching built in conventions for handling credentials java compilation supports release flag for more details see starting a gradle daemon incompatible daemon could not be reused use status for details task aws analytics pinpoint checkstyle c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided androidx test core app applicationprovider getapplicationcontext c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert assertfalse c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert asserttrue c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert fail c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito inorder c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito mock c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito never c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito verify task aws analytics pinpoint checkstyle failed failure build failed with an exception what went wrong execution failed for task aws analytics pinpoint checkstyle checkstyle rule violations were found see the report at file c mydata amplify amplify android aws analytics pinpoint build reports checkstyle checkstyle html checkstyle files with violations checkstyle violations by severity try run with info or debug option to get more log output run with scan to get full insights exception is org gradle api tasks taskexecutionexception execution failed for task aws analytics pinpoint checkstyle at org gradle api internal tasks execution executeactionstaskexecuter lambda executeifvalid executeactionstaskexecuter java at org gradle internal try failure ifsuccessfulorelse try java at org gradle api internal tasks execution executeactionstaskexecuter executeifvalid executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter execute executeactionstaskexecuter java at org gradle api internal tasks execution cleanupstaleoutputsexecuter execute cleanupstaleoutputsexecuter java at org gradle api internal tasks execution finalizepropertiestaskexecuter execute finalizepropertiestaskexecuter java at org gradle api internal tasks execution resolvetaskexecutionmodeexecuter execute resolvetaskexecutionmodeexecuter java at org gradle api internal tasks execution skiptaskwithnoactionsexecuter execute skiptaskwithnoactionsexecuter java at org gradle api internal tasks execution skiponlyiftaskexecuter execute skiponlyiftaskexecuter java at org gradle api internal tasks execution catchexceptiontaskexecuter execute catchexceptiontaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter executetask eventfiringtaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter call eventfiringtaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter call eventfiringtaskexecuter java at org gradle internal operations defaultbuildoperationexecutor callablebuildoperationworker execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor callablebuildoperationworker execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle internal operations delegatingbuildoperationexecutor call delegatingbuildoperationexecutor java at org gradle api internal tasks execution eventfiringtaskexecuter execute eventfiringtaskexecuter java at org gradle execution plan localtasknodeexecutor execute localtasknodeexecutor java at org gradle execution taskgraph defaulttaskexecutiongraph invokenodeexecutorsaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph invokenodeexecutorsaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph buildoperationawareexecutionaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph buildoperationawareexecutionaction execute defaulttaskexecutiongraph java at org gradle execution plan defaultplanexecutor executorworker lambda run defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker execute defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker executenextnode defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker run defaultplanexecutor java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java at org gradle internal concurrent threadfactoryimpl managedthreadrunnable run threadfactoryimpl java caused by org gradle api gradleexception checkstyle rule violations were found see the report at file c mydata amplify amplify android aws analytics pinp oint build reports checkstyle checkstyle html checkstyle files with violations checkstyle violations by severity at org gradle api plugins quality internal checkstyleinvoker invoke docall checkstyleinvoker groovy at org gradle util closurebackedaction execute closurebackedaction java at org gradle util closurebackedaction execute closurebackedaction java at org gradle api internal project antbuilder defaultisolatedantbuilder execute defaultisolatedantbuilder java at org gradle api internal project antbuilder defaultisolatedantbuilder execute defaultisolatedantbuilder java at org gradle api internal project antbuilder classpathtoclassloadercache withcachedclassloader classpathtoclassloadercache java at org gradle api internal project antbuilder defaultisolatedantbuilder execute defaultisolatedantbuilder java at org gradle api internal project isolatedantbuilder execute call unknown source at org gradle api plugins quality internal checkstyleinvoker invoke checkstyleinvoker groovy at org gradle api plugins quality checkstyle run checkstyle java at org gradle internal reflect javamethod invoke javamethod java at org gradle api internal project taskfactory standardtaskaction doexecute standardtaskaction java at org gradle api internal project taskfactory standardtaskaction execute standardtaskaction java at org gradle api internal project taskfactory standardtaskaction execute standardtaskaction java at org gradle api internal tasks execution executeactionstaskexecuter run executeactionstaskexecuter java at org gradle internal operations defaultbuildoperationexecutor runnablebuildoperationworker execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor runnablebuildoperationworker execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor run defaultbuildoperationexecutor java at org gradle internal operations delegatingbuildoperationexecutor run delegatingbuildoperationexecutor java at org gradle api internal tasks execution executeactionstaskexecuter executeaction executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter executeactions executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter access executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter taskexecution executewithpreviousoutputfiles executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter taskexecution execute executeactionstaskexecuter java at org gradle internal execution steps executestep lambda execute executestep java at org gradle internal execution steps executestep execute executestep java at org gradle internal execution steps executestep execute executestep java at org gradle internal execution steps cleanupoutputsstep execute cleanupoutputsstep java at org gradle internal execution steps cleanupoutputsstep execute cleanupoutputsstep java at org gradle internal execution steps resolveinputchangesstep execute resolveinputchangesstep java at org gradle internal execution steps resolveinputchangesstep execute resolveinputchangesstep java at org gradle internal execution steps cancelexecutionstep execute cancelexecutionstep java at org gradle internal execution steps timeoutstep executewithouttimeout timeoutstep java at org gradle internal execution steps timeoutstep execute timeoutstep java at org gradle internal execution steps catchexceptionstep execute catchexceptionstep java at org gradle internal execution steps createoutputsstep execute createoutputsstep java at org gradle internal execution steps snapshotoutputsstep execute snapshotoutputsstep java at org gradle internal execution steps snapshotoutputsstep execute snapshotoutputsstep java at org gradle internal execution steps broadcastchangingoutputsstep execute broadcastchangingoutputsstep java at org gradle internal execution steps cachestep executewithoutcache cachestep java at org gradle internal execution steps cachestep execute cachestep java at org gradle internal execution steps cachestep execute cachestep java at org gradle internal execution steps storeexecutionstatestep execute storeexecutionstatestep java at org gradle internal execution steps storeexecutionstatestep execute storeexecutionstatestep java at org gradle internal execution steps recordoutputsstep execute recordoutputsstep java at org gradle internal execution steps recordoutputsstep execute recordoutputsstep java at org gradle internal execution steps skipuptodatestep executebecause skipuptodatestep java at org gradle internal execution steps skipuptodatestep lambda execute skipuptodatestep java at org gradle internal execution steps skipuptodatestep execute skipuptodatestep java at org gradle internal execution steps skipuptodatestep execute skipuptodatestep java at org gradle internal execution steps resolvechangesstep execute resolvechangesstep java at org gradle internal execution steps resolvechangesstep execute resolvechangesstep java at org gradle internal execution steps legacy marksnapshottinginputsfinishedstep execute marksnapshottinginputsfinishedstep java at org gradle internal execution steps legacy marksnapshottinginputsfinishedstep execute marksnapshottinginputsfinishedstep java at org gradle internal execution steps resolvecachingstatestep execute resolvecachingstatestep java at org gradle internal execution steps resolvecachingstatestep execute resolvecachingstatestep java at org gradle internal execution steps capturestatebeforeexecutionstep execute capturestatebeforeexecutionstep java at org gradle internal execution steps capturestatebeforeexecutionstep execute capturestatebeforeexecutionstep java at org gradle internal execution steps validatestep execute validatestep java at org gradle internal execution steps skipemptyworkstep lambda execute skipemptyworkstep java at org gradle internal execution steps skipemptyworkstep execute skipemptyworkstep java at org gradle internal execution steps skipemptyworkstep execute skipemptyworkstep java at org gradle internal execution steps legacy marksnapshottinginputsstartedstep execute marksnapshottinginputsstartedstep java at org gradle internal execution steps loadexecutionstatestep execute loadexecutionstatestep java at org gradle internal execution steps loadexecutionstatestep execute loadexecutionstatestep java at org gradle internal execution impl defaultworkexecutor execute defaultworkexecutor java at org gradle api internal tasks execution executeactionstaskexecuter executeifvalid executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter execute executeactionstaskexecuter java at org gradle api internal tasks execution cleanupstaleoutputsexecuter execute cleanupstaleoutputsexecuter java at org gradle api internal tasks execution finalizepropertiestaskexecuter execute finalizepropertiestaskexecuter java at org gradle api internal tasks execution resolvetaskexecutionmodeexecuter execute resolvetaskexecutionmodeexecuter java at org gradle api internal tasks execution skiptaskwithnoactionsexecuter execute skiptaskwithnoactionsexecuter java at org gradle api internal tasks execution skiponlyiftaskexecuter execute skiponlyiftaskexecuter java at org gradle api internal tasks execution catchexceptiontaskexecuter execute catchexceptiontaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter executetask eventfiringtaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter call eventfiringtaskexecuter java at org gradle api internal tasks execution eventfiringtaskexecuter call eventfiringtaskexecuter java at org gradle internal operations defaultbuildoperationexecutor callablebuildoperationworker execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor callablebuildoperationworker execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor execute defaultbuildoperationexecutor java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle internal operations delegatingbuildoperationexecutor call delegatingbuildoperationexecutor java at org gradle api internal tasks execution eventfiringtaskexecuter execute eventfiringtaskexecuter java at org gradle execution plan localtasknodeexecutor execute localtasknodeexecutor java at org gradle execution taskgraph defaulttaskexecutiongraph invokenodeexecutorsaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph invokenodeexecutorsaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph buildoperationawareexecutionaction execute defaulttaskexecutiongraph java at org gradle execution taskgraph defaulttaskexecutiongraph buildoperationawareexecutionaction execute defaulttaskexecutiongraph java at org gradle execution plan defaultplanexecutor executorworker lambda run defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker execute defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker executenextnode defaultplanexecutor java at org gradle execution plan defaultplanexecutor executorworker run defaultplanexecutor java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java at org gradle internal concurrent threadfactoryimpl managedthreadrunnable run threadfactoryimpl java get more help at deprecated gradle features were used in this build making it incompatible with gradle use warning mode all to show the individual deprecation warnings see build failed in actionable tasks executed c mydata amplify original amplify android gradlew build task aws analytics pinpoint checkstyle c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided androidx test core app applicationprovider getapplicationcontext c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert assertfalse c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert asserttrue c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert fail c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito inorder c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito mock c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito never c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito verify task aws analytics pinpoint checkstyle failed failure build failed with an exception what went wrong execution failed for task aws analytics pinpoint checkstyle checkstyle rule violations were found see the report at file c mydata amplify amplify android aws analytics pinpoint build reports checkstyle checkstyle html checkstyle files with violations checkstyle violations by severity try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights get more help at deprecated gradle features were used in this build making it incompatible with gradle use warning mode all to show the individual deprecation warnings see build failed in actionable tasks executed up to date c mydata amplify original amplify android gradlew build scan task aws analytics pinpoint checkstyle c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided androidx test core app applicationprovider getapplicationcontext c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert assertfalse c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert asserttrue c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java using a static member import should be avoided org junit assert fail c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointins trumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito inorder c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito mock c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito never c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mockito mockito verify task aws analytics pinpoint checkstyle failed task aws api checkstyle c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java usi ng a static member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java usi ng a static member import should be avoided org junit assert asserttrue c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java is a magic number c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java is a magic number c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java is a magic number c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java is a magic number c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java is a magic number c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws graphqlinstrumentationtest java using a st atic member import should be avoided androidx test core app applicationprovider getapplicationcontext c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws graphqlinstrumentationtest java using a st atic member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws restapiinstrumentationtest java using a st atic member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws restapiinstrumentationtest java using a st atic member import should be avoided org junit assert assertfalse c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws restapiinstrumentationtest java using a st atic member import should be avoided org junit assert assertnotnull c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws restapiinstrumentationtest java using a st atic member import should be avoided org junit assert asserttrue c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws subscriptionendpointtest java using a stat ic member import should be avoided androidx test core app applicationprovider getapplicationcontext c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws subscriptionendpointtest java using a stat ic member import should be avoided org junit assert assertnotequals c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws subscriptionendpointtest java using a stat ic member import should be avoided org junit assert assertnotnull c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws testapicategory java using a static member import should be avoided androidx test core app applicationprovider getapplicationcontext c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginconfigurationreadertest java using a static member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginconfigurationreadertest java using a static member import should be avoided org junit assert assertnotnull c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginconfigurationreadertest java using a static member import should be avoided org junit assert asserttrue c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member impor t should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member impor t should be avoided org junit assert assertfalse c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member impor t should be avoided org junit assert assertnotnull c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member impor t should be avoided org junit assert asserttrue c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member impor t should be avoided org mockito mockito mock c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member impor t should be avoided org mockito mockito when c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java is a magic numbe r c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java using a static mem ber import should be avoided androidx test core app applicationprovider getapplicationcontext c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java using a static mem ber import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java using a static mem ber import should be avoided org junit assert assertnotnull c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java using a static mem ber import should be avoided org junit assert asserttrue c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java is a magi c number c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java is a magi c number c mydata amplify original amplify android aws api src test java com amplifyframework api aws apiselectortest java using a static member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws appsyncgraphqlrequestfactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws appsyncgraphqlrequestfactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws appsyncgraphqlrequestfactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws appsyncgraphqlrequestfactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws java usi ng a static member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java using a stati c member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java using a stati c member import should be avoided org junit assert assertfalse c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java using a stati c member import should be avoided org junit assert assertnotnull c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsonutiltest java using a static member import sh ould be avoided org junit assert assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsonutiltest java is a magic number m agicnumber c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsonutiltest java is a magic number magi cnumber c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsonutiltest java is a magic number m agicnumber c mydata amplify original amplify android aws api src test java com amplifyframework api aws subscriptionauthorizertest java using a static me mber import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws timeoutwatchdogtest java using a static member im port should be avoided org mockito mockito mock c mydata amplify original amplify android aws api src test java com amplifyframework api aws timeoutwatchdogtest java using a static member im port should be avoided org mockito mockito verify c mydata amplify original amplify android aws api src test java com amplifyframework api aws timeoutwatchdogtest java using a static member im port should be avoided org mockito mockito verifynointeractions c mydata amplify original amplify android aws api src test java com amplifyframework api aws typemakertest java using a static member import s hould be avoided org junit assert assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws typemakertest java using a static member import s hould be avoided org junit assert assertthrows c mydata amplify original amplify android aws api src test java com amplifyframework api aws utils restrequestfactorytest java using a static member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws utils restrequestfactorytest java using a static member import should be avoided org junit assert assertnotnull task aws api checkstyle failed failure build completed with failures task failed with an exception what went wrong execution failed for task aws analytics pinpoint checkstyle checkstyle rule violations were found see the report at file c mydata amplify amplify android aws analytics pinpoint build reports checkstyle checkstyle html checkstyle files with violations checkstyle violations by severity try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights task failed with an exception what went wrong execution failed for task aws api checkstyle checkstyle rule violations were found see the report at file c mydata amplify amplify android aws api build reports checkstyle checkstyle html checkstyle files with violations checkstyle violations by severity try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights get more help at deprecated gradle features were used in this build making it incompatible with gradle use warning mode all to show the individual deprecation warnings see publishing a build scan to scans gradle com requires accepting the gradle terms of service defined at do you accept these terms ygradle terms of service accepted publishing build scan c mydata amplify original amplify android gradlew build status pid status info idle idle only daemons for the current gradle version are displayed see c mydata amplify original amplify android clear clear is not recognized as an internal or external command operable program or batch file c mydata amplify original amplify android cr cr is not recognized as an internal or external command operable program or batch file c mydata amplify original amplify android cls c mydata amplify original amplify android gradlew build scan task aws analytics pinpoint checkstyle failed c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java using a static member import should b e avoided androidx test core app applicationprovider getapplicationcontext c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java using a static member import should b e avoided org junit assert assertequals c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java using a static member import should b e avoided org junit assert assertfalse c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java using a static member import should b e avoided org junit assert asserttrue c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java using a static member import should b e avoided org junit assert fail c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java is a magic number magicn umber c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java is a magic number magicnum ber c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java is a magic number magicn umber c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java is a magic number magicnum ber c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java is a magic number magicnumber c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java is a magic number magicnumber c mydata amplify original amplify android aws analytics pinpoint src androidtest java com amplifyframework analytics pinpoint analyticspinpointinstrumentedtest java is a magic number c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mo ckito mockito inorder c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mo ckito mockito mock c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mo ckito mockito never c mydata amplify original amplify android aws analytics pinpoint src test java com amplifyframework analytics pinpoint autosessiontrackertest java using a static member import should be avoided org mo ckito mockito verify task aws api appsync checkstyle c mydata amplify original amplify android aws api appsync src test java com amplifyframework api aws appsyncgraphqlrequesttest java using a static member import should be avoided org junit assert asse rtequals c mydata amplify original amplify android aws api appsync src test java com amplifyframework api aws appsyncgraphqlrequesttest java missing a javadoc comment c mydata amplify original amplify android aws api appsync src test java com amplifyframework api aws defaultgraphqlrequestoptions java missing a javadoc comment c mydata amplify original amplify android aws api appsync src test java com amplifyframework api aws selectionsettest java using a static member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api appsync src test java com amplifyframework api aws selectionsettest java missing a javadoc comment c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetest java using a static member import should be avoided org junit assert a ssertequals c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetest java using a static member import should be avoided org junit assert a ssertthrows c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetimetest java using a static member import should be avoided org junit asse rt assertequals c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaldatetimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimetest java using a static member import should be avoided org junit assert a ssertequals c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimetest java using a static member import should be avoided org junit assert a ssertthrows c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimetest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimestamptest java using a static member import should be avoided org junit ass ert assertequals c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimestamptest java using a static member import should be avoided org junit ass ert asserttrue c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimestamptest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimestamptest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimestamptest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimestamptest java is a magic number c mydata amplify original amplify android aws api appsync src test java com amplifyframework core model temporal temporaltimestamptest java is a magic number task aws api appsync checkstyle failed task aws api checkstyle c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java using a static member import should be avoided org junit asse rt assertequals c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java using a static member import should be avoided org junit asse rt asserttrue c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java is a magic number c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java is a magic number c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java is a magic number c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java is a magic number c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws codegenerationinstrumentationtest java is a magic number c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws graphqlinstrumentationtest java using a static member import should be avoided androidx test core ap p applicationprovider getapplicationcontext c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws graphqlinstrumentationtest java using a static member import should be avoided org junit assert asse rtequals c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws restapiinstrumentationtest java using a static member import should be avoided org junit assert asse rtequals c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws restapiinstrumentationtest java using a static member import should be avoided org junit assert asse rtfalse c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws restapiinstrumentationtest java using a static member import should be avoided org junit assert asse rtnotnull c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws restapiinstrumentationtest java using a static member import should be avoided org junit assert asse rttrue c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws subscriptionendpointtest java using a static member import should be avoided androidx test core app applicationprovider getapplicationcontext c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws subscriptionendpointtest java using a static member import should be avoided org junit assert assert notequals c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws subscriptionendpointtest java using a static member import should be avoided org junit assert assert notnull c mydata amplify original amplify android aws api src androidtest java com amplifyframework api aws testapicategory java using a static member import should be avoided androidx test core app applicati onprovider getapplicationcontext c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginconfigurationreadertest java using a static member import should be avoided org junit assert as sertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginconfigurationreadertest java using a static member import should be avoided org junit assert as sertnotnull c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginconfigurationreadertest java using a static member import should be avoided org junit assert as serttrue c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member import should be avoided org junit assert assertequals avoids taticimport c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member import should be avoided org junit assert assertfalse avoidst aticimport c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member import should be avoided org junit assert assertnotnull avoid staticimport c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member import should be avoided org junit assert asserttrue avoidsta ticimport c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member import should be avoided org mockito mockito mock avoidstatic import c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java using a static member import should be avoided org mockito mockito when avoidstatic import c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapiplugintest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java using a static member import should be avoided androidx test core app applic ationprovider getapplicationcontext c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java using a static member import should be avoided org junit assert assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java using a static member import should be avoided org junit assert assertnotnul l c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java using a static member import should be avoided org junit assert asserttrue c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws awsapipluginuseragenttest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws apiselectortest java using a static member import should be avoided org junit assert assertequals avoidst aticimport c mydata amplify original amplify android aws api src test java com amplifyframework api aws appsyncgraphqlrequestfactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws appsyncgraphqlrequestfactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws appsyncgraphqlrequestfactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws appsyncgraphqlrequestfactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws java using a static member import should be avoided org junit asse rt assertequals c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java using a static member import should be avoided org junit assert asserte quals c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java using a static member import should be avoided org junit assert assertf alse c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java using a static member import should be avoided org junit assert assertn otnull c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsongraphqlresponsefactorytest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsonutiltest java using a static member import should be avoided org junit assert assertequals avoidstati cimport c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsonutiltest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsonutiltest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws gsonutiltest java is a magic number c mydata amplify original amplify android aws api src test java com amplifyframework api aws subscriptionauthorizertest java using a static member import should be avoided org junit assert assertequal s c mydata amplify original amplify android aws api src test java com amplifyframework api aws timeoutwatchdogtest java using a static member import should be avoided org mockito mockito mock avoidsta ticimport c mydata amplify original amplify android aws api src test java com amplifyframework api aws timeoutwatchdogtest java using a static member import should be avoided org mockito mockito verify avoids taticimport c mydata amplify original amplify android aws api src test java com amplifyframework api aws timeoutwatchdogtest java using a static member import should be avoided org mockito mockito verifynointerac tions c mydata amplify original amplify android aws api src test java com amplifyframework api aws typemakertest java using a static member import should be avoided org junit assert assertequals avoidstat icimport c mydata amplify original amplify android aws api src test java com amplifyframework api aws typemakertest java using a static member import should be avoided org junit assert assertthrows avoidstat icimport c mydata amplify original amplify android aws api src test java com amplifyframework api aws utils restrequestfactorytest java using a static member import should be avoided org junit assert assertequ als c mydata amplify original amplify android aws api src test java com amplifyframework api aws utils restrequestfactorytest java using a static member import should be avoided org junit assert assertnot null task aws api checkstyle failed failure build completed with failures task failed with an exception what went wrong execution failed for task aws analytics pinpoint checkstyle checkstyle rule violations were found see the report at file c mydata amplify amplify android aws analytics pinpoint build reports checkstyle checkstyle html checkstyle files with violations checkstyle violations by severity try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights task failed with an exception what went wrong execution failed for task aws api appsync checkstyle checkstyle rule violations were found see the report at file c mydata amplify amplify android aws api appsync build reports checkstyle checkstyle html checkstyle files with violations checkstyle violations by severity try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights task failed with an exception what went wrong execution failed for task aws api checkstyle checkstyle rule violations were found see the report at file c mydata amplify amplify android aws api build reports checkstyle checkstyle html checkstyle files with violations checkstyle violations by severity try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights get more help at deprecated gradle features were used in this build making it incompatible with gradle use warning mode all to show the individual deprecation warnings see build failed in actionable tasks executed up to date publishing a build scan to scans gradle com requires accepting the gradle terms of service defined at do you accept these terms yes gradle terms of service accepted publishing build scan waiting idle idle idle idle c mydata amplify original amplify android
0
259,384
27,621,853,952
IssuesEvent
2023-03-10 01:16:36
nidhi7598/linux-3.0.35
https://api.github.com/repos/nidhi7598/linux-3.0.35
closed
CVE-2014-3185 (High) detected in linuxlinux-3.0.40 - autoclosed
Mend: dependency security vulnerability
## CVE-2014-3185 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35/commit/4cc6d4a22f88b8effe1090492c1a242ce587b492">4cc6d4a22f88b8effe1090492c1a242ce587b492</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/serial/whiteheat.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/serial/whiteheat.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/serial/whiteheat.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Multiple buffer overflows in the command_port_read_callback function in drivers/usb/serial/whiteheat.c in the Whiteheat USB Serial Driver in the Linux kernel before 3.16.2 allow physically proximate attackers to execute arbitrary code or cause a denial of service (memory corruption and system crash) via a crafted device that provides a large amount of (1) EHCI or (2) XHCI data associated with a bulk response. <p>Publish Date: 2014-09-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-3185>CVE-2014-3185</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-3185">https://nvd.nist.gov/vuln/detail/CVE-2014-3185</a></p> <p>Release Date: 2014-09-28</p> <p>Fix Resolution: 3.16.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2014-3185 (High) detected in linuxlinux-3.0.40 - autoclosed - ## CVE-2014-3185 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35/commit/4cc6d4a22f88b8effe1090492c1a242ce587b492">4cc6d4a22f88b8effe1090492c1a242ce587b492</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/serial/whiteheat.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/serial/whiteheat.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/serial/whiteheat.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Multiple buffer overflows in the command_port_read_callback function in drivers/usb/serial/whiteheat.c in the Whiteheat USB Serial Driver in the Linux kernel before 3.16.2 allow physically proximate attackers to execute arbitrary code or cause a denial of service (memory corruption and system crash) via a crafted device that provides a large amount of (1) EHCI or (2) XHCI data associated with a bulk response. <p>Publish Date: 2014-09-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-3185>CVE-2014-3185</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2014-3185">https://nvd.nist.gov/vuln/detail/CVE-2014-3185</a></p> <p>Release Date: 2014-09-28</p> <p>Fix Resolution: 3.16.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files drivers usb serial whiteheat c drivers usb serial whiteheat c drivers usb serial whiteheat c vulnerability details multiple buffer overflows in the command port read callback function in drivers usb serial whiteheat c in the whiteheat usb serial driver in the linux kernel before allow physically proximate attackers to execute arbitrary code or cause a denial of service memory corruption and system crash via a crafted device that provides a large amount of ehci or xhci data associated with a bulk response publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
2,903
5,889,357,347
IssuesEvent
2017-05-17 12:42:48
LOVDnl/LOVD3
https://api.github.com/repos/LOVDnl/LOVD3
opened
Allow submitter to delete data
cat: submission process feature request
Submitters are currently not allowed to delete data. It would be useful if this is added. Depending on the status of the data: - If the data is in progress or pending, the data hasn't been reviewed yet, and the submitter should be allowed to delete the entire submission or part of it (removing a phenotype entry, removing a screening w/o leaving variants orphaned, removing variants as long as some variants are left). It is important to note that it should not be possible to leave a incomplete submission as pending. - If the data has already been checked by the curator, the submitter should be able to retract the entire submission, but this will result in the data being taken offline, and a request being sent to the curator to perform the actual delete.
1.0
Allow submitter to delete data - Submitters are currently not allowed to delete data. It would be useful if this is added. Depending on the status of the data: - If the data is in progress or pending, the data hasn't been reviewed yet, and the submitter should be allowed to delete the entire submission or part of it (removing a phenotype entry, removing a screening w/o leaving variants orphaned, removing variants as long as some variants are left). It is important to note that it should not be possible to leave a incomplete submission as pending. - If the data has already been checked by the curator, the submitter should be able to retract the entire submission, but this will result in the data being taken offline, and a request being sent to the curator to perform the actual delete.
process
allow submitter to delete data submitters are currently not allowed to delete data it would be useful if this is added depending on the status of the data if the data is in progress or pending the data hasn t been reviewed yet and the submitter should be allowed to delete the entire submission or part of it removing a phenotype entry removing a screening w o leaving variants orphaned removing variants as long as some variants are left it is important to note that it should not be possible to leave a incomplete submission as pending if the data has already been checked by the curator the submitter should be able to retract the entire submission but this will result in the data being taken offline and a request being sent to the curator to perform the actual delete
1
1,956
2,649,231,499
IssuesEvent
2015-03-14 18:17:36
NebulousLabs/Sia
https://api.github.com/repos/NebulousLabs/Sia
closed
Proposal: Testing conventions
documentation long term
All packages should have internal testing that brings the coverage to 100% (except intentionally unreachable code). All testing should be done using exported functions only. Furthermore, these functions should only be exported if they are intended to be used by other packages, meaning we plan to maintain support for the function calls.
1.0
Proposal: Testing conventions - All packages should have internal testing that brings the coverage to 100% (except intentionally unreachable code). All testing should be done using exported functions only. Furthermore, these functions should only be exported if they are intended to be used by other packages, meaning we plan to maintain support for the function calls.
non_process
proposal testing conventions all packages should have internal testing that brings the coverage to except intentionally unreachable code all testing should be done using exported functions only furthermore these functions should only be exported if they are intended to be used by other packages meaning we plan to maintain support for the function calls
0
45,388
12,758,755,956
IssuesEvent
2020-06-29 03:28:07
SasView/sasview
https://api.github.com/repos/SasView/sasview
opened
Problem loading canSAS data into Invariant
defect major
the `latex_smeared.xml` data set in the test folder includes 2 separate curves: a slit smeared low Q USANS data set and a high Q pinhole smeared SANS data set. These are loaded as two separate files into the data manager. Checking only one of the data sets allows only that data to be sent to be plotted or to fitting and once fitted and plotted with the fit only that data set is shown. however, when sent to the invariant both the checked AND unchecked data sets are plotted in the invariant plot though it appears the Invariant is only calculated from the chosen data set? At least the default qmin and qmax come from that data set and the extrapolations appear to be made from that data set.
1.0
Problem loading canSAS data into Invariant - the `latex_smeared.xml` data set in the test folder includes 2 separate curves: a slit smeared low Q USANS data set and a high Q pinhole smeared SANS data set. These are loaded as two separate files into the data manager. Checking only one of the data sets allows only that data to be sent to be plotted or to fitting and once fitted and plotted with the fit only that data set is shown. however, when sent to the invariant both the checked AND unchecked data sets are plotted in the invariant plot though it appears the Invariant is only calculated from the chosen data set? At least the default qmin and qmax come from that data set and the extrapolations appear to be made from that data set.
non_process
problem loading cansas data into invariant the latex smeared xml data set in the test folder includes separate curves a slit smeared low q usans data set and a high q pinhole smeared sans data set these are loaded as two separate files into the data manager checking only one of the data sets allows only that data to be sent to be plotted or to fitting and once fitted and plotted with the fit only that data set is shown however when sent to the invariant both the checked and unchecked data sets are plotted in the invariant plot though it appears the invariant is only calculated from the chosen data set at least the default qmin and qmax come from that data set and the extrapolations appear to be made from that data set
0
19,850
26,252,297,956
IssuesEvent
2023-01-05 20:30:25
unicode-org/icu4x
https://api.github.com/repos/unicode-org/icu4x
closed
Decide on data file versioning policy
T-docs-tests C-process C-data-infra S-small
One of the value propositions driving ICU4X is the ability to share one data file across multiple ICU4X instances. Design doc: https://docs.google.com/document/d/1yg_2l5FFo0aAuNi4jpgcIhIYjHqJyUoJWtMduyQ0vR8/edit# Seeking feedback from: - [x] @nciric - [x] @Manishearth - [x] @markusicu Leaving comments in the doc is fine. Thanks!
1.0
Decide on data file versioning policy - One of the value propositions driving ICU4X is the ability to share one data file across multiple ICU4X instances. Design doc: https://docs.google.com/document/d/1yg_2l5FFo0aAuNi4jpgcIhIYjHqJyUoJWtMduyQ0vR8/edit# Seeking feedback from: - [x] @nciric - [x] @Manishearth - [x] @markusicu Leaving comments in the doc is fine. Thanks!
process
decide on data file versioning policy one of the value propositions driving is the ability to share one data file across multiple instances design doc seeking feedback from nciric manishearth markusicu leaving comments in the doc is fine thanks
1
13,928
16,685,412,732
IssuesEvent
2021-06-08 07:31:24
bisq-network/bisq
https://api.github.com/repos/bisq-network/bisq
closed
DAO out of sync causes Bisq fee to be validated incorrectly
in:dao in:trade-process
Since 1.6.0, Bisq has had [this feature](https://github.com/bisq-network/bisq/issues/5119) to validate trading fees paid. With this model, Bisq trading partners check each other's transactions thus ensuring fees are paid (which is what keeps Bisq running). However it is possible for a Bisq node's DAO to be out of sync and thus have a differing view on what the sufficient fee should be. For example user Alice's node expected trading partner Bob to have paid a fee of 2.11 BSQ; in fact Bob had paid the correct fee of 1.82 BSQ. In this case Alice's node had an out of sync DAO which told them that Bob should have used a fee rate from two months ago which is higher than today's fee rate. Bisq complains to Alice that her counterparty did not pay enough fees and that a support ticket should be opened. There are already [leniency checks](https://github.com/bisq-network/bisq/pull/5347) built onto the fee validation to allow historical fee rates. But in this case the appropriate fee rate was unknown to Alice since her DAO was stuck, **and this was only evident by looking at DAO -> Dao Status screen**. The solution for Alice was to re-sync the DAO. --- Ideally we would not have the situation where the DAO can silently get stuck. Related issue: https://github.com/bisq-network/bisq/issues/5329
1.0
DAO out of sync causes Bisq fee to be validated incorrectly - Since 1.6.0, Bisq has had [this feature](https://github.com/bisq-network/bisq/issues/5119) to validate trading fees paid. With this model, Bisq trading partners check each other's transactions thus ensuring fees are paid (which is what keeps Bisq running). However it is possible for a Bisq node's DAO to be out of sync and thus have a differing view on what the sufficient fee should be. For example user Alice's node expected trading partner Bob to have paid a fee of 2.11 BSQ; in fact Bob had paid the correct fee of 1.82 BSQ. In this case Alice's node had an out of sync DAO which told them that Bob should have used a fee rate from two months ago which is higher than today's fee rate. Bisq complains to Alice that her counterparty did not pay enough fees and that a support ticket should be opened. There are already [leniency checks](https://github.com/bisq-network/bisq/pull/5347) built onto the fee validation to allow historical fee rates. But in this case the appropriate fee rate was unknown to Alice since her DAO was stuck, **and this was only evident by looking at DAO -> Dao Status screen**. The solution for Alice was to re-sync the DAO. --- Ideally we would not have the situation where the DAO can silently get stuck. Related issue: https://github.com/bisq-network/bisq/issues/5329
process
dao out of sync causes bisq fee to be validated incorrectly since bisq has had to validate trading fees paid with this model bisq trading partners check each other s transactions thus ensuring fees are paid which is what keeps bisq running however it is possible for a bisq node s dao to be out of sync and thus have a differing view on what the sufficient fee should be for example user alice s node expected trading partner bob to have paid a fee of bsq in fact bob had paid the correct fee of bsq in this case alice s node had an out of sync dao which told them that bob should have used a fee rate from two months ago which is higher than today s fee rate bisq complains to alice that her counterparty did not pay enough fees and that a support ticket should be opened there are already built onto the fee validation to allow historical fee rates but in this case the appropriate fee rate was unknown to alice since her dao was stuck and this was only evident by looking at dao dao status screen the solution for alice was to re sync the dao ideally we would not have the situation where the dao can silently get stuck related issue
1
17,646
23,468,805,389
IssuesEvent
2022-08-16 19:30:28
googleapis/cloud-trace-nodejs
https://api.github.com/repos/googleapis/cloud-trace-nodejs
closed
Your .repo-metadata.json file has a problem ๐Ÿค’
type: process api: cloudtrace repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan ๐Ÿ“ˆ: * api_shortname 'trace' invalid in .repo-metadata.json โ˜๏ธ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem ๐Ÿค’ - You have a problem with your .repo-metadata.json file: Result of scan ๐Ÿ“ˆ: * api_shortname 'trace' invalid in .repo-metadata.json โ˜๏ธ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem ๐Ÿค’ you have a problem with your repo metadata json file result of scan ๐Ÿ“ˆ api shortname trace invalid in repo metadata json โ˜๏ธ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
7,819
10,981,329,068
IssuesEvent
2019-11-30 20:56:28
deeplearningunb/stonks
https://api.github.com/repos/deeplearningunb/stonks
closed
Processing training dataset data
data pre processing dataset enhancement
**Is your feature request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> I need training dataset data to be processed before being used. **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> Create a code named stonks.py and processing the data. This includes erasing empty csv information and normalizing the values. **Describe alternatives you've considered** <!-- A clear and concise description of any alternative solutions or features you've considered. --> Search about processing data. <!-- **Additional context** --> <!-- Add any other context or screenshots about the feature request here. --> **Checklist** - [x] The issue has a meaningful name. - [x] The issue has a meaningful description. - [ ] The issue has screenshots when needed. - [x] The issue has labels.
1.0
Processing training dataset data - **Is your feature request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> I need training dataset data to be processed before being used. **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> Create a code named stonks.py and processing the data. This includes erasing empty csv information and normalizing the values. **Describe alternatives you've considered** <!-- A clear and concise description of any alternative solutions or features you've considered. --> Search about processing data. <!-- **Additional context** --> <!-- Add any other context or screenshots about the feature request here. --> **Checklist** - [x] The issue has a meaningful name. - [x] The issue has a meaningful description. - [ ] The issue has screenshots when needed. - [x] The issue has labels.
process
processing training dataset data is your feature request related to a problem please describe i need training dataset data to be processed before being used describe the solution you d like create a code named stonks py and processing the data this includes erasing empty csv information and normalizing the values describe alternatives you ve considered search about processing data checklist the issue has a meaningful name the issue has a meaningful description the issue has screenshots when needed the issue has labels
1
673,227
22,953,359,859
IssuesEvent
2022-07-19 09:22:06
FlyBase/drosophila-anatomy-developmental-ontology
https://api.github.com/repos/FlyBase/drosophila-anatomy-developmental-ontology
closed
FAFB Shiu et al. (2022)
VFB data EM priority for next release FAFB
Taste quality and hunger interactions in a feeding sensorimotor circuit https://elifesciences.org/articles/79887 doi:10.7554/eLife.79887 (not fully published yet)
1.0
FAFB Shiu et al. (2022) - Taste quality and hunger interactions in a feeding sensorimotor circuit https://elifesciences.org/articles/79887 doi:10.7554/eLife.79887 (not fully published yet)
non_process
fafb shiu et al taste quality and hunger interactions in a feeding sensorimotor circuit doi elife not fully published yet
0
20,050
26,538,952,535
IssuesEvent
2023-01-19 17:38:56
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Release checklist 0.71
enhancement process
### Problem We need a checklist to verify the release is rolled out successfully. ### Solution ## Preparation - [x] Milestone field populated on relevant [issues](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aclosed+no%3Amilestone+sort%3Aupdated-desc) - [x] Nothing open for [milestone](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aopen+sort%3Aupdated-desc+milestone%3A0.71.0) - [x] GitHub checks for branch are passing - [x] Automated Kubernetes deployment successful - [x] Tag release - [x] Upload release artifacts - [ ] Manual Submission for GCP Marketplace verification by google - [ ] Publish marketplace release - [x] Publish release ## Performance - [x] Deploy to Kubernetes - [x] Deploy to VM - [x] gRPC API performance tests - [x] Importer performance tests - [x] REST API performance tests ## Previewnet - [x] Deploy to Kubernetes ## Staging - [x] Deploy to Kubernetes ## Testnet - [x] Deploy to VM ## Mainnet - [x] Deploy to Kubernetes EU - [x] Deploy to Kubernetes NA - [x] Deploy to VM - [x] Deploy to ETL ### Alternatives _No response_
1.0
Release checklist 0.71 - ### Problem We need a checklist to verify the release is rolled out successfully. ### Solution ## Preparation - [x] Milestone field populated on relevant [issues](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aclosed+no%3Amilestone+sort%3Aupdated-desc) - [x] Nothing open for [milestone](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aopen+sort%3Aupdated-desc+milestone%3A0.71.0) - [x] GitHub checks for branch are passing - [x] Automated Kubernetes deployment successful - [x] Tag release - [x] Upload release artifacts - [ ] Manual Submission for GCP Marketplace verification by google - [ ] Publish marketplace release - [x] Publish release ## Performance - [x] Deploy to Kubernetes - [x] Deploy to VM - [x] gRPC API performance tests - [x] Importer performance tests - [x] REST API performance tests ## Previewnet - [x] Deploy to Kubernetes ## Staging - [x] Deploy to Kubernetes ## Testnet - [x] Deploy to VM ## Mainnet - [x] Deploy to Kubernetes EU - [x] Deploy to Kubernetes NA - [x] Deploy to VM - [x] Deploy to ETL ### Alternatives _No response_
process
release checklist problem we need a checklist to verify the release is rolled out successfully solution preparation milestone field populated on relevant nothing open for github checks for branch are passing automated kubernetes deployment successful tag release upload release artifacts manual submission for gcp marketplace verification by google publish marketplace release publish release performance deploy to kubernetes deploy to vm grpc api performance tests importer performance tests rest api performance tests previewnet deploy to kubernetes staging deploy to kubernetes testnet deploy to vm mainnet deploy to kubernetes eu deploy to kubernetes na deploy to vm deploy to etl alternatives no response
1