Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
44,598
13,060,610,356
IssuesEvent
2020-07-30 12:42:56
jgeraigery/frost-gs-spring-boot-docker
https://api.github.com/repos/jgeraigery/frost-gs-spring-boot-docker
opened
CVE-2019-17531 (High) detected in jackson-databind-2.9.9.jar
security vulnerability
## CVE-2019-17531 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/frost-gs-spring-boot-docker/initial/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9/d6eb9817d9c7289a91f043ac5ee02a6b3cc86238/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/frost-gs-spring-boot-docker/complete/target/dependency/BOOT-INF/lib/jackson-databind-2.9.9.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/frost-gs-spring-boot-docker/commit/2913b67e67d02acdd30a738e35187b8c7922ed4d">2913b67e67d02acdd30a738e35187b8c7922ed4d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload. <p>Publish Date: 2019-10-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531>CVE-2019-17531</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531</a></p> <p>Release Date: 2019-10-12</p> <p>Fix Resolution: 2.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.10"}],"vulnerabilityIdentifier":"CVE-2019-17531","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-17531 (High) detected in jackson-databind-2.9.9.jar - ## CVE-2019-17531 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/frost-gs-spring-boot-docker/initial/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9/d6eb9817d9c7289a91f043ac5ee02a6b3cc86238/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/frost-gs-spring-boot-docker/complete/target/dependency/BOOT-INF/lib/jackson-databind-2.9.9.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/frost-gs-spring-boot-docker/commit/2913b67e67d02acdd30a738e35187b8c7922ed4d">2913b67e67d02acdd30a738e35187b8c7922ed4d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload. <p>Publish Date: 2019-10-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531>CVE-2019-17531</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531</a></p> <p>Release Date: 2019-10-12</p> <p>Fix Resolution: 2.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.10"}],"vulnerabilityIdentifier":"CVE-2019-17531","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm frost gs spring boot docker initial build gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar frost gs spring boot docker complete target dependency boot inf lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the apache extra version x jar in the classpath and an attacker can provide a jndi service to access it is possible to make the service execute a malicious payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the apache extra version x jar in the classpath and an attacker can provide a jndi service to access it is possible to make the service execute a malicious payload vulnerabilityurl
0
183,867
6,694,303,128
IssuesEvent
2017-10-10 01:01:03
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
hangouts.google.com - see bug description
browser-firefox priority-critical status-incomplete
<!-- @browser: Firefox 57.0 --> <!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:57.0) Gecko/20100101 Firefox/57.0 --> <!-- @reported_with: web --> **URL**: https://hangouts.google.com/ **Browser / Version**: Firefox 57.0 **Operating System**: Linux **Tested Another Browser**: Unknown **Problem type**: Something else **Description**: Fail loading API Google Hangout JavaScript no load. **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2017/10/1d569467-b425-460f-8f88-05bafcd75e87-thumb.jpg)](https://webcompat.com/uploads/2017/10/1d569467-b425-460f-8f88-05bafcd75e87.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
hangouts.google.com - see bug description - <!-- @browser: Firefox 57.0 --> <!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:57.0) Gecko/20100101 Firefox/57.0 --> <!-- @reported_with: web --> **URL**: https://hangouts.google.com/ **Browser / Version**: Firefox 57.0 **Operating System**: Linux **Tested Another Browser**: Unknown **Problem type**: Something else **Description**: Fail loading API Google Hangout JavaScript no load. **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2017/10/1d569467-b425-460f-8f88-05bafcd75e87-thumb.jpg)](https://webcompat.com/uploads/2017/10/1d569467-b425-460f-8f88-05bafcd75e87.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
hangouts google com see bug description url browser version firefox operating system linux tested another browser unknown problem type something else description fail loading api google hangout javascript no load steps to reproduce from with ❤️
0
5,987
8,805,374,570
IssuesEvent
2018-12-26 19:13:58
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Incorrect processing of subject scheme maps
bug preprocess preprocess/filtering preprocess/keyref priority/medium stale
## General description DITA-OT 1.7 does not handle subject scheme maps properly in regard to validating the attribute values against the specified subject schemes. While the simplest test yields the expected warning messages when the content contains attribute values not specified in the subject scheme, it requires that the subjectScheme map be referenced with type="subjectScheme" specified. (The DITA 1.2 spec does not suggest this behavior; if it is a DITA-OT requirement, we need to document it.) More realistic test cases using schemeref elements simply fail, however, regardless of whether type="subjectScheme" is specified. The DITA maps and topic are located at https://gist.github.com/keberlein/5237276 . ## Test content The following subject definitions and enumeration: ``` <subjectdef keys="operating-systems"> <subjectdef keys="linux"> <subjectdef keys="redhat"/> <subjectdef keys="suse"/> </subjectdef> <subjectdef keys="unix"/> <subjectdef keys="windows"> <subjectdef keys="windows7"/> <subjectdef keys="windows8"/> </subjectdef> </subjectdef> ``` One DITA topic with following values for the platform attribute: ubuntu, mandriva, solaris ## Test 1 **Test-1.ditamap** contains the following content: ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd"> <map> <title>Example of a map referencing a subject scheme</title> <mapref href="baseScheme-with-enumeration.ditamap" type="subjectScheme"/> <topicref href="t-creating-controlled-values.dita"/> ``` </map> **The expected results are generated -- warning messages** ``` [filter] [DOTJ049W][WARN] The attribute value platform="mandriva" on element "step" does not comply with the specified subject scheme. According to the subject scheme map, the following values are valid for the platform attribute: redhat,windows8,windows7,windows,suse,linux,unix,operating-systems … [filter] [DOTJ049W][WARN] The attribute value platform="solaris" on element "step" does not comply with the specified subject scheme. According to the subject scheme map, the following values are valid for the platform attribute: redhat,windows8,windows7,windows,suse,linux,unix,operating-systems ``` Unfortunately, the DITA-OT is incorrectly building the list of valid labels; it should NOT include operating-systems, according to the examples and guidance provided in the DITA 1.2 spec. See the examples in http://docs.oasis-open.org/dita/v1.2/os/spec/langref/subjectScheme.html#subjectScheme , as well as http://docs.oasis-open.org/dita/v1.2/os/spec/langref/enumerationdef.html#enumerationdef ## Test 2 I removed type=’’subjectScheme” from the <mapref> element Test-2.ditamap. **As a result, no warning messages are generated.** ## Test 3 Here the master map references a subjectScheme map which in turn references another subjectScheme map. This is standard use case that is outlined in the DITA 1.2 spec. **Test-3.ditamap** contains the following content: ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd"> <map> <title>Example of a map referencing a subject scheme</title> <mapref href="extensionScheme.ditamap" type="subjectScheme"/> <topicref href="t-creating-controlled-values.dita"/> ``` </map> **extensionScheme.ditamap** contains the following content: ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE subjectScheme PUBLIC "-//OASIS//DTD DITA Subject Scheme Map//EN" "subjectScheme.dtd"> <subjectScheme> <schemeref href="baseScheme.ditamap"/> <enumerationdef> <attributedef name="platform"/> <subjectdef keyref="operating-systems"/> </enumerationdef> </subjectScheme> ``` **Results:** No warning messages. ## Test 4 I modified the <schemeref> element in the extension scheme to see if it also needed type=”subjectScheme” to be specified. **extensionScheme2.ditamap** contains the following content: ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE subjectScheme PUBLIC "-//OASIS//DTD DITA Subject Scheme Map//EN" "subjectScheme.dtd"> <subjectScheme> <schemeref href="baseScheme.ditamap" type="subjectScheme"/> <enumerationdef> <attributedef name="platform"/> <subjectdef keyref="operating-systems"/> </enumerationdef> </subjectScheme> ``` **Results:** No warning messages.
3.0
Incorrect processing of subject scheme maps - ## General description DITA-OT 1.7 does not handle subject scheme maps properly in regard to validating the attribute values against the specified subject schemes. While the simplest test yields the expected warning messages when the content contains attribute values not specified in the subject scheme, it requires that the subjectScheme map be referenced with type="subjectScheme" specified. (The DITA 1.2 spec does not suggest this behavior; if it is a DITA-OT requirement, we need to document it.) More realistic test cases using schemeref elements simply fail, however, regardless of whether type="subjectScheme" is specified. The DITA maps and topic are located at https://gist.github.com/keberlein/5237276 . ## Test content The following subject definitions and enumeration: ``` <subjectdef keys="operating-systems"> <subjectdef keys="linux"> <subjectdef keys="redhat"/> <subjectdef keys="suse"/> </subjectdef> <subjectdef keys="unix"/> <subjectdef keys="windows"> <subjectdef keys="windows7"/> <subjectdef keys="windows8"/> </subjectdef> </subjectdef> ``` One DITA topic with following values for the platform attribute: ubuntu, mandriva, solaris ## Test 1 **Test-1.ditamap** contains the following content: ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd"> <map> <title>Example of a map referencing a subject scheme</title> <mapref href="baseScheme-with-enumeration.ditamap" type="subjectScheme"/> <topicref href="t-creating-controlled-values.dita"/> ``` </map> **The expected results are generated -- warning messages** ``` [filter] [DOTJ049W][WARN] The attribute value platform="mandriva" on element "step" does not comply with the specified subject scheme. According to the subject scheme map, the following values are valid for the platform attribute: redhat,windows8,windows7,windows,suse,linux,unix,operating-systems … [filter] [DOTJ049W][WARN] The attribute value platform="solaris" on element "step" does not comply with the specified subject scheme. According to the subject scheme map, the following values are valid for the platform attribute: redhat,windows8,windows7,windows,suse,linux,unix,operating-systems ``` Unfortunately, the DITA-OT is incorrectly building the list of valid labels; it should NOT include operating-systems, according to the examples and guidance provided in the DITA 1.2 spec. See the examples in http://docs.oasis-open.org/dita/v1.2/os/spec/langref/subjectScheme.html#subjectScheme , as well as http://docs.oasis-open.org/dita/v1.2/os/spec/langref/enumerationdef.html#enumerationdef ## Test 2 I removed type=’’subjectScheme” from the <mapref> element Test-2.ditamap. **As a result, no warning messages are generated.** ## Test 3 Here the master map references a subjectScheme map which in turn references another subjectScheme map. This is standard use case that is outlined in the DITA 1.2 spec. **Test-3.ditamap** contains the following content: ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd"> <map> <title>Example of a map referencing a subject scheme</title> <mapref href="extensionScheme.ditamap" type="subjectScheme"/> <topicref href="t-creating-controlled-values.dita"/> ``` </map> **extensionScheme.ditamap** contains the following content: ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE subjectScheme PUBLIC "-//OASIS//DTD DITA Subject Scheme Map//EN" "subjectScheme.dtd"> <subjectScheme> <schemeref href="baseScheme.ditamap"/> <enumerationdef> <attributedef name="platform"/> <subjectdef keyref="operating-systems"/> </enumerationdef> </subjectScheme> ``` **Results:** No warning messages. ## Test 4 I modified the <schemeref> element in the extension scheme to see if it also needed type=”subjectScheme” to be specified. **extensionScheme2.ditamap** contains the following content: ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE subjectScheme PUBLIC "-//OASIS//DTD DITA Subject Scheme Map//EN" "subjectScheme.dtd"> <subjectScheme> <schemeref href="baseScheme.ditamap" type="subjectScheme"/> <enumerationdef> <attributedef name="platform"/> <subjectdef keyref="operating-systems"/> </enumerationdef> </subjectScheme> ``` **Results:** No warning messages.
process
incorrect processing of subject scheme maps general description dita ot does not handle subject scheme maps properly in regard to validating the attribute values against the specified subject schemes while the simplest test yields the expected warning messages when the content contains attribute values not specified in the subject scheme it requires that the subjectscheme map be referenced with type subjectscheme specified the dita spec does not suggest this behavior if it is a dita ot requirement we need to document it more realistic test cases using schemeref elements simply fail however regardless of whether type subjectscheme is specified the dita maps and topic are located at test content the following subject definitions and enumeration one dita topic with following values for the platform attribute ubuntu mandriva solaris test test ditamap contains the following content example of a map referencing a subject scheme the expected results are generated warning messages the attribute value platform mandriva on element step does not comply with the specified subject scheme according to the subject scheme map the following values are valid for the platform attribute redhat windows suse linux unix operating systems … the attribute value platform solaris on element step does not comply with the specified subject scheme according to the subject scheme map the following values are valid for the platform attribute redhat windows suse linux unix operating systems unfortunately the dita ot is incorrectly building the list of valid labels it should not include operating systems according to the examples and guidance provided in the dita spec see the examples in as well as test i removed type ’’subjectscheme” from the element test ditamap as a result no warning messages are generated test here the master map references a subjectscheme map which in turn references another subjectscheme map this is standard use case that is outlined in the dita spec test ditamap contains the following content example of a map referencing a subject scheme extensionscheme ditamap contains the following content results no warning messages test i modified the element in the extension scheme to see if it also needed type ”subjectscheme” to be specified ditamap contains the following content results no warning messages
1
2,398
5,192,325,198
IssuesEvent
2017-01-22 07:14:47
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
Props warning from ToolBar in v3.0.0-beta.6
enhancement inprocess
> "Failed prop type: The prop `contentLabel` is marked as required in `Modal`, but its value is `undefined`." Since React modal v1.6 there are two required props `contentLabel` and `isOpen`: https://github.com/reactjs/react-modal#usage
1.0
Props warning from ToolBar in v3.0.0-beta.6 - > "Failed prop type: The prop `contentLabel` is marked as required in `Modal`, but its value is `undefined`." Since React modal v1.6 there are two required props `contentLabel` and `isOpen`: https://github.com/reactjs/react-modal#usage
process
props warning from toolbar in beta failed prop type the prop contentlabel is marked as required in modal but its value is undefined since react modal there are two required props contentlabel and isopen
1
67,839
13,039,490,562
IssuesEvent
2020-07-28 16:50:30
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[4.0] - [API] User Groups and Access Levels should be accessible via API!
No Code Attached Yet
### Steps to reproduce the issue There is NO API that let's us access the names of user groups and access levels for the site. com_content API returns the ID's of access level but there is no way to turn those ID's into readable text. This is also useful while creating new users and giving the ability to select User Groups for the user. ### Expected result User Groups and Access Levels should be available via the API Even if creation is not possible, it's fine but listing of ID's with names should be possible. ### Actual result No such API exists! ### System information (as much as possible) Joomla 4 Beta 2. ### Additional comments If needed, most likely for @alikon
1.0
[4.0] - [API] User Groups and Access Levels should be accessible via API! - ### Steps to reproduce the issue There is NO API that let's us access the names of user groups and access levels for the site. com_content API returns the ID's of access level but there is no way to turn those ID's into readable text. This is also useful while creating new users and giving the ability to select User Groups for the user. ### Expected result User Groups and Access Levels should be available via the API Even if creation is not possible, it's fine but listing of ID's with names should be possible. ### Actual result No such API exists! ### System information (as much as possible) Joomla 4 Beta 2. ### Additional comments If needed, most likely for @alikon
non_process
user groups and access levels should be accessible via api steps to reproduce the issue there is no api that let s us access the names of user groups and access levels for the site com content api returns the id s of access level but there is no way to turn those id s into readable text this is also useful while creating new users and giving the ability to select user groups for the user expected result user groups and access levels should be available via the api even if creation is not possible it s fine but listing of id s with names should be possible actual result no such api exists system information as much as possible joomla beta additional comments if needed most likely for alikon
0
67,474
7,048,366,680
IssuesEvent
2018-01-02 17:22:02
vmware/vic
https://api.github.com/repos/vmware/vic
opened
nightly 12/28: 5-15-NFS-Datastore panics on VCH delete
component/test kind/nightly-blocker priority/high team/lifecycle
In test cleanup, VCH delete panics and CDE. `vic-machine.log` contains a stack trace: ``` Dec 29 2017 03:00:02.143-06:00 INFO op=10546.1: ### Removing VCH #### Dec 29 2017 03:30:02.143-06:00 ERROR op=10546.1: vic/pkg/trace.(*Operation).Err: vic-machine-linux error: context deadline exceeded vic/cmd/vic-machine/delete.(*Uninstall).Run:92 vic-machine-linux vic/cmd/vic-machine/common.NewOperation:27 vic-machine-linux Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: vic/pkg/trace.(*Operation).Err: vic-machine-linux error: context deadline exceeded ...... Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: -------------------- Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: vic-machine-linux delete failed: Delete timed out: use --timeout to add more time goroutine 1 [running]: runtime/debug.Stack(0xc42001c5f0, 0xfcc546, 0x2f) /usr/local/go/src/runtime/debug/stack.go:24 +0x79 main.main.func1(0xc420067520, 0xc420084288) /go/src/github.com/vmware/vic/cmd/vic-machine/main.go:160 +0x298 panic(0xda3b60, 0x1885f90) /usr/local/go/src/runtime/panic.go:489 +0x2cf github.com/vmware/vic/pkg/vsphere/session.(*Session).Populate(0xc420206a00, 0x189bda0, 0xc4201edb80, 0xd, 0x1, 0x0) /go/src/github.com/vmware/vic/pkg/vsphere/session/session.go:358 +0xf8a github.com/vmware/vic/lib/install/validate.NewValidator(0x189bda0, 0xc4201edb80, 0xc4201df400, 0x0, 0x0, 0x0) /go/src/github.com/vmware/vic/lib/install/validate/validator.go:159 +0x978 github.com/vmware/vic/cmd/vic-machine/delete.(*Uninstall).Run(0xc4201d58f0, 0xc42017db80, 0x1895de0, 0xc4204766e0) /go/src/github.com/vmware/vic/cmd/vic-machine/delete/delete.go:112 +0x2d6 github.com/vmware/vic/cmd/vic-machine/delete.(*Uninstall).Run-fm(0xc42017db80, 0x1010101010100, 0xc42017db80) /go/src/github.com/vmware/vic/cmd/vic-machine/main.go:75 +0x34 github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.HandleAction(0xd56cc0, 0xc4201d5940, 0xc42017db80, 0xc42006b100, 0x0) /go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/app.go:485 +0xd4 github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.Command.Run(0xf8ef69, 0x6, 0x0, 0x0, 0x0, 0x0, 0x0, 0xfbf004, 0x23, 0x0, ...) /go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/command.go:193 +0xb72 github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.(*App).Run(0xc420067520, 0xc420088000, 0xa, 0xa, 0x0, 0x0) /go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/app.go:250 +0x7d0 main.main() /go/src/github.com/vmware/vic/cmd/vic-machine/main.go:165 +0x2233 ``` From the stack trace, the panic happens during validation session populate, where it grabs VM folder from datacenter using govmomi: ``` if s.Datacenter != nil { folders, err := s.Datacenter.Folders(op) if err != nil { errs = append(errs, fmt.Sprintf("Failure finding folders (%s): %s", s.DatacenterPath, err.Error())) } else { op.Debugf("Cached folders: %s", s.DatacenterPath) } s.VMFolder = folders.VmFolder } ``` In line `s.VMFolder = folders.VmFolder`. https://github.com/vmware/vic/blob/master/pkg/vsphere/session/session.go#L358 Log: [5-15-NFS-Datastore.zip](https://github.com/vmware/vic/files/1598194/5-15-NFS-Datastore.zip) Cc: @dougm @hickeng
1.0
nightly 12/28: 5-15-NFS-Datastore panics on VCH delete - In test cleanup, VCH delete panics and CDE. `vic-machine.log` contains a stack trace: ``` Dec 29 2017 03:00:02.143-06:00 INFO op=10546.1: ### Removing VCH #### Dec 29 2017 03:30:02.143-06:00 ERROR op=10546.1: vic/pkg/trace.(*Operation).Err: vic-machine-linux error: context deadline exceeded vic/cmd/vic-machine/delete.(*Uninstall).Run:92 vic-machine-linux vic/cmd/vic-machine/common.NewOperation:27 vic-machine-linux Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: vic/pkg/trace.(*Operation).Err: vic-machine-linux error: context deadline exceeded ...... Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: -------------------- Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: vic-machine-linux delete failed: Delete timed out: use --timeout to add more time goroutine 1 [running]: runtime/debug.Stack(0xc42001c5f0, 0xfcc546, 0x2f) /usr/local/go/src/runtime/debug/stack.go:24 +0x79 main.main.func1(0xc420067520, 0xc420084288) /go/src/github.com/vmware/vic/cmd/vic-machine/main.go:160 +0x298 panic(0xda3b60, 0x1885f90) /usr/local/go/src/runtime/panic.go:489 +0x2cf github.com/vmware/vic/pkg/vsphere/session.(*Session).Populate(0xc420206a00, 0x189bda0, 0xc4201edb80, 0xd, 0x1, 0x0) /go/src/github.com/vmware/vic/pkg/vsphere/session/session.go:358 +0xf8a github.com/vmware/vic/lib/install/validate.NewValidator(0x189bda0, 0xc4201edb80, 0xc4201df400, 0x0, 0x0, 0x0) /go/src/github.com/vmware/vic/lib/install/validate/validator.go:159 +0x978 github.com/vmware/vic/cmd/vic-machine/delete.(*Uninstall).Run(0xc4201d58f0, 0xc42017db80, 0x1895de0, 0xc4204766e0) /go/src/github.com/vmware/vic/cmd/vic-machine/delete/delete.go:112 +0x2d6 github.com/vmware/vic/cmd/vic-machine/delete.(*Uninstall).Run-fm(0xc42017db80, 0x1010101010100, 0xc42017db80) /go/src/github.com/vmware/vic/cmd/vic-machine/main.go:75 +0x34 github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.HandleAction(0xd56cc0, 0xc4201d5940, 0xc42017db80, 0xc42006b100, 0x0) /go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/app.go:485 +0xd4 github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.Command.Run(0xf8ef69, 0x6, 0x0, 0x0, 0x0, 0x0, 0x0, 0xfbf004, 0x23, 0x0, ...) /go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/command.go:193 +0xb72 github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.(*App).Run(0xc420067520, 0xc420088000, 0xa, 0xa, 0x0, 0x0) /go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/app.go:250 +0x7d0 main.main() /go/src/github.com/vmware/vic/cmd/vic-machine/main.go:165 +0x2233 ``` From the stack trace, the panic happens during validation session populate, where it grabs VM folder from datacenter using govmomi: ``` if s.Datacenter != nil { folders, err := s.Datacenter.Folders(op) if err != nil { errs = append(errs, fmt.Sprintf("Failure finding folders (%s): %s", s.DatacenterPath, err.Error())) } else { op.Debugf("Cached folders: %s", s.DatacenterPath) } s.VMFolder = folders.VmFolder } ``` In line `s.VMFolder = folders.VmFolder`. https://github.com/vmware/vic/blob/master/pkg/vsphere/session/session.go#L358 Log: [5-15-NFS-Datastore.zip](https://github.com/vmware/vic/files/1598194/5-15-NFS-Datastore.zip) Cc: @dougm @hickeng
non_process
nightly nfs datastore panics on vch delete in test cleanup vch delete panics and cde vic machine log contains a stack trace dec info op removing vch dec error op vic pkg trace operation err vic machine linux error context deadline exceeded vic cmd vic machine delete uninstall run vic machine linux vic cmd vic machine common newoperation vic machine linux dec error op vic pkg trace operation err vic machine linux error context deadline exceeded dec error op dec error op vic machine linux delete failed delete timed out use timeout to add more time goroutine runtime debug stack usr local go src runtime debug stack go main main go src github com vmware vic cmd vic machine main go panic usr local go src runtime panic go github com vmware vic pkg vsphere session session populate go src github com vmware vic pkg vsphere session session go github com vmware vic lib install validate newvalidator go src github com vmware vic lib install validate validator go github com vmware vic cmd vic machine delete uninstall run go src github com vmware vic cmd vic machine delete delete go github com vmware vic cmd vic machine delete uninstall run fm go src github com vmware vic cmd vic machine main go github com vmware vic vendor gopkg in urfave cli handleaction go src github com vmware vic vendor gopkg in urfave cli app go github com vmware vic vendor gopkg in urfave cli command run go src github com vmware vic vendor gopkg in urfave cli command go github com vmware vic vendor gopkg in urfave cli app run go src github com vmware vic vendor gopkg in urfave cli app go main main go src github com vmware vic cmd vic machine main go from the stack trace the panic happens during validation session populate where it grabs vm folder from datacenter using govmomi if s datacenter nil folders err s datacenter folders op if err nil errs append errs fmt sprintf failure finding folders s s s datacenterpath err error else op debugf cached folders s s datacenterpath s vmfolder folders vmfolder in line s vmfolder folders vmfolder log cc dougm hickeng
0
293,240
25,279,058,570
IssuesEvent
2022-11-16 14:36:21
cockroachdb/pebble
https://api.github.com/repos/cockroachdb/pebble
closed
github.com/cockroachdb/pebble/internal/metamorphic: TestMeta failed
O-robot C-test-failure metamorphic-failure branch-refs/heads/master T-storage A-storage
github.com/cockroachdb/pebble/internal/metamorphic.TestMeta [failed](https://buildConfiguration/Cockroach_Nightlies_Pebble_Metamorphic/7423668?buildTab=log) with [artifacts](https://buildConfiguration/Cockroach_Nightlies_Pebble_Metamorphic/7423668?buildTab=artifacts#meta) on refs/heads/master @ [7b30bd86ff65](https://github.com/cockroachdb/pebble/commits/7b30bd86ff65): ``` === CONT TestMeta/execution/standard-023 === RUN TestMeta/compare/standard-012 === RUN TestMeta/execution === RUN TestMeta/execution/standard-010 === PAUSE TestMeta/execution/standard-010 === CONT TestMeta/execution/standard-010 === RUN TestMeta/execution/random-024 === PAUSE TestMeta/execution/random-024 === CONT TestMeta/execution/random-024 === RUN TestMeta/execution/random-013 === PAUSE TestMeta/execution/random-013 === CONT TestMeta/execution/random-013 === RUN TestMeta/execution/random-023 === PAUSE TestMeta/execution/random-023 === CONT TestMeta/execution/random-023 === RUN TestMeta/execution/random-015 === PAUSE TestMeta/execution/random-015 === CONT TestMeta/execution/random-015 === RUN TestMeta/execution/random-020 === PAUSE TestMeta/execution/random-020 === CONT TestMeta/execution/random-020 === RUN TestMeta/execution/standard-021 === PAUSE TestMeta/execution/standard-021 === CONT TestMeta/execution/standard-021 === RUN TestMeta/execution/random-007 === PAUSE TestMeta/execution/random-007 === CONT TestMeta/execution/random-007 === RUN TestMeta/execution/random-012 === PAUSE TestMeta/execution/random-012 === CONT TestMeta/execution/random-012 === RUN TestMeta/execution/random-006 === PAUSE TestMeta/execution/random-006 === CONT TestMeta/execution/random-006 === RUN TestMeta/execution/standard-011 === PAUSE TestMeta/execution/standard-011 === CONT TestMeta/execution/standard-011 === RUN TestMeta/execution/standard-020 === PAUSE TestMeta/execution/standard-020 === CONT TestMeta/execution/standard-020 === RUN TestMeta/compare/standard-007 === RUN TestMeta/compare/standard-008 === RUN TestMeta/execution/standard-002 === PAUSE TestMeta/execution/standard-002 === CONT TestMeta/execution/standard-002 === RUN TestMeta/execution/random-008 === PAUSE TestMeta/execution/random-008 === CONT TestMeta/execution/random-008 === RUN TestMeta/execution/random-014 === PAUSE TestMeta/execution/random-014 === CONT TestMeta/execution/random-014 ``` <details><summary>Help</summary> <p> To reproduce, try: ```bash go test -mod=vendor -tags 'invariants' -exec 'stress -p 1' -timeout 0 -test.v -run TestMeta$ ./internal/metamorphic -seed 1668072880985143374 -ops "uniform:5000-10000" ``` </p> </details> <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestMeta.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
github.com/cockroachdb/pebble/internal/metamorphic: TestMeta failed - github.com/cockroachdb/pebble/internal/metamorphic.TestMeta [failed](https://buildConfiguration/Cockroach_Nightlies_Pebble_Metamorphic/7423668?buildTab=log) with [artifacts](https://buildConfiguration/Cockroach_Nightlies_Pebble_Metamorphic/7423668?buildTab=artifacts#meta) on refs/heads/master @ [7b30bd86ff65](https://github.com/cockroachdb/pebble/commits/7b30bd86ff65): ``` === CONT TestMeta/execution/standard-023 === RUN TestMeta/compare/standard-012 === RUN TestMeta/execution === RUN TestMeta/execution/standard-010 === PAUSE TestMeta/execution/standard-010 === CONT TestMeta/execution/standard-010 === RUN TestMeta/execution/random-024 === PAUSE TestMeta/execution/random-024 === CONT TestMeta/execution/random-024 === RUN TestMeta/execution/random-013 === PAUSE TestMeta/execution/random-013 === CONT TestMeta/execution/random-013 === RUN TestMeta/execution/random-023 === PAUSE TestMeta/execution/random-023 === CONT TestMeta/execution/random-023 === RUN TestMeta/execution/random-015 === PAUSE TestMeta/execution/random-015 === CONT TestMeta/execution/random-015 === RUN TestMeta/execution/random-020 === PAUSE TestMeta/execution/random-020 === CONT TestMeta/execution/random-020 === RUN TestMeta/execution/standard-021 === PAUSE TestMeta/execution/standard-021 === CONT TestMeta/execution/standard-021 === RUN TestMeta/execution/random-007 === PAUSE TestMeta/execution/random-007 === CONT TestMeta/execution/random-007 === RUN TestMeta/execution/random-012 === PAUSE TestMeta/execution/random-012 === CONT TestMeta/execution/random-012 === RUN TestMeta/execution/random-006 === PAUSE TestMeta/execution/random-006 === CONT TestMeta/execution/random-006 === RUN TestMeta/execution/standard-011 === PAUSE TestMeta/execution/standard-011 === CONT TestMeta/execution/standard-011 === RUN TestMeta/execution/standard-020 === PAUSE TestMeta/execution/standard-020 === CONT TestMeta/execution/standard-020 === RUN TestMeta/compare/standard-007 === RUN TestMeta/compare/standard-008 === RUN TestMeta/execution/standard-002 === PAUSE TestMeta/execution/standard-002 === CONT TestMeta/execution/standard-002 === RUN TestMeta/execution/random-008 === PAUSE TestMeta/execution/random-008 === CONT TestMeta/execution/random-008 === RUN TestMeta/execution/random-014 === PAUSE TestMeta/execution/random-014 === CONT TestMeta/execution/random-014 ``` <details><summary>Help</summary> <p> To reproduce, try: ```bash go test -mod=vendor -tags 'invariants' -exec 'stress -p 1' -timeout 0 -test.v -run TestMeta$ ./internal/metamorphic -seed 1668072880985143374 -ops "uniform:5000-10000" ``` </p> </details> <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestMeta.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_process
github com cockroachdb pebble internal metamorphic testmeta failed github com cockroachdb pebble internal metamorphic testmeta with on refs heads master cont testmeta execution standard run testmeta compare standard run testmeta execution run testmeta execution standard pause testmeta execution standard cont testmeta execution standard run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution standard pause testmeta execution standard cont testmeta execution standard run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution standard pause testmeta execution standard cont testmeta execution standard run testmeta execution standard pause testmeta execution standard cont testmeta execution standard run testmeta compare standard run testmeta compare standard run testmeta execution standard pause testmeta execution standard cont testmeta execution standard run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random help to reproduce try bash go test mod vendor tags invariants exec stress p timeout test v run testmeta internal metamorphic seed ops uniform
0
390,566
26,866,379,752
IssuesEvent
2023-02-04 00:35:15
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[docs] [infra] allow for unversioned content areas
area/documentation
Relates to #10790... If we move version-independent sections (like cloud, faq, releases, and others) into a separate folder, we need to figure out to build a complete left nav that incorporates those version-independent sections. The left nav should pull its content from the version you're currently in (latest, stable, v2.6, etc.), and default to pulling from latest if you're currently IN an unversioned section.
1.0
[docs] [infra] allow for unversioned content areas - Relates to #10790... If we move version-independent sections (like cloud, faq, releases, and others) into a separate folder, we need to figure out to build a complete left nav that incorporates those version-independent sections. The left nav should pull its content from the version you're currently in (latest, stable, v2.6, etc.), and default to pulling from latest if you're currently IN an unversioned section.
non_process
allow for unversioned content areas relates to if we move version independent sections like cloud faq releases and others into a separate folder we need to figure out to build a complete left nav that incorporates those version independent sections the left nav should pull its content from the version you re currently in latest stable etc and default to pulling from latest if you re currently in an unversioned section
0
15,615
19,753,063,200
IssuesEvent
2022-01-15 09:03:41
googleapis/java-pubsublite-spark
https://api.github.com/repos/googleapis/java-pubsublite-spark
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'pubsublite-spark' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'pubsublite-spark' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname pubsublite spark invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
22,066
30,591,680,572
IssuesEvent
2023-07-21 17:37:40
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
transform process won't load in metrics pipeline since v0.71.0+
bug processor/transform needs triage
### Component(s) processor/transform ### What happened? ## Description The [transform](https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/transformprocessor) processor won't load into the `metrics` pipeline since version 0.71.0. ## Steps to Reproduce 1. Copy config below into `agent.yaml` 2. Run `otelcol-contrib --config agent.yaml` 3. Check the logs and see that `transform` is not loaded *NOTES*: The `filter` processor is loaded but not the `transform` processor. ## Expected Result Transform processor should be loaded. ## Actual Result Transform processor didn't load. ### Collector version 0.81.0 ### Environment information ## Environment OS: macOS Ventura 13.4 Also tried the container and Ubuntu, same issue. ### OpenTelemetry Collector configuration ```yaml receivers: hostmetrics: collection_interval: 30s scrapers: load: processors: filter: metrics: include: match_type: regexp metric_names: - jfrt_.* transform: metric_statements: - context: resource statements: - delete_key(attributes, "service.instance.id") - context: datapoint statements: - delete_key(attributes, "service.instance.id") exporters: logging: service: pipelines: metrics: receivers: - hostmetrics processors: - filter - transform exporters: - logging ``` ### Log output ```shell 2023-07-20T14:03:05.008-0400 info service/telemetry.go:81 Setting up own telemetry... 2023-07-20T14:03:05.008-0400 info service/telemetry.go:104 Serving Prometheus metrics {"address": ":8888", "level": "Basic"} 2023-07-20T14:03:05.008-0400 info exporter@v0.81.0/exporter.go:275 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-20T14:03:05.008-0400 info filterprocessor@v0.81.0/metrics.go:89 Metric filter configured {"kind": "processor", "name": "filter", "pipeline": "metrics", "include match_type": "regexp", "in clude expressions": [], "include metric names": ["jfrt_.*"], "include metrics with resource attributes": null, "exclude match_type": "", "exclude expressions": [], "exclude metric names": [], "exclude metrics w ith resource attributes": null} 2023-07-20T14:03:05.008-0400 info service/service.go:131 Starting otelcol-contrib... {"Version": "0.81.0", "NumCPU": 8} 2023-07-20T14:03:05.008-0400 info extensions/extensions.go:30 Starting extensions... 2023-07-20T14:03:05.008-0400 info service/service.go:148 Everything is ready. Begin running and processing data. ``` ### Additional context Below is the log for version 0.70.0 and `transform` processor is loaded. ```console 2023-07-20T14:09:38.895-0400 info service/telemetry.go:90 Setting up own telemetry... 2023-07-20T14:09:38.895-0400 info service/telemetry.go:116 Serving Prometheus metrics {"address": ":8888", "level": "Basic"} 2023-07-20T14:09:38.895-0400 info exporter/exporter.go:290 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-20T14:09:38.896-0400 info filterprocessor@v0.70.0/metrics.go:97 Metric filter configured {"kind": "processor", "name": "filter", "pipeline": "metrics", "include match_type": "regexp", "in clude expressions": [], "include metric names": ["jfrt_.*"], "include metrics with resource attributes": null, "exclude match_type": "", "exclude expressions": [], "exclude metric names": [], "exclude metrics w ith resource attributes": null} 2023-07-20T14:09:38.897-0400 info service/service.go:128 Starting otelcol-contrib... {"Version": "0.70.0", "NumCPU": 8} 2023-07-20T14:09:38.897-0400 info extensions/extensions.go:41 Starting extensions... 2023-07-20T14:09:38.897-0400 info service/pipelines.go:86 Starting exporters... 2023-07-20T14:09:38.897-0400 info service/pipelines.go:90 Exporter is starting... {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:94 Exporter started. {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:98 Starting processors... 2023-07-20T14:09:38.897-0400 info service/pipelines.go:102 Processor is starting... {"kind": "processor", "name": "transform", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:106 Processor started. {"kind": "processor", "name": "transform", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:102 Processor is starting... {"kind": "processor", "name": "filter", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:106 Processor started. {"kind": "processor", "name": "filter", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:110 Starting receivers... 2023-07-20T14:09:38.897-0400 info service/pipelines.go:114 Receiver is starting... {"kind": "receiver", "name": "hostmetrics", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:118 Receiver started. {"kind": "receiver", "name": "hostmetrics", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/service.go:145 Everything is ready. Begin running and processing data. ```
1.0
transform process won't load in metrics pipeline since v0.71.0+ - ### Component(s) processor/transform ### What happened? ## Description The [transform](https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/transformprocessor) processor won't load into the `metrics` pipeline since version 0.71.0. ## Steps to Reproduce 1. Copy config below into `agent.yaml` 2. Run `otelcol-contrib --config agent.yaml` 3. Check the logs and see that `transform` is not loaded *NOTES*: The `filter` processor is loaded but not the `transform` processor. ## Expected Result Transform processor should be loaded. ## Actual Result Transform processor didn't load. ### Collector version 0.81.0 ### Environment information ## Environment OS: macOS Ventura 13.4 Also tried the container and Ubuntu, same issue. ### OpenTelemetry Collector configuration ```yaml receivers: hostmetrics: collection_interval: 30s scrapers: load: processors: filter: metrics: include: match_type: regexp metric_names: - jfrt_.* transform: metric_statements: - context: resource statements: - delete_key(attributes, "service.instance.id") - context: datapoint statements: - delete_key(attributes, "service.instance.id") exporters: logging: service: pipelines: metrics: receivers: - hostmetrics processors: - filter - transform exporters: - logging ``` ### Log output ```shell 2023-07-20T14:03:05.008-0400 info service/telemetry.go:81 Setting up own telemetry... 2023-07-20T14:03:05.008-0400 info service/telemetry.go:104 Serving Prometheus metrics {"address": ":8888", "level": "Basic"} 2023-07-20T14:03:05.008-0400 info exporter@v0.81.0/exporter.go:275 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-20T14:03:05.008-0400 info filterprocessor@v0.81.0/metrics.go:89 Metric filter configured {"kind": "processor", "name": "filter", "pipeline": "metrics", "include match_type": "regexp", "in clude expressions": [], "include metric names": ["jfrt_.*"], "include metrics with resource attributes": null, "exclude match_type": "", "exclude expressions": [], "exclude metric names": [], "exclude metrics w ith resource attributes": null} 2023-07-20T14:03:05.008-0400 info service/service.go:131 Starting otelcol-contrib... {"Version": "0.81.0", "NumCPU": 8} 2023-07-20T14:03:05.008-0400 info extensions/extensions.go:30 Starting extensions... 2023-07-20T14:03:05.008-0400 info service/service.go:148 Everything is ready. Begin running and processing data. ``` ### Additional context Below is the log for version 0.70.0 and `transform` processor is loaded. ```console 2023-07-20T14:09:38.895-0400 info service/telemetry.go:90 Setting up own telemetry... 2023-07-20T14:09:38.895-0400 info service/telemetry.go:116 Serving Prometheus metrics {"address": ":8888", "level": "Basic"} 2023-07-20T14:09:38.895-0400 info exporter/exporter.go:290 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-20T14:09:38.896-0400 info filterprocessor@v0.70.0/metrics.go:97 Metric filter configured {"kind": "processor", "name": "filter", "pipeline": "metrics", "include match_type": "regexp", "in clude expressions": [], "include metric names": ["jfrt_.*"], "include metrics with resource attributes": null, "exclude match_type": "", "exclude expressions": [], "exclude metric names": [], "exclude metrics w ith resource attributes": null} 2023-07-20T14:09:38.897-0400 info service/service.go:128 Starting otelcol-contrib... {"Version": "0.70.0", "NumCPU": 8} 2023-07-20T14:09:38.897-0400 info extensions/extensions.go:41 Starting extensions... 2023-07-20T14:09:38.897-0400 info service/pipelines.go:86 Starting exporters... 2023-07-20T14:09:38.897-0400 info service/pipelines.go:90 Exporter is starting... {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:94 Exporter started. {"kind": "exporter", "data_type": "metrics", "name": "logging"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:98 Starting processors... 2023-07-20T14:09:38.897-0400 info service/pipelines.go:102 Processor is starting... {"kind": "processor", "name": "transform", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:106 Processor started. {"kind": "processor", "name": "transform", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:102 Processor is starting... {"kind": "processor", "name": "filter", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:106 Processor started. {"kind": "processor", "name": "filter", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:110 Starting receivers... 2023-07-20T14:09:38.897-0400 info service/pipelines.go:114 Receiver is starting... {"kind": "receiver", "name": "hostmetrics", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/pipelines.go:118 Receiver started. {"kind": "receiver", "name": "hostmetrics", "pipeline": "metrics"} 2023-07-20T14:09:38.897-0400 info service/service.go:145 Everything is ready. Begin running and processing data. ```
process
transform process won t load in metrics pipeline since component s processor transform what happened description the processor won t load into the metrics pipeline since version steps to reproduce copy config below into agent yaml run otelcol contrib config agent yaml check the logs and see that transform is not loaded notes the filter processor is loaded but not the transform processor expected result transform processor should be loaded actual result transform processor didn t load collector version environment information environment os macos ventura also tried the container and ubuntu same issue opentelemetry collector configuration yaml receivers hostmetrics collection interval scrapers load processors filter metrics include match type regexp metric names jfrt transform metric statements context resource statements delete key attributes service instance id context datapoint statements delete key attributes service instance id exporters logging service pipelines metrics receivers hostmetrics processors filter transform exporters logging log output shell info service telemetry go setting up own telemetry info service telemetry go serving prometheus metrics address level basic info exporter exporter go development component may change in the future kind exporter data type metrics name logging info filterprocessor metrics go metric filter configured kind processor name filter pipeline metrics include match type regexp in clude expressions include metric names include metrics with resource attributes null exclude match type exclude expressions exclude metric names exclude metrics w ith resource attributes null info service service go starting otelcol contrib version numcpu info extensions extensions go starting extensions info service service go everything is ready begin running and processing data additional context below is the log for version and transform processor is loaded console info service telemetry go setting up own telemetry info service telemetry go serving prometheus metrics address level basic info exporter exporter go development component may change in the future kind exporter data type metrics name logging info filterprocessor metrics go metric filter configured kind processor name filter pipeline metrics include match type regexp in clude expressions include metric names include metrics with resource attributes null exclude match type exclude expressions exclude metric names exclude metrics w ith resource attributes null info service service go starting otelcol contrib version numcpu info extensions extensions go starting extensions info service pipelines go starting exporters info service pipelines go exporter is starting kind exporter data type metrics name logging info service pipelines go exporter started kind exporter data type metrics name logging info service pipelines go starting processors info service pipelines go processor is starting kind processor name transform pipeline metrics info service pipelines go processor started kind processor name transform pipeline metrics info service pipelines go processor is starting kind processor name filter pipeline metrics info service pipelines go processor started kind processor name filter pipeline metrics info service pipelines go starting receivers info service pipelines go receiver is starting kind receiver name hostmetrics pipeline metrics info service pipelines go receiver started kind receiver name hostmetrics pipeline metrics info service service go everything is ready begin running and processing data
1
14,462
17,568,534,799
IssuesEvent
2021-08-14 07:19:31
Geoxor/Sakuria
https://api.github.com/repos/Geoxor/Sakuria
opened
Improve the GIF Encoder's performance
bug image processors
This shit is really bottlenecking the render speed ![image](https://user-images.githubusercontent.com/34042825/129438540-a433689e-0131-4a87-a018-1e1aa1a657ce.png)
1.0
Improve the GIF Encoder's performance - This shit is really bottlenecking the render speed ![image](https://user-images.githubusercontent.com/34042825/129438540-a433689e-0131-4a87-a018-1e1aa1a657ce.png)
process
improve the gif encoder s performance this shit is really bottlenecking the render speed
1
16,942
22,294,718,295
IssuesEvent
2022-06-12 21:56:48
0xffset/rOSt
https://api.github.com/repos/0xffset/rOSt
closed
Memory freeing on process termination
memory processes
Currently when a process is created the frames allocated to it are never freed back. We need to add freeing the process's memory after it's terminated. To do that we would probably need to rewrite the frame allocator to support deallocation.
1.0
Memory freeing on process termination - Currently when a process is created the frames allocated to it are never freed back. We need to add freeing the process's memory after it's terminated. To do that we would probably need to rewrite the frame allocator to support deallocation.
process
memory freeing on process termination currently when a process is created the frames allocated to it are never freed back we need to add freeing the process s memory after it s terminated to do that we would probably need to rewrite the frame allocator to support deallocation
1
3,720
6,732,884,771
IssuesEvent
2017-10-18 13:11:09
lockedata/rcms
https://api.github.com/repos/lockedata/rcms
opened
Build agenda
conference team odoo processes
## Detailed task - Create a schedule over multiple rooms (and days if required) - Publish agenda ## Assessing the task Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks. Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback. ## Extra Info - Site: [odoo](//http://188.166.159.192:8069) - System documentation: [odoo docs](https://www.odoo.com/page/docs) - Role: Conference team - Area: Processes
1.0
Build agenda - ## Detailed task - Create a schedule over multiple rooms (and days if required) - Publish agenda ## Assessing the task Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks. Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback. ## Extra Info - Site: [odoo](//http://188.166.159.192:8069) - System documentation: [odoo docs](https://www.odoo.com/page/docs) - Role: Conference team - Area: Processes
process
build agenda detailed task create a schedule over multiple rooms and days if required publish agenda assessing the task try to perform the task use google and the system documentation to help part of what we re trying to assess how easy it is for people to work out how to do tasks use a 👍 reaction to this task if you were able to perform the task use a 👎 reaction to the task if you could not complete it add a reply with any comments or feedback extra info site system documentation role conference team area processes
1
171,174
20,948,546,491
IssuesEvent
2022-03-26 08:18:02
npenin/thewheel-dotnet
https://api.github.com/repos/npenin/thewheel-dotnet
opened
CVE-2017-0249 (High) detected in system.net.http.4.3.0.nupkg
security vulnerability
## CVE-2017-0249 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary> <p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p> <p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p> <p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p> <p> Dependency Hierarchy: - mstest.testframework.2.2.3.nupkg (Root Library) - netstandard.library.1.6.1.nupkg - :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/682d2be920def33815c7fdb6916d426a96c8df76">682d2be920def33815c7fdb6916d426a96c8df76</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An elevation of privilege vulnerability exists when the ASP.NET Core fails to properly sanitize web requests. <p>Publish Date: 2017-05-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-0249>CVE-2017-0249</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/aspnet/Announcements/issues/239">https://github.com/aspnet/Announcements/issues/239</a></p> <p>Release Date: 2017-05-12</p> <p>Fix Resolution: System.Text.Encodings.Web - 4.0.1,4.3.1;System.Net.Http - 4.1.2,4.3.2;System.Net.Http.WinHttpHandler - 4.0.2,4.3.1;System.Net.Security - 4.0.1,4.3.1;System.Net.WebSockets.Client - 4.0.1,4.3.1;Microsoft.AspNetCore.Mvc - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Core - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Abstractions - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ApiExplorer - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Cors - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.DataAnnotations - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Json - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Xml - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Localization - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor.Host - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.TagHelpers - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ViewFeatures - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.WebApiCompatShim - 1.0.4,1.1.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-0249 (High) detected in system.net.http.4.3.0.nupkg - ## CVE-2017-0249 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary> <p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p> <p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p> <p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p> <p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p> <p> Dependency Hierarchy: - mstest.testframework.2.2.3.nupkg (Root Library) - netstandard.library.1.6.1.nupkg - :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/682d2be920def33815c7fdb6916d426a96c8df76">682d2be920def33815c7fdb6916d426a96c8df76</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An elevation of privilege vulnerability exists when the ASP.NET Core fails to properly sanitize web requests. <p>Publish Date: 2017-05-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-0249>CVE-2017-0249</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/aspnet/Announcements/issues/239">https://github.com/aspnet/Announcements/issues/239</a></p> <p>Release Date: 2017-05-12</p> <p>Fix Resolution: System.Text.Encodings.Web - 4.0.1,4.3.1;System.Net.Http - 4.1.2,4.3.2;System.Net.Http.WinHttpHandler - 4.0.2,4.3.1;System.Net.Security - 4.0.1,4.3.1;System.Net.WebSockets.Client - 4.0.1,4.3.1;Microsoft.AspNetCore.Mvc - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Core - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Abstractions - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ApiExplorer - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Cors - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.DataAnnotations - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Json - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Xml - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Localization - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor.Host - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.TagHelpers - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ViewFeatures - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.WebApiCompatShim - 1.0.4,1.1.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in system net http nupkg cve high severity vulnerability vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that library home page a href path to dependency file thewheel tests thewheel tests csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg dependency hierarchy mstest testframework nupkg root library netstandard library nupkg x system net http nupkg vulnerable library found in head commit a href found in base branch master vulnerability details an elevation of privilege vulnerability exists when the asp net core fails to properly sanitize web requests publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution system text encodings web system net http system net http winhttphandler system net security system net websockets client microsoft aspnetcore mvc microsoft aspnetcore mvc core microsoft aspnetcore mvc abstractions microsoft aspnetcore mvc apiexplorer microsoft aspnetcore mvc cors microsoft aspnetcore mvc dataannotations microsoft aspnetcore mvc formatters json microsoft aspnetcore mvc formatters xml microsoft aspnetcore mvc localization microsoft aspnetcore mvc razor host microsoft aspnetcore mvc razor microsoft aspnetcore mvc taghelpers microsoft aspnetcore mvc viewfeatures microsoft aspnetcore mvc webapicompatshim step up your open source security game with whitesource
0
16,859
22,139,931,280
IssuesEvent
2022-06-03 05:21:57
deepset-ai/haystack
https://api.github.com/repos/deepset-ai/haystack
closed
split_length not working as expected
type:question topic:preprocessing journey:intermediate
**Question** I am trying to split my documents such that I have multiple sentences with 500 words each. But after executing the `PreProcessor` function, I get the warning: ``` > 0%| | 0/1052 [00:00<?, ?docs/s]WARNING - haystack.nodes.preprocessor.preprocessor - One or more sentence found with word count higher than the split length. ``` Upon checking the count of sentences, I see that nothing was split. I fed in 1052 documents and the resulting list also has 1052 elements. What am I doing wrong here? This is the code I am using: ``` def pre_process(source): all_docs = convert_files_to_docs(dir_path=source) preprocessor = PreProcessor( clean_empty_lines=True, clean_whitespace=True, clean_header_footer=False, split_by="word", split_length=500, split_respect_sentence_boundary=True, ) processed_docs = preprocessor.process(all_docs) print(f"Number of input files: {len(all_docs)}\nNumber of output files: {len(processed_docs)}") return [item.content for item in processed_docs] ``` **FAQ Check** - [x] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)?
1.0
split_length not working as expected - **Question** I am trying to split my documents such that I have multiple sentences with 500 words each. But after executing the `PreProcessor` function, I get the warning: ``` > 0%| | 0/1052 [00:00<?, ?docs/s]WARNING - haystack.nodes.preprocessor.preprocessor - One or more sentence found with word count higher than the split length. ``` Upon checking the count of sentences, I see that nothing was split. I fed in 1052 documents and the resulting list also has 1052 elements. What am I doing wrong here? This is the code I am using: ``` def pre_process(source): all_docs = convert_files_to_docs(dir_path=source) preprocessor = PreProcessor( clean_empty_lines=True, clean_whitespace=True, clean_header_footer=False, split_by="word", split_length=500, split_respect_sentence_boundary=True, ) processed_docs = preprocessor.process(all_docs) print(f"Number of input files: {len(all_docs)}\nNumber of output files: {len(processed_docs)}") return [item.content for item in processed_docs] ``` **FAQ Check** - [x] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)?
process
split length not working as expected question i am trying to split my documents such that i have multiple sentences with words each but after executing the preprocessor function i get the warning warning haystack nodes preprocessor preprocessor one or more sentence found with word count higher than the split length upon checking the count of sentences i see that nothing was split i fed in documents and the resulting list also has elements what am i doing wrong here this is the code i am using def pre process source all docs convert files to docs dir path source preprocessor preprocessor clean empty lines true clean whitespace true clean header footer false split by word split length split respect sentence boundary true processed docs preprocessor process all docs print f number of input files len all docs nnumber of output files len processed docs return faq check have you had a look at
1
17,796
23,723,833,162
IssuesEvent
2022-08-30 17:38:59
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Release checklist 0.64
enhancement process
### Problem We need a checklist to verify the release is rolled out successfully. ### Solution ## Preparation - [x] Milestone field populated on relevant [issues](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aclosed+no%3Amilestone+sort%3Aupdated-desc) - [x] Nothing open for [milestone](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aopen+sort%3Aupdated-desc+milestone%3A0.64.0) - [x] GitHub checks for branch are passing - [x] Automated Kubernetes deployment successful - [x] Tag release - [x] Upload release artifacts - [x] Publish release ## Integration - [ ] Deploy to VM ## Performance - [ ] Deploy to Kubernetes - [ ] Deploy to VM - [ ] gRPC API performance tests - [ ] Importer performance tests - [ ] REST API performance tests - [ ] Migrations tested against mainnet clone ## Previewnet - [ ] Deploy to VM ## Staging - [ ] Deploy to Kubernetes EU - [ ] Deploy to Kubernetes NA ## Testnet - [ ] Deploy to VM ## Mainnet - [ ] Deploy to Kubernetes EU - [ ] Deploy to Kubernetes NA - [ ] Deploy to VM - [ ] Deploy to ETL ### Alternatives _No response_
1.0
Release checklist 0.64 - ### Problem We need a checklist to verify the release is rolled out successfully. ### Solution ## Preparation - [x] Milestone field populated on relevant [issues](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aclosed+no%3Amilestone+sort%3Aupdated-desc) - [x] Nothing open for [milestone](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aopen+sort%3Aupdated-desc+milestone%3A0.64.0) - [x] GitHub checks for branch are passing - [x] Automated Kubernetes deployment successful - [x] Tag release - [x] Upload release artifacts - [x] Publish release ## Integration - [ ] Deploy to VM ## Performance - [ ] Deploy to Kubernetes - [ ] Deploy to VM - [ ] gRPC API performance tests - [ ] Importer performance tests - [ ] REST API performance tests - [ ] Migrations tested against mainnet clone ## Previewnet - [ ] Deploy to VM ## Staging - [ ] Deploy to Kubernetes EU - [ ] Deploy to Kubernetes NA ## Testnet - [ ] Deploy to VM ## Mainnet - [ ] Deploy to Kubernetes EU - [ ] Deploy to Kubernetes NA - [ ] Deploy to VM - [ ] Deploy to ETL ### Alternatives _No response_
process
release checklist problem we need a checklist to verify the release is rolled out successfully solution preparation milestone field populated on relevant nothing open for github checks for branch are passing automated kubernetes deployment successful tag release upload release artifacts publish release integration deploy to vm performance deploy to kubernetes deploy to vm grpc api performance tests importer performance tests rest api performance tests migrations tested against mainnet clone previewnet deploy to vm staging deploy to kubernetes eu deploy to kubernetes na testnet deploy to vm mainnet deploy to kubernetes eu deploy to kubernetes na deploy to vm deploy to etl alternatives no response
1
725,085
24,950,885,612
IssuesEvent
2022-11-01 07:06:35
AY2223S1-CS2103T-T12-3/tp
https://api.github.com/repos/AY2223S1-CS2103T-T12-3/tp
closed
[PE-D][Tester B] Difference in date behavior in UG and in programme
bug priority.High
In the user guide, it is stated that only one date can be added, while multiple tags can be added: ![image.png](https://raw.githubusercontent.com/farisanadia/ped/main/files/6a513176-bc88-481b-bf73-b151055a4c97.png) However, if multiple dates are added, the hobby will adopt multiple dates: ![image.png](https://raw.githubusercontent.com/farisanadia/ped/main/files/29b24936-7bb3-49ce-a59b-4d54b25f9dc2.png) Select also shows the first date added, as opposed to the last date (in this case feb 2 2022 is shown as opposed to feb 1 2022: ![image.png](https://raw.githubusercontent.com/farisanadia/ped/main/files/0d0dcf06-292d-4034-a5b7-6fc55b7f9f21.png) <!--session: 1666944900526-b54b72e7-f84e-408d-9338-6c15585ac9ca--> <!--Version: Web v3.4.4--> ------------- Labels: `type.FunctionalityBug` `severity.High` original: farisanadia/ped#1
1.0
[PE-D][Tester B] Difference in date behavior in UG and in programme - In the user guide, it is stated that only one date can be added, while multiple tags can be added: ![image.png](https://raw.githubusercontent.com/farisanadia/ped/main/files/6a513176-bc88-481b-bf73-b151055a4c97.png) However, if multiple dates are added, the hobby will adopt multiple dates: ![image.png](https://raw.githubusercontent.com/farisanadia/ped/main/files/29b24936-7bb3-49ce-a59b-4d54b25f9dc2.png) Select also shows the first date added, as opposed to the last date (in this case feb 2 2022 is shown as opposed to feb 1 2022: ![image.png](https://raw.githubusercontent.com/farisanadia/ped/main/files/0d0dcf06-292d-4034-a5b7-6fc55b7f9f21.png) <!--session: 1666944900526-b54b72e7-f84e-408d-9338-6c15585ac9ca--> <!--Version: Web v3.4.4--> ------------- Labels: `type.FunctionalityBug` `severity.High` original: farisanadia/ped#1
non_process
difference in date behavior in ug and in programme in the user guide it is stated that only one date can be added while multiple tags can be added however if multiple dates are added the hobby will adopt multiple dates select also shows the first date added as opposed to the last date in this case feb is shown as opposed to feb labels type functionalitybug severity high original farisanadia ped
0
747,953
26,102,415,301
IssuesEvent
2022-12-27 08:55:46
bounswe/bounswe2022group1
https://api.github.com/repos/bounswe/bounswe2022group1
opened
Checking PR and Merging Frontend Branch to Master
Priority: Critical Type: Task Status: In Progress Frontend
**Issue Description:** We are going to check PR from `frontend` to `master` for latest version of our frontend app, resolve conflicts and merge this branch. Thanks all of my teaammates for their efforts, health to our hands 🤝 **Tasks to Do:** - [ ] add issue labels - [ ] add related links - [ ] check PR - [ ] resolve conflicts - [ ] merge `frontend` to `master` *Task Deadline: 27/12/2022 12:15 pm* *Final Situation:*
1.0
Checking PR and Merging Frontend Branch to Master - **Issue Description:** We are going to check PR from `frontend` to `master` for latest version of our frontend app, resolve conflicts and merge this branch. Thanks all of my teaammates for their efforts, health to our hands 🤝 **Tasks to Do:** - [ ] add issue labels - [ ] add related links - [ ] check PR - [ ] resolve conflicts - [ ] merge `frontend` to `master` *Task Deadline: 27/12/2022 12:15 pm* *Final Situation:*
non_process
checking pr and merging frontend branch to master issue description we are going to check pr from frontend to master for latest version of our frontend app resolve conflicts and merge this branch thanks all of my teaammates for their efforts health to our hands 🤝 tasks to do add issue labels add related links check pr resolve conflicts merge frontend to master task deadline pm final situation
0
511,368
14,859,018,510
IssuesEvent
2021-01-18 17:43:33
Sphereserver/Source-X
https://api.github.com/repos/Sphereserver/Source-X
closed
Server Loop if wrong .set <SKILL NUMBER> Value
Priority: Medium Status-Bug: Confirmed Status: Done - waiting feedback
If for mistake try to use command .set <SKILLID> whit invalid skill ID, server enter in loop error. **How reproduce:** - .set 100 - Target self **Console error:** ``` 17:27:1:'a' commands 'set 100'=1 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CNetworkManager::tick | +0 17:27:DEBUG:(03172) T_Main | 1 | CNetworkThread::tick | +9 17:27:DEBUG:(03172) T_Main | 2 | CNetworkInput::processInput | +0 17:27:DEBUG:(03172) T_Main | 3 | CNetworkInput::processData | +1487 17:27:DEBUG:(03172) T_Main | 4 | CNetworkInput::processData | +0 17:27:DEBUG:(03172) T_Main | 5 | CNetworkInput::processGameClientData | +1 17:27:DEBUG:(03172) T_Main | 6 | PacketTarget::onReceive | +1 17:27:DEBUG:(03172) T_Main | 7 | CClient::Event_Target | +1 17:27:DEBUG:(03172) T_Main | 8 | CClient::OnTarg_Obj_Set | +1 17:27:DEBUG:(03172) T_Main | 9 | CChar::r_Verb | +7 17:27:DEBUG:(03172) T_Main | 10 | CObjBase::r_Verb | +3 17:27:DEBUG:(03172) T_Main | 11 | CScriptObj::r_Verb | +3 17:27:DEBUG:(03172) T_Main | 12 | CChar::r_LoadVal | +3 17:27:DEBUG:(03172) T_Main | 13 | CChar::Skill_SetBase | +5 17:27:DEBUG:(03172) T_Main | 14 | CClient::addSkillWindow | +0 17:27:DEBUG:(03172) T_Main | 15 | PacketSkills::PacketSkills | +1 <-- last function call (stack unwinding began here) 17:27:CRITICAL:"Access Violation" (0x1b2074), in ExcType=CSError catched in CChar::LoadVal() #3 "Keyword" 17:27:DEBUG:command '100' args '' 17:27:'a' commands uid=0307 (B) to '100'=0 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +9 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +1 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +2 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +8 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +1 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +12 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:CRITICAL:"Access Violation" (0x1ae76b), in ExcType=CSError catched in CChar::OnTickPeriodic() #1 "update stats" 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::CheckLocation | +10 17:27:DEBUG:(03172) T_Main | 3 | CItemMultiCustom::GetDesignArea | +3268 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +15 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +20 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +3 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +16 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:CRITICAL:"Access Violation" (0x1ae76b), in ExcType=CSError catched in CChar::OnTickPeriodic() #1 "update stats" 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +16 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +2 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) ``` **Extra info:** - Tested on build 3381 - Tested whit clear script pack
1.0
Server Loop if wrong .set <SKILL NUMBER> Value - If for mistake try to use command .set <SKILLID> whit invalid skill ID, server enter in loop error. **How reproduce:** - .set 100 - Target self **Console error:** ``` 17:27:1:'a' commands 'set 100'=1 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CNetworkManager::tick | +0 17:27:DEBUG:(03172) T_Main | 1 | CNetworkThread::tick | +9 17:27:DEBUG:(03172) T_Main | 2 | CNetworkInput::processInput | +0 17:27:DEBUG:(03172) T_Main | 3 | CNetworkInput::processData | +1487 17:27:DEBUG:(03172) T_Main | 4 | CNetworkInput::processData | +0 17:27:DEBUG:(03172) T_Main | 5 | CNetworkInput::processGameClientData | +1 17:27:DEBUG:(03172) T_Main | 6 | PacketTarget::onReceive | +1 17:27:DEBUG:(03172) T_Main | 7 | CClient::Event_Target | +1 17:27:DEBUG:(03172) T_Main | 8 | CClient::OnTarg_Obj_Set | +1 17:27:DEBUG:(03172) T_Main | 9 | CChar::r_Verb | +7 17:27:DEBUG:(03172) T_Main | 10 | CObjBase::r_Verb | +3 17:27:DEBUG:(03172) T_Main | 11 | CScriptObj::r_Verb | +3 17:27:DEBUG:(03172) T_Main | 12 | CChar::r_LoadVal | +3 17:27:DEBUG:(03172) T_Main | 13 | CChar::Skill_SetBase | +5 17:27:DEBUG:(03172) T_Main | 14 | CClient::addSkillWindow | +0 17:27:DEBUG:(03172) T_Main | 15 | PacketSkills::PacketSkills | +1 <-- last function call (stack unwinding began here) 17:27:CRITICAL:"Access Violation" (0x1b2074), in ExcType=CSError catched in CChar::LoadVal() #3 "Keyword" 17:27:DEBUG:command '100' args '' 17:27:'a' commands uid=0307 (B) to '100'=0 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +9 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +1 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +2 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +8 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +1 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +12 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:CRITICAL:"Access Violation" (0x1ae76b), in ExcType=CSError catched in CChar::OnTickPeriodic() #1 "update stats" 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::CheckLocation | +10 17:27:DEBUG:(03172) T_Main | 3 | CItemMultiCustom::GetDesignArea | +3268 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +15 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +20 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +3 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +16 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:CRITICAL:"Access Violation" (0x1ae76b), in ExcType=CSError catched in CChar::OnTickPeriodic() #1 "update stats" 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +16 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here) 17:27:DEBUG:Printing STACK TRACE for debugging purposes. 17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start 17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0 17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +2 17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13 17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0 17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0 17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here) ``` **Extra info:** - Tested on build 3381 - Tested whit clear script pack
non_process
server loop if wrong set value if for mistake try to use command set whit invalid skill id server enter in loop error how reproduce set target self console error a commands set debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cnetworkmanager tick debug t main cnetworkthread tick debug t main cnetworkinput processinput debug t main cnetworkinput processdata debug t main cnetworkinput processdata debug t main cnetworkinput processgameclientdata debug t main packettarget onreceive debug t main cclient event target debug t main cclient ontarg obj set debug t main cchar r verb debug t main cobjbase r verb debug t main cscriptobj r verb debug t main cchar r loadval debug t main cchar skill setbase debug t main cclient addskillwindow debug t main packetskills packetskills last function call stack unwinding began here critical access violation in exctype cserror catched in cchar loadval keyword debug command args a commands uid b to debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here critical access violation in exctype cserror catched in cchar ontickperiodic update stats debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar checklocation debug t main citemmulticustom getdesignarea last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here critical access violation in exctype cserror catched in cchar ontickperiodic update stats debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here extra info tested on build tested whit clear script pack
0
112,328
14,238,928,303
IssuesEvent
2020-11-18 19:21:50
Subscribie/subscribie
https://api.github.com/repos/Subscribie/subscribie
opened
Stripe payment confirmation for one off item
needs-design needs-marketing needs-qa needs-user-story
Just tested a one off product, the confirmation page is blank and the auto confirmation email is the same as it would be for an ongoing subscription product.
1.0
Stripe payment confirmation for one off item - Just tested a one off product, the confirmation page is blank and the auto confirmation email is the same as it would be for an ongoing subscription product.
non_process
stripe payment confirmation for one off item just tested a one off product the confirmation page is blank and the auto confirmation email is the same as it would be for an ongoing subscription product
0
5,501
8,366,664,110
IssuesEvent
2018-10-04 09:46:50
linnovate/root
https://api.github.com/repos/linnovate/root
closed
in every entity, after selecting one of the entites ( a task, project, discussion, etc...) using multiple choice mode and trying to add a new tag, every time you hover over the update button with your mouse it adds the same tag over and over again
Process bug bug
in every entity, after selecting one of the entites ( a task, project, discussion, etc...) using multiple choice mode and trying to add a new tag, every time you hover over the update button with your mouse it adds the same tag over and over again
1.0
in every entity, after selecting one of the entites ( a task, project, discussion, etc...) using multiple choice mode and trying to add a new tag, every time you hover over the update button with your mouse it adds the same tag over and over again - in every entity, after selecting one of the entites ( a task, project, discussion, etc...) using multiple choice mode and trying to add a new tag, every time you hover over the update button with your mouse it adds the same tag over and over again
process
in every entity after selecting one of the entites a task project discussion etc using multiple choice mode and trying to add a new tag every time you hover over the update button with your mouse it adds the same tag over and over again in every entity after selecting one of the entites a task project discussion etc using multiple choice mode and trying to add a new tag every time you hover over the update button with your mouse it adds the same tag over and over again
1
17,814
23,741,281,459
IssuesEvent
2022-08-31 12:39:45
km4ack/patmenu2
https://api.github.com/repos/km4ack/patmenu2
closed
VARA modem isn't closed
bug in process
The VARA modem doesn't close when using the "Stop Modem" button in Pat
1.0
VARA modem isn't closed - The VARA modem doesn't close when using the "Stop Modem" button in Pat
process
vara modem isn t closed the vara modem doesn t close when using the stop modem button in pat
1
22,232
30,782,372,881
IssuesEvent
2023-07-31 10:55:19
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
opened
Kalman filter
enhancement preprocessor
Here's a nice paper on how a Kalman filter is used as a post-processing (in their case) tool, in our case would be a nice preprocessor. I vaguely remember someone asking about a Kalman filter but I can't find an issue about it, so am plopping it here. I've not done any preprocs dev in a while, if peeps think it's useful I might do it meself, am getting a bit rusty these days, too much admin and maintenance stuff :grin: https://gmd.copernicus.org/articles/16/4233/2023/
1.0
Kalman filter - Here's a nice paper on how a Kalman filter is used as a post-processing (in their case) tool, in our case would be a nice preprocessor. I vaguely remember someone asking about a Kalman filter but I can't find an issue about it, so am plopping it here. I've not done any preprocs dev in a while, if peeps think it's useful I might do it meself, am getting a bit rusty these days, too much admin and maintenance stuff :grin: https://gmd.copernicus.org/articles/16/4233/2023/
process
kalman filter here s a nice paper on how a kalman filter is used as a post processing in their case tool in our case would be a nice preprocessor i vaguely remember someone asking about a kalman filter but i can t find an issue about it so am plopping it here i ve not done any preprocs dev in a while if peeps think it s useful i might do it meself am getting a bit rusty these days too much admin and maintenance stuff grin
1
694
3,184,492,529
IssuesEvent
2015-09-27 12:31:18
nodejs/node
https://api.github.com/repos/nodejs/node
opened
process.versions.icu
feature request intl process
This would be useful to test for the presence of ICU and to implement version-specific behaviour. I've been trying to add [`U_ICU_VERSION`](https://ssl.icu-project.org/apiref/icu4c/uvernum_8h.html), but haven't gotten it working yet, If someone wants to take a shot. Related: #3007
1.0
process.versions.icu - This would be useful to test for the presence of ICU and to implement version-specific behaviour. I've been trying to add [`U_ICU_VERSION`](https://ssl.icu-project.org/apiref/icu4c/uvernum_8h.html), but haven't gotten it working yet, If someone wants to take a shot. Related: #3007
process
process versions icu this would be useful to test for the presence of icu and to implement version specific behaviour i ve been trying to add but haven t gotten it working yet if someone wants to take a shot related
1
676,916
23,142,559,762
IssuesEvent
2022-07-28 20:03:42
Cheos137/ArmorpointsPlusplus
https://api.github.com/repos/Cheos137/ArmorpointsPlusplus
closed
[Bug]: max health below 20 still draws heart borders
bug minecraft: all supported future update priority: medium
### Is there an existing issue for this? - [X] I have searched the existing issues ### Description When having a maximum health below 20, all heart borders are drawn. However, vanilla doesn't draw unnecessary heart borders - and as this mod tries to follow vanilla behavior as closely as possible, this should be corrected. Though, a config option should be added to re-enable the current behavior. ### Version Other (this list is not up to date - again...) ### Version 3.0.0 for all supported minecraft versions ### To Reproduce reduce max-health to a value below 20 (19 is not really useful for testing, too) ### Expexted Behavior Unnecessary heart borders are not drawn. ### Additional Context _No response_ ### Screenshots _No response_ ### Relevant Log / Crash-Report Output _No response_
1.0
[Bug]: max health below 20 still draws heart borders - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Description When having a maximum health below 20, all heart borders are drawn. However, vanilla doesn't draw unnecessary heart borders - and as this mod tries to follow vanilla behavior as closely as possible, this should be corrected. Though, a config option should be added to re-enable the current behavior. ### Version Other (this list is not up to date - again...) ### Version 3.0.0 for all supported minecraft versions ### To Reproduce reduce max-health to a value below 20 (19 is not really useful for testing, too) ### Expexted Behavior Unnecessary heart borders are not drawn. ### Additional Context _No response_ ### Screenshots _No response_ ### Relevant Log / Crash-Report Output _No response_
non_process
max health below still draws heart borders is there an existing issue for this i have searched the existing issues description when having a maximum health below all heart borders are drawn however vanilla doesn t draw unnecessary heart borders and as this mod tries to follow vanilla behavior as closely as possible this should be corrected though a config option should be added to re enable the current behavior version other this list is not up to date again version for all supported minecraft versions to reproduce reduce max health to a value below is not really useful for testing too expexted behavior unnecessary heart borders are not drawn additional context no response screenshots no response relevant log crash report output no response
0
196,097
14,799,050,401
IssuesEvent
2021-01-13 01:19:05
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Make sure Scrollable.ensureVisible works with children before the center
P4 a: tests f: scrolling found in release: 1.22 framework has reproducible steps
See "ListView ensureVisible negative child" in `ensure_visible_test.dart`.
1.0
Make sure Scrollable.ensureVisible works with children before the center - See "ListView ensureVisible negative child" in `ensure_visible_test.dart`.
non_process
make sure scrollable ensurevisible works with children before the center see listview ensurevisible negative child in ensure visible test dart
0
16,348
21,006,561,479
IssuesEvent
2022-03-29 23:26:46
grpc/grpc-ios
https://api.github.com/repos/grpc/grpc-ios
opened
gRPC C++ build instructions for iOS
area/c-core kind/process
Add goto doc for how to build gRPC native c-core for iOS.
1.0
gRPC C++ build instructions for iOS - Add goto doc for how to build gRPC native c-core for iOS.
process
grpc c build instructions for ios add goto doc for how to build grpc native c core for ios
1
11,991
9,546,237,724
IssuesEvent
2019-05-01 19:20:49
trilinos/Trilinos
https://api.github.com/repos/trilinos/Trilinos
closed
KokkosKernels_graph tests timing out in ATDM builds
ATDM Sev: Nonblocker PA: Data Services client: ATDM pkg: KokkosKernels type: bug
CC: @trilinos/kokkoskernels, @kddevin (Trilinos Data Services Product Lead), @bartlettroscoe, @fryeguy52 <Checklist> <???: Add label "ATDM"> <???: Add label "bug"?> <???: Add label for affected packages (e.g. "MueLu", "Tpetra", "Kokkos", etc.)> <???: Add milestone "Initial cleanup of new ATDM builds of Trilinos" or "Keep promoted ATDM builds of Trilinos clean"> <???: Once GitHub Issue is created, add entries for tests to TrilinosATDMStatus/*.csv files> <???: Add label "PA: ???Project Area???" (e.g. "PA: Linear Solvers", "PA: Data Services")> ## Next Action Status <status-and-or-first-action> ## Description As shown in [this query](https://testing.sandia.gov/cdash/queryTests.php?project=Trilinos&filtercombine=and&filtercombine=&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercount=6&showfilters=1&filtercombine=and&field1=buildname&compare1=65&value1=Trilinos-atdm-&field2=testname&compare2=65&value2=KokkosKernels_graph_&field3=testname&compare3=66&value3=_MPI_1&field4=status&compare4=62&value4=Passed&field5=buildstarttime&compare5=83&value5=2019-04-07&field6=buildstarttime&compare6=84&value6=2019-04-10) the tests: * KokkosKernels_graph_openmp_MPI_1 * KokkosKernels_graph_serial_MPI_1 started timing out on 2019-04-07 in the builds: * Trilinos-atdm-hansen-shiller-gnu-debug-serial * Trilinos-atdm-waterman-cuda-9.2-debug * Trilinos-atdm-white-ride-cuda-9.2-gnu-7.2.0-debug * Trilinos-atdm-white-ride-gnu-7.2.0-openmp-debug ## Current Status on CDash Failures for the current testing day can be found [here](https://testing.sandia.gov/cdash/queryTests.php?project=Trilinos&filtercombine=and&filtercombine=&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercount=6&showfilters=1&filtercombine=and&field1=buildname&compare1=65&value1=Trilinos-atdm-&field2=testname&compare2=65&value2=KokkosKernels_graph_&field3=testname&compare3=66&value3=_MPI_1&field4=status&compare4=62&value4=Passed&field5=buildstarttime&compare5=83&value5=yesterday&field6=buildstarttime&compare6=84&value6=today) ## Steps to Reproduce One should be able to reproduce this failure on ride or white as described in: * https://github.com/trilinos/Trilinos/blob/develop/cmake/std/atdm/README.md More specifically, the commands given for ride or white are provided at: * https://github.com/trilinos/Trilinos/blob/develop/cmake/std/atdm/README.md#ridewhite The exact commands to reproduce this issue should be: ``` $ cd <some_build_dir>/ $ source $TRILINOS_DIR/cmake/std/atdm/load-env.sh Trilinos-atdm-white-ride-gnu-7.2.0-openmp-debug $ cmake \ -GNinja \ -DTrilinos_CONFIGURE_OPTIONS_FILE:STRING=cmake/std/atdm/ATDMDevEnv.cmake \ -DTrilinos_ENABLE_TESTS=ON -DTrilinos_ENABLE_KokkosKernels=ON \ $TRILINOS_DIR $ make NP=16 $ bsub -x -Is -q rhel7F -n 16 ctest -j16 ```
1.0
KokkosKernels_graph tests timing out in ATDM builds - CC: @trilinos/kokkoskernels, @kddevin (Trilinos Data Services Product Lead), @bartlettroscoe, @fryeguy52 <Checklist> <???: Add label "ATDM"> <???: Add label "bug"?> <???: Add label for affected packages (e.g. "MueLu", "Tpetra", "Kokkos", etc.)> <???: Add milestone "Initial cleanup of new ATDM builds of Trilinos" or "Keep promoted ATDM builds of Trilinos clean"> <???: Once GitHub Issue is created, add entries for tests to TrilinosATDMStatus/*.csv files> <???: Add label "PA: ???Project Area???" (e.g. "PA: Linear Solvers", "PA: Data Services")> ## Next Action Status <status-and-or-first-action> ## Description As shown in [this query](https://testing.sandia.gov/cdash/queryTests.php?project=Trilinos&filtercombine=and&filtercombine=&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercount=6&showfilters=1&filtercombine=and&field1=buildname&compare1=65&value1=Trilinos-atdm-&field2=testname&compare2=65&value2=KokkosKernels_graph_&field3=testname&compare3=66&value3=_MPI_1&field4=status&compare4=62&value4=Passed&field5=buildstarttime&compare5=83&value5=2019-04-07&field6=buildstarttime&compare6=84&value6=2019-04-10) the tests: * KokkosKernels_graph_openmp_MPI_1 * KokkosKernels_graph_serial_MPI_1 started timing out on 2019-04-07 in the builds: * Trilinos-atdm-hansen-shiller-gnu-debug-serial * Trilinos-atdm-waterman-cuda-9.2-debug * Trilinos-atdm-white-ride-cuda-9.2-gnu-7.2.0-debug * Trilinos-atdm-white-ride-gnu-7.2.0-openmp-debug ## Current Status on CDash Failures for the current testing day can be found [here](https://testing.sandia.gov/cdash/queryTests.php?project=Trilinos&filtercombine=and&filtercombine=&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercount=6&showfilters=1&filtercombine=and&field1=buildname&compare1=65&value1=Trilinos-atdm-&field2=testname&compare2=65&value2=KokkosKernels_graph_&field3=testname&compare3=66&value3=_MPI_1&field4=status&compare4=62&value4=Passed&field5=buildstarttime&compare5=83&value5=yesterday&field6=buildstarttime&compare6=84&value6=today) ## Steps to Reproduce One should be able to reproduce this failure on ride or white as described in: * https://github.com/trilinos/Trilinos/blob/develop/cmake/std/atdm/README.md More specifically, the commands given for ride or white are provided at: * https://github.com/trilinos/Trilinos/blob/develop/cmake/std/atdm/README.md#ridewhite The exact commands to reproduce this issue should be: ``` $ cd <some_build_dir>/ $ source $TRILINOS_DIR/cmake/std/atdm/load-env.sh Trilinos-atdm-white-ride-gnu-7.2.0-openmp-debug $ cmake \ -GNinja \ -DTrilinos_CONFIGURE_OPTIONS_FILE:STRING=cmake/std/atdm/ATDMDevEnv.cmake \ -DTrilinos_ENABLE_TESTS=ON -DTrilinos_ENABLE_KokkosKernels=ON \ $TRILINOS_DIR $ make NP=16 $ bsub -x -Is -q rhel7F -n 16 ctest -j16 ```
non_process
kokkoskernels graph tests timing out in atdm builds cc trilinos kokkoskernels kddevin trilinos data services product lead bartlettroscoe next action status description as shown in the tests kokkoskernels graph openmp mpi kokkoskernels graph serial mpi started timing out on in the builds trilinos atdm hansen shiller gnu debug serial trilinos atdm waterman cuda debug trilinos atdm white ride cuda gnu debug trilinos atdm white ride gnu openmp debug current status on cdash failures for the current testing day can be found steps to reproduce one should be able to reproduce this failure on ride or white as described in more specifically the commands given for ride or white are provided at the exact commands to reproduce this issue should be cd source trilinos dir cmake std atdm load env sh trilinos atdm white ride gnu openmp debug cmake gninja dtrilinos configure options file string cmake std atdm atdmdevenv cmake dtrilinos enable tests on dtrilinos enable kokkoskernels on trilinos dir make np bsub x is q n ctest
0
883
3,348,219,913
IssuesEvent
2015-11-17 00:25:16
beesmart-it/trend-hrm
https://api.github.com/repos/beesmart-it/trend-hrm
opened
Can't delete team member if owning processes
company requirement selection process
Can't delete team member if owning processes. One option is to ask user to re-assign all processes to other user.
1.0
Can't delete team member if owning processes - Can't delete team member if owning processes. One option is to ask user to re-assign all processes to other user.
process
can t delete team member if owning processes can t delete team member if owning processes one option is to ask user to re assign all processes to other user
1
422,290
12,269,443,616
IssuesEvent
2020-05-07 14:05:08
googleapis/java-iam
https://api.github.com/repos/googleapis/java-iam
opened
Synthesis failed for java-iam
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate java-iam. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to a new branch 'autosynth' Cloning into '/tmpfs/tmp/tmpd1dezaou/synthtool'... Switched to branch 'autosynth-self' Note: checking out '81a7470ee9633ff052ccb679cbe37b0079ee51ef'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2) Note: checking out '04cb397eb7590ea1e6c10a39fca3d8fe0fb3d256'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 04cb397 chore: add labels with context info (#518) Switched to a new branch 'autosynth-self-1' 2020-05-07 07:00:56 [INFO] Running synthtool 2020-05-07 07:00:56 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-05-07 07:00:56,457 synthtool > Executing /tmpfs/src/github/synthtool/working_repo/synth.py. On branch autosynth-self-1 nothing to commit, working tree clean 2020-05-07 07:00:56,589 synthtool > Ensuring dependencies. 2020-05-07 07:00:56,601 synthtool > Cloning googleapis. 2020-05-07 07:00:58,052 synthtool > Generating code for: //google/iam/v1:google-iam-v1-java. 2020-05-07 07:00:58,168 synthtool > Failed executing bazel build //google/iam/v1:google-iam-v1-java: Starting local Bazel server and connecting to it... Server crashed during startup. Now printing /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/server/jvm.out OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x0000000715200000, 247463936, 0) failed; error='Not enough space' (errno=12) # # There is insufficient memory for the Java Runtime Environment to continue. # Native memory allocation (mmap) failed to map 247463936 bytes for committing reserved memory. # An error report file with more information is saved as: # /home/kbuilder/.cache/synthtool/googleapis/hs_err_pid15482.log 2020-05-07 07:00:58,168 synthtool > Wrote metadata to synth.metadata. Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/tmpfs/src/github/synthtool/working_repo/synth.py", line 31, in <module> bazel_target=f'//google/iam/{version}:google-iam-{version}-java', File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 63, in java_library return self._generate_code(service, version, "java", **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 177, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', 'build', '//google/iam/v1:google-iam-v1-java']' returned non-zero exit status 37. 2020-05-07 07:00:58 [ERROR] Synthesis failed HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2) Switched to branch 'autosynth-self' Note: checking out '81a7470ee9633ff052ccb679cbe37b0079ee51ef'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2) Previous HEAD position was 04cb397 chore: add labels with context info (#518) HEAD is now at be74d3e build: do not fail builds on codecov errors (#528) Switched to a new branch 'autosynth-8' 2020-05-07 07:00:58 [INFO] Running synthtool 2020-05-07 07:00:58 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-05-07 07:00:58,447 synthtool > Executing /tmpfs/src/github/synthtool/working_repo/synth.py. On branch autosynth-8 nothing to commit, working tree clean 2020-05-07 07:00:58,581 synthtool > Ensuring dependencies. 2020-05-07 07:00:58,593 synthtool > Cloning googleapis. 2020-05-07 07:00:58,986 synthtool > Generating code for: //google/iam/v1:google-iam-v1-java. 2020-05-07 07:00:59,149 synthtool > Failed executing bazel build //google/iam/v1:google-iam-v1-java: Starting local Bazel server and connecting to it... Server crashed during startup. Now printing /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/server/jvm.out OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x0000000715200000, 247463936, 0) failed; error='Not enough space' (errno=12) # # There is insufficient memory for the Java Runtime Environment to continue. # Native memory allocation (mmap) failed to map 247463936 bytes for committing reserved memory. # An error report file with more information is saved as: # /home/kbuilder/.cache/synthtool/googleapis/hs_err_pid15513.log 2020-05-07 07:00:59,150 synthtool > Wrote metadata to synth.metadata. Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/tmpfs/src/github/synthtool/working_repo/synth.py", line 31, in <module> bazel_target=f'//google/iam/{version}:google-iam-{version}-java', File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 63, in java_library return self._generate_code(service, version, "java", **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 177, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', 'build', '//google/iam/v1:google-iam-v1-java']' returned non-zero exit status 37. 2020-05-07 07:00:59 [ERROR] Synthesis failed HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2) Switched to branch 'autosynth' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 584, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 465, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 574, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 360, in synthesize_loop synthesize_inner_loop(toolbox, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 370, in synthesize_inner_loop synthesizer, len(toolbox.versions) - 1 File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 259, in synthesize_version_in_new_branch synthesizer.synthesize(self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 115, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/f9c87743-f01c-4b38-93fb-8cba46610304).
1.0
Synthesis failed for java-iam - Hello! Autosynth couldn't regenerate java-iam. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to a new branch 'autosynth' Cloning into '/tmpfs/tmp/tmpd1dezaou/synthtool'... Switched to branch 'autosynth-self' Note: checking out '81a7470ee9633ff052ccb679cbe37b0079ee51ef'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2) Note: checking out '04cb397eb7590ea1e6c10a39fca3d8fe0fb3d256'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 04cb397 chore: add labels with context info (#518) Switched to a new branch 'autosynth-self-1' 2020-05-07 07:00:56 [INFO] Running synthtool 2020-05-07 07:00:56 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-05-07 07:00:56,457 synthtool > Executing /tmpfs/src/github/synthtool/working_repo/synth.py. On branch autosynth-self-1 nothing to commit, working tree clean 2020-05-07 07:00:56,589 synthtool > Ensuring dependencies. 2020-05-07 07:00:56,601 synthtool > Cloning googleapis. 2020-05-07 07:00:58,052 synthtool > Generating code for: //google/iam/v1:google-iam-v1-java. 2020-05-07 07:00:58,168 synthtool > Failed executing bazel build //google/iam/v1:google-iam-v1-java: Starting local Bazel server and connecting to it... Server crashed during startup. Now printing /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/server/jvm.out OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x0000000715200000, 247463936, 0) failed; error='Not enough space' (errno=12) # # There is insufficient memory for the Java Runtime Environment to continue. # Native memory allocation (mmap) failed to map 247463936 bytes for committing reserved memory. # An error report file with more information is saved as: # /home/kbuilder/.cache/synthtool/googleapis/hs_err_pid15482.log 2020-05-07 07:00:58,168 synthtool > Wrote metadata to synth.metadata. Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/tmpfs/src/github/synthtool/working_repo/synth.py", line 31, in <module> bazel_target=f'//google/iam/{version}:google-iam-{version}-java', File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 63, in java_library return self._generate_code(service, version, "java", **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 177, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', 'build', '//google/iam/v1:google-iam-v1-java']' returned non-zero exit status 37. 2020-05-07 07:00:58 [ERROR] Synthesis failed HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2) Switched to branch 'autosynth-self' Note: checking out '81a7470ee9633ff052ccb679cbe37b0079ee51ef'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2) Previous HEAD position was 04cb397 chore: add labels with context info (#518) HEAD is now at be74d3e build: do not fail builds on codecov errors (#528) Switched to a new branch 'autosynth-8' 2020-05-07 07:00:58 [INFO] Running synthtool 2020-05-07 07:00:58 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--'] 2020-05-07 07:00:58,447 synthtool > Executing /tmpfs/src/github/synthtool/working_repo/synth.py. On branch autosynth-8 nothing to commit, working tree clean 2020-05-07 07:00:58,581 synthtool > Ensuring dependencies. 2020-05-07 07:00:58,593 synthtool > Cloning googleapis. 2020-05-07 07:00:58,986 synthtool > Generating code for: //google/iam/v1:google-iam-v1-java. 2020-05-07 07:00:59,149 synthtool > Failed executing bazel build //google/iam/v1:google-iam-v1-java: Starting local Bazel server and connecting to it... Server crashed during startup. Now printing /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/server/jvm.out OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x0000000715200000, 247463936, 0) failed; error='Not enough space' (errno=12) # # There is insufficient memory for the Java Runtime Environment to continue. # Native memory allocation (mmap) failed to map 247463936 bytes for committing reserved memory. # An error report file with more information is saved as: # /home/kbuilder/.cache/synthtool/googleapis/hs_err_pid15513.log 2020-05-07 07:00:59,150 synthtool > Wrote metadata to synth.metadata. Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/tmpfs/src/github/synthtool/working_repo/synth.py", line 31, in <module> bazel_target=f'//google/iam/{version}:google-iam-{version}-java', File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 63, in java_library return self._generate_code(service, version, "java", **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 177, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', 'build', '//google/iam/v1:google-iam-v1-java']' returned non-zero exit status 37. 2020-05-07 07:00:59 [ERROR] Synthesis failed HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2) Switched to branch 'autosynth' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 584, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 465, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 574, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 360, in synthesize_loop synthesize_inner_loop(toolbox, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 370, in synthesize_inner_loop synthesizer, len(toolbox.versions) - 1 File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 259, in synthesize_version_in_new_branch synthesizer.synthesize(self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 115, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/f9c87743-f01c-4b38-93fb-8cba46610304).
non_process
synthesis failed for java iam hello autosynth couldn t regenerate java iam broken heart here s the output from running synth py cloning into working repo switched to a new branch autosynth cloning into tmpfs tmp synthtool switched to branch autosynth self note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at deps update dependency com google api grpc proto google common protos to note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at chore add labels with context info switched to a new branch autosynth self running synthtool synthtool executing tmpfs src github synthtool working repo synth py on branch autosynth self nothing to commit working tree clean synthtool ensuring dependencies synthtool cloning googleapis synthtool generating code for google iam google iam java synthtool failed executing bazel build google iam google iam java starting local bazel server and connecting to it server crashed during startup now printing home kbuilder cache bazel bazel kbuilder server jvm out openjdk bit server vm warning info os commit memory failed error not enough space errno there is insufficient memory for the java runtime environment to continue native memory allocation mmap failed to map bytes for committing reserved memory an error report file with more information is saved as home kbuilder cache synthtool googleapis hs err log synthtool wrote metadata to synth metadata traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file tmpfs src github synthtool working repo synth py line in bazel target f google iam version google iam version java file tmpfs src github synthtool synthtool gcp gapic bazel py line in java library return self generate code service version java kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in generate code shell run bazel run args file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status synthesis failed head is now at deps update dependency com google api grpc proto google common protos to switched to branch autosynth self note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at deps update dependency com google api grpc proto google common protos to previous head position was chore add labels with context info head is now at build do not fail builds on codecov errors switched to a new branch autosynth running synthtool synthtool executing tmpfs src github synthtool working repo synth py on branch autosynth nothing to commit working tree clean synthtool ensuring dependencies synthtool cloning googleapis synthtool generating code for google iam google iam java synthtool failed executing bazel build google iam google iam java starting local bazel server and connecting to it server crashed during startup now printing home kbuilder cache bazel bazel kbuilder server jvm out openjdk bit server vm warning info os commit memory failed error not enough space errno there is insufficient memory for the java runtime environment to continue native memory allocation mmap failed to map bytes for committing reserved memory an error report file with more information is saved as home kbuilder cache synthtool googleapis hs err log synthtool wrote metadata to synth metadata traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file tmpfs src github synthtool working repo synth py line in bazel target f google iam version google iam version java file tmpfs src github synthtool synthtool gcp gapic bazel py line in java library return self generate code service version java kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in generate code shell run bazel run args file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status synthesis failed head is now at deps update dependency com google api grpc proto google common protos to switched to branch autosynth traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop synthesize inner loop toolbox synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize inner loop synthesizer len toolbox versions file tmpfs src github synthtool autosynth synth py line in synthesize version in new branch synthesizer synthesize self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
1,798
4,539,979,563
IssuesEvent
2016-09-09 13:14:01
openvstorage/alba
https://api.github.com/repos/openvstorage/alba
closed
ASDs can refuse to start up after power loss
process_wontfix type_bug
``` 2016/05/12 14:39:55 259633 +0200 - ovs03 - 8878/0 - alba/asd - 0 - info - ETCD: etcdctl --peers=127.0.0.1:2379 get ovs/alba/asds/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX/config 2016/05/12 14:39:55 265048 +0200 - ovs03 - 8878/0 - alba/asd - 1 - info - Found the following config: {"asd_id": "HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX", "log_level": "info", "ips": ["172.19.197.3"], "node_id": "LxQuz79r7xe7k4gFhPUqJMGs9Cs7PxF6", "home": "/mnt/alba-asd/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX", "port": 8602} 2016/05/12 14:39:55 265150 +0200 - ovs03 - 8878/0 - alba/asd - 2 - info - Interpreted the config as: { Asd_config.Config.ips = ["172.19.197.3"]; port = (Some 8602);; node_id = "LxQuz79r7xe7k4gFhPUqJMGs9Cs7PxF6";; home = "/mnt/alba-asd/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX";; log_level = "info"; asd_id = (Some "HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX");; __sync_dont_use = true; limit = 99L; buffer_size = 786432;; multicast = (Some 10.); tls = None;; tcp_keepalive = { Tcp_keepalive.enable_tcp_keepalive = true;; tcp_keepalive_time = 20; tcp_keepalive_intvl = 20;; tcp_keepalive_probes = 3 };; __warranty_void__write_blobs = true; use_fadvise = true;; use_fallocate = true; rocksdb_block_cache_size = None } 2016/05/12 14:39:55 265172 +0200 - ovs03 - 8878/0 - alba/asd - 3 - info - asd_server version:0.9.3-0-gd6d4a3d alba: internal error, uncaught exception: (Failure "Corruption: checksum mismatch") Raised at file "src/core/lwt.ml", line 789, characters 22-23 Called from file "src/unix/lwt_main.ml", line 34, characters 8-18 ``` The rocksdb log shows this: ``` 2016/05/17-17:51:34.132241 7f2446538940 Recovered from manifest file:/mnt/alba-asd/7aSGYFzBuY7uFODPysPntsDijG7ZxtPm/db/MANIFEST-004971 succeeded,manifest_file_number is 4971, next_file_number is 5136, last_sequence is 47018602, log_number is 0,prev_log_number is 0,max_column_family is 0 2016/05/17-17:51:34.132256 7f2446538940 Column family [default] (ID 0), log number is 5133 2016/05/17-17:51:34.132418 7f2446538940 EVENT_LOG_v1 {"time_micros": 1463487694132404, "job": 1, "event": "recovery_started", "log_files": [5133]} 2016/05/17-17:51:34.132433 7f2446538940 Recovering log #5133 mode 0 skip-recovery 0 2016/05/17-17:51:34.163110 7f2446538940 [WARN] /mnt/alba-asd/7aSGYFzBuY7uFODPysPntsDijG7ZxtPm/db/005133.log: dropping 5792 bytes; Corruption: checksum mismatch ``` We started seeing this after we started using the rocksdb recycle_log_file_name option, so it will probably be solved by https://github.com/facebook/rocksdb/pull/881.
1.0
ASDs can refuse to start up after power loss - ``` 2016/05/12 14:39:55 259633 +0200 - ovs03 - 8878/0 - alba/asd - 0 - info - ETCD: etcdctl --peers=127.0.0.1:2379 get ovs/alba/asds/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX/config 2016/05/12 14:39:55 265048 +0200 - ovs03 - 8878/0 - alba/asd - 1 - info - Found the following config: {"asd_id": "HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX", "log_level": "info", "ips": ["172.19.197.3"], "node_id": "LxQuz79r7xe7k4gFhPUqJMGs9Cs7PxF6", "home": "/mnt/alba-asd/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX", "port": 8602} 2016/05/12 14:39:55 265150 +0200 - ovs03 - 8878/0 - alba/asd - 2 - info - Interpreted the config as: { Asd_config.Config.ips = ["172.19.197.3"]; port = (Some 8602);; node_id = "LxQuz79r7xe7k4gFhPUqJMGs9Cs7PxF6";; home = "/mnt/alba-asd/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX";; log_level = "info"; asd_id = (Some "HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX");; __sync_dont_use = true; limit = 99L; buffer_size = 786432;; multicast = (Some 10.); tls = None;; tcp_keepalive = { Tcp_keepalive.enable_tcp_keepalive = true;; tcp_keepalive_time = 20; tcp_keepalive_intvl = 20;; tcp_keepalive_probes = 3 };; __warranty_void__write_blobs = true; use_fadvise = true;; use_fallocate = true; rocksdb_block_cache_size = None } 2016/05/12 14:39:55 265172 +0200 - ovs03 - 8878/0 - alba/asd - 3 - info - asd_server version:0.9.3-0-gd6d4a3d alba: internal error, uncaught exception: (Failure "Corruption: checksum mismatch") Raised at file "src/core/lwt.ml", line 789, characters 22-23 Called from file "src/unix/lwt_main.ml", line 34, characters 8-18 ``` The rocksdb log shows this: ``` 2016/05/17-17:51:34.132241 7f2446538940 Recovered from manifest file:/mnt/alba-asd/7aSGYFzBuY7uFODPysPntsDijG7ZxtPm/db/MANIFEST-004971 succeeded,manifest_file_number is 4971, next_file_number is 5136, last_sequence is 47018602, log_number is 0,prev_log_number is 0,max_column_family is 0 2016/05/17-17:51:34.132256 7f2446538940 Column family [default] (ID 0), log number is 5133 2016/05/17-17:51:34.132418 7f2446538940 EVENT_LOG_v1 {"time_micros": 1463487694132404, "job": 1, "event": "recovery_started", "log_files": [5133]} 2016/05/17-17:51:34.132433 7f2446538940 Recovering log #5133 mode 0 skip-recovery 0 2016/05/17-17:51:34.163110 7f2446538940 [WARN] /mnt/alba-asd/7aSGYFzBuY7uFODPysPntsDijG7ZxtPm/db/005133.log: dropping 5792 bytes; Corruption: checksum mismatch ``` We started seeing this after we started using the rocksdb recycle_log_file_name option, so it will probably be solved by https://github.com/facebook/rocksdb/pull/881.
process
asds can refuse to start up after power loss alba asd info etcd etcdctl peers get ovs alba asds config alba asd info found the following config asd id log level info ips node id home mnt alba asd port alba asd info interpreted the config as asd config config ips port some node id home mnt alba asd log level info asd id some sync dont use true limit buffer size multicast some tls none tcp keepalive tcp keepalive enable tcp keepalive true tcp keepalive time tcp keepalive intvl tcp keepalive probes warranty void write blobs true use fadvise true use fallocate true rocksdb block cache size none alba asd info asd server version alba internal error uncaught exception failure corruption checksum mismatch raised at file src core lwt ml line characters called from file src unix lwt main ml line characters the rocksdb log shows this recovered from manifest file mnt alba asd db manifest succeeded manifest file number is next file number is last sequence is log number is prev log number is max column family is column family id log number is event log time micros job event recovery started log files recovering log mode skip recovery mnt alba asd db log dropping bytes corruption checksum mismatch we started seeing this after we started using the rocksdb recycle log file name option so it will probably be solved by
1
132,257
28,128,141,000
IssuesEvent
2023-03-31 19:43:26
creativecommons/cc-resource-archive
https://api.github.com/repos/creativecommons/cc-resource-archive
closed
[Feature] Adding Footer
🟩 priority: low ⛔️ status: discarded 🚦 status: awaiting triage ✨ goal: improvement 💻 aspect: code
## Problem There is No footer Section ## Description Adding footer for showing contact information about Creative Commons. ## Implementation ![image](https://user-images.githubusercontent.com/65482186/226956610-c19e440b-63f0-4155-a087-581a6c6c00c7.png) <!-- Replace the [ ] with [x] to check the box. --> - [x] I would be interested in implementing this feature.
1.0
[Feature] Adding Footer - ## Problem There is No footer Section ## Description Adding footer for showing contact information about Creative Commons. ## Implementation ![image](https://user-images.githubusercontent.com/65482186/226956610-c19e440b-63f0-4155-a087-581a6c6c00c7.png) <!-- Replace the [ ] with [x] to check the box. --> - [x] I would be interested in implementing this feature.
non_process
adding footer problem there is no footer section description adding footer for showing contact information about creative commons implementation i would be interested in implementing this feature
0
11,114
13,957,681,476
IssuesEvent
2020-10-24 08:07:26
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
DE: request for a new harvesting
DE - Germany Geoportal Harvesting process
Dear Geoportal Helpdesk, As mentioned in Roberts Mail from 2020/03/02 we would like to initiate a new push of our metadata records to the EU Geoportal. For this reason we kindly ask you to start a new harvesting of our catalogue instance and publish them for us in the Geoportal harvesting &quot;sandbox&quot;, please. Also we kindly ask you, if you could provide us two or three original csw-requests (for an internal validation/review on our side), which you are using to get the metadata records from our catalogue instance. Thanks in advance and best regards, Anja (on behalf of Coordination Office SDI Germany)
1.0
DE: request for a new harvesting - Dear Geoportal Helpdesk, As mentioned in Roberts Mail from 2020/03/02 we would like to initiate a new push of our metadata records to the EU Geoportal. For this reason we kindly ask you to start a new harvesting of our catalogue instance and publish them for us in the Geoportal harvesting &quot;sandbox&quot;, please. Also we kindly ask you, if you could provide us two or three original csw-requests (for an internal validation/review on our side), which you are using to get the metadata records from our catalogue instance. Thanks in advance and best regards, Anja (on behalf of Coordination Office SDI Germany)
process
de request for a new harvesting dear geoportal helpdesk as mentioned in roberts mail from we would like to initiate a new push of our metadata records to the eu geoportal for this reason we kindly ask you to start a new harvesting of our catalogue instance and publish them for us in the geoportal harvesting quot sandbox quot please also we kindly ask you if you could provide us two or three original csw requests for an internal validation review on our side which you are using to get the metadata records from our catalogue instance thanks in advance and best regards anja on behalf of coordination office sdi germany
1
18,344
24,467,515,861
IssuesEvent
2022-10-07 16:19:43
bondaleksey/credit-card-fraud-detection
https://api.github.com/repos/bondaleksey/credit-card-fraud-detection
opened
daily data generation
work plan data preprocessing
- Write a script that generates a new chunk of data and saves it to HDFS. - Automate your regular launch with AirFlow.
1.0
daily data generation - - Write a script that generates a new chunk of data and saves it to HDFS. - Automate your regular launch with AirFlow.
process
daily data generation write a script that generates a new chunk of data and saves it to hdfs automate your regular launch with airflow
1
277,176
24,054,501,787
IssuesEvent
2022-09-16 15:32:31
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
closed
Release 4.3.8 - Release Candidate 1
release test/4.3.8
The following issue will gather all the info regarding testing and fixing in order to validate this release candidate. The definition of done for this one is the validation from the product owner of each QA analysis and the acceptance of the implemented fixes implemented, all the below issues must be closed in order to close this one. ## Release candidate info |Project|Main issue|Version|Release candidate #|Tag|Previous RC issue|Next RC issue| |---|---|---|---|---|---|---| |[v4.3.8](https://github.com/orgs/wazuh/projects/20/views/1)|[#14826](https://github.com/wazuh/wazuh/issues/14826)|4.3.8|RC1|[v4.3.8-rc1](https://github.com/wazuh/wazuh/tree/v4.3.8-rc1)|-|-| ## QA testing issues In order to move to a new release candidate or the GA version, all tests and metrics analyses below must be in Closed status. | Name | Issue | Status |DRI| |-----------------------------|-----------------------------------------------|-------------|---| | C unit | https://github.com/wazuh/wazuh/issues/14879 | 🟣 Completed |@wazuh/core| | Python unit | https://github.com/wazuh/wazuh/issues/14870 | 🟣 Completed |@wazuh/framework| | Footprint metrics | https://github.com/wazuh/wazuh/issues/14859 | 🟣 Completed |@wazuh/cicd| | Workload benchmarks metrics | https://github.com/wazuh/wazuh/issues/14860 | 🟣 Completed |@wazuh/framework| | Integration | https://github.com/wazuh/wazuh/issues/14877 | 🟣 Completed |@wazuh/qa| | API integration | https://github.com/wazuh/wazuh/issues/14871 | 🟣 Completed |@wazuh/framework| | System | https://github.com/wazuh/wazuh/issues/14887 | 🟣 Completed | @wazuh/framework| | External integrations modules | - | ⚪ Skipped | @wazuh/framework| | Demo uses cases | https://github.com/wazuh/wazuh/issues/14921 | 🟣 Completed |@wazuh/cicd| | Packages | https://github.com/wazuh/wazuh/issues/14862 | 🟣 Completed |@wazuh/cicd| | Coverity scan | https://github.com/wazuh/wazuh/issues/14881 | 🟣 Completed |@wazuh/core| | Ruleset | https://github.com/wazuh/wazuh/issues/14872 | 🟣 Completed |@wazuh/threat-intel| | Kibana UI regression | https://github.com/wazuh/wazuh-kibana-app/issues/4535 | 🟣 Completed |@wazuh/frontend| | Splunk UI regression | https://github.com/wazuh/wazuh-splunk/issues/1366 | 🟣 Completed |@wazuh/frontend| | WPK Upgrade |https://github.com/wazuh/wazuh/issues/14861 | 🟣 Completed |@wazuh/core| | E2E UX | https://github.com/wazuh/wazuh/issues/14828 | 🟣 Completed |@wazuh| ⚫ _Not started: The tasks didn't start yet._ 🟡 _In progress: The team is already working on it._ 🟢 _Ready to review: The product owner must audit and validate the results._ ⚪ _Skipped: The task has been deemed not necessary for this RC._ 🟣 _Completed: Task finished. Nothing to do here._ 🔴 _Completed with failures: Some issues raised here._ ## Auditors' validation In order to close and proceed with the release or the next candidate version, the following auditors must give the green light to this RC. - [x] @davidjiglesias
1.0
Release 4.3.8 - Release Candidate 1 - The following issue will gather all the info regarding testing and fixing in order to validate this release candidate. The definition of done for this one is the validation from the product owner of each QA analysis and the acceptance of the implemented fixes implemented, all the below issues must be closed in order to close this one. ## Release candidate info |Project|Main issue|Version|Release candidate #|Tag|Previous RC issue|Next RC issue| |---|---|---|---|---|---|---| |[v4.3.8](https://github.com/orgs/wazuh/projects/20/views/1)|[#14826](https://github.com/wazuh/wazuh/issues/14826)|4.3.8|RC1|[v4.3.8-rc1](https://github.com/wazuh/wazuh/tree/v4.3.8-rc1)|-|-| ## QA testing issues In order to move to a new release candidate or the GA version, all tests and metrics analyses below must be in Closed status. | Name | Issue | Status |DRI| |-----------------------------|-----------------------------------------------|-------------|---| | C unit | https://github.com/wazuh/wazuh/issues/14879 | 🟣 Completed |@wazuh/core| | Python unit | https://github.com/wazuh/wazuh/issues/14870 | 🟣 Completed |@wazuh/framework| | Footprint metrics | https://github.com/wazuh/wazuh/issues/14859 | 🟣 Completed |@wazuh/cicd| | Workload benchmarks metrics | https://github.com/wazuh/wazuh/issues/14860 | 🟣 Completed |@wazuh/framework| | Integration | https://github.com/wazuh/wazuh/issues/14877 | 🟣 Completed |@wazuh/qa| | API integration | https://github.com/wazuh/wazuh/issues/14871 | 🟣 Completed |@wazuh/framework| | System | https://github.com/wazuh/wazuh/issues/14887 | 🟣 Completed | @wazuh/framework| | External integrations modules | - | ⚪ Skipped | @wazuh/framework| | Demo uses cases | https://github.com/wazuh/wazuh/issues/14921 | 🟣 Completed |@wazuh/cicd| | Packages | https://github.com/wazuh/wazuh/issues/14862 | 🟣 Completed |@wazuh/cicd| | Coverity scan | https://github.com/wazuh/wazuh/issues/14881 | 🟣 Completed |@wazuh/core| | Ruleset | https://github.com/wazuh/wazuh/issues/14872 | 🟣 Completed |@wazuh/threat-intel| | Kibana UI regression | https://github.com/wazuh/wazuh-kibana-app/issues/4535 | 🟣 Completed |@wazuh/frontend| | Splunk UI regression | https://github.com/wazuh/wazuh-splunk/issues/1366 | 🟣 Completed |@wazuh/frontend| | WPK Upgrade |https://github.com/wazuh/wazuh/issues/14861 | 🟣 Completed |@wazuh/core| | E2E UX | https://github.com/wazuh/wazuh/issues/14828 | 🟣 Completed |@wazuh| ⚫ _Not started: The tasks didn't start yet._ 🟡 _In progress: The team is already working on it._ 🟢 _Ready to review: The product owner must audit and validate the results._ ⚪ _Skipped: The task has been deemed not necessary for this RC._ 🟣 _Completed: Task finished. Nothing to do here._ 🔴 _Completed with failures: Some issues raised here._ ## Auditors' validation In order to close and proceed with the release or the next candidate version, the following auditors must give the green light to this RC. - [x] @davidjiglesias
non_process
release release candidate the following issue will gather all the info regarding testing and fixing in order to validate this release candidate the definition of done for this one is the validation from the product owner of each qa analysis and the acceptance of the implemented fixes implemented all the below issues must be closed in order to close this one release candidate info project main issue version release candidate tag previous rc issue next rc issue qa testing issues in order to move to a new release candidate or the ga version all tests and metrics analyses below must be in closed status name issue status dri c unit 🟣 completed wazuh core python unit 🟣 completed wazuh framework footprint metrics 🟣 completed wazuh cicd workload benchmarks metrics 🟣 completed wazuh framework integration 🟣 completed wazuh qa api integration 🟣 completed wazuh framework system 🟣 completed wazuh framework external integrations modules ⚪ skipped wazuh framework demo uses cases 🟣 completed wazuh cicd packages 🟣 completed wazuh cicd coverity scan 🟣 completed wazuh core ruleset 🟣 completed wazuh threat intel kibana ui regression 🟣 completed wazuh frontend splunk ui regression 🟣 completed wazuh frontend wpk upgrade 🟣 completed wazuh core ux 🟣 completed wazuh ⚫ not started the tasks didn t start yet 🟡 in progress the team is already working on it 🟢 ready to review the product owner must audit and validate the results ⚪ skipped the task has been deemed not necessary for this rc 🟣 completed task finished nothing to do here 🔴 completed with failures some issues raised here auditors validation in order to close and proceed with the release or the next candidate version the following auditors must give the green light to this rc davidjiglesias
0
5,027
7,849,611,670
IssuesEvent
2018-06-20 04:51:03
kerubistan/kerub
https://api.github.com/repos/kerubistan/kerub
opened
planner: filter out repeating inverses
component:data processing performance
> filter out repeating inverses: when the plan has already steps "power host-1 up" and then later "power host-1 down" then any new "power host-1 up" steps could be filtered out - and let's see how this effects performance
1.0
planner: filter out repeating inverses - > filter out repeating inverses: when the plan has already steps "power host-1 up" and then later "power host-1 down" then any new "power host-1 up" steps could be filtered out - and let's see how this effects performance
process
planner filter out repeating inverses filter out repeating inverses when the plan has already steps power host up and then later power host down then any new power host up steps could be filtered out and let s see how this effects performance
1
368,600
25,799,866,232
IssuesEvent
2022-12-10 22:51:58
EliasSchaut/Movie-Monday-Manager
https://api.github.com/repos/EliasSchaut/Movie-Monday-Manager
closed
The Dev Update
documentation enhancement
- Dev Documentation - Api Documentation - Type definitions for data transmision objects - Improve set up and installation
1.0
The Dev Update - - Dev Documentation - Api Documentation - Type definitions for data transmision objects - Improve set up and installation
non_process
the dev update dev documentation api documentation type definitions for data transmision objects improve set up and installation
0
69,806
7,164,635,682
IssuesEvent
2018-01-29 11:56:38
smap-consulting/smapserver
https://api.github.com/repos/smap-consulting/smapserver
closed
BUG - 17.12 - Help link in sessions goes to wrong help folder
In Test bug
### SmapServer ### Version 17.12 ### Expected Behaviour: Link must goes to: [Here](https://drive.google.com/drive/folders/0B966-DLxVMApbDJzd1NVVUMyNms) ### Actual Behaviour link goes to: [SMAP](https://drive.google.com/drive/folders/0B5_SmpWlQYxvNExwSFV5X051Yjg) ![image](https://user-images.githubusercontent.com/29519901/35484517-e2321ef2-041e-11e8-941f-9b84b3f84e86.png)
1.0
BUG - 17.12 - Help link in sessions goes to wrong help folder - ### SmapServer ### Version 17.12 ### Expected Behaviour: Link must goes to: [Here](https://drive.google.com/drive/folders/0B966-DLxVMApbDJzd1NVVUMyNms) ### Actual Behaviour link goes to: [SMAP](https://drive.google.com/drive/folders/0B5_SmpWlQYxvNExwSFV5X051Yjg) ![image](https://user-images.githubusercontent.com/29519901/35484517-e2321ef2-041e-11e8-941f-9b84b3f84e86.png)
non_process
bug help link in sessions goes to wrong help folder smapserver version expected behaviour link must goes to actual behaviour link goes to
0
1,963
4,779,096,737
IssuesEvent
2016-10-27 21:19:45
nodejs/node
https://api.github.com/repos/nodejs/node
closed
doc: need better link to signals list in process documentation
doc good first contribution process
* **Version**: all * **Platform**: n/a * **Subsystem**: doc In the 'Signal Events' section of the `process` documentation, there is a link to a man page that supposedly lists the names of the various signals that can be listened for. However, the page that is currently linked ([sigaction(2)](http://man7.org/linux/man-pages/man2/sigaction.2.html)) doesn't show all of the valid signals and the ones that it does show are not formatted very well. I think a better link would be to [signal(7)](http://man7.org/linux/man-pages/man7/signal.7.html), which *does* list all of the signal names and in an easy-to-spot/read format.
1.0
doc: need better link to signals list in process documentation - * **Version**: all * **Platform**: n/a * **Subsystem**: doc In the 'Signal Events' section of the `process` documentation, there is a link to a man page that supposedly lists the names of the various signals that can be listened for. However, the page that is currently linked ([sigaction(2)](http://man7.org/linux/man-pages/man2/sigaction.2.html)) doesn't show all of the valid signals and the ones that it does show are not formatted very well. I think a better link would be to [signal(7)](http://man7.org/linux/man-pages/man7/signal.7.html), which *does* list all of the signal names and in an easy-to-spot/read format.
process
doc need better link to signals list in process documentation version all platform n a subsystem doc in the signal events section of the process documentation there is a link to a man page that supposedly lists the names of the various signals that can be listened for however the page that is currently linked doesn t show all of the valid signals and the ones that it does show are not formatted very well i think a better link would be to which does list all of the signal names and in an easy to spot read format
1
26,457
2,684,552,301
IssuesEvent
2015-03-29 03:23:59
cs2103jan2015-w13-1j/main
https://api.github.com/repos/cs2103jan2015-w13-1j/main
closed
A user can remove multiple tasks by tags
priority.medium
... so that I do not need to remove them individually
1.0
A user can remove multiple tasks by tags - ... so that I do not need to remove them individually
non_process
a user can remove multiple tasks by tags so that i do not need to remove them individually
0
21,920
30,446,458,712
IssuesEvent
2023-07-15 18:31:11
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
pyutils 0.0.1a1 has 2 GuardDog issues
guarddog typosquatting silent-process-execution
https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1a1", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils", "silent-process-execution": [ { "location": "pyutils-0.0.1a1/src/pyutils/exec_utils.py:200", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpxi7_due4/pyutils" } }```
1.0
pyutils 0.0.1a1 has 2 GuardDog issues - https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1a1", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils", "silent-process-execution": [ { "location": "pyutils-0.0.1a1/src/pyutils/exec_utils.py:200", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpxi7_due4/pyutils" } }```
process
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt pytils python utils silent process execution location pyutils src pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp pyutils
1
4,499
7,348,662,047
IssuesEvent
2018-03-08 07:40:11
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Missed help for all Processing Graphics algorithms
Processing
Moved from QGIS redmine https://hub.qgis.org/issues/12702 There are no descriptions for Bar plot, Mean and std.dev plot, ~~Polar plot~~ and Vector layer scatterplot algorithms
1.0
Missed help for all Processing Graphics algorithms - Moved from QGIS redmine https://hub.qgis.org/issues/12702 There are no descriptions for Bar plot, Mean and std.dev plot, ~~Polar plot~~ and Vector layer scatterplot algorithms
process
missed help for all processing graphics algorithms moved from qgis redmine there are no descriptions for bar plot mean and std dev plot polar plot and vector layer scatterplot algorithms
1
38,358
2,846,359,873
IssuesEvent
2015-05-29 10:55:55
WarEmu/WarBugs
https://api.github.com/repos/WarEmu/WarBugs
closed
Relogging onto another character makes the previous characters abilities appear on the action bar
Ability Emulator Interface Low Priority
When you relog a character to another one. The abilities from the previous character are shown on the second character's hot bar. For example: Relogging from an Archmage to a Sword master. You get the abilities of the Archmage on the Sword master's hot bar. ![relogging char hot bar ability crossing](https://cloud.githubusercontent.com/assets/7492543/2883322/b5cfbb84-d499-11e3-8530-a5b0669027b9.png)
1.0
Relogging onto another character makes the previous characters abilities appear on the action bar - When you relog a character to another one. The abilities from the previous character are shown on the second character's hot bar. For example: Relogging from an Archmage to a Sword master. You get the abilities of the Archmage on the Sword master's hot bar. ![relogging char hot bar ability crossing](https://cloud.githubusercontent.com/assets/7492543/2883322/b5cfbb84-d499-11e3-8530-a5b0669027b9.png)
non_process
relogging onto another character makes the previous characters abilities appear on the action bar when you relog a character to another one the abilities from the previous character are shown on the second character s hot bar for example relogging from an archmage to a sword master you get the abilities of the archmage on the sword master s hot bar
0
9,552
12,514,450,665
IssuesEvent
2020-06-03 05:17:58
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Aggregate tool not working
Bug Processing Regression
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** Aggregate tool doesn't work **How to Reproduce** 1. Go to Toolbox 2. Click on Aggregate tool 3. Python error message pops and the tools fails to "start" 4. See error: `Traceback (most recent call last): File "/usr/share/qgis/python/plugins/processing/gui/ProcessingToolbox.py", line 262, in executeAlgorithm dlg = AlgorithmDialog(alg, self.in_place_mode, iface.mainWindow()) File "/usr/share/qgis/python/plugins/processing/gui/AlgorithmDialog.py", line 67, in __init__ self.setMainWidget(self.getParametersPanel(alg, self)) File "/usr/share/qgis/python/plugins/processing/gui/AlgorithmDialog.py", line 87, in getParametersPanel panel = ParametersPanel(parent, alg, self.in_place) File "/usr/share/qgis/python/plugins/processing/gui/ParametersPanel.py", line 70, in __init__ self.initWidgets() File "/usr/share/qgis/python/plugins/processing/gui/ParametersPanel.py", line 104, in initWidgets wrapper = WidgetWrapperFactory.create_wrapper(param, self.parent()) File "/usr/share/qgis/python/plugins/processing/gui/wrappers.py", line 1845, in create_wrapper return WidgetWrapperFactory.create_wrapper_from_metadata(param, dialog, row, col) File "/usr/share/qgis/python/plugins/processing/gui/wrappers.py", line 1876, in create_wrapper_from_metadata mod = __import__('.'.join(tokens[:-1]), fromlist=[tokens[-1]]) File "/usr/lib/python3/dist-packages/qgis/utils.py", line 743, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/usr/share/qgis/python/plugins/processing/algs/qgis/ui/AggregatesPanel.py", line 45, in from processing.algs.qgis.ui.FieldsMappingPanel import ( ImportError: cannot import name 'ExpressionDelegate' from 'processing.algs.qgis.ui.FieldsMappingPanel' (/usr/share/qgis/python/plugins/processing/algs/qgis/ui/FieldsMappingPanel.py) ` **QGIS and OS versions** QGIS 3.13 Master PopOS 20.04 / Ubuntu 20.04
1.0
Aggregate tool not working - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** Aggregate tool doesn't work **How to Reproduce** 1. Go to Toolbox 2. Click on Aggregate tool 3. Python error message pops and the tools fails to "start" 4. See error: `Traceback (most recent call last): File "/usr/share/qgis/python/plugins/processing/gui/ProcessingToolbox.py", line 262, in executeAlgorithm dlg = AlgorithmDialog(alg, self.in_place_mode, iface.mainWindow()) File "/usr/share/qgis/python/plugins/processing/gui/AlgorithmDialog.py", line 67, in __init__ self.setMainWidget(self.getParametersPanel(alg, self)) File "/usr/share/qgis/python/plugins/processing/gui/AlgorithmDialog.py", line 87, in getParametersPanel panel = ParametersPanel(parent, alg, self.in_place) File "/usr/share/qgis/python/plugins/processing/gui/ParametersPanel.py", line 70, in __init__ self.initWidgets() File "/usr/share/qgis/python/plugins/processing/gui/ParametersPanel.py", line 104, in initWidgets wrapper = WidgetWrapperFactory.create_wrapper(param, self.parent()) File "/usr/share/qgis/python/plugins/processing/gui/wrappers.py", line 1845, in create_wrapper return WidgetWrapperFactory.create_wrapper_from_metadata(param, dialog, row, col) File "/usr/share/qgis/python/plugins/processing/gui/wrappers.py", line 1876, in create_wrapper_from_metadata mod = __import__('.'.join(tokens[:-1]), fromlist=[tokens[-1]]) File "/usr/lib/python3/dist-packages/qgis/utils.py", line 743, in _import mod = _builtin_import(name, globals, locals, fromlist, level) File "/usr/share/qgis/python/plugins/processing/algs/qgis/ui/AggregatesPanel.py", line 45, in from processing.algs.qgis.ui.FieldsMappingPanel import ( ImportError: cannot import name 'ExpressionDelegate' from 'processing.algs.qgis.ui.FieldsMappingPanel' (/usr/share/qgis/python/plugins/processing/algs/qgis/ui/FieldsMappingPanel.py) ` **QGIS and OS versions** QGIS 3.13 Master PopOS 20.04 / Ubuntu 20.04
process
aggregate tool not working bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug aggregate tool doesn t work how to reproduce go to toolbox click on aggregate tool python error message pops and the tools fails to start see error traceback most recent call last file usr share qgis python plugins processing gui processingtoolbox py line in executealgorithm dlg algorithmdialog alg self in place mode iface mainwindow file usr share qgis python plugins processing gui algorithmdialog py line in init self setmainwidget self getparameterspanel alg self file usr share qgis python plugins processing gui algorithmdialog py line in getparameterspanel panel parameterspanel parent alg self in place file usr share qgis python plugins processing gui parameterspanel py line in init self initwidgets file usr share qgis python plugins processing gui parameterspanel py line in initwidgets wrapper widgetwrapperfactory create wrapper param self parent file usr share qgis python plugins processing gui wrappers py line in create wrapper return widgetwrapperfactory create wrapper from metadata param dialog row col file usr share qgis python plugins processing gui wrappers py line in create wrapper from metadata mod import join tokens fromlist file usr lib dist packages qgis utils py line in import mod builtin import name globals locals fromlist level file usr share qgis python plugins processing algs qgis ui aggregatespanel py line in from processing algs qgis ui fieldsmappingpanel import importerror cannot import name expressiondelegate from processing algs qgis ui fieldsmappingpanel usr share qgis python plugins processing algs qgis ui fieldsmappingpanel py qgis and os versions qgis master popos ubuntu
1
619
3,086,178,612
IssuesEvent
2015-08-25 00:39:48
e107inc/e107
https://api.github.com/repos/e107inc/e107
closed
alpha 2 upgrade database update
bug upgrade process 1.x to 2.x
How many times will the database update run? Is there any way I can check progress. I have clicked the update tab about 50 times now and it still says it is functioning correctly and the query stats vary slightly each time suggesting they are doing different work, but there is no end in site. It will not let me continue on to add additional/missing core files until this step completes.
1.0
alpha 2 upgrade database update - How many times will the database update run? Is there any way I can check progress. I have clicked the update tab about 50 times now and it still says it is functioning correctly and the query stats vary slightly each time suggesting they are doing different work, but there is no end in site. It will not let me continue on to add additional/missing core files until this step completes.
process
alpha upgrade database update how many times will the database update run is there any way i can check progress i have clicked the update tab about times now and it still says it is functioning correctly and the query stats vary slightly each time suggesting they are doing different work but there is no end in site it will not let me continue on to add additional missing core files until this step completes
1
20,988
27,853,626,269
IssuesEvent
2023-03-20 20:46:01
dDevTech/tapas-top-frontend
https://api.github.com/repos/dDevTech/tapas-top-frontend
opened
Panel usuario
in process
Diseño de panel con información resumida del perfil También incluir estadísticas - Número de degustaciones - Locales nuevos ult.7 días con opción de ver todo
1.0
Panel usuario - Diseño de panel con información resumida del perfil También incluir estadísticas - Número de degustaciones - Locales nuevos ult.7 días con opción de ver todo
process
panel usuario diseño de panel con información resumida del perfil también incluir estadísticas número de degustaciones locales nuevos ult días con opción de ver todo
1
15,052
18,762,899,670
IssuesEvent
2021-11-05 18:46:39
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
Moisture in combustion air
Process Heating
Field in Flue Gas calculator This is implemented for Solid/Liquid Fuels but not gas fuels. I think it is in the math for gas fuels and will be needed for the ESC 31 calc #4060 Hopefully just need the field visible with help, etc.
1.0
Moisture in combustion air - Field in Flue Gas calculator This is implemented for Solid/Liquid Fuels but not gas fuels. I think it is in the math for gas fuels and will be needed for the ESC 31 calc #4060 Hopefully just need the field visible with help, etc.
process
moisture in combustion air field in flue gas calculator this is implemented for solid liquid fuels but not gas fuels i think it is in the math for gas fuels and will be needed for the esc calc hopefully just need the field visible with help etc
1
437,307
12,576,526,931
IssuesEvent
2020-06-09 08:02:54
kubeflow/manifests
https://api.github.com/repos/kubeflow/manifests
closed
application CR's istio-install and knative-serving-install do not deploy
area/kfctl kind/bug lifecycle/stale priority/p1
both fail for different reasons knative-serving-install deployment seems to wait forever istio-install seems to not resolve all $(generateName)'s
1.0
application CR's istio-install and knative-serving-install do not deploy - both fail for different reasons knative-serving-install deployment seems to wait forever istio-install seems to not resolve all $(generateName)'s
non_process
application cr s istio install and knative serving install do not deploy both fail for different reasons knative serving install deployment seems to wait forever istio install seems to not resolve all generatename s
0
1,895
4,724,814,764
IssuesEvent
2016-10-18 02:36:35
webshell/materia-designer
https://api.github.com/repos/webshell/materia-designer
closed
Ghost created when Materia designer is restarted from the Chrome dev tools
area/materia-server-process type/bug
### Expected behavior Clear the process before the restart ### Actual behavior Restart without deleting internal process. ### Step to reproduce the behavior * Navigate in Materia designer a while * Open the Chrome dev tools * Go in the **Network** tab => it asks you to refresh with `Command+R` * Hit `Command+R` => Materia designer restart => a ghost of the latest Materia-server process is still in memory. --- * Materia Server Version: 0.1.1 * Materia Designer Version: 0.1.0-rc.3
1.0
Ghost created when Materia designer is restarted from the Chrome dev tools - ### Expected behavior Clear the process before the restart ### Actual behavior Restart without deleting internal process. ### Step to reproduce the behavior * Navigate in Materia designer a while * Open the Chrome dev tools * Go in the **Network** tab => it asks you to refresh with `Command+R` * Hit `Command+R` => Materia designer restart => a ghost of the latest Materia-server process is still in memory. --- * Materia Server Version: 0.1.1 * Materia Designer Version: 0.1.0-rc.3
process
ghost created when materia designer is restarted from the chrome dev tools expected behavior clear the process before the restart actual behavior restart without deleting internal process step to reproduce the behavior navigate in materia designer a while open the chrome dev tools go in the network tab it asks you to refresh with command r hit command r materia designer restart a ghost of the latest materia server process is still in memory materia server version materia designer version rc
1
2,109
4,944,604,554
IssuesEvent
2016-11-30 08:32:59
mathiasbynens/es-regexp-dotall-flag
https://api.github.com/repos/mathiasbynens/es-regexp-dotall-flag
closed
Advance to stage 1
process
Criteria taken from [the TC39 process document](https://tc39.github.io/process-document/): > - [x] Identified “champion” who will advance the addition Done: @bterlson. > - [x] Prose outlining the problem or need and the general shape of a solution https://github.com/mathiasbynens/es-regexp-singleline-flag#motivation > - [x] Illustrative examples of usage https://github.com/mathiasbynens/es-regexp-dotall-flag#high-level-api > - [x] High-level API https://github.com/mathiasbynens/es-regexp-singleline-flag#high-level-api > - [x] Discussion of key algorithms, abstractions and semantics https://github.com/mathiasbynens/es-regexp-singleline-flag#faq > - [x] Identification of potential “cross-cutting” concerns and implementation challenges/complexity There should be none.
1.0
Advance to stage 1 - Criteria taken from [the TC39 process document](https://tc39.github.io/process-document/): > - [x] Identified “champion” who will advance the addition Done: @bterlson. > - [x] Prose outlining the problem or need and the general shape of a solution https://github.com/mathiasbynens/es-regexp-singleline-flag#motivation > - [x] Illustrative examples of usage https://github.com/mathiasbynens/es-regexp-dotall-flag#high-level-api > - [x] High-level API https://github.com/mathiasbynens/es-regexp-singleline-flag#high-level-api > - [x] Discussion of key algorithms, abstractions and semantics https://github.com/mathiasbynens/es-regexp-singleline-flag#faq > - [x] Identification of potential “cross-cutting” concerns and implementation challenges/complexity There should be none.
process
advance to stage criteria taken from identified “champion” who will advance the addition done bterlson prose outlining the problem or need and the general shape of a solution illustrative examples of usage high level api discussion of key algorithms abstractions and semantics identification of potential “cross cutting” concerns and implementation challenges complexity there should be none
1
20,036
26,520,442,149
IssuesEvent
2023-01-19 01:47:18
DIGGSml/Geophysics
https://api.github.com/repos/DIGGSml/Geophysics
opened
General Topic: Spatial-temporal domain data
Data acquisition Processed data
The Test measurement structure currently only supports data in a spatial domain, where all coordinate axes are spatial. To support spatial-temporal results, where the "locatioN" of a data value has a temporal axis, likely can be handled by providing reference compound spatial reference systems and/or the ability for the user to define a custom spatial-temporal compound coordinate For reference material on compound CRS's: https://docs.opengeospatial.org/is/18-010r7/18-010r7.html.
1.0
General Topic: Spatial-temporal domain data - The Test measurement structure currently only supports data in a spatial domain, where all coordinate axes are spatial. To support spatial-temporal results, where the "locatioN" of a data value has a temporal axis, likely can be handled by providing reference compound spatial reference systems and/or the ability for the user to define a custom spatial-temporal compound coordinate For reference material on compound CRS's: https://docs.opengeospatial.org/is/18-010r7/18-010r7.html.
process
general topic spatial temporal domain data the test measurement structure currently only supports data in a spatial domain where all coordinate axes are spatial to support spatial temporal results where the location of a data value has a temporal axis likely can be handled by providing reference compound spatial reference systems and or the ability for the user to define a custom spatial temporal compound coordinate for reference material on compound crs s
1
211,373
23,808,128,890
IssuesEvent
2022-09-04 11:03:09
neptune-mutual-blue/protocol
https://api.github.com/repos/neptune-mutual-blue/protocol
closed
Decouple pause and unpause ACL logic in the store contract
enhancement security rc2
Unlike the `unpause` feature, which is solely available to the owner, the `pause` feature of the store should be accessible to a separate account in order to programmatically pause the contract. In a hosted environment, this helps to avoid using the owner's private key.
True
Decouple pause and unpause ACL logic in the store contract - Unlike the `unpause` feature, which is solely available to the owner, the `pause` feature of the store should be accessible to a separate account in order to programmatically pause the contract. In a hosted environment, this helps to avoid using the owner's private key.
non_process
decouple pause and unpause acl logic in the store contract unlike the unpause feature which is solely available to the owner the pause feature of the store should be accessible to a separate account in order to programmatically pause the contract in a hosted environment this helps to avoid using the owner s private key
0
7,321
10,454,860,287
IssuesEvent
2019-09-19 19:53:16
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
Inconsistencies between Efficiency Improvement Calculator and O2 Enrichment calculator
Process Heating bug
User reported differences via email: Consider the basic problem below: Natural Gas Furnace Current Flue Gas Oxygen = New Flue Gas Oxygen = 4% Current Flue Gas Temp = New Flue Gas Temp = 500 F Current Combustion Air = 80 F New Combustion Air = 300 F The Fuel Savings (%) calculated by the "Efficiency Improvement Calculator" does not match with the Fuel Savings calculated by the "Oxygen Enrichment Calculator". The latter is closer to PHAST's calculations. Can you look into why the two calculators give different answers?
1.0
Inconsistencies between Efficiency Improvement Calculator and O2 Enrichment calculator - User reported differences via email: Consider the basic problem below: Natural Gas Furnace Current Flue Gas Oxygen = New Flue Gas Oxygen = 4% Current Flue Gas Temp = New Flue Gas Temp = 500 F Current Combustion Air = 80 F New Combustion Air = 300 F The Fuel Savings (%) calculated by the "Efficiency Improvement Calculator" does not match with the Fuel Savings calculated by the "Oxygen Enrichment Calculator". The latter is closer to PHAST's calculations. Can you look into why the two calculators give different answers?
process
inconsistencies between efficiency improvement calculator and enrichment calculator user reported differences via email consider the basic problem below natural gas furnace current flue gas oxygen new flue gas oxygen current flue gas temp new flue gas temp f current combustion air f new combustion air f the fuel savings calculated by the efficiency improvement calculator does not match with the fuel savings calculated by the oxygen enrichment calculator the latter is closer to phast s calculations can you look into why the two calculators give different answers
1
296,215
22,292,681,854
IssuesEvent
2022-06-12 15:45:15
noradream/SwimSetGenerator
https://api.github.com/repos/noradream/SwimSetGenerator
opened
Improve Code Readibility
documentation
JavaDoc & Other comments are very clunky right now, and need an overhaul.
1.0
Improve Code Readibility - JavaDoc & Other comments are very clunky right now, and need an overhaul.
non_process
improve code readibility javadoc other comments are very clunky right now and need an overhaul
0
17,290
23,098,664,137
IssuesEvent
2022-07-26 22:41:33
maticnetwork/miden
https://api.github.com/repos/maticnetwork/miden
closed
MPVERIFY instruction potential refactoring
assembly instruction set processor
`MPVERIFY` computes the Merkle root of a path for a given specified node in the stack. The stack is arranged as follows (from the top): - depth of the path, 1 element. - index of the node, 1 element. - value of the node, 4 elements. - root of the tree, 4 elements. Currently, post this operation, we are replacing the node value with the newly computed root calculated by the `hasher` co-processor. The stack will look like this post this operation: - depth of the path, 1 element. - index of the node, 1 element. - computed root of the path, 4 elements. - root of the tree, 4 elements. We are already asserting the computed root of the path with root of the tree and repeating this in assembly(`mtree.get`) wastes couple of clock cycles. As noted by @bobbinth: > "Instead of replacing the node with the newly computed root, we should just copy over the stack to the next cycle. The instruction should fail if the roots are not the same". Operations relying on `MPVERIFY` might get a bit simplified due to fewer VM cycles needed now. edit: The instruction also needs to be populated with six helper registers which are as follows: - row address of the start of the computation in hasher chiplet - 1 element - the least significant bit of the node index - 1 element - sibling node of the leaf - 4 elements The `mtree.get` assembly instruction also needs to be updated to be in line with the new changes made in `MPVERIFY`as there are some potential optimization possibilities..
1.0
MPVERIFY instruction potential refactoring - `MPVERIFY` computes the Merkle root of a path for a given specified node in the stack. The stack is arranged as follows (from the top): - depth of the path, 1 element. - index of the node, 1 element. - value of the node, 4 elements. - root of the tree, 4 elements. Currently, post this operation, we are replacing the node value with the newly computed root calculated by the `hasher` co-processor. The stack will look like this post this operation: - depth of the path, 1 element. - index of the node, 1 element. - computed root of the path, 4 elements. - root of the tree, 4 elements. We are already asserting the computed root of the path with root of the tree and repeating this in assembly(`mtree.get`) wastes couple of clock cycles. As noted by @bobbinth: > "Instead of replacing the node with the newly computed root, we should just copy over the stack to the next cycle. The instruction should fail if the roots are not the same". Operations relying on `MPVERIFY` might get a bit simplified due to fewer VM cycles needed now. edit: The instruction also needs to be populated with six helper registers which are as follows: - row address of the start of the computation in hasher chiplet - 1 element - the least significant bit of the node index - 1 element - sibling node of the leaf - 4 elements The `mtree.get` assembly instruction also needs to be updated to be in line with the new changes made in `MPVERIFY`as there are some potential optimization possibilities..
process
mpverify instruction potential refactoring mpverify computes the merkle root of a path for a given specified node in the stack the stack is arranged as follows from the top depth of the path element index of the node element value of the node elements root of the tree elements currently post this operation we are replacing the node value with the newly computed root calculated by the hasher co processor the stack will look like this post this operation depth of the path element index of the node element computed root of the path elements root of the tree elements we are already asserting the computed root of the path with root of the tree and repeating this in assembly mtree get wastes couple of clock cycles as noted by bobbinth instead of replacing the node with the newly computed root we should just copy over the stack to the next cycle the instruction should fail if the roots are not the same operations relying on mpverify might get a bit simplified due to fewer vm cycles needed now edit the instruction also needs to be populated with six helper registers which are as follows row address of the start of the computation in hasher chiplet element the least significant bit of the node index element sibling node of the leaf elements the mtree get assembly instruction also needs to be updated to be in line with the new changes made in mpverify as there are some potential optimization possibilities
1
80,880
15,589,336,045
IssuesEvent
2021-03-18 07:54:04
AlchTech/JeremySalettes-Wozniak_4_05032021
https://api.github.com/repos/AlchTech/JeremySalettes-Wozniak_4_05032021
closed
CVE-2016-10735 (Medium) detected in bootstrap-3.3.5.min.js
security vulnerability
## CVE-2016-10735 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.5.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js</a></p> <p>Path to dependency file: JeremySalettes-Wozniak_4_05032021/index.html</p> <p>Path to vulnerable library: JeremySalettes-Wozniak_4_05032021/js/bootstrap.js,JeremySalettes-Wozniak_4_05032021/./js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.5.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlchTech/JeremySalettes-Wozniak_4_05032021/commit/6596720eb100c60b452ca547f2a47690766344f7">6596720eb100c60b452ca547f2a47690766344f7</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-10735 (Medium) detected in bootstrap-3.3.5.min.js - ## CVE-2016-10735 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.5.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js</a></p> <p>Path to dependency file: JeremySalettes-Wozniak_4_05032021/index.html</p> <p>Path to vulnerable library: JeremySalettes-Wozniak_4_05032021/js/bootstrap.js,JeremySalettes-Wozniak_4_05032021/./js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.5.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlchTech/JeremySalettes-Wozniak_4_05032021/commit/6596720eb100c60b452ca547f2a47690766344f7">6596720eb100c60b452ca547f2a47690766344f7</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file jeremysalettes wozniak index html path to vulnerable library jeremysalettes wozniak js bootstrap js jeremysalettes wozniak js bootstrap js dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch main vulnerability details in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
16,097
20,268,502,150
IssuesEvent
2022-02-15 14:17:24
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[FLINK-26161][BUG] PulsarSource: Split should be added to finishedSplit if the the split is empty from beginning.
compute/data-processing type/bug
Currently we add a split to finishedSplit in the read for loop, but if the split is empty and never enters for loop it should be checked whether to add to finishedSplit as well. This is an edge case and priority is low. ``` // current if (stopCursor.shouldStop(message)) { builder.addFinishedSplit(splitId); break; } ```
1.0
[FLINK-26161][BUG] PulsarSource: Split should be added to finishedSplit if the the split is empty from beginning. - Currently we add a split to finishedSplit in the read for loop, but if the split is empty and never enters for loop it should be checked whether to add to finishedSplit as well. This is an edge case and priority is low. ``` // current if (stopCursor.shouldStop(message)) { builder.addFinishedSplit(splitId); break; } ```
process
pulsarsource split should be added to finishedsplit if the the split is empty from beginning currently we add a split to finishedsplit in the read for loop but if the split is empty and never enters for loop it should be checked whether to add to finishedsplit as well this is an edge case and priority is low current if stopcursor shouldstop message builder addfinishedsplit splitid break
1
9,589
12,540,472,181
IssuesEvent
2020-06-05 10:25:56
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
Anomalies preprocessor with period 'full' removes important metadata from cube
bug preprocessor
**Describe the bug** When the `anomalies` preprocessor function is applied with `period=full`, it removes important metadata such as the variable name, all attributes, etc. This doesn't happen for other periods like `month` or `day`. Consequently, subsequent preprocessors or diagnostics fail. I think it's happening [here](https://github.com/ESMValGroup/ESMValCore/blob/8306a6561a011f61dec949dc56889568b1af815e/esmvalcore/preprocessor/_time.py#L463): ``` cube = cube - reference ``` and this seems to be the intended behaviour in Iris, as it is documented [here](https://scitools.org.uk/iris/docs/latest/userguide/cube_maths.html#calculating-a-cube-anomaly). Perhaps EMSValTool should only subtract the data instead of the full cube? **Example** `print(cube)`: ``` air_temperature / (K) (time: 1668) Dimension coordinates: time x Auxiliary coordinates: day_of_year x month_number x Scalar coordinates: height: 1.5 m latitude: 0.0 degrees, bound=(-90.0, 90.0) degrees longitude: 179.0625 degrees, bound=(-0.9375, 359.0625) degrees Attributes: Conventions: CF-1.7 associated_files: baseURL: http://cmip-pcmdi.llnl.gov/CMIP5/dataLocation gridspecFile: gridspec_atmos_fx_ACCESS1-0_rcp45_r0i0p0.nc... branch_time: 732311.0;109207.0;732311.0 cmor_version: 2.8.0 contact: The ACCESS wiki: http://wiki.csiro.au/confluence/display/ACCESS/Home. Contact... experiment: RCP4.5;historical;RCP4.5 experiment_id: rcp45;historical;rcp45 forcing: GHG, Oz, SA, Sl, Vl, BC, OC, (GHG = CO2, N2O, CH4, CFC11, CFC12, CFC113,... frequency: mon initialization_method: 1 institute_id: CSIRO-BOM institution: CSIRO (Commonwealth Scientific and Industrial Research Organisation, Australia),... model_id: ACCESS1-0 modeling_realm: atmos parent_experiment: historical;pre-industrial control;historical parent_experiment_id: historical;piControl;historical parent_experiment_rip: r1i1p1 physics_version: 1 product: output project_id: CMIP5 realization: 1 references: See http://wiki.csiro.au/confluence/display/ACCESS/ACCESS+Publications source: ACCESS1-0 2011. Atmosphere: AGCM v1.0 (N96 grid-point, 1.875 degrees EW... table_id: Table Amon (27 April 2011) 9c851218e3842df9a62ef38b1e2575bb title: ACCESS1-0 model output prepared for CMIP5 RCP4.5;ACCESS1-0 model output... version_number: v20120115 Cell methods: mean: time mean: longitude, latitude ``` ``` anoms = anomalies(cube, period='full', reference=dict(start_year=1980, start_month=1, start_day=1, end_year=2009, end_month=12, end_day=31)) print(anoms) ``` gives ``` unknown / (K) (time: 1668) Dimension coordinates: time x Auxiliary coordinates: month_number x Scalar coordinates: height: 1.5 m latitude: 0.0 degrees, bound=(-90.0, 90.0) degrees longitude: 179.0625 degrees, bound=(-0.9375, 359.0625) degrees ```
1.0
Anomalies preprocessor with period 'full' removes important metadata from cube - **Describe the bug** When the `anomalies` preprocessor function is applied with `period=full`, it removes important metadata such as the variable name, all attributes, etc. This doesn't happen for other periods like `month` or `day`. Consequently, subsequent preprocessors or diagnostics fail. I think it's happening [here](https://github.com/ESMValGroup/ESMValCore/blob/8306a6561a011f61dec949dc56889568b1af815e/esmvalcore/preprocessor/_time.py#L463): ``` cube = cube - reference ``` and this seems to be the intended behaviour in Iris, as it is documented [here](https://scitools.org.uk/iris/docs/latest/userguide/cube_maths.html#calculating-a-cube-anomaly). Perhaps EMSValTool should only subtract the data instead of the full cube? **Example** `print(cube)`: ``` air_temperature / (K) (time: 1668) Dimension coordinates: time x Auxiliary coordinates: day_of_year x month_number x Scalar coordinates: height: 1.5 m latitude: 0.0 degrees, bound=(-90.0, 90.0) degrees longitude: 179.0625 degrees, bound=(-0.9375, 359.0625) degrees Attributes: Conventions: CF-1.7 associated_files: baseURL: http://cmip-pcmdi.llnl.gov/CMIP5/dataLocation gridspecFile: gridspec_atmos_fx_ACCESS1-0_rcp45_r0i0p0.nc... branch_time: 732311.0;109207.0;732311.0 cmor_version: 2.8.0 contact: The ACCESS wiki: http://wiki.csiro.au/confluence/display/ACCESS/Home. Contact... experiment: RCP4.5;historical;RCP4.5 experiment_id: rcp45;historical;rcp45 forcing: GHG, Oz, SA, Sl, Vl, BC, OC, (GHG = CO2, N2O, CH4, CFC11, CFC12, CFC113,... frequency: mon initialization_method: 1 institute_id: CSIRO-BOM institution: CSIRO (Commonwealth Scientific and Industrial Research Organisation, Australia),... model_id: ACCESS1-0 modeling_realm: atmos parent_experiment: historical;pre-industrial control;historical parent_experiment_id: historical;piControl;historical parent_experiment_rip: r1i1p1 physics_version: 1 product: output project_id: CMIP5 realization: 1 references: See http://wiki.csiro.au/confluence/display/ACCESS/ACCESS+Publications source: ACCESS1-0 2011. Atmosphere: AGCM v1.0 (N96 grid-point, 1.875 degrees EW... table_id: Table Amon (27 April 2011) 9c851218e3842df9a62ef38b1e2575bb title: ACCESS1-0 model output prepared for CMIP5 RCP4.5;ACCESS1-0 model output... version_number: v20120115 Cell methods: mean: time mean: longitude, latitude ``` ``` anoms = anomalies(cube, period='full', reference=dict(start_year=1980, start_month=1, start_day=1, end_year=2009, end_month=12, end_day=31)) print(anoms) ``` gives ``` unknown / (K) (time: 1668) Dimension coordinates: time x Auxiliary coordinates: month_number x Scalar coordinates: height: 1.5 m latitude: 0.0 degrees, bound=(-90.0, 90.0) degrees longitude: 179.0625 degrees, bound=(-0.9375, 359.0625) degrees ```
process
anomalies preprocessor with period full removes important metadata from cube describe the bug when the anomalies preprocessor function is applied with period full it removes important metadata such as the variable name all attributes etc this doesn t happen for other periods like month or day consequently subsequent preprocessors or diagnostics fail i think it s happening cube cube reference and this seems to be the intended behaviour in iris as it is documented perhaps emsvaltool should only subtract the data instead of the full cube example print cube air temperature k time dimension coordinates time x auxiliary coordinates day of year x month number x scalar coordinates height m latitude degrees bound degrees longitude degrees bound degrees attributes conventions cf associated files baseurl gridspecfile gridspec atmos fx nc branch time cmor version contact the access wiki contact experiment historical experiment id historical forcing ghg oz sa sl vl bc oc ghg frequency mon initialization method institute id csiro bom institution csiro commonwealth scientific and industrial research organisation australia model id modeling realm atmos parent experiment historical pre industrial control historical parent experiment id historical picontrol historical parent experiment rip physics version product output project id realization references see source atmosphere agcm grid point degrees ew table id table amon april title model output prepared for model output version number cell methods mean time mean longitude latitude anoms anomalies cube period full reference dict start year start month start day end year end month end day print anoms gives unknown k time dimension coordinates time x auxiliary coordinates month number x scalar coordinates height m latitude degrees bound degrees longitude degrees bound degrees
1
8,628
11,782,266,773
IssuesEvent
2020-03-17 01:17:55
googleapis/java-storage-nio
https://api.github.com/repos/googleapis/java-storage-nio
closed
org.mockito.matcher + org.junit.Assert + org.junit.rules.ExpectedException has been deprecated
api: storage type: cleanup type: process
``` [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileSystemTest.java:[61,60] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[106,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[125,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[126,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[127,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[128,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[135,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[136,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[137,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[138,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[139,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[140,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[145,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[146,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[147,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[148,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[149,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[150,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[151,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[156,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[162,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[163,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[167,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[168,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[169,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[175,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[176,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[177,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[178,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[184,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[186,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadTest.java:[58,60] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[50,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[177,5] verifyZeroInteractions(java.lang.Object...) in org.mockito.Mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStoragePathTest.java:[43,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[106,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[126,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[127,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[128,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[129,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[139,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[140,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[141,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[143,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[144,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[145,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[150,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[151,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[152,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[153,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[154,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[155,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[156,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[161,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[163,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[165,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[166,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[175,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[176,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[178,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[179,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[180,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[187,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[188,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[189,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[196,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[197,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[199,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[200,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[201,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[45,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[98,5] verifyZeroInteractions(java.lang.Object...) in org.mockito.Mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/UnixPathTest.java:[34,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageLateInitializationTest.java:[39,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileSystemProviderTest.java:[88,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileAttributeViewTest.java:[45,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageConfigurationTest.java:[34,66] none() in org.junit.rules.ExpectedException has been deprecated ```
1.0
org.mockito.matcher + org.junit.Assert + org.junit.rules.ExpectedException has been deprecated - ``` [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileSystemTest.java:[61,60] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[106,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[125,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[126,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[127,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[128,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[135,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[136,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[137,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[138,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[139,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[140,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[145,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[146,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[147,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[148,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[149,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[150,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[151,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[156,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[162,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[163,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[167,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[168,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[169,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[175,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[176,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[177,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[178,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[184,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[186,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadTest.java:[58,60] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[50,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[177,5] verifyZeroInteractions(java.lang.Object...) in org.mockito.Mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStoragePathTest.java:[43,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[106,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[126,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[127,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[128,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[129,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[139,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[140,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[141,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[143,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[144,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[145,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[150,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[151,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[152,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[153,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[154,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[155,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[156,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[161,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[163,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[165,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[166,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[175,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[176,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[178,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[179,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[180,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[187,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[188,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[189,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[196,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[197,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[199,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[200,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[201,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[45,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[98,5] verifyZeroInteractions(java.lang.Object...) in org.mockito.Mockito has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/UnixPathTest.java:[34,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageLateInitializationTest.java:[39,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileSystemProviderTest.java:[88,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileAttributeViewTest.java:[45,66] none() in org.junit.rules.ExpectedException has been deprecated [WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageConfigurationTest.java:[34,66] none() in org.junit.rules.ExpectedException has been deprecated ```
process
org mockito matcher org junit assert org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragefilesystemtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java verifyzerointeractions java lang object in org mockito mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragepathtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java verifyzerointeractions java lang object in org mockito mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio unixpathtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragelateinitializationtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragefilesystemprovidertest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragefileattributeviewtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstorageconfigurationtest java none in org junit rules expectedexception has been deprecated
1
13,419
15,880,355,727
IssuesEvent
2021-04-09 13:36:34
department-of-veterans-affairs/notification-api
https://api.github.com/repos/department-of-veterans-affairs/notification-api
closed
Get Amazon SES into production
Process Task
# Value Statement **As** VANotify platform developer **I want to** onboard Amazon SES as a provider into production **So that**I can use it to send emails and begin to "warm up" # Acceptance Criteria **GIVEN** a VANotify platform developer **WHEN** is viewing email providers in the VANotify platform **THEN** Amazon SES is an option with which to send emails # Checklist - [x] AWS Support request to get dedicated IP addresses - [x] AWS Support request for sending limits? - [x] AWS Support request to take us out of sandbox - [x] Submit forms with main ISPs to clear IP history - [x] Make sure account suppression list is on # Assumptions - # Additional Info/Resources - # Out of Scope - # Open Questions -
1.0
Get Amazon SES into production - # Value Statement **As** VANotify platform developer **I want to** onboard Amazon SES as a provider into production **So that**I can use it to send emails and begin to "warm up" # Acceptance Criteria **GIVEN** a VANotify platform developer **WHEN** is viewing email providers in the VANotify platform **THEN** Amazon SES is an option with which to send emails # Checklist - [x] AWS Support request to get dedicated IP addresses - [x] AWS Support request for sending limits? - [x] AWS Support request to take us out of sandbox - [x] Submit forms with main ISPs to clear IP history - [x] Make sure account suppression list is on # Assumptions - # Additional Info/Resources - # Out of Scope - # Open Questions -
process
get amazon ses into production value statement as vanotify platform developer i want to onboard amazon ses as a provider into production so that i can use it to send emails and begin to warm up acceptance criteria given a vanotify platform developer when is viewing email providers in the vanotify platform then amazon ses is an option with which to send emails checklist aws support request to get dedicated ip addresses aws support request for sending limits aws support request to take us out of sandbox submit forms with main isps to clear ip history make sure account suppression list is on assumptions additional info resources out of scope open questions
1
12,552
14,976,925,172
IssuesEvent
2021-01-28 08:48:01
threefoldtech/js-sdk
https://api.github.com/repos/threefoldtech/js-sdk
closed
failed to deploy mattermost
process_wontfix
![image](https://user-images.githubusercontent.com/43240801/106015344-b021f400-60be-11eb-926f-f590f15bd61b.png) failed to deploy mattermost after providing my MySQL username and password
1.0
failed to deploy mattermost - ![image](https://user-images.githubusercontent.com/43240801/106015344-b021f400-60be-11eb-926f-f590f15bd61b.png) failed to deploy mattermost after providing my MySQL username and password
process
failed to deploy mattermost failed to deploy mattermost after providing my mysql username and password
1
809,158
30,177,465,968
IssuesEvent
2023-07-04 06:23:40
channel-io/bezier-react
https://api.github.com/repos/channel-io/bezier-react
closed
Improve `Tooltip`'s ghosting
enhancement priority:A
### Description 툴팁의 잔상 현상을 개선합니다. ### Reasons for suggestion 다수의 툴팁이 있는 경우, 툴팁의 딜레이가 현재 `0` 으로 설정되어 있기 때문에 마우스 포인터 이동에 불필요하게 많은 툴팁이 보이는 경우가 있습니다. ### Proposed solution 1. `delayShow`, `skipDelayShow` 옵션을 적절히 조절합니다. 2. 애니메이션의 영향일 수 있습니다. 1번 안에 더해, 애니메이션을 제거한 상태에서 동작을 점검해보면 좋겠습니다. 필요에 따라 애니메이션을 제거합니다. ### References - [Feedback](https://desk.channel.io/root/groups/WebDesk-113051/649e88b11b0ee9ec8965)
1.0
Improve `Tooltip`'s ghosting - ### Description 툴팁의 잔상 현상을 개선합니다. ### Reasons for suggestion 다수의 툴팁이 있는 경우, 툴팁의 딜레이가 현재 `0` 으로 설정되어 있기 때문에 마우스 포인터 이동에 불필요하게 많은 툴팁이 보이는 경우가 있습니다. ### Proposed solution 1. `delayShow`, `skipDelayShow` 옵션을 적절히 조절합니다. 2. 애니메이션의 영향일 수 있습니다. 1번 안에 더해, 애니메이션을 제거한 상태에서 동작을 점검해보면 좋겠습니다. 필요에 따라 애니메이션을 제거합니다. ### References - [Feedback](https://desk.channel.io/root/groups/WebDesk-113051/649e88b11b0ee9ec8965)
non_process
improve tooltip s ghosting description 툴팁의 잔상 현상을 개선합니다 reasons for suggestion 다수의 툴팁이 있는 경우 툴팁의 딜레이가 현재 으로 설정되어 있기 때문에 마우스 포인터 이동에 불필요하게 많은 툴팁이 보이는 경우가 있습니다 proposed solution delayshow skipdelayshow 옵션을 적절히 조절합니다 애니메이션의 영향일 수 있습니다 안에 더해 애니메이션을 제거한 상태에서 동작을 점검해보면 좋겠습니다 필요에 따라 애니메이션을 제거합니다 references
0
26,505
2,684,633,813
IssuesEvent
2015-03-29 06:08:13
gtcasl/gpuocelot
https://api.github.com/repos/gtcasl/gpuocelot
opened
Build Error: llvm/Assembly/Parser.h: No such file or directory
bug imported Priority-Medium
_From [RichardH...@gmail.com](https://code.google.com/u/107171517134746479087/) on November 11, 2014 16:57:28_ What steps will reproduce the problem? 1. Build Ocelot using the build.py script: What is the expected output? What do you see instead? expect: build successfully build error: ocelot/ocelot/ir/implementation/ExternalFunctionSet.cpp:28:34: fatal error: llvm/Assembly/Parser.h: No such file or directory What version of the product are you using? On what operating system? I checkout the ocelot from svn. Also, checkout the llvm from svn. Please provide any additional information below. _Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=100_
1.0
Build Error: llvm/Assembly/Parser.h: No such file or directory - _From [RichardH...@gmail.com](https://code.google.com/u/107171517134746479087/) on November 11, 2014 16:57:28_ What steps will reproduce the problem? 1. Build Ocelot using the build.py script: What is the expected output? What do you see instead? expect: build successfully build error: ocelot/ocelot/ir/implementation/ExternalFunctionSet.cpp:28:34: fatal error: llvm/Assembly/Parser.h: No such file or directory What version of the product are you using? On what operating system? I checkout the ocelot from svn. Also, checkout the llvm from svn. Please provide any additional information below. _Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=100_
non_process
build error llvm assembly parser h no such file or directory from on november what steps will reproduce the problem build ocelot using the build py script what is the expected output what do you see instead expect build successfully build error ocelot ocelot ir implementation externalfunctionset cpp fatal error llvm assembly parser h no such file or directory what version of the product are you using on what operating system i checkout the ocelot from svn also checkout the llvm from svn please provide any additional information below original issue
0
238,405
7,779,261,895
IssuesEvent
2018-06-05 16:16:32
emory-libraries/ezpaarse-platforms
https://api.github.com/repos/emory-libraries/ezpaarse-platforms
opened
Cambridge Companions Online
High Priority
### Example:star::star: : https://proxy.library.emory.edu/login?url=http://cco.cambridge.org/uid=2892/private_home ### Priority: High ### Subscriber (Library): Theology
1.0
Cambridge Companions Online - ### Example:star::star: : https://proxy.library.emory.edu/login?url=http://cco.cambridge.org/uid=2892/private_home ### Priority: High ### Subscriber (Library): Theology
non_process
cambridge companions online example star star priority high subscriber library theology
0
272,791
23,703,651,805
IssuesEvent
2022-08-29 21:39:44
KaTeX/KaTeX
https://api.github.com/repos/KaTeX/KaTeX
closed
Update `toSplitInto` matcher to accept the same delimiter object type that's used in renderMathInElement
testing good first issue
Right now `toSplitInto` uses `display: false` for all tests.
1.0
Update `toSplitInto` matcher to accept the same delimiter object type that's used in renderMathInElement - Right now `toSplitInto` uses `display: false` for all tests.
non_process
update tosplitinto matcher to accept the same delimiter object type that s used in rendermathinelement right now tosplitinto uses display false for all tests
0
239,157
18,264,849,594
IssuesEvent
2021-10-04 07:09:40
ayushs1ngh/Fun-and-Tricky-Programs
https://api.github.com/repos/ayushs1ngh/Fun-and-Tricky-Programs
closed
Add fun/tricky/ uncommon programs
documentation help wanted good first issue question hacktoberfest up-for-grabs
Add a PR with any tricky or uncommon coding problem. Add coding problems and solutions in the respective language folder. if that language folder doesn't exist then create it. Check other folders in Java or C/CPP how it is done in the rest of the file. E.g Add a folder in the Java folder with the Name of Problem Title. Inside that folder add the solution ProblemTitle.java and a README.md Make sure the class name is ProblemTitle that is the name of the file. README.md must contain a detailed description of the problem with the test cases or example input/output.
1.0
Add fun/tricky/ uncommon programs - Add a PR with any tricky or uncommon coding problem. Add coding problems and solutions in the respective language folder. if that language folder doesn't exist then create it. Check other folders in Java or C/CPP how it is done in the rest of the file. E.g Add a folder in the Java folder with the Name of Problem Title. Inside that folder add the solution ProblemTitle.java and a README.md Make sure the class name is ProblemTitle that is the name of the file. README.md must contain a detailed description of the problem with the test cases or example input/output.
non_process
add fun tricky uncommon programs add a pr with any tricky or uncommon coding problem add coding problems and solutions in the respective language folder if that language folder doesn t exist then create it check other folders in java or c cpp how it is done in the rest of the file e g add a folder in the java folder with the name of problem title inside that folder add the solution problemtitle java and a readme md make sure the class name is problemtitle that is the name of the file readme md must contain a detailed description of the problem with the test cases or example input output
0
967
2,522,132,609
IssuesEvent
2015-01-19 19:38:07
YaleSTC/shifts
https://api.github.com/repos/YaleSTC/shifts
closed
Fix fixtures
testing
Currently, the RSpec suite (#281) runs on fixtures. [Whether or not this is an antipattern remains a matter of taste.](https://whatdoitest.com/getting-friendly-with-fixtures) Since we have fixtures, not factories, fixing them is worth investigating. (And if fixing fixtures is too taxing, migrating to factories will make #325 easier.)
1.0
Fix fixtures - Currently, the RSpec suite (#281) runs on fixtures. [Whether or not this is an antipattern remains a matter of taste.](https://whatdoitest.com/getting-friendly-with-fixtures) Since we have fixtures, not factories, fixing them is worth investigating. (And if fixing fixtures is too taxing, migrating to factories will make #325 easier.)
non_process
fix fixtures currently the rspec suite runs on fixtures since we have fixtures not factories fixing them is worth investigating and if fixing fixtures is too taxing migrating to factories will make easier
0
45,793
13,133,247,902
IssuesEvent
2020-08-06 20:31:48
shaundmorris/ddf
https://api.github.com/repos/shaundmorris/ddf
closed
CVE-2016-1000343 High Severity Vulnerability detected by WhiteSource
security vulnerability wontfix
## CVE-2016-1000343 - High Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.54.jar</b></p></summary> <p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p> <p>path: /root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar</p> <p> <p>Library home page: <a href=http://www.bouncycastle.org/java.html>http://www.bouncycastle.org/java.html</a></p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **bcprov-jdk15on-1.54.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Bouncy Castle JCE Provider version 1.55 and earlier the DSA key pair generator generates a weak private key if used with default values. If the JCA key pair generator is not explicitly initialised with DSA parameters, 1.55 and earlier generates a private value assuming a 1024 bit key size. In earlier releases this can be dealt with by explicitly passing parameters to the key pair generator. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-1000343>CVE-2016-1000343</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/bcgit/bc-java/commit/50a53068c094d6cff37659da33c9b4505becd389#diff-5578e61500abb2b87b300d3114bdfd7d">https://github.com/bcgit/bc-java/commit/50a53068c094d6cff37659da33c9b4505becd389#diff-5578e61500abb2b87b300d3114bdfd7d</a></p> <p>Release Date: 2016-11-03</p> <p>Fix Resolution: Replace or update the following files: KeyPairGeneratorSpi.java, DSATest.java</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-1000343 High Severity Vulnerability detected by WhiteSource - ## CVE-2016-1000343 - High Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.54.jar</b></p></summary> <p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p> <p>path: /root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar</p> <p> <p>Library home page: <a href=http://www.bouncycastle.org/java.html>http://www.bouncycastle.org/java.html</a></p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **bcprov-jdk15on-1.54.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Bouncy Castle JCE Provider version 1.55 and earlier the DSA key pair generator generates a weak private key if used with default values. If the JCA key pair generator is not explicitly initialised with DSA parameters, 1.55 and earlier generates a private value assuming a 1024 bit key size. In earlier releases this can be dealt with by explicitly passing parameters to the key pair generator. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-1000343>CVE-2016-1000343</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/bcgit/bc-java/commit/50a53068c094d6cff37659da33c9b4505becd389#diff-5578e61500abb2b87b300d3114bdfd7d">https://github.com/bcgit/bc-java/commit/50a53068c094d6cff37659da33c9b4505becd389#diff-5578e61500abb2b87b300d3114bdfd7d</a></p> <p>Release Date: 2016-11-03</p> <p>Fix Resolution: Replace or update the following files: KeyPairGeneratorSpi.java, DSATest.java</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high severity vulnerability detected by whitesource cve high severity vulnerability vulnerable library bcprov jar the bouncy castle crypto package is a java implementation of cryptographic algorithms this jar contains jce provider and lightweight api for the bouncy castle cryptography apis for jdk to jdk path root repository org bouncycastle bcprov bcprov jar root repository org bouncycastle bcprov bcprov jar root repository org bouncycastle bcprov bcprov jar root repository org bouncycastle bcprov bcprov jar root repository org bouncycastle bcprov bcprov jar library home page a href dependency hierarchy tika parsers jar root library x bcprov jar vulnerable library vulnerability details in the bouncy castle jce provider version and earlier the dsa key pair generator generates a weak private key if used with default values if the jca key pair generator is not explicitly initialised with dsa parameters and earlier generates a private value assuming a bit key size in earlier releases this can be dealt with by explicitly passing parameters to the key pair generator publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following files keypairgeneratorspi java dsatest java step up your open source security game with whitesource
0
507,406
14,679,979,421
IssuesEvent
2020-12-31 08:40:45
k8smeetup/website-tasks
https://api.github.com/repos/k8smeetup/website-tasks
opened
/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md
lang/zh priority/P0 sync/update version/master welcome
Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md](https://github.com/kubernetes/website/blob/master/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md) Diff 命令参考: ```bash # 查看原始文档与翻译文档更新差异 git diff --no-index -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md content/zh/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md # 跨分支持查看原始文档更新差异 git diff release-1.19 master -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md ```
1.0
/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md - Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md](https://github.com/kubernetes/website/blob/master/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md) Diff 命令参考: ```bash # 查看原始文档与翻译文档更新差异 git diff --no-index -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md content/zh/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md # 跨分支持查看原始文档更新差异 git diff release-1.19 master -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md ```
non_process
docs reference setup tools kubeadm generated kubeadm alpha selfhosting md source file diff 命令参考 bash 查看原始文档与翻译文档更新差异 git diff no index content en docs reference setup tools kubeadm generated kubeadm alpha selfhosting md content zh docs reference setup tools kubeadm generated kubeadm alpha selfhosting md 跨分支持查看原始文档更新差异 git diff release master content en docs reference setup tools kubeadm generated kubeadm alpha selfhosting md
0
57,785
3,083,786,505
IssuesEvent
2015-08-24 11:18:54
MinetestForFun/minetest-minetestforfun-server
https://api.github.com/repos/MinetestForFun/minetest-minetestforfun-server
opened
Some Maize node are not animated
Modding ➤ BugFix Priority: Low
![capture du 2015-08-24 13 11 04](https://cloud.githubusercontent.com/assets/12512762/9438895/699b2d62-4a62-11e5-9327-6a7c76573080.png) Very small maize is animated, and the biggest one too. But not some intermediaries. Very important. Or not. But it should be investigated once.
1.0
Some Maize node are not animated - ![capture du 2015-08-24 13 11 04](https://cloud.githubusercontent.com/assets/12512762/9438895/699b2d62-4a62-11e5-9327-6a7c76573080.png) Very small maize is animated, and the biggest one too. But not some intermediaries. Very important. Or not. But it should be investigated once.
non_process
some maize node are not animated very small maize is animated and the biggest one too but not some intermediaries very important or not but it should be investigated once
0
3,178
6,254,039,060
IssuesEvent
2017-07-14 00:17:20
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
opened
UAPAOT: System.Diagnostics.Tests.ProcessStartInfoTests/StartInfo_NotepadWithContent(useShellExecute: False)
area-System.Diagnostics.Process os-windows-uwp test-run-uwp-coreclr
Failed in Outerloop during validation for #22237: https://mc.dot.net/#/user/CIPop/pr~2Fjenkins~2Fdotnet~2Fcorefx~2Fmaster~2F/test~2Ffunctional~2Fcli~2F/343581b48553b789c3a92222e109b9d4b07927c1/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessStartInfoTests~2FStartInfo_NotepadWithContent(useShellExecute:%20False) ``` Windows.10.Amd64.ClientRS2.Open-Debug-x64 Unhandled Exception of Type System.InvalidOperationException Message : System.InvalidOperationException : WaitForInputIdle failed. This could be because the process does not have a graphical interface. Stack Trace : at System.Diagnostics.Process.WaitForInputIdleCore(Int32 milliseconds) at System.Diagnostics.Process.WaitForInputIdle(Int32 milliseconds) at System.Diagnostics.Process.WaitForInputIdle() at System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_NotepadWithContent(Boolean useShellExecute) ```
1.0
UAPAOT: System.Diagnostics.Tests.ProcessStartInfoTests/StartInfo_NotepadWithContent(useShellExecute: False) - Failed in Outerloop during validation for #22237: https://mc.dot.net/#/user/CIPop/pr~2Fjenkins~2Fdotnet~2Fcorefx~2Fmaster~2F/test~2Ffunctional~2Fcli~2F/343581b48553b789c3a92222e109b9d4b07927c1/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessStartInfoTests~2FStartInfo_NotepadWithContent(useShellExecute:%20False) ``` Windows.10.Amd64.ClientRS2.Open-Debug-x64 Unhandled Exception of Type System.InvalidOperationException Message : System.InvalidOperationException : WaitForInputIdle failed. This could be because the process does not have a graphical interface. Stack Trace : at System.Diagnostics.Process.WaitForInputIdleCore(Int32 milliseconds) at System.Diagnostics.Process.WaitForInputIdle(Int32 milliseconds) at System.Diagnostics.Process.WaitForInputIdle() at System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_NotepadWithContent(Boolean useShellExecute) ```
process
uapaot system diagnostics tests processstartinfotests startinfo notepadwithcontent useshellexecute false failed in outerloop during validation for windows open debug unhandled exception of type system invalidoperationexception message system invalidoperationexception waitforinputidle failed this could be because the process does not have a graphical interface stack trace at system diagnostics process waitforinputidlecore milliseconds at system diagnostics process waitforinputidle milliseconds at system diagnostics process waitforinputidle at system diagnostics tests processstartinfotests startinfo notepadwithcontent boolean useshellexecute
1
19,676
26,031,724,271
IssuesEvent
2022-12-21 22:06:36
biocodellc/localcontexts_db
https://api.github.com/repos/biocodellc/localcontexts_db
closed
Registration: register button disable on submit
registration process
Disable button on registration page on submit while the page loads so that duplicate user profiles are not created.
1.0
Registration: register button disable on submit - Disable button on registration page on submit while the page loads so that duplicate user profiles are not created.
process
registration register button disable on submit disable button on registration page on submit while the page loads so that duplicate user profiles are not created
1
85,336
3,689,595,090
IssuesEvent
2016-02-25 16:57:22
icatproject/topcat
https://api.github.com/repos/icatproject/topcat
closed
Availability wording needs improving
Priority-Low TopCAT v2
The use of ONLINE, ARCHIVED etc may not be obvious to users. It is described in the help but many users will not use that. These values should probably be configurable with maybe a description given as a tooltip.
1.0
Availability wording needs improving - The use of ONLINE, ARCHIVED etc may not be obvious to users. It is described in the help but many users will not use that. These values should probably be configurable with maybe a description given as a tooltip.
non_process
availability wording needs improving the use of online archived etc may not be obvious to users it is described in the help but many users will not use that these values should probably be configurable with maybe a description given as a tooltip
0
110,229
16,977,084,746
IssuesEvent
2021-06-30 01:42:13
SmartBear/readyapi-swagger-assertion-plugin
https://api.github.com/repos/SmartBear/readyapi-swagger-assertion-plugin
closed
CVE-2014-0114 (High) detected in commons-beanutils-1.8.0.jar - autoclosed
security vulnerability
## CVE-2014-0114 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.8.0.jar</b></p></summary> <p>BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p>Path to dependency file: readyapi-swagger-assertion-plugin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.0/commons-beanutils-1.8.0.jar</p> <p> Dependency Hierarchy: - ready-api-soapui-pro-1.7.0.jar (Root Library) - jasperreports-4.0.1.jar - :x: **commons-beanutils-1.8.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/SmartBear/readyapi-swagger-assertion-plugin/commit/dc701947b2244c9b5dff8c9a75808fab2d107d60">dc701947b2244c9b5dff8c9a75808fab2d107d60</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1. <p>Publish Date: 2014-04-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114>CVE-2014-0114</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p> <p>Release Date: 2014-04-30</p> <p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-beanutils","packageName":"commons-beanutils","packageVersion":"1.8.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;net.sf.jasperreports:jasperreports:4.0.1;commons-beanutils:commons-beanutils:1.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2014-0114","vulnerabilityDetails":"Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to \"manipulate\" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> -->
True
CVE-2014-0114 (High) detected in commons-beanutils-1.8.0.jar - autoclosed - ## CVE-2014-0114 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.8.0.jar</b></p></summary> <p>BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p>Path to dependency file: readyapi-swagger-assertion-plugin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.0/commons-beanutils-1.8.0.jar</p> <p> Dependency Hierarchy: - ready-api-soapui-pro-1.7.0.jar (Root Library) - jasperreports-4.0.1.jar - :x: **commons-beanutils-1.8.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/SmartBear/readyapi-swagger-assertion-plugin/commit/dc701947b2244c9b5dff8c9a75808fab2d107d60">dc701947b2244c9b5dff8c9a75808fab2d107d60</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1. <p>Publish Date: 2014-04-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114>CVE-2014-0114</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p> <p>Release Date: 2014-04-30</p> <p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-beanutils","packageName":"commons-beanutils","packageVersion":"1.8.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;net.sf.jasperreports:jasperreports:4.0.1;commons-beanutils:commons-beanutils:1.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2014-0114","vulnerabilityDetails":"Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to \"manipulate\" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> -->
non_process
cve high detected in commons beanutils jar autoclosed cve high severity vulnerability vulnerable library commons beanutils jar beanutils provides an easy to use but flexible wrapper around reflection and introspection path to dependency file readyapi swagger assertion plugin pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy ready api soapui pro jar root library jasperreports jar x commons beanutils jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons beanutils as distributed in lib commons beanutils jar in apache struts x through and in other products requiring commons beanutils through does not suppress the class property which allows remote attackers to manipulate the classloader and execute arbitrary code via the class parameter as demonstrated by the passing of this parameter to the getclass method of the actionform object in struts publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution commons beanutils commons beanutils org apache struts core isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com smartbear ready api soapui pro net sf jasperreports jasperreports commons beanutils commons beanutils isminimumfixversionavailable true minimumfixversion commons beanutils commons beanutils org apache struts core basebranches vulnerabilityidentifier cve vulnerabilitydetails apache commons beanutils as distributed in lib commons beanutils jar in apache struts x through and in other products requiring commons beanutils through does not suppress the class property which allows remote attackers to manipulate the classloader and execute arbitrary code via the class parameter as demonstrated by the passing of this parameter to the getclass method of the actionform object in struts vulnerabilityurl
0
15,305
19,347,233,234
IssuesEvent
2021-12-15 12:09:32
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
opened
Explicitly **not** using any fx variable is not possible in preprocessor
preprocessor
Due to the implementation of the check for the `fx_variables` option in the preprocessors, i.e. basically something like ```py if not step_settings.get('fx_variables'): # add default fx_vars else: # use fx_variables that have been specified ``` it is currently not possible to explicitly **not** use any fx variable. Empty dictionaries, lists, or `None` will all trigger the first `if`, and the default values are automatically added. <details> https://github.com/ESMValGroup/ESMValCore/blob/5131ffc4cdf5aa5c46215aea659740ea442769dc/esmvalcore/_recipe.py#L506-L551 (specifically line 515) </details> I think by explicitly stating something like ```yml area_statistics: operator: mean fx_variables: {} # or fx_variables: [] or fx_variables: null ``` **no** fx variables should be used. Only if `fx_variables` is not used at all, the default values should be used. Any opinions on this @ESMValGroup/esmvaltool-coreteam ?
1.0
Explicitly **not** using any fx variable is not possible in preprocessor - Due to the implementation of the check for the `fx_variables` option in the preprocessors, i.e. basically something like ```py if not step_settings.get('fx_variables'): # add default fx_vars else: # use fx_variables that have been specified ``` it is currently not possible to explicitly **not** use any fx variable. Empty dictionaries, lists, or `None` will all trigger the first `if`, and the default values are automatically added. <details> https://github.com/ESMValGroup/ESMValCore/blob/5131ffc4cdf5aa5c46215aea659740ea442769dc/esmvalcore/_recipe.py#L506-L551 (specifically line 515) </details> I think by explicitly stating something like ```yml area_statistics: operator: mean fx_variables: {} # or fx_variables: [] or fx_variables: null ``` **no** fx variables should be used. Only if `fx_variables` is not used at all, the default values should be used. Any opinions on this @ESMValGroup/esmvaltool-coreteam ?
process
explicitly not using any fx variable is not possible in preprocessor due to the implementation of the check for the fx variables option in the preprocessors i e basically something like py if not step settings get fx variables add default fx vars else use fx variables that have been specified it is currently not possible to explicitly not use any fx variable empty dictionaries lists or none will all trigger the first if and the default values are automatically added specifically line i think by explicitly stating something like yml area statistics operator mean fx variables or fx variables or fx variables null no fx variables should be used only if fx variables is not used at all the default values should be used any opinions on this esmvalgroup esmvaltool coreteam
1
168
2,586,801,090
IssuesEvent
2015-02-17 14:39:51
tinkerpop/tinkerpop3
https://api.github.com/repos/tinkerpop/tinkerpop3
closed
We need a couple more Scope.local steps.
enhancement process
We have `order(local)`. But we also need: * `dedup(local)` : if the current object is a collection, dedup it. * `sample(local)`: if the current object is a collection, sample it. I think thats it. Look at `OrderLocalStep` and `OrderGlobalStep` to see why its important that they are different classes --- one is a barrier, the other is not.
1.0
We need a couple more Scope.local steps. - We have `order(local)`. But we also need: * `dedup(local)` : if the current object is a collection, dedup it. * `sample(local)`: if the current object is a collection, sample it. I think thats it. Look at `OrderLocalStep` and `OrderGlobalStep` to see why its important that they are different classes --- one is a barrier, the other is not.
process
we need a couple more scope local steps we have order local but we also need dedup local if the current object is a collection dedup it sample local if the current object is a collection sample it i think thats it look at orderlocalstep and orderglobalstep to see why its important that they are different classes one is a barrier the other is not
1
10,354
13,179,289,990
IssuesEvent
2020-08-12 10:38:44
shirou/gopsutil
https://api.github.com/repos/shirou/gopsutil
closed
func Processes() ([]*Process, error) can pass parameters
package:process
func Processes() ([]*Process, error) This function, whether it is possible to pass parameters, once called, will collect all the process information of the system, which consumes more CPU, if you can selectively collect process information, it will be more it is good. Hope to adopt **Additional context** [Add any other context or screenshots about the feature request here.]
1.0
func Processes() ([]*Process, error) can pass parameters - func Processes() ([]*Process, error) This function, whether it is possible to pass parameters, once called, will collect all the process information of the system, which consumes more CPU, if you can selectively collect process information, it will be more it is good. Hope to adopt **Additional context** [Add any other context or screenshots about the feature request here.]
process
func processes process error can pass parameters func processes process error this function whether it is possible to pass parameters once called will collect all the process information of the system which consumes more cpu if you can selectively collect process information it will be more it is good hope to adopt additional context
1
21,382
29,202,230,096
IssuesEvent
2023-05-21 00:37:14
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Hibrido / ] Fullstack Developer (Sênior - Híbrido - Belo Horizonte) na Coodesh
SALVADOR PJ JAVA MYSQL JAVASCRIPT FULL-STACK SENIOR PRIMEFACES JSF SPRING SQL GIT HIBERNATE MAVEN REST SOAP JSON ANGULAR REQUISITOS NGINX PROCESSOS INOVAÇÃO GITHUB APACHE UMA C DOCUMENTAÇÃO WILDFLY HTTP MANUTENÇÃO HIBRIDO ALOCADO Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/fullstack-developer-senior-hibrido-belo-horizonte-200652198?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Prime Results </strong>está buscando <strong><ins>Fullstack Developer</ins></strong> para compor seu time!</p> <p>Acreditamos no poder de transformação social realizado pelas empresas. Acreditamos no poder transformador das pessoas, aliado à gestão e tecnologia. Compartilhamos nosso conhecimento para solucionar problemas complexos e gerar valor para nossos clientes.</p> <p><strong>Responsabilidades:</strong></p> <ul> <li>Desenvolvimento/implementação e manutenção de aplicações;</li> <li>Participar da análise e execução dos projetos e execução dos tickets;</li> <li>Definir as atividades necessárias para a realização de projetos, analisando os impactos em sistemas e processos através do entendimento da necessidade, conhecimento técnico e arquitetônico dos sistemas;</li> <li>Desenvolver códigos para atendimento às áreas e empresas clientes, proporcionando o esclarecimento de dúvidas relacionados ao projeto, contribuindo para uma melhor análise de impactos de processos e sistemas sob sua responsabilidade;&nbsp;</li> <li>Participar das atividades de planejamento para a liberação do produto para homologação e produção, por meio da validação de testes de aceite, assim como documentação de não conformidades avaliando e planejando a execução das correções reportadas;</li> <li>Participar da rotina de SQUADs.&nbsp;</li> </ul> <p></p> ## Prime Results : <p>O Best Seller Simon Sinek, diz que a maioria das empresas sabem o que fazem, porém não sabem por que o fazem. Não é o nosso caso. A Prime Results é uma empresa especializada em gestão organizacional que usa seu potencial de transformação em empresas que geram impacto positivo na sociedade. Nossos clientes hoje, fazem a diferença na vida de mais de 250.000 brasileiros, nas áreas de proteção patrimonial, saúde e assistência 24 horas.&nbsp;</p> <p>Nosso objetivo central é criar um ambiente criativo, dinâmico e engajado, sempre aliados a métodos, processos inteligentes e muita inovação.</p><a href='https://coodesh.com/empresas/prime-results'>Veja mais no site</a> ## Habilidades: - Java - Hibernate - Angular - Javascript - JSON - Apache - MySQL - Microsoft SQL Server - Spring ## Local: undefined ## Requisitos: - Experiência em Java: JSF, Spring, PrimeFaces, Hibernate, JasperReports; - Conhecimentos em modelagem e desenvolvimento de Bancos de Dados relacionais: MySQL, SQL Server; - Conhecimentos em Tecnologias Web: HTML5, CSS e Frameworks JavaScript, Angular; - Conhecimento de Arquiteturas Web e Serviços (HTTP, SOAP, REST ou JSON); - Conhecimentos nas ferramentas: GIT e Maven; - Conhecimentos técnicos em servidores de aplicação (Wildfly - J2EE), servidores web (Apache e NGINX) e Spring Boot. ## Benefícios: - GymPass; - Assistência Médica após o período de experiência. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (Sênior - Híbrido - Belo Horizonte) na Prime Results ](https://coodesh.com/vagas/fullstack-developer-senior-hibrido-belo-horizonte-200652198?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime PJ #### Categoria Full-Stack
1.0
[Hibrido / ] Fullstack Developer (Sênior - Híbrido - Belo Horizonte) na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/fullstack-developer-senior-hibrido-belo-horizonte-200652198?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Prime Results </strong>está buscando <strong><ins>Fullstack Developer</ins></strong> para compor seu time!</p> <p>Acreditamos no poder de transformação social realizado pelas empresas. Acreditamos no poder transformador das pessoas, aliado à gestão e tecnologia. Compartilhamos nosso conhecimento para solucionar problemas complexos e gerar valor para nossos clientes.</p> <p><strong>Responsabilidades:</strong></p> <ul> <li>Desenvolvimento/implementação e manutenção de aplicações;</li> <li>Participar da análise e execução dos projetos e execução dos tickets;</li> <li>Definir as atividades necessárias para a realização de projetos, analisando os impactos em sistemas e processos através do entendimento da necessidade, conhecimento técnico e arquitetônico dos sistemas;</li> <li>Desenvolver códigos para atendimento às áreas e empresas clientes, proporcionando o esclarecimento de dúvidas relacionados ao projeto, contribuindo para uma melhor análise de impactos de processos e sistemas sob sua responsabilidade;&nbsp;</li> <li>Participar das atividades de planejamento para a liberação do produto para homologação e produção, por meio da validação de testes de aceite, assim como documentação de não conformidades avaliando e planejando a execução das correções reportadas;</li> <li>Participar da rotina de SQUADs.&nbsp;</li> </ul> <p></p> ## Prime Results : <p>O Best Seller Simon Sinek, diz que a maioria das empresas sabem o que fazem, porém não sabem por que o fazem. Não é o nosso caso. A Prime Results é uma empresa especializada em gestão organizacional que usa seu potencial de transformação em empresas que geram impacto positivo na sociedade. Nossos clientes hoje, fazem a diferença na vida de mais de 250.000 brasileiros, nas áreas de proteção patrimonial, saúde e assistência 24 horas.&nbsp;</p> <p>Nosso objetivo central é criar um ambiente criativo, dinâmico e engajado, sempre aliados a métodos, processos inteligentes e muita inovação.</p><a href='https://coodesh.com/empresas/prime-results'>Veja mais no site</a> ## Habilidades: - Java - Hibernate - Angular - Javascript - JSON - Apache - MySQL - Microsoft SQL Server - Spring ## Local: undefined ## Requisitos: - Experiência em Java: JSF, Spring, PrimeFaces, Hibernate, JasperReports; - Conhecimentos em modelagem e desenvolvimento de Bancos de Dados relacionais: MySQL, SQL Server; - Conhecimentos em Tecnologias Web: HTML5, CSS e Frameworks JavaScript, Angular; - Conhecimento de Arquiteturas Web e Serviços (HTTP, SOAP, REST ou JSON); - Conhecimentos nas ferramentas: GIT e Maven; - Conhecimentos técnicos em servidores de aplicação (Wildfly - J2EE), servidores web (Apache e NGINX) e Spring Boot. ## Benefícios: - GymPass; - Assistência Médica após o período de experiência. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (Sênior - Híbrido - Belo Horizonte) na Prime Results ](https://coodesh.com/vagas/fullstack-developer-senior-hibrido-belo-horizonte-200652198?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime PJ #### Categoria Full-Stack
process
fullstack developer sênior híbrido belo horizonte na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a prime results está buscando fullstack developer para compor seu time acreditamos no poder de transformação social realizado pelas empresas acreditamos no poder transformador das pessoas aliado à gestão e tecnologia compartilhamos nosso conhecimento para solucionar problemas complexos e gerar valor para nossos clientes responsabilidades desenvolvimento implementação e manutenção de aplicações participar da análise e execução dos projetos e execução dos tickets definir as atividades necessárias para a realização de projetos analisando os impactos em sistemas e processos através do entendimento da necessidade conhecimento técnico e arquitetônico dos sistemas desenvolver códigos para atendimento às áreas e empresas clientes proporcionando o esclarecimento de dúvidas relacionados ao projeto contribuindo para uma melhor análise de impactos de processos e sistemas sob sua responsabilidade nbsp participar das atividades de planejamento para a liberação do produto para homologação e produção por meio da validação de testes de aceite assim como documentação de não conformidades avaliando e planejando a execução das correções reportadas participar da rotina de squads nbsp prime results o best seller simon sinek diz que a maioria das empresas sabem o que fazem porém não sabem por que o fazem não é o nosso caso a prime results é uma empresa especializada em gestão organizacional que usa seu potencial de transformação em empresas que geram impacto positivo na sociedade nossos clientes hoje fazem a diferença na vida de mais de brasileiros nas áreas de proteção patrimonial saúde e assistência horas nbsp nosso objetivo central é criar um ambiente criativo dinâmico e engajado sempre aliados a métodos processos inteligentes e muita inovação habilidades java hibernate angular javascript json apache mysql microsoft sql server spring local undefined requisitos experiência em java jsf spring primefaces hibernate jasperreports conhecimentos em modelagem e desenvolvimento de bancos de dados relacionais mysql sql server conhecimentos em tecnologias web css e frameworks javascript angular conhecimento de arquiteturas web e serviços http soap rest ou json conhecimentos nas ferramentas git e maven conhecimentos técnicos em servidores de aplicação wildfly servidores web apache e nginx e spring boot benefícios gympass assistência médica após o período de experiência como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado regime pj categoria full stack
1
116,747
24,983,260,346
IssuesEvent
2022-11-02 13:22:48
arduino/arduino-cli
https://api.github.com/repos/arduino/arduino-cli
opened
Previous custom discovery tool process left running after update
topic: code type: imperfection
### Describe the problem In addition to the always available [**serial-discovery**](https://github.com/arduino/serial-discovery) and [**mdns-discovery**](https://github.com/arduino/mdns-discovery) [pluggable discovery](https://arduino.github.io/arduino-cli/dev/pluggable-discovery-specification/) tools of the `builtin` package, dependencies on custom discovery tools [can be specified](https://arduino.github.io/arduino-cli/dev/package_index_json-specification/#tools-definitions:~:text=discoveryDependencies%3A%20the%20Pluggable%20Discoveries%20needed%20by%20this%20platform) by boards platforms. [The latest version of the custom discovery dependency is installed](https://arduino.github.io/arduino-cli/dev/package_index_json-specification/#tools-definitions:~:text=the%20latest%20installed%20discovery%20tool%20will%20always%20be%20used) when installing or updating that platform. 🐛 The process for the previous version of a custom discovery tool is left running after a boards platform update that updates the discovery tool. ### To reproduce #### Setup ```text $ arduino-cli version arduino-cli Version: git-snapshot Commit: 732bdc54 Date: 2022-09-07T09:00:01Z $ export ARDUINO_DIRECTORIES_DATA="/tmp/arduino-cli-directories/data" $ export ARDUINO_BOARD_MANAGER_ADDITIONAL_URLS="https://www.pjrc.com/teensy/package_teensy_index.json" $ arduino-cli core update-index $ arduino-cli core install teensy:avr@1.57.0 # Install outdated version of a platform with custom pluggable discovery tool dependency [...] Installing teensy:teensy-discovery@1.57.1... teensy:teensy-discovery@1.57.1 installed [...] $ mv "${ARDUINO_DIRECTORIES_DATA}/packages/teensy/tools/teensy-discovery/1.57.1" "${ARDUINO_DIRECTORIES_DATA}/packages/teensy/tools/teensy-discovery/1.57.0" # Simulate outdated discovery tool $ arduino-cli daemon --debug ``` #### Demo run the following [`grpcurl`](https://github.com/fullstorydev/grpcurl) commands in another terminal: ```text $ grpcurl \ -plaintext \ -import-path ./rpc \ -proto cc/arduino/cli/commands/v1/commands.proto \ 127.0.0.1:50051 \ cc.arduino.cli.commands.v1.ArduinoCoreService.Create { "instance": { "id": 1 } } $ grpcurl \ -plaintext \ -import-path ./rpc \ -proto cc/arduino/cli/commands/v1/commands.proto \ -d '{"instance": {"id": 1}}' \ 127.0.0.1:50051 \ cc.arduino.cli.commands.v1.ArduinoCoreService.Init $ pgrep -a -f 'teensy-discovery' $ grpcurl \ -plaintext \ -import-path ./rpc \ -proto cc/arduino/cli/commands/v1/commands.proto \ -d '{"instance": {"id": 1}}' \ 127.0.0.1:50051 \ cc.arduino.cli.commands.v1.ArduinoCoreService.BoardListWatch [...] $ pgrep -a -f 'teensy-discovery' 142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery $ grpcurl \ -plaintext \ -import-path ./rpc \ -proto cc/arduino/cli/commands/v1/commands.proto \ -d '{"instance": {"id": 1}, "platform_package": "teensy", "architecture": "avr"}' \ 127.0.0.1:50051 \ cc.arduino.cli.commands.v1.ArduinoCoreService.PlatformUpgrade [...] { "taskProgress": { "name": "Installing teensy:teensy-discovery@1.57.1" } } [...] $ pgrep -a -f 'teensy-discovery' 142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery $ grpcurl \ -plaintext \ -import-path ./rpc \ -proto cc/arduino/cli/commands/v1/commands.proto \ -d '{"instance": {"id": 1}}' \ 127.0.0.1:50051 \ cc.arduino.cli.commands.v1.ArduinoCoreService.BoardListWatch [...] $ pgrep -a -f 'teensy-discovery' 142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery 142363 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.1/teensy-discovery ``` 🐛 There is a leftover `teensy:teensy-discovery@1.57.0` process still running in addition to `teensy:teensy-discovery@1.57.1`. ### Expected behavior Discovery tool processes are cleaned up. ### Arduino CLI version 732bdc54 ### Operating system Windows, Ubuntu ### Operating system version Windows 10, Ubuntu 20.04 ### Additional context Originally reported by @PaulStoffregen at https://forum.pjrc.com/threads/71370-Teensy4-1-reading-from-Serial-Monitor-of-Arduino-IDE?p=315184#post315184 --- I can also reproduce the issue using the equivalent command line interface commands. I chose to use the gRPC interface for the demo because the command line version seems unlikely to occur in real world usage, while the gRPC version is a simplification of an existing real world occurrence in Arduino IDE 2.x. --- I suspected a similar problem might apply to the pluggable monitor tools (e.g., **teensy-monitor**). I did a quick check using Arduino IDE 2.x and found that only the previously running **teensy-monitor** process was running after the update. After I closed and reopened Serial Monitor, only a single **teensy-monitor** process for the updated version was running. ### Issue checklist - [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-cli/issues?q=) - [X] I verified the problem still occurs when using the [nightly build](https://arduino.github.io/arduino-cli/dev/installation/#nightly-builds) - [X] My report contains all necessary details
1.0
Previous custom discovery tool process left running after update - ### Describe the problem In addition to the always available [**serial-discovery**](https://github.com/arduino/serial-discovery) and [**mdns-discovery**](https://github.com/arduino/mdns-discovery) [pluggable discovery](https://arduino.github.io/arduino-cli/dev/pluggable-discovery-specification/) tools of the `builtin` package, dependencies on custom discovery tools [can be specified](https://arduino.github.io/arduino-cli/dev/package_index_json-specification/#tools-definitions:~:text=discoveryDependencies%3A%20the%20Pluggable%20Discoveries%20needed%20by%20this%20platform) by boards platforms. [The latest version of the custom discovery dependency is installed](https://arduino.github.io/arduino-cli/dev/package_index_json-specification/#tools-definitions:~:text=the%20latest%20installed%20discovery%20tool%20will%20always%20be%20used) when installing or updating that platform. 🐛 The process for the previous version of a custom discovery tool is left running after a boards platform update that updates the discovery tool. ### To reproduce #### Setup ```text $ arduino-cli version arduino-cli Version: git-snapshot Commit: 732bdc54 Date: 2022-09-07T09:00:01Z $ export ARDUINO_DIRECTORIES_DATA="/tmp/arduino-cli-directories/data" $ export ARDUINO_BOARD_MANAGER_ADDITIONAL_URLS="https://www.pjrc.com/teensy/package_teensy_index.json" $ arduino-cli core update-index $ arduino-cli core install teensy:avr@1.57.0 # Install outdated version of a platform with custom pluggable discovery tool dependency [...] Installing teensy:teensy-discovery@1.57.1... teensy:teensy-discovery@1.57.1 installed [...] $ mv "${ARDUINO_DIRECTORIES_DATA}/packages/teensy/tools/teensy-discovery/1.57.1" "${ARDUINO_DIRECTORIES_DATA}/packages/teensy/tools/teensy-discovery/1.57.0" # Simulate outdated discovery tool $ arduino-cli daemon --debug ``` #### Demo run the following [`grpcurl`](https://github.com/fullstorydev/grpcurl) commands in another terminal: ```text $ grpcurl \ -plaintext \ -import-path ./rpc \ -proto cc/arduino/cli/commands/v1/commands.proto \ 127.0.0.1:50051 \ cc.arduino.cli.commands.v1.ArduinoCoreService.Create { "instance": { "id": 1 } } $ grpcurl \ -plaintext \ -import-path ./rpc \ -proto cc/arduino/cli/commands/v1/commands.proto \ -d '{"instance": {"id": 1}}' \ 127.0.0.1:50051 \ cc.arduino.cli.commands.v1.ArduinoCoreService.Init $ pgrep -a -f 'teensy-discovery' $ grpcurl \ -plaintext \ -import-path ./rpc \ -proto cc/arduino/cli/commands/v1/commands.proto \ -d '{"instance": {"id": 1}}' \ 127.0.0.1:50051 \ cc.arduino.cli.commands.v1.ArduinoCoreService.BoardListWatch [...] $ pgrep -a -f 'teensy-discovery' 142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery $ grpcurl \ -plaintext \ -import-path ./rpc \ -proto cc/arduino/cli/commands/v1/commands.proto \ -d '{"instance": {"id": 1}, "platform_package": "teensy", "architecture": "avr"}' \ 127.0.0.1:50051 \ cc.arduino.cli.commands.v1.ArduinoCoreService.PlatformUpgrade [...] { "taskProgress": { "name": "Installing teensy:teensy-discovery@1.57.1" } } [...] $ pgrep -a -f 'teensy-discovery' 142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery $ grpcurl \ -plaintext \ -import-path ./rpc \ -proto cc/arduino/cli/commands/v1/commands.proto \ -d '{"instance": {"id": 1}}' \ 127.0.0.1:50051 \ cc.arduino.cli.commands.v1.ArduinoCoreService.BoardListWatch [...] $ pgrep -a -f 'teensy-discovery' 142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery 142363 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.1/teensy-discovery ``` 🐛 There is a leftover `teensy:teensy-discovery@1.57.0` process still running in addition to `teensy:teensy-discovery@1.57.1`. ### Expected behavior Discovery tool processes are cleaned up. ### Arduino CLI version 732bdc54 ### Operating system Windows, Ubuntu ### Operating system version Windows 10, Ubuntu 20.04 ### Additional context Originally reported by @PaulStoffregen at https://forum.pjrc.com/threads/71370-Teensy4-1-reading-from-Serial-Monitor-of-Arduino-IDE?p=315184#post315184 --- I can also reproduce the issue using the equivalent command line interface commands. I chose to use the gRPC interface for the demo because the command line version seems unlikely to occur in real world usage, while the gRPC version is a simplification of an existing real world occurrence in Arduino IDE 2.x. --- I suspected a similar problem might apply to the pluggable monitor tools (e.g., **teensy-monitor**). I did a quick check using Arduino IDE 2.x and found that only the previously running **teensy-monitor** process was running after the update. After I closed and reopened Serial Monitor, only a single **teensy-monitor** process for the updated version was running. ### Issue checklist - [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-cli/issues?q=) - [X] I verified the problem still occurs when using the [nightly build](https://arduino.github.io/arduino-cli/dev/installation/#nightly-builds) - [X] My report contains all necessary details
non_process
previous custom discovery tool process left running after update describe the problem in addition to the always available and tools of the builtin package dependencies on custom discovery tools by boards platforms when installing or updating that platform 🐛 the process for the previous version of a custom discovery tool is left running after a boards platform update that updates the discovery tool to reproduce setup text arduino cli version arduino cli version git snapshot commit date export arduino directories data tmp arduino cli directories data export arduino board manager additional urls arduino cli core update index arduino cli core install teensy avr install outdated version of a platform with custom pluggable discovery tool dependency installing teensy teensy discovery teensy teensy discovery installed mv arduino directories data packages teensy tools teensy discovery arduino directories data packages teensy tools teensy discovery simulate outdated discovery tool arduino cli daemon debug demo run the following commands in another terminal text grpcurl plaintext import path rpc proto cc arduino cli commands commands proto cc arduino cli commands arduinocoreservice create instance id grpcurl plaintext import path rpc proto cc arduino cli commands commands proto d instance id cc arduino cli commands arduinocoreservice init pgrep a f teensy discovery grpcurl plaintext import path rpc proto cc arduino cli commands commands proto d instance id cc arduino cli commands arduinocoreservice boardlistwatch pgrep a f teensy discovery tmp arduino cli directories data packages teensy tools teensy discovery teensy discovery grpcurl plaintext import path rpc proto cc arduino cli commands commands proto d instance id platform package teensy architecture avr cc arduino cli commands arduinocoreservice platformupgrade taskprogress name installing teensy teensy discovery pgrep a f teensy discovery tmp arduino cli directories data packages teensy tools teensy discovery teensy discovery grpcurl plaintext import path rpc proto cc arduino cli commands commands proto d instance id cc arduino cli commands arduinocoreservice boardlistwatch pgrep a f teensy discovery tmp arduino cli directories data packages teensy tools teensy discovery teensy discovery tmp arduino cli directories data packages teensy tools teensy discovery teensy discovery 🐛 there is a leftover teensy teensy discovery process still running in addition to teensy teensy discovery expected behavior discovery tool processes are cleaned up arduino cli version operating system windows ubuntu operating system version windows ubuntu additional context originally reported by paulstoffregen at i can also reproduce the issue using the equivalent command line interface commands i chose to use the grpc interface for the demo because the command line version seems unlikely to occur in real world usage while the grpc version is a simplification of an existing real world occurrence in arduino ide x i suspected a similar problem might apply to the pluggable monitor tools e g teensy monitor i did a quick check using arduino ide x and found that only the previously running teensy monitor process was running after the update after i closed and reopened serial monitor only a single teensy monitor process for the updated version was running issue checklist i searched for previous reports in i verified the problem still occurs when using the my report contains all necessary details
0
6,316
9,329,627,188
IssuesEvent
2019-03-28 03:11:50
brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
https://api.github.com/repos/brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
closed
Реализовать часть программы
C++ Work in process
# Написать часть программы. ### * В этой части должен содержаться хотя бы ввод исходной строки. Пример кода. `#include <iostream>` `int main ()` `{` `string ishodnik;` `cin >> ishodnik;` `return 0;` `}`
1.0
Реализовать часть программы - # Написать часть программы. ### * В этой части должен содержаться хотя бы ввод исходной строки. Пример кода. `#include <iostream>` `int main ()` `{` `string ishodnik;` `cin >> ishodnik;` `return 0;` `}`
process
реализовать часть программы написать часть программы в этой части должен содержаться хотя бы ввод исходной строки пример кода include int main string ishodnik cin ishodnik return
1
91,000
11,456,395,416
IssuesEvent
2020-02-06 21:08:07
tokio-rs/tracing
https://api.github.com/repos/tokio-rs/tracing
closed
trace: Determine new syntax for uninitialized fields
crate/tracing kind/rfc needs/design
Followup from #1062 and https://github.com/tokio-rs/tokio/pull/1103#issuecomment-500019388. Prior to #1103, the syntax `field,` was used by the `tokio-trace` macros to indicate span fields which have yet to have a value recorded. In order to support shorthand for using local variables as fields, we changed the meaning of this syntax, since local variable shorthand is likely to be a more common use-case (see #1103). This leaves uninitialized fields without an appropriate macro syntax. There have been some suggestions, of which I think the strongest thus far are `let field,` and `field = _,`. `let field` nicely mirrors the way let bindings with unset values are created in Rust; it does, however, seem somewhat inconsistent with the existing syntax, since other fields do not begin with `let`. On the other hand, `field = _` fits in with the existing field syntax, but @carllerche has pointed out that it uses the assignment syntax to indicate something which is _unassigned_, which seems confusing. We should try to agree on a new syntax for uninitialized fields, and re-enable support for them in the macros.
1.0
trace: Determine new syntax for uninitialized fields - Followup from #1062 and https://github.com/tokio-rs/tokio/pull/1103#issuecomment-500019388. Prior to #1103, the syntax `field,` was used by the `tokio-trace` macros to indicate span fields which have yet to have a value recorded. In order to support shorthand for using local variables as fields, we changed the meaning of this syntax, since local variable shorthand is likely to be a more common use-case (see #1103). This leaves uninitialized fields without an appropriate macro syntax. There have been some suggestions, of which I think the strongest thus far are `let field,` and `field = _,`. `let field` nicely mirrors the way let bindings with unset values are created in Rust; it does, however, seem somewhat inconsistent with the existing syntax, since other fields do not begin with `let`. On the other hand, `field = _` fits in with the existing field syntax, but @carllerche has pointed out that it uses the assignment syntax to indicate something which is _unassigned_, which seems confusing. We should try to agree on a new syntax for uninitialized fields, and re-enable support for them in the macros.
non_process
trace determine new syntax for uninitialized fields followup from and prior to the syntax field was used by the tokio trace macros to indicate span fields which have yet to have a value recorded in order to support shorthand for using local variables as fields we changed the meaning of this syntax since local variable shorthand is likely to be a more common use case see this leaves uninitialized fields without an appropriate macro syntax there have been some suggestions of which i think the strongest thus far are let field and field let field nicely mirrors the way let bindings with unset values are created in rust it does however seem somewhat inconsistent with the existing syntax since other fields do not begin with let on the other hand field fits in with the existing field syntax but carllerche has pointed out that it uses the assignment syntax to indicate something which is unassigned which seems confusing we should try to agree on a new syntax for uninitialized fields and re enable support for them in the macros
0
11,573
14,442,128,623
IssuesEvent
2020-12-07 17:42:50
Arch666Angel/mods
https://api.github.com/repos/Arch666Angel/mods
opened
Module tech icons wrong size
Angels Bio Processing Impact: Bug
**Describe the bug** Module tech icons have the wrong size **Screenshots** If applicable, add screenshots to help explain your problem. ![image](https://user-images.githubusercontent.com/26593477/101385334-f4152b80-38bb-11eb-8e88-91af94dc19b2.png) **Additional context** This is already fixed for the player crafting menu group, however the tech icons I overlooked: https://github.com/Arch666Angel/mods/blob/b45c76fe1eb650ea5acf3a1230fbdb2c356c2743/angelsbioprocessing/prototypes/overrides/bio-processing-override-bob.lua#L160-L163 https://github.com/Arch666Angel/mods/blob/b45c76fe1eb650ea5acf3a1230fbdb2c356c2743/angelsbioprocessing/prototypes/overrides/bio-processing-override-bob.lua#L226-L229
1.0
Module tech icons wrong size - **Describe the bug** Module tech icons have the wrong size **Screenshots** If applicable, add screenshots to help explain your problem. ![image](https://user-images.githubusercontent.com/26593477/101385334-f4152b80-38bb-11eb-8e88-91af94dc19b2.png) **Additional context** This is already fixed for the player crafting menu group, however the tech icons I overlooked: https://github.com/Arch666Angel/mods/blob/b45c76fe1eb650ea5acf3a1230fbdb2c356c2743/angelsbioprocessing/prototypes/overrides/bio-processing-override-bob.lua#L160-L163 https://github.com/Arch666Angel/mods/blob/b45c76fe1eb650ea5acf3a1230fbdb2c356c2743/angelsbioprocessing/prototypes/overrides/bio-processing-override-bob.lua#L226-L229
process
module tech icons wrong size describe the bug module tech icons have the wrong size screenshots if applicable add screenshots to help explain your problem additional context this is already fixed for the player crafting menu group however the tech icons i overlooked
1
13,418
23,061,384,452
IssuesEvent
2022-07-25 10:14:48
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
opened
npm package latest version not seen
type:bug status:requirements priority-5-triage
### How are you running Renovate? Self-hosted ### If you're self-hosting Renovate, tell us what version of Renovate you run. 32.126.2 ### Please select which platform you are using if self-hosting. GitHub Enterprise Server ### If you're self-hosting Renovate, tell us what version of the platform you run. 3.3.8 ### Was this something which used to work for you, and then stopped? It used to work, and then stopped ### Describe the bug Renovate doesn't create PRs for npm packages. For example, there is `3.8.2` version of chart.js available (we have `3.8.0`) but PR is not created. When running renovate with debug logging, we can see it fetches information from our NPM source (Artifactory) and that the Artifactoy URL returns the expected JSON (so `3.8.2` is included in the versions). Please note that we performed the update from 2.x to 3x manually but it shouldn't be any problem. ![image](https://user-images.githubusercontent.com/273052/180752951-a420c82a-0c2b-46fa-b064-85160bc4ebaa.png) ### Relevant debug logs <details><summary>Logs</summary> ``` INFO: Dependency extraction complete (repository=project/ui) "baseBranch": "master", "stats": { "managers": { "kubernetes": {"fileCount": 1, "depCount": 4}, "npm": {"fileCount": 2, "depCount": 48} }, "total": {"fileCount": 3, "depCount": 52} } { "depType": "dependencies", "depName": "chart.js", "currentValue": "^3.8.0", "datasource": "npm", "prettyDepType": "dependency", "lockedVersion": "3.8.0", "depIndex": 4, "updates": [], "warnings": [], "versioning": "npm", "sourceUrl": "https://github.com/chartjs/Chart.js", "homepage": "https://www.chartjs.org", "currentVersion": "3.8.0", "fixedVersion": "3.8.0" }, DEBUG: http statistics (repository=project/ui) "urls": { "https://artifactory.our.company.com/artifactory/api/npm/registry.npmjs.org/chart.js (GET,200)": 1, ``` </details> ### Have you created a minimal reproduction repository? No reproduction repository
1.0
npm package latest version not seen - ### How are you running Renovate? Self-hosted ### If you're self-hosting Renovate, tell us what version of Renovate you run. 32.126.2 ### Please select which platform you are using if self-hosting. GitHub Enterprise Server ### If you're self-hosting Renovate, tell us what version of the platform you run. 3.3.8 ### Was this something which used to work for you, and then stopped? It used to work, and then stopped ### Describe the bug Renovate doesn't create PRs for npm packages. For example, there is `3.8.2` version of chart.js available (we have `3.8.0`) but PR is not created. When running renovate with debug logging, we can see it fetches information from our NPM source (Artifactory) and that the Artifactoy URL returns the expected JSON (so `3.8.2` is included in the versions). Please note that we performed the update from 2.x to 3x manually but it shouldn't be any problem. ![image](https://user-images.githubusercontent.com/273052/180752951-a420c82a-0c2b-46fa-b064-85160bc4ebaa.png) ### Relevant debug logs <details><summary>Logs</summary> ``` INFO: Dependency extraction complete (repository=project/ui) "baseBranch": "master", "stats": { "managers": { "kubernetes": {"fileCount": 1, "depCount": 4}, "npm": {"fileCount": 2, "depCount": 48} }, "total": {"fileCount": 3, "depCount": 52} } { "depType": "dependencies", "depName": "chart.js", "currentValue": "^3.8.0", "datasource": "npm", "prettyDepType": "dependency", "lockedVersion": "3.8.0", "depIndex": 4, "updates": [], "warnings": [], "versioning": "npm", "sourceUrl": "https://github.com/chartjs/Chart.js", "homepage": "https://www.chartjs.org", "currentVersion": "3.8.0", "fixedVersion": "3.8.0" }, DEBUG: http statistics (repository=project/ui) "urls": { "https://artifactory.our.company.com/artifactory/api/npm/registry.npmjs.org/chart.js (GET,200)": 1, ``` </details> ### Have you created a minimal reproduction repository? No reproduction repository
non_process
npm package latest version not seen how are you running renovate self hosted if you re self hosting renovate tell us what version of renovate you run please select which platform you are using if self hosting github enterprise server if you re self hosting renovate tell us what version of the platform you run was this something which used to work for you and then stopped it used to work and then stopped describe the bug renovate doesn t create prs for npm packages for example there is version of chart js available we have but pr is not created when running renovate with debug logging we can see it fetches information from our npm source artifactory and that the artifactoy url returns the expected json so is included in the versions please note that we performed the update from x to manually but it shouldn t be any problem relevant debug logs logs info dependency extraction complete repository project ui basebranch master stats managers kubernetes filecount depcount npm filecount depcount total filecount depcount deptype dependencies depname chart js currentvalue datasource npm prettydeptype dependency lockedversion depindex updates warnings versioning npm sourceurl homepage currentversion fixedversion debug http statistics repository project ui urls get have you created a minimal reproduction repository no reproduction repository
0
53,975
6,777,427,927
IssuesEvent
2017-10-27 22:12:54
Automattic/jetpack
https://api.github.com/repos/Automattic/jetpack
opened
Connect Banner styles off
Admin Page General [Status] Design Polish [Type] Bug [Type] Good First Bug
#### Steps to reproduce the issue Run the latest version of Jetpack #### What I expected Things to look more styled. #### What happened instead The banner looks brokern see "By connecting your site you agree to our fascinatin" ![screen shot 2017-10-27 at 3 10 38 pm](https://user-images.githubusercontent.com/115071/32127067-1b768124-bb29-11e7-87a0-39001d859e8f.png) <!-- PLEASE NOTE - These comments won't show up when you submit the issue. - Everything is optional, but try to add as many details as possible. - If requesting a new feature, explain why you'd like to see it added. - This issue tracker is not for support. If you have questions about Jetpack, you can [start a new thread in the Jetpack support forums](https://wordpress.org/support/plugin/jetpack#new-post), or [send us an email](http://jetpack.com/contact-support/). - Do not report potential security vulnerabilities here. For responsible disclosure of security issues and to be eligible for our bug bounty program, please submit your report via [the HackerOne portal](https://hackerone.com/automattic). -->
1.0
Connect Banner styles off - #### Steps to reproduce the issue Run the latest version of Jetpack #### What I expected Things to look more styled. #### What happened instead The banner looks brokern see "By connecting your site you agree to our fascinatin" ![screen shot 2017-10-27 at 3 10 38 pm](https://user-images.githubusercontent.com/115071/32127067-1b768124-bb29-11e7-87a0-39001d859e8f.png) <!-- PLEASE NOTE - These comments won't show up when you submit the issue. - Everything is optional, but try to add as many details as possible. - If requesting a new feature, explain why you'd like to see it added. - This issue tracker is not for support. If you have questions about Jetpack, you can [start a new thread in the Jetpack support forums](https://wordpress.org/support/plugin/jetpack#new-post), or [send us an email](http://jetpack.com/contact-support/). - Do not report potential security vulnerabilities here. For responsible disclosure of security issues and to be eligible for our bug bounty program, please submit your report via [the HackerOne portal](https://hackerone.com/automattic). -->
non_process
connect banner styles off steps to reproduce the issue run the latest version of jetpack what i expected things to look more styled what happened instead the banner looks brokern see by connecting your site you agree to our fascinatin please note these comments won t show up when you submit the issue everything is optional but try to add as many details as possible if requesting a new feature explain why you d like to see it added this issue tracker is not for support if you have questions about jetpack you can or do not report potential security vulnerabilities here for responsible disclosure of security issues and to be eligible for our bug bounty program please submit your report via
0
12,822
15,196,541,024
IssuesEvent
2021-02-16 08:26:31
threefoldtech/js-sdk
https://api.github.com/repos/threefoldtech/js-sdk
closed
Blog can't get verified certificate
process_wontfix type_bug
### Description While testing, I found that Blog solution can't get SSL verified certificate. ### Steps to reproduce - Deploy VDC - Deploy Blog solution - Go to URL domain ![image](https://user-images.githubusercontent.com/11272864/107213138-a4b0c000-6a10-11eb-9018-630343a33a0b.png) ![image](https://user-images.githubusercontent.com/11272864/107213198-bbefad80-6a10-11eb-9187-7e523b55e2b5.png)
1.0
Blog can't get verified certificate - ### Description While testing, I found that Blog solution can't get SSL verified certificate. ### Steps to reproduce - Deploy VDC - Deploy Blog solution - Go to URL domain ![image](https://user-images.githubusercontent.com/11272864/107213138-a4b0c000-6a10-11eb-9018-630343a33a0b.png) ![image](https://user-images.githubusercontent.com/11272864/107213198-bbefad80-6a10-11eb-9187-7e523b55e2b5.png)
process
blog can t get verified certificate description while testing i found that blog solution can t get ssl verified certificate steps to reproduce deploy vdc deploy blog solution go to url domain
1
5,404
27,115,681,186
IssuesEvent
2023-02-15 18:22:31
VA-Explorer/va_explorer
https://api.github.com/repos/VA-Explorer/va_explorer
closed
Calculate and highlight outlier data within VA trends
Type: Maintainance good first issue Domain: Frontend Status: Inactive
**What is the expected state?** As a data analyst I expect to be able to tell at a glance if my trends data contains outliers so I can quickly identify them or further investigate them. I would like the trends tab chart to somehow highlight data that qualifies as an outlier. **What is the actual state?** The trends chart currently shows data without regard to outliers. It doesn't calculate them and it doesn't show them. **Relevant context** - `va_analytics/dash_apps/va_dashboard.py`
True
Calculate and highlight outlier data within VA trends - **What is the expected state?** As a data analyst I expect to be able to tell at a glance if my trends data contains outliers so I can quickly identify them or further investigate them. I would like the trends tab chart to somehow highlight data that qualifies as an outlier. **What is the actual state?** The trends chart currently shows data without regard to outliers. It doesn't calculate them and it doesn't show them. **Relevant context** - `va_analytics/dash_apps/va_dashboard.py`
non_process
calculate and highlight outlier data within va trends what is the expected state as a data analyst i expect to be able to tell at a glance if my trends data contains outliers so i can quickly identify them or further investigate them i would like the trends tab chart to somehow highlight data that qualifies as an outlier what is the actual state the trends chart currently shows data without regard to outliers it doesn t calculate them and it doesn t show them relevant context va analytics dash apps va dashboard py
0
72,532
9,596,906,567
IssuesEvent
2019-05-09 19:52:21
biotaphy/analyses
https://api.github.com/repos/biotaphy/analyses
opened
Document how to use phylo beta diversity
documentation help wanted
This includes how to use the script for performing the computations as well as to what the actual computation is and why someone would want to do it.
1.0
Document how to use phylo beta diversity - This includes how to use the script for performing the computations as well as to what the actual computation is and why someone would want to do it.
non_process
document how to use phylo beta diversity this includes how to use the script for performing the computations as well as to what the actual computation is and why someone would want to do it
0
47,432
19,650,833,723
IssuesEvent
2022-01-10 06:49:01
IBM-Cloud/terraform-provider-ibm
https://api.github.com/repos/IBM-Cloud/terraform-provider-ibm
closed
Add Placement Group support for PowerVS
enhancement service/Power Systems
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description In June 2021 the PowerVS REST API added support for Placement Groups: * https://cloud.ibm.com/apidocs/power-cloud#release-notes Documented here: * https://cloud.ibm.com/apidocs/power-cloud#pcloud-placementgroups-post These would be a useful addition to the Terraform provider. ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * ibm_pi_placement_group * ibm_pi_instance ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> Proposed new resource: ```hcl resource "ibm_pi_placement_group" "placement_group" { pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60" pi_placement_name = "some_group_name" pi_placement_policy = "affinity" } ``` Proposed new data source: ```hcl data "ibm_pi_placement_group" "ds_placement_group" { pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60" pi_placement_group_name = "some_group_name" } ``` Addition to existing ibm_pi_instance resource: ```hcl resource "ibm_pi_instance" "test-instance" { pi_memory = "4" pi_processors = "2" pi_instance_name = "test-vm" pi_proc_type = "shared" pi_image_id = "${data.ibm_pi_image.powerimages.id}" pi_key_pair_name = ibm_pi_key.key.key_id pi_sys_type = "s922" pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60" pi_pin_policy = "none" pi_health_status = "WARNING" pi_network { network_id = data.ibm_pi_public_network.dsnetwork.id } # NEW VARIABLE pi_placement_group_id = "49fba6c9-23f8-40bc-9899-aca322ee7d5b" } ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? ---> * https://cloud.ibm.com/apidocs/power-cloud#release-notes * https://cloud.ibm.com/apidocs/power-cloud#pcloud-placementgroups-post
1.0
Add Placement Group support for PowerVS - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description In June 2021 the PowerVS REST API added support for Placement Groups: * https://cloud.ibm.com/apidocs/power-cloud#release-notes Documented here: * https://cloud.ibm.com/apidocs/power-cloud#pcloud-placementgroups-post These would be a useful addition to the Terraform provider. ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * ibm_pi_placement_group * ibm_pi_instance ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> Proposed new resource: ```hcl resource "ibm_pi_placement_group" "placement_group" { pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60" pi_placement_name = "some_group_name" pi_placement_policy = "affinity" } ``` Proposed new data source: ```hcl data "ibm_pi_placement_group" "ds_placement_group" { pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60" pi_placement_group_name = "some_group_name" } ``` Addition to existing ibm_pi_instance resource: ```hcl resource "ibm_pi_instance" "test-instance" { pi_memory = "4" pi_processors = "2" pi_instance_name = "test-vm" pi_proc_type = "shared" pi_image_id = "${data.ibm_pi_image.powerimages.id}" pi_key_pair_name = ibm_pi_key.key.key_id pi_sys_type = "s922" pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60" pi_pin_policy = "none" pi_health_status = "WARNING" pi_network { network_id = data.ibm_pi_public_network.dsnetwork.id } # NEW VARIABLE pi_placement_group_id = "49fba6c9-23f8-40bc-9899-aca322ee7d5b" } ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? ---> * https://cloud.ibm.com/apidocs/power-cloud#release-notes * https://cloud.ibm.com/apidocs/power-cloud#pcloud-placementgroups-post
non_process
add placement group support for powervs community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description in june the powervs rest api added support for placement groups documented here these would be a useful addition to the terraform provider new or affected resource s ibm pi placement group ibm pi instance potential terraform configuration proposed new resource hcl resource ibm pi placement group placement group pi cloud instance id bcbe pi placement name some group name pi placement policy affinity proposed new data source hcl data ibm pi placement group ds placement group pi cloud instance id bcbe pi placement group name some group name addition to existing ibm pi instance resource hcl resource ibm pi instance test instance pi memory pi processors pi instance name test vm pi proc type shared pi image id data ibm pi image powerimages id pi key pair name ibm pi key key key id pi sys type pi cloud instance id bcbe pi pin policy none pi health status warning pi network network id data ibm pi public network dsnetwork id new variable pi placement group id references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation
0
64,044
14,657,149,446
IssuesEvent
2020-12-28 14:59:02
turkdevops/grafana
https://api.github.com/repos/turkdevops/grafana
opened
CVE-2019-0205 (High) detected in github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9, github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9
security vulnerability
## CVE-2019-0205 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b>, <b>github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary> <p> <details><summary><b>github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary> <p>Jaeger Bindings for Go OpenTracing API.</p> <p> Dependency Hierarchy: - github.com/grafana/grafana/pkg/infra/tracing-375e8e4fd051a5c821c2b477634c73166d0694ab (Root Library) - github.com/uber/jaeger-client-go/config-fe3fa553c313b32f58cc684a59a4d48f03e07df9 - :x: **github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9** (Vulnerable Library) </details> <details><summary><b>github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary> <p>Jaeger Bindings for Go OpenTracing API.</p> <p> Dependency Hierarchy: - github.com/grafana/grafana/pkg/infra/tracing-375e8e4fd051a5c821c2b477634c73166d0694ab (Root Library) - github.com/uber/jaeger-client-go/config-fe3fa553c313b32f58cc684a59a4d48f03e07df9 - github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9 - github.com/uber/jaeger-client-go/thrift-gen/agent-fe3fa553c313b32f58cc684a59a4d48f03e07df9 - :x: **github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/eaebfb49103a255dc2d75f384b343f6a7f708816">eaebfb49103a255dc2d75f384b343f6a7f708816</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings. <p>Publish Date: 2019-10-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205>CVE-2019-0205</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205</a></p> <p>Release Date: 2019-10-29</p> <p>Fix Resolution: org.apache.thrift:libthrift:0.13.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-0205 (High) detected in github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9, github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9 - ## CVE-2019-0205 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b>, <b>github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary> <p> <details><summary><b>github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary> <p>Jaeger Bindings for Go OpenTracing API.</p> <p> Dependency Hierarchy: - github.com/grafana/grafana/pkg/infra/tracing-375e8e4fd051a5c821c2b477634c73166d0694ab (Root Library) - github.com/uber/jaeger-client-go/config-fe3fa553c313b32f58cc684a59a4d48f03e07df9 - :x: **github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9** (Vulnerable Library) </details> <details><summary><b>github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary> <p>Jaeger Bindings for Go OpenTracing API.</p> <p> Dependency Hierarchy: - github.com/grafana/grafana/pkg/infra/tracing-375e8e4fd051a5c821c2b477634c73166d0694ab (Root Library) - github.com/uber/jaeger-client-go/config-fe3fa553c313b32f58cc684a59a4d48f03e07df9 - github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9 - github.com/uber/jaeger-client-go/thrift-gen/agent-fe3fa553c313b32f58cc684a59a4d48f03e07df9 - :x: **github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/eaebfb49103a255dc2d75f384b343f6a7f708816">eaebfb49103a255dc2d75f384b343f6a7f708816</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings. <p>Publish Date: 2019-10-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205>CVE-2019-0205</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205</a></p> <p>Release Date: 2019-10-29</p> <p>Fix Resolution: org.apache.thrift:libthrift:0.13.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in github com uber jaeger client go github com uber jaeger client go thrift cve high severity vulnerability vulnerable libraries github com uber jaeger client go github com uber jaeger client go thrift github com uber jaeger client go jaeger bindings for go opentracing api dependency hierarchy github com grafana grafana pkg infra tracing root library github com uber jaeger client go config x github com uber jaeger client go vulnerable library github com uber jaeger client go thrift jaeger bindings for go opentracing api dependency hierarchy github com grafana grafana pkg infra tracing root library github com uber jaeger client go config github com uber jaeger client go github com uber jaeger client go thrift gen agent x github com uber jaeger client go thrift vulnerable library found in head commit a href found in base branch master vulnerability details in apache thrift all versions up to and including a server or client may run into an endless loop when feed with specific input data because the issue had already been partially fixed in version depending on the installed version it affects only certain language bindings publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache thrift libthrift step up your open source security game with whitesource
0
8,608
11,764,726,822
IssuesEvent
2020-03-14 14:07:00
googleapis/google-cloud-cpp
https://api.github.com/repos/googleapis/google-cloud-cpp
closed
Create a CI build using macOS+CMake.
type: process
We need a build on macOS that uses CMake. There are instructions on the README, and I use CMake routinely, but we also need a CI for it. - [x] Create new scripts and test them locally - [x] Enable **only** the continuous cmake and bazel builds on Kokoro. - [x] Troubleshoot - [x] Enable presubmit builds and disable old Kokoro builds. - [x] Remove old scripts, copy `-new` files to their final names. - [x] Configure Kokoro to use the final version of the files. - [x] Remove the `-new` files.
1.0
Create a CI build using macOS+CMake. - We need a build on macOS that uses CMake. There are instructions on the README, and I use CMake routinely, but we also need a CI for it. - [x] Create new scripts and test them locally - [x] Enable **only** the continuous cmake and bazel builds on Kokoro. - [x] Troubleshoot - [x] Enable presubmit builds and disable old Kokoro builds. - [x] Remove old scripts, copy `-new` files to their final names. - [x] Configure Kokoro to use the final version of the files. - [x] Remove the `-new` files.
process
create a ci build using macos cmake we need a build on macos that uses cmake there are instructions on the readme and i use cmake routinely but we also need a ci for it create new scripts and test them locally enable only the continuous cmake and bazel builds on kokoro troubleshoot enable presubmit builds and disable old kokoro builds remove old scripts copy new files to their final names configure kokoro to use the final version of the files remove the new files
1
26,771
20,677,559,511
IssuesEvent
2022-03-10 10:43:36
airyhq/airy
https://api.github.com/repos/airyhq/airy
closed
Inconsistencies with the backup/restore process
infrastructure docs
## This is how we could improve the [documentation](https://docs.airy.co): Because backups are usually performed on a running Kafka instance, sometimes the restored Kafka can have inconsistencies, particularly in the internal topics. We should add to the `backup and restore` documentation how to reset the apps and fix these issues.
1.0
Inconsistencies with the backup/restore process - ## This is how we could improve the [documentation](https://docs.airy.co): Because backups are usually performed on a running Kafka instance, sometimes the restored Kafka can have inconsistencies, particularly in the internal topics. We should add to the `backup and restore` documentation how to reset the apps and fix these issues.
non_process
inconsistencies with the backup restore process this is how we could improve the because backups are usually performed on a running kafka instance sometimes the restored kafka can have inconsistencies particularly in the internal topics we should add to the backup and restore documentation how to reset the apps and fix these issues
0
50
2,513,878,262
IssuesEvent
2015-01-15 04:33:35
GsDevKit/zinc
https://api.github.com/repos/GsDevKit/zinc
closed
NeoJSON loads from master and not from gs_master ?
inprocess
Why is NeoJSON referenced via "github://GsDevKit/NeoJSON:master/repository" and not via the published "github://GsDevKit/NeoJSON:gs_master/repository" ? All the newly Unicode stuff is then missing ... Marten
1.0
NeoJSON loads from master and not from gs_master ? - Why is NeoJSON referenced via "github://GsDevKit/NeoJSON:master/repository" and not via the published "github://GsDevKit/NeoJSON:gs_master/repository" ? All the newly Unicode stuff is then missing ... Marten
process
neojson loads from master and not from gs master why is neojson referenced via github gsdevkit neojson master repository and not via the published github gsdevkit neojson gs master repository all the newly unicode stuff is then missing marten
1
7,639
10,736,553,887
IssuesEvent
2019-10-29 11:08:29
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
opened
Proposal: remove EF Core provider for Spanner
type: process
We currently have an alpha version of an Entity Framework Core provider for Spanner, along with a GA version of an ADO.NET provider. The EF Core provider has not been reviewed and maintained as thoroughly as the ADO.NET provider, and probably needs significant effort to bring it from alpha to beta and then GA. At the moment, we have no plans to put in that effort. Proposal: - Delist the packages in NuGet - Remove the source code from GitHub in a single commit (so we can easily revert it later if we decide to bring the project back) cc @meteatamel, @JustinBeckwith, @chrisdunelm, @SurferJeffAtGoogle, @jsimonweb, @amanda-tarafa
1.0
Proposal: remove EF Core provider for Spanner - We currently have an alpha version of an Entity Framework Core provider for Spanner, along with a GA version of an ADO.NET provider. The EF Core provider has not been reviewed and maintained as thoroughly as the ADO.NET provider, and probably needs significant effort to bring it from alpha to beta and then GA. At the moment, we have no plans to put in that effort. Proposal: - Delist the packages in NuGet - Remove the source code from GitHub in a single commit (so we can easily revert it later if we decide to bring the project back) cc @meteatamel, @JustinBeckwith, @chrisdunelm, @SurferJeffAtGoogle, @jsimonweb, @amanda-tarafa
process
proposal remove ef core provider for spanner we currently have an alpha version of an entity framework core provider for spanner along with a ga version of an ado net provider the ef core provider has not been reviewed and maintained as thoroughly as the ado net provider and probably needs significant effort to bring it from alpha to beta and then ga at the moment we have no plans to put in that effort proposal delist the packages in nuget remove the source code from github in a single commit so we can easily revert it later if we decide to bring the project back cc meteatamel justinbeckwith chrisdunelm surferjeffatgoogle jsimonweb amanda tarafa
1
10,095
13,044,162,081
IssuesEvent
2020-07-29 03:47:29
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `AddDurationAndString` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `AddDurationAndString` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `AddDurationAndString` from TiDB - ## Description Port the scalar function `AddDurationAndString` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function adddurationandstring from tidb description port the scalar function adddurationandstring from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
1
7,385
10,660,745,520
IssuesEvent
2019-10-18 10:36:59
VincenzoVisconti16/Social-Music
https://api.github.com/repos/VincenzoVisconti16/Social-Music
opened
evaluation system
functional requirement
The system shall implement an evaluation system*, for rating people samples and answers.
1.0
evaluation system - The system shall implement an evaluation system*, for rating people samples and answers.
non_process
evaluation system the system shall implement an evaluation system for rating people samples and answers
0
173,943
6,534,736,894
IssuesEvent
2017-08-31 12:06:34
gaocegege/Processing.R
https://api.github.com/repos/gaocegege/Processing.R
closed
Allow failure in macOS to solve the queueing problem
difficulty/low for-new-contributors priority/p2 size/small status/WIP type/enhancement
Now we could not get the macOS environment as soon as possible so I think we could set the build to https://docs.travis-ci.com/user/customizing-the-build#Rows-that-are-Allowed-to-Fail ![image](https://user-images.githubusercontent.com/5100735/29920173-dae4cca0-8e7e-11e7-9fb2-25b0bca49d31.png)
1.0
Allow failure in macOS to solve the queueing problem - Now we could not get the macOS environment as soon as possible so I think we could set the build to https://docs.travis-ci.com/user/customizing-the-build#Rows-that-are-Allowed-to-Fail ![image](https://user-images.githubusercontent.com/5100735/29920173-dae4cca0-8e7e-11e7-9fb2-25b0bca49d31.png)
non_process
allow failure in macos to solve the queueing problem now we could not get the macos environment as soon as possible so i think we could set the build to
0
796,790
28,127,879,532
IssuesEvent
2023-03-31 19:28:19
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
reopened
speech.snippets.quickstart_v2_test: test_quickstart_v2 failed
priority: p1 type: bug api: speech samples flakybot: issue flakybot: flaky
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 2b37b6a745661f0a80a75a5ae92077db9d69cb6c buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2e840773-8a31-4c25-9f03-affdb423d789), [Sponge](http://sponge2/2e840773-8a31-4c25-9f03-affdb423d789) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/grpc/_channel.py", line 1030, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking raise _InactiveRpcError(state) # pytype: disable=not-instantiable grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAVAILABLE details = "The service is currently unavailable." debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.132.95:443 {grpc_message:"The service is currently unavailable.", grpc_status:14, created_time:"2023-03-31T13:53:04.232449854+00:00"}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/speech/snippets/quickstart_v2_test.py", line 37, in test_quickstart_v2 response = quickstart_v2.quickstart_v2( File "/workspace/speech/snippets/quickstart_v2.py", line 36, in quickstart_v2 operation = client.create_recognizer(request=request) File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/cloud/speech_v2/services/speech/client.py", line 692, in create_recognizer response = rpc( File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.ServiceUnavailable: 503 The service is currently unavailable.</pre></details>
1.0
speech.snippets.quickstart_v2_test: test_quickstart_v2 failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 2b37b6a745661f0a80a75a5ae92077db9d69cb6c buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2e840773-8a31-4c25-9f03-affdb423d789), [Sponge](http://sponge2/2e840773-8a31-4c25-9f03-affdb423d789) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/grpc/_channel.py", line 1030, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking raise _InactiveRpcError(state) # pytype: disable=not-instantiable grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAVAILABLE details = "The service is currently unavailable." debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.132.95:443 {grpc_message:"The service is currently unavailable.", grpc_status:14, created_time:"2023-03-31T13:53:04.232449854+00:00"}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/speech/snippets/quickstart_v2_test.py", line 37, in test_quickstart_v2 response = quickstart_v2.quickstart_v2( File "/workspace/speech/snippets/quickstart_v2.py", line 36, in quickstart_v2 operation = client.create_recognizer(request=request) File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/cloud/speech_v2/services/speech/client.py", line 692, in create_recognizer response = rpc( File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.ServiceUnavailable: 503 The service is currently unavailable.</pre></details>
non_process
speech snippets quickstart test test quickstart failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace speech snippets nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace speech snippets nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace speech snippets nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state pytype disable not instantiable grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode unavailable details the service is currently unavailable debug error string unknown error received from peer grpc message the service is currently unavailable grpc status created time the above exception was the direct cause of the following exception traceback most recent call last file workspace speech snippets quickstart test py line in test quickstart response quickstart quickstart file workspace speech snippets quickstart py line in quickstart operation client create recognizer request request file workspace speech snippets nox py lib site packages google cloud speech services speech client py line in create recognizer response rpc file workspace speech snippets nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace speech snippets nox py lib site packages google api core grpc helpers py line in error remapped callable raise exceptions from grpc error exc from exc google api core exceptions serviceunavailable the service is currently unavailable
0
22,460
31,237,150,275
IssuesEvent
2023-08-20 12:21:53
Warzone2100/map-submission
https://api.github.com/repos/Warzone2100/map-submission
opened
[MAP]: 10c-UNPNTW5vs5
map unprocessed
### Upload Map [10c-UNPNTW5vs5V2.zip](https://github.com/Warzone2100/map-submission/files/12388133/10c-UNPNTW5vs5V2.zip) ### Authorship Mine: I am the author of this map ### Map Description (optional) ```text New NTW with large battlefield, scavs and circular mountains. ``` ### Notes for Reviewers (optional) _No response_
1.0
[MAP]: 10c-UNPNTW5vs5 - ### Upload Map [10c-UNPNTW5vs5V2.zip](https://github.com/Warzone2100/map-submission/files/12388133/10c-UNPNTW5vs5V2.zip) ### Authorship Mine: I am the author of this map ### Map Description (optional) ```text New NTW with large battlefield, scavs and circular mountains. ``` ### Notes for Reviewers (optional) _No response_
process
upload map authorship mine i am the author of this map map description optional text new ntw with large battlefield scavs and circular mountains notes for reviewers optional no response
1
775,858
27,239,234,432
IssuesEvent
2023-02-21 18:50:28
GoogleChrome/lighthouse
https://api.github.com/repos/GoogleChrome/lighthouse
closed
SyntaxError when attempting to use user-flow from lighthouse v10
bug pending-close needs-priority
### FAQ - [X] Yes, my issue is not about [variability](https://github.com/GoogleChrome/lighthouse/blob/main/docs/variability.md) or [throttling](https://github.com/GoogleChrome/lighthouse/blob/main/docs/throttling.md). - [X] Yes, my issue is not about a specific accessibility audit (file with [axe-core](https://github.com/dequelabs/axe-core) instead). ### URL https://www.example.com ### What happened? I am not able to use use user-flow from lighthouse version 10. I followed the example from the repo [lighthouse/docs/user-flows.md](https://github.com/GoogleChrome/lighthouse/blob/main/docs/user-flows.md#navigations-in-the-nodejs-api) but it throws and error. Here is a [minimal reproducible example](https://github.com/ChristopherPHolder/mr-lh-uf-v10). ### What did you expect? Usual behaviour. I expected lighthouse user-flow to run a navigation audit. ### What have you tried? I am using it inside of a project where I wrote a mini jest test to insure its working, that trow an error so I decided to make a minimal reproducible example repo and that throw the same error. The error from jest was: ```ts /Users/christopherholder/Applications/ufo/node_modules/lighthouse/core/index.js:7 import log from 'lighthouse-logger'; ^^^^^^ SyntaxError: Cannot use import statement outside a module > 1 | import { startFlow } from 'lighthouse'; | ^ 2 | import { launch } from 'puppeteer'; 3 | import { ResultReports } from 'shared'; 4 | at Runtime.createScriptFromCode (../../node_modules/jest-runtime/build/index.js:1796:14) at Object.<anonymous> (src/lib/user-flow-runner.ts:1:1) ``` The error is the same from the example repo `SyntaxError: Cannot use import statement outside a module` but jest seems to provided more information about its origin. Error in minimal reproduction: ``` SyntaxError: Cannot use import statement outside a module at Object.compileFunction (node:vm:360:18) at wrapSafe (node:internal/modules/cjs/loader:1088:15) at Module._compile (node:internal/modules/cjs/loader:1123:27) at Module._extensions..js (node:internal/modules/cjs/loader:1213:10) at Module.load (node:internal/modules/cjs/loader:1037:32) at Module._load (node:internal/modules/cjs/loader:878:12) at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:81:12) at node:internal/main/run_main_module:23:47 ``` ### How were you running Lighthouse? node ### Lighthouse Version v10.0.1 ### Chrome Version _No response_ ### Node Version v18.12.1 & v16.16.0 ### OS Mac ### Relevant log output ```shell // In Minimal reproduction example repo SyntaxError: Cannot use import statement outside a module at Object.compileFunction (node:vm:360:18) at wrapSafe (node:internal/modules/cjs/loader:1088:15) at Module._compile (node:internal/modules/cjs/loader:1123:27) at Module._extensions..js (node:internal/modules/cjs/loader:1213:10) at Module.load (node:internal/modules/cjs/loader:1037:32) at Module._load (node:internal/modules/cjs/loader:878:12) at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:81:12) at node:internal/main/run_main_module:23:47 ```
1.0
SyntaxError when attempting to use user-flow from lighthouse v10 - ### FAQ - [X] Yes, my issue is not about [variability](https://github.com/GoogleChrome/lighthouse/blob/main/docs/variability.md) or [throttling](https://github.com/GoogleChrome/lighthouse/blob/main/docs/throttling.md). - [X] Yes, my issue is not about a specific accessibility audit (file with [axe-core](https://github.com/dequelabs/axe-core) instead). ### URL https://www.example.com ### What happened? I am not able to use use user-flow from lighthouse version 10. I followed the example from the repo [lighthouse/docs/user-flows.md](https://github.com/GoogleChrome/lighthouse/blob/main/docs/user-flows.md#navigations-in-the-nodejs-api) but it throws and error. Here is a [minimal reproducible example](https://github.com/ChristopherPHolder/mr-lh-uf-v10). ### What did you expect? Usual behaviour. I expected lighthouse user-flow to run a navigation audit. ### What have you tried? I am using it inside of a project where I wrote a mini jest test to insure its working, that trow an error so I decided to make a minimal reproducible example repo and that throw the same error. The error from jest was: ```ts /Users/christopherholder/Applications/ufo/node_modules/lighthouse/core/index.js:7 import log from 'lighthouse-logger'; ^^^^^^ SyntaxError: Cannot use import statement outside a module > 1 | import { startFlow } from 'lighthouse'; | ^ 2 | import { launch } from 'puppeteer'; 3 | import { ResultReports } from 'shared'; 4 | at Runtime.createScriptFromCode (../../node_modules/jest-runtime/build/index.js:1796:14) at Object.<anonymous> (src/lib/user-flow-runner.ts:1:1) ``` The error is the same from the example repo `SyntaxError: Cannot use import statement outside a module` but jest seems to provided more information about its origin. Error in minimal reproduction: ``` SyntaxError: Cannot use import statement outside a module at Object.compileFunction (node:vm:360:18) at wrapSafe (node:internal/modules/cjs/loader:1088:15) at Module._compile (node:internal/modules/cjs/loader:1123:27) at Module._extensions..js (node:internal/modules/cjs/loader:1213:10) at Module.load (node:internal/modules/cjs/loader:1037:32) at Module._load (node:internal/modules/cjs/loader:878:12) at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:81:12) at node:internal/main/run_main_module:23:47 ``` ### How were you running Lighthouse? node ### Lighthouse Version v10.0.1 ### Chrome Version _No response_ ### Node Version v18.12.1 & v16.16.0 ### OS Mac ### Relevant log output ```shell // In Minimal reproduction example repo SyntaxError: Cannot use import statement outside a module at Object.compileFunction (node:vm:360:18) at wrapSafe (node:internal/modules/cjs/loader:1088:15) at Module._compile (node:internal/modules/cjs/loader:1123:27) at Module._extensions..js (node:internal/modules/cjs/loader:1213:10) at Module.load (node:internal/modules/cjs/loader:1037:32) at Module._load (node:internal/modules/cjs/loader:878:12) at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:81:12) at node:internal/main/run_main_module:23:47 ```
non_process
syntaxerror when attempting to use user flow from lighthouse faq yes my issue is not about or yes my issue is not about a specific accessibility audit file with instead url what happened i am not able to use use user flow from lighthouse version i followed the example from the repo but it throws and error here is a what did you expect usual behaviour i expected lighthouse user flow to run a navigation audit what have you tried i am using it inside of a project where i wrote a mini jest test to insure its working that trow an error so i decided to make a minimal reproducible example repo and that throw the same error the error from jest was ts users christopherholder applications ufo node modules lighthouse core index js import log from lighthouse logger syntaxerror cannot use import statement outside a module import startflow from lighthouse import launch from puppeteer import resultreports from shared at runtime createscriptfromcode node modules jest runtime build index js at object src lib user flow runner ts the error is the same from the example repo syntaxerror cannot use import statement outside a module but jest seems to provided more information about its origin error in minimal reproduction syntaxerror cannot use import statement outside a module at object compilefunction node vm at wrapsafe node internal modules cjs loader at module compile node internal modules cjs loader at module extensions js node internal modules cjs loader at module load node internal modules cjs loader at module load node internal modules cjs loader at function executeuserentrypoint node internal modules run main at node internal main run main module how were you running lighthouse node lighthouse version chrome version no response node version os mac relevant log output shell in minimal reproduction example repo syntaxerror cannot use import statement outside a module at object compilefunction node vm at wrapsafe node internal modules cjs loader at module compile node internal modules cjs loader at module extensions js node internal modules cjs loader at module load node internal modules cjs loader at module load node internal modules cjs loader at function executeuserentrypoint node internal modules run main at node internal main run main module
0
4,454
7,320,588,318
IssuesEvent
2018-03-02 08:05:27
EBWiki/EBWiki
https://api.github.com/repos/EBWiki/EBWiki
closed
Spec coverage needs to cover failure cases
bug process
While we have decent spec coverage, we are not covering failure cases enough, where the user enters invalid or incomplete information.
1.0
Spec coverage needs to cover failure cases - While we have decent spec coverage, we are not covering failure cases enough, where the user enters invalid or incomplete information.
process
spec coverage needs to cover failure cases while we have decent spec coverage we are not covering failure cases enough where the user enters invalid or incomplete information
1