Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
44,598
| 13,060,610,356
|
IssuesEvent
|
2020-07-30 12:42:56
|
jgeraigery/frost-gs-spring-boot-docker
|
https://api.github.com/repos/jgeraigery/frost-gs-spring-boot-docker
|
opened
|
CVE-2019-17531 (High) detected in jackson-databind-2.9.9.jar
|
security vulnerability
|
## CVE-2019-17531 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/frost-gs-spring-boot-docker/initial/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9/d6eb9817d9c7289a91f043ac5ee02a6b3cc86238/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/frost-gs-spring-boot-docker/complete/target/dependency/BOOT-INF/lib/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/frost-gs-spring-boot-docker/commit/2913b67e67d02acdd30a738e35187b8c7922ed4d">2913b67e67d02acdd30a738e35187b8c7922ed4d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload.
<p>Publish Date: 2019-10-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531>CVE-2019-17531</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531</a></p>
<p>Release Date: 2019-10-12</p>
<p>Fix Resolution: 2.10</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.10"}],"vulnerabilityIdentifier":"CVE-2019-17531","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-17531 (High) detected in jackson-databind-2.9.9.jar - ## CVE-2019-17531 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/frost-gs-spring-boot-docker/initial/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9/d6eb9817d9c7289a91f043ac5ee02a6b3cc86238/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar,/frost-gs-spring-boot-docker/complete/target/dependency/BOOT-INF/lib/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/frost-gs-spring-boot-docker/commit/2913b67e67d02acdd30a738e35187b8c7922ed4d">2913b67e67d02acdd30a738e35187b8c7922ed4d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload.
<p>Publish Date: 2019-10-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531>CVE-2019-17531</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531</a></p>
<p>Release Date: 2019-10-12</p>
<p>Fix Resolution: 2.10</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.9","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.10"}],"vulnerabilityIdentifier":"CVE-2019-17531","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm frost gs spring boot docker initial build gradle path to vulnerable library home wss scanner gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar frost gs spring boot docker complete target dependency boot inf lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the apache extra version x jar in the classpath and an attacker can provide a jndi service to access it is possible to make the service execute a malicious payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the apache extra version x jar in the classpath and an attacker can provide a jndi service to access it is possible to make the service execute a malicious payload vulnerabilityurl
| 0
|
183,867
| 6,694,303,128
|
IssuesEvent
|
2017-10-10 01:01:03
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
hangouts.google.com - see bug description
|
browser-firefox priority-critical status-incomplete
|
<!-- @browser: Firefox 57.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:57.0) Gecko/20100101 Firefox/57.0 -->
<!-- @reported_with: web -->
**URL**: https://hangouts.google.com/
**Browser / Version**: Firefox 57.0
**Operating System**: Linux
**Tested Another Browser**: Unknown
**Problem type**: Something else
**Description**: Fail loading API Google Hangout JavaScript no load.
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2017/10/1d569467-b425-460f-8f88-05bafcd75e87.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
hangouts.google.com - see bug description - <!-- @browser: Firefox 57.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:57.0) Gecko/20100101 Firefox/57.0 -->
<!-- @reported_with: web -->
**URL**: https://hangouts.google.com/
**Browser / Version**: Firefox 57.0
**Operating System**: Linux
**Tested Another Browser**: Unknown
**Problem type**: Something else
**Description**: Fail loading API Google Hangout JavaScript no load.
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2017/10/1d569467-b425-460f-8f88-05bafcd75e87.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
hangouts google com see bug description url browser version firefox operating system linux tested another browser unknown problem type something else description fail loading api google hangout javascript no load steps to reproduce from with ❤️
| 0
|
5,987
| 8,805,374,570
|
IssuesEvent
|
2018-12-26 19:13:58
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Incorrect processing of subject scheme maps
|
bug preprocess preprocess/filtering preprocess/keyref priority/medium stale
|
## General description
DITA-OT 1.7 does not handle subject scheme maps properly in regard to validating the attribute values against the specified subject schemes. While the simplest test yields the expected warning messages when the content contains attribute values not specified in the subject scheme, it requires that the subjectScheme map be referenced with type="subjectScheme" specified. (The DITA 1.2 spec does not suggest this behavior; if it is a DITA-OT requirement, we need to document it.)
More realistic test cases using schemeref elements simply fail, however, regardless of whether type="subjectScheme" is specified.
The DITA maps and topic are located at https://gist.github.com/keberlein/5237276 .
## Test content
The following subject definitions and enumeration:
```
<subjectdef keys="operating-systems">
<subjectdef keys="linux">
<subjectdef keys="redhat"/>
<subjectdef keys="suse"/>
</subjectdef>
<subjectdef keys="unix"/>
<subjectdef keys="windows">
<subjectdef keys="windows7"/>
<subjectdef keys="windows8"/>
</subjectdef>
</subjectdef>
```
One DITA topic with following values for the platform attribute: ubuntu, mandriva, solaris
## Test 1
**Test-1.ditamap** contains the following content:
```
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd">
<map>
<title>Example of a map referencing a subject scheme</title>
<mapref href="baseScheme-with-enumeration.ditamap" type="subjectScheme"/>
<topicref href="t-creating-controlled-values.dita"/>
```
</map>
**The expected results are generated -- warning messages**
```
[filter] [DOTJ049W][WARN] The attribute value platform="mandriva" on element "step" does not comply with the specified subject scheme. According to the subject scheme map, the following values are valid for the platform attribute: redhat,windows8,windows7,windows,suse,linux,unix,operating-systems
…
[filter] [DOTJ049W][WARN] The attribute value platform="solaris" on element "step" does not comply with the specified subject scheme. According to the subject scheme map, the following values are valid for the platform attribute: redhat,windows8,windows7,windows,suse,linux,unix,operating-systems
```
Unfortunately, the DITA-OT is incorrectly building the list of valid labels; it should NOT include operating-systems, according to the examples and guidance provided in the DITA 1.2 spec. See the examples in http://docs.oasis-open.org/dita/v1.2/os/spec/langref/subjectScheme.html#subjectScheme , as well as http://docs.oasis-open.org/dita/v1.2/os/spec/langref/enumerationdef.html#enumerationdef
## Test 2
I removed type=’’subjectScheme” from the <mapref> element Test-2.ditamap. **As a result, no warning messages are generated.**
## Test 3
Here the master map references a subjectScheme map which in turn references another subjectScheme map. This is standard use case that is outlined in the DITA 1.2 spec.
**Test-3.ditamap** contains the following content:
```
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd">
<map>
<title>Example of a map referencing a subject scheme</title>
<mapref href="extensionScheme.ditamap" type="subjectScheme"/>
<topicref href="t-creating-controlled-values.dita"/>
```
</map>
**extensionScheme.ditamap** contains the following content:
```
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE subjectScheme PUBLIC "-//OASIS//DTD DITA Subject Scheme Map//EN" "subjectScheme.dtd">
<subjectScheme>
<schemeref href="baseScheme.ditamap"/>
<enumerationdef>
<attributedef name="platform"/>
<subjectdef keyref="operating-systems"/>
</enumerationdef>
</subjectScheme>
```
**Results:** No warning messages.
## Test 4
I modified the <schemeref> element in the extension scheme to see if it also needed type=”subjectScheme” to be specified.
**extensionScheme2.ditamap** contains the following content:
```
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE subjectScheme PUBLIC "-//OASIS//DTD DITA Subject Scheme Map//EN" "subjectScheme.dtd">
<subjectScheme>
<schemeref href="baseScheme.ditamap" type="subjectScheme"/>
<enumerationdef>
<attributedef name="platform"/>
<subjectdef keyref="operating-systems"/>
</enumerationdef>
</subjectScheme>
```
**Results:** No warning messages.
|
3.0
|
Incorrect processing of subject scheme maps - ## General description
DITA-OT 1.7 does not handle subject scheme maps properly in regard to validating the attribute values against the specified subject schemes. While the simplest test yields the expected warning messages when the content contains attribute values not specified in the subject scheme, it requires that the subjectScheme map be referenced with type="subjectScheme" specified. (The DITA 1.2 spec does not suggest this behavior; if it is a DITA-OT requirement, we need to document it.)
More realistic test cases using schemeref elements simply fail, however, regardless of whether type="subjectScheme" is specified.
The DITA maps and topic are located at https://gist.github.com/keberlein/5237276 .
## Test content
The following subject definitions and enumeration:
```
<subjectdef keys="operating-systems">
<subjectdef keys="linux">
<subjectdef keys="redhat"/>
<subjectdef keys="suse"/>
</subjectdef>
<subjectdef keys="unix"/>
<subjectdef keys="windows">
<subjectdef keys="windows7"/>
<subjectdef keys="windows8"/>
</subjectdef>
</subjectdef>
```
One DITA topic with following values for the platform attribute: ubuntu, mandriva, solaris
## Test 1
**Test-1.ditamap** contains the following content:
```
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd">
<map>
<title>Example of a map referencing a subject scheme</title>
<mapref href="baseScheme-with-enumeration.ditamap" type="subjectScheme"/>
<topicref href="t-creating-controlled-values.dita"/>
```
</map>
**The expected results are generated -- warning messages**
```
[filter] [DOTJ049W][WARN] The attribute value platform="mandriva" on element "step" does not comply with the specified subject scheme. According to the subject scheme map, the following values are valid for the platform attribute: redhat,windows8,windows7,windows,suse,linux,unix,operating-systems
…
[filter] [DOTJ049W][WARN] The attribute value platform="solaris" on element "step" does not comply with the specified subject scheme. According to the subject scheme map, the following values are valid for the platform attribute: redhat,windows8,windows7,windows,suse,linux,unix,operating-systems
```
Unfortunately, the DITA-OT is incorrectly building the list of valid labels; it should NOT include operating-systems, according to the examples and guidance provided in the DITA 1.2 spec. See the examples in http://docs.oasis-open.org/dita/v1.2/os/spec/langref/subjectScheme.html#subjectScheme , as well as http://docs.oasis-open.org/dita/v1.2/os/spec/langref/enumerationdef.html#enumerationdef
## Test 2
I removed type=’’subjectScheme” from the <mapref> element Test-2.ditamap. **As a result, no warning messages are generated.**
## Test 3
Here the master map references a subjectScheme map which in turn references another subjectScheme map. This is standard use case that is outlined in the DITA 1.2 spec.
**Test-3.ditamap** contains the following content:
```
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd">
<map>
<title>Example of a map referencing a subject scheme</title>
<mapref href="extensionScheme.ditamap" type="subjectScheme"/>
<topicref href="t-creating-controlled-values.dita"/>
```
</map>
**extensionScheme.ditamap** contains the following content:
```
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE subjectScheme PUBLIC "-//OASIS//DTD DITA Subject Scheme Map//EN" "subjectScheme.dtd">
<subjectScheme>
<schemeref href="baseScheme.ditamap"/>
<enumerationdef>
<attributedef name="platform"/>
<subjectdef keyref="operating-systems"/>
</enumerationdef>
</subjectScheme>
```
**Results:** No warning messages.
## Test 4
I modified the <schemeref> element in the extension scheme to see if it also needed type=”subjectScheme” to be specified.
**extensionScheme2.ditamap** contains the following content:
```
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE subjectScheme PUBLIC "-//OASIS//DTD DITA Subject Scheme Map//EN" "subjectScheme.dtd">
<subjectScheme>
<schemeref href="baseScheme.ditamap" type="subjectScheme"/>
<enumerationdef>
<attributedef name="platform"/>
<subjectdef keyref="operating-systems"/>
</enumerationdef>
</subjectScheme>
```
**Results:** No warning messages.
|
process
|
incorrect processing of subject scheme maps general description dita ot does not handle subject scheme maps properly in regard to validating the attribute values against the specified subject schemes while the simplest test yields the expected warning messages when the content contains attribute values not specified in the subject scheme it requires that the subjectscheme map be referenced with type subjectscheme specified the dita spec does not suggest this behavior if it is a dita ot requirement we need to document it more realistic test cases using schemeref elements simply fail however regardless of whether type subjectscheme is specified the dita maps and topic are located at test content the following subject definitions and enumeration one dita topic with following values for the platform attribute ubuntu mandriva solaris test test ditamap contains the following content example of a map referencing a subject scheme the expected results are generated warning messages the attribute value platform mandriva on element step does not comply with the specified subject scheme according to the subject scheme map the following values are valid for the platform attribute redhat windows suse linux unix operating systems … the attribute value platform solaris on element step does not comply with the specified subject scheme according to the subject scheme map the following values are valid for the platform attribute redhat windows suse linux unix operating systems unfortunately the dita ot is incorrectly building the list of valid labels it should not include operating systems according to the examples and guidance provided in the dita spec see the examples in as well as test i removed type ’’subjectscheme” from the element test ditamap as a result no warning messages are generated test here the master map references a subjectscheme map which in turn references another subjectscheme map this is standard use case that is outlined in the dita spec test ditamap contains the following content example of a map referencing a subject scheme extensionscheme ditamap contains the following content results no warning messages test i modified the element in the extension scheme to see if it also needed type ”subjectscheme” to be specified ditamap contains the following content results no warning messages
| 1
|
2,398
| 5,192,325,198
|
IssuesEvent
|
2017-01-22 07:14:47
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
Props warning from ToolBar in v3.0.0-beta.6
|
enhancement inprocess
|
> "Failed prop type: The prop `contentLabel` is marked as required in `Modal`, but its value is `undefined`."
Since React modal v1.6 there are two required props `contentLabel` and `isOpen`: https://github.com/reactjs/react-modal#usage
|
1.0
|
Props warning from ToolBar in v3.0.0-beta.6 - > "Failed prop type: The prop `contentLabel` is marked as required in `Modal`, but its value is `undefined`."
Since React modal v1.6 there are two required props `contentLabel` and `isOpen`: https://github.com/reactjs/react-modal#usage
|
process
|
props warning from toolbar in beta failed prop type the prop contentlabel is marked as required in modal but its value is undefined since react modal there are two required props contentlabel and isopen
| 1
|
67,839
| 13,039,490,562
|
IssuesEvent
|
2020-07-28 16:50:30
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
[4.0] - [API] User Groups and Access Levels should be accessible via API!
|
No Code Attached Yet
|
### Steps to reproduce the issue
There is NO API that let's us access the names of user groups and access levels for the site. com_content API returns the ID's of access level but there is no way to turn those ID's into readable text. This is also useful while creating new users and giving the ability to select User Groups for the user.
### Expected result
User Groups and Access Levels should be available via the API
Even if creation is not possible, it's fine but listing of ID's with names should be possible.
### Actual result
No such API exists!
### System information (as much as possible)
Joomla 4 Beta 2.
### Additional comments
If needed, most likely for @alikon
|
1.0
|
[4.0] - [API] User Groups and Access Levels should be accessible via API! - ### Steps to reproduce the issue
There is NO API that let's us access the names of user groups and access levels for the site. com_content API returns the ID's of access level but there is no way to turn those ID's into readable text. This is also useful while creating new users and giving the ability to select User Groups for the user.
### Expected result
User Groups and Access Levels should be available via the API
Even if creation is not possible, it's fine but listing of ID's with names should be possible.
### Actual result
No such API exists!
### System information (as much as possible)
Joomla 4 Beta 2.
### Additional comments
If needed, most likely for @alikon
|
non_process
|
user groups and access levels should be accessible via api steps to reproduce the issue there is no api that let s us access the names of user groups and access levels for the site com content api returns the id s of access level but there is no way to turn those id s into readable text this is also useful while creating new users and giving the ability to select user groups for the user expected result user groups and access levels should be available via the api even if creation is not possible it s fine but listing of id s with names should be possible actual result no such api exists system information as much as possible joomla beta additional comments if needed most likely for alikon
| 0
|
67,474
| 7,048,366,680
|
IssuesEvent
|
2018-01-02 17:22:02
|
vmware/vic
|
https://api.github.com/repos/vmware/vic
|
opened
|
nightly 12/28: 5-15-NFS-Datastore panics on VCH delete
|
component/test kind/nightly-blocker priority/high team/lifecycle
|
In test cleanup, VCH delete panics and CDE.
`vic-machine.log` contains a stack trace:
```
Dec 29 2017 03:00:02.143-06:00 INFO op=10546.1: ### Removing VCH ####
Dec 29 2017 03:30:02.143-06:00 ERROR op=10546.1: vic/pkg/trace.(*Operation).Err: vic-machine-linux error: context deadline exceeded
vic/cmd/vic-machine/delete.(*Uninstall).Run:92 vic-machine-linux
vic/cmd/vic-machine/common.NewOperation:27 vic-machine-linux
Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: vic/pkg/trace.(*Operation).Err: vic-machine-linux error: context deadline exceeded
......
Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: --------------------
Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: vic-machine-linux delete failed: Delete timed out: use --timeout to add more time
goroutine 1 [running]:
runtime/debug.Stack(0xc42001c5f0, 0xfcc546, 0x2f)
/usr/local/go/src/runtime/debug/stack.go:24 +0x79
main.main.func1(0xc420067520, 0xc420084288)
/go/src/github.com/vmware/vic/cmd/vic-machine/main.go:160 +0x298
panic(0xda3b60, 0x1885f90)
/usr/local/go/src/runtime/panic.go:489 +0x2cf
github.com/vmware/vic/pkg/vsphere/session.(*Session).Populate(0xc420206a00, 0x189bda0, 0xc4201edb80, 0xd, 0x1, 0x0)
/go/src/github.com/vmware/vic/pkg/vsphere/session/session.go:358 +0xf8a
github.com/vmware/vic/lib/install/validate.NewValidator(0x189bda0, 0xc4201edb80, 0xc4201df400, 0x0, 0x0, 0x0)
/go/src/github.com/vmware/vic/lib/install/validate/validator.go:159 +0x978
github.com/vmware/vic/cmd/vic-machine/delete.(*Uninstall).Run(0xc4201d58f0, 0xc42017db80, 0x1895de0, 0xc4204766e0)
/go/src/github.com/vmware/vic/cmd/vic-machine/delete/delete.go:112 +0x2d6
github.com/vmware/vic/cmd/vic-machine/delete.(*Uninstall).Run-fm(0xc42017db80, 0x1010101010100, 0xc42017db80)
/go/src/github.com/vmware/vic/cmd/vic-machine/main.go:75 +0x34
github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.HandleAction(0xd56cc0, 0xc4201d5940, 0xc42017db80, 0xc42006b100, 0x0)
/go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/app.go:485 +0xd4
github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.Command.Run(0xf8ef69, 0x6, 0x0, 0x0, 0x0, 0x0, 0x0, 0xfbf004, 0x23, 0x0, ...)
/go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/command.go:193 +0xb72
github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.(*App).Run(0xc420067520, 0xc420088000, 0xa, 0xa, 0x0, 0x0)
/go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/app.go:250 +0x7d0
main.main()
/go/src/github.com/vmware/vic/cmd/vic-machine/main.go:165 +0x2233
```
From the stack trace, the panic happens during validation session populate, where it grabs VM folder from datacenter using govmomi:
```
if s.Datacenter != nil {
folders, err := s.Datacenter.Folders(op)
if err != nil {
errs = append(errs, fmt.Sprintf("Failure finding folders (%s): %s", s.DatacenterPath, err.Error()))
} else {
op.Debugf("Cached folders: %s", s.DatacenterPath)
}
s.VMFolder = folders.VmFolder
}
```
In line `s.VMFolder = folders.VmFolder`.
https://github.com/vmware/vic/blob/master/pkg/vsphere/session/session.go#L358
Log: [5-15-NFS-Datastore.zip](https://github.com/vmware/vic/files/1598194/5-15-NFS-Datastore.zip)
Cc: @dougm @hickeng
|
1.0
|
nightly 12/28: 5-15-NFS-Datastore panics on VCH delete - In test cleanup, VCH delete panics and CDE.
`vic-machine.log` contains a stack trace:
```
Dec 29 2017 03:00:02.143-06:00 INFO op=10546.1: ### Removing VCH ####
Dec 29 2017 03:30:02.143-06:00 ERROR op=10546.1: vic/pkg/trace.(*Operation).Err: vic-machine-linux error: context deadline exceeded
vic/cmd/vic-machine/delete.(*Uninstall).Run:92 vic-machine-linux
vic/cmd/vic-machine/common.NewOperation:27 vic-machine-linux
Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: vic/pkg/trace.(*Operation).Err: vic-machine-linux error: context deadline exceeded
......
Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: --------------------
Dec 29 2017 03:30:02.145-06:00 ERROR op=10546.1: vic-machine-linux delete failed: Delete timed out: use --timeout to add more time
goroutine 1 [running]:
runtime/debug.Stack(0xc42001c5f0, 0xfcc546, 0x2f)
/usr/local/go/src/runtime/debug/stack.go:24 +0x79
main.main.func1(0xc420067520, 0xc420084288)
/go/src/github.com/vmware/vic/cmd/vic-machine/main.go:160 +0x298
panic(0xda3b60, 0x1885f90)
/usr/local/go/src/runtime/panic.go:489 +0x2cf
github.com/vmware/vic/pkg/vsphere/session.(*Session).Populate(0xc420206a00, 0x189bda0, 0xc4201edb80, 0xd, 0x1, 0x0)
/go/src/github.com/vmware/vic/pkg/vsphere/session/session.go:358 +0xf8a
github.com/vmware/vic/lib/install/validate.NewValidator(0x189bda0, 0xc4201edb80, 0xc4201df400, 0x0, 0x0, 0x0)
/go/src/github.com/vmware/vic/lib/install/validate/validator.go:159 +0x978
github.com/vmware/vic/cmd/vic-machine/delete.(*Uninstall).Run(0xc4201d58f0, 0xc42017db80, 0x1895de0, 0xc4204766e0)
/go/src/github.com/vmware/vic/cmd/vic-machine/delete/delete.go:112 +0x2d6
github.com/vmware/vic/cmd/vic-machine/delete.(*Uninstall).Run-fm(0xc42017db80, 0x1010101010100, 0xc42017db80)
/go/src/github.com/vmware/vic/cmd/vic-machine/main.go:75 +0x34
github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.HandleAction(0xd56cc0, 0xc4201d5940, 0xc42017db80, 0xc42006b100, 0x0)
/go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/app.go:485 +0xd4
github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.Command.Run(0xf8ef69, 0x6, 0x0, 0x0, 0x0, 0x0, 0x0, 0xfbf004, 0x23, 0x0, ...)
/go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/command.go:193 +0xb72
github.com/vmware/vic/vendor/gopkg.in/urfave/cli%2ev1.(*App).Run(0xc420067520, 0xc420088000, 0xa, 0xa, 0x0, 0x0)
/go/src/github.com/vmware/vic/vendor/gopkg.in/urfave/cli.v1/app.go:250 +0x7d0
main.main()
/go/src/github.com/vmware/vic/cmd/vic-machine/main.go:165 +0x2233
```
From the stack trace, the panic happens during validation session populate, where it grabs VM folder from datacenter using govmomi:
```
if s.Datacenter != nil {
folders, err := s.Datacenter.Folders(op)
if err != nil {
errs = append(errs, fmt.Sprintf("Failure finding folders (%s): %s", s.DatacenterPath, err.Error()))
} else {
op.Debugf("Cached folders: %s", s.DatacenterPath)
}
s.VMFolder = folders.VmFolder
}
```
In line `s.VMFolder = folders.VmFolder`.
https://github.com/vmware/vic/blob/master/pkg/vsphere/session/session.go#L358
Log: [5-15-NFS-Datastore.zip](https://github.com/vmware/vic/files/1598194/5-15-NFS-Datastore.zip)
Cc: @dougm @hickeng
|
non_process
|
nightly nfs datastore panics on vch delete in test cleanup vch delete panics and cde vic machine log contains a stack trace dec info op removing vch dec error op vic pkg trace operation err vic machine linux error context deadline exceeded vic cmd vic machine delete uninstall run vic machine linux vic cmd vic machine common newoperation vic machine linux dec error op vic pkg trace operation err vic machine linux error context deadline exceeded dec error op dec error op vic machine linux delete failed delete timed out use timeout to add more time goroutine runtime debug stack usr local go src runtime debug stack go main main go src github com vmware vic cmd vic machine main go panic usr local go src runtime panic go github com vmware vic pkg vsphere session session populate go src github com vmware vic pkg vsphere session session go github com vmware vic lib install validate newvalidator go src github com vmware vic lib install validate validator go github com vmware vic cmd vic machine delete uninstall run go src github com vmware vic cmd vic machine delete delete go github com vmware vic cmd vic machine delete uninstall run fm go src github com vmware vic cmd vic machine main go github com vmware vic vendor gopkg in urfave cli handleaction go src github com vmware vic vendor gopkg in urfave cli app go github com vmware vic vendor gopkg in urfave cli command run go src github com vmware vic vendor gopkg in urfave cli command go github com vmware vic vendor gopkg in urfave cli app run go src github com vmware vic vendor gopkg in urfave cli app go main main go src github com vmware vic cmd vic machine main go from the stack trace the panic happens during validation session populate where it grabs vm folder from datacenter using govmomi if s datacenter nil folders err s datacenter folders op if err nil errs append errs fmt sprintf failure finding folders s s s datacenterpath err error else op debugf cached folders s s datacenterpath s vmfolder folders vmfolder in line s vmfolder folders vmfolder log cc dougm hickeng
| 0
|
293,240
| 25,279,058,570
|
IssuesEvent
|
2022-11-16 14:36:21
|
cockroachdb/pebble
|
https://api.github.com/repos/cockroachdb/pebble
|
closed
|
github.com/cockroachdb/pebble/internal/metamorphic: TestMeta failed
|
O-robot C-test-failure metamorphic-failure branch-refs/heads/master T-storage A-storage
|
github.com/cockroachdb/pebble/internal/metamorphic.TestMeta [failed](https://buildConfiguration/Cockroach_Nightlies_Pebble_Metamorphic/7423668?buildTab=log) with [artifacts](https://buildConfiguration/Cockroach_Nightlies_Pebble_Metamorphic/7423668?buildTab=artifacts#meta) on refs/heads/master @ [7b30bd86ff65](https://github.com/cockroachdb/pebble/commits/7b30bd86ff65):
```
=== CONT TestMeta/execution/standard-023
=== RUN TestMeta/compare/standard-012
=== RUN TestMeta/execution
=== RUN TestMeta/execution/standard-010
=== PAUSE TestMeta/execution/standard-010
=== CONT TestMeta/execution/standard-010
=== RUN TestMeta/execution/random-024
=== PAUSE TestMeta/execution/random-024
=== CONT TestMeta/execution/random-024
=== RUN TestMeta/execution/random-013
=== PAUSE TestMeta/execution/random-013
=== CONT TestMeta/execution/random-013
=== RUN TestMeta/execution/random-023
=== PAUSE TestMeta/execution/random-023
=== CONT TestMeta/execution/random-023
=== RUN TestMeta/execution/random-015
=== PAUSE TestMeta/execution/random-015
=== CONT TestMeta/execution/random-015
=== RUN TestMeta/execution/random-020
=== PAUSE TestMeta/execution/random-020
=== CONT TestMeta/execution/random-020
=== RUN TestMeta/execution/standard-021
=== PAUSE TestMeta/execution/standard-021
=== CONT TestMeta/execution/standard-021
=== RUN TestMeta/execution/random-007
=== PAUSE TestMeta/execution/random-007
=== CONT TestMeta/execution/random-007
=== RUN TestMeta/execution/random-012
=== PAUSE TestMeta/execution/random-012
=== CONT TestMeta/execution/random-012
=== RUN TestMeta/execution/random-006
=== PAUSE TestMeta/execution/random-006
=== CONT TestMeta/execution/random-006
=== RUN TestMeta/execution/standard-011
=== PAUSE TestMeta/execution/standard-011
=== CONT TestMeta/execution/standard-011
=== RUN TestMeta/execution/standard-020
=== PAUSE TestMeta/execution/standard-020
=== CONT TestMeta/execution/standard-020
=== RUN TestMeta/compare/standard-007
=== RUN TestMeta/compare/standard-008
=== RUN TestMeta/execution/standard-002
=== PAUSE TestMeta/execution/standard-002
=== CONT TestMeta/execution/standard-002
=== RUN TestMeta/execution/random-008
=== PAUSE TestMeta/execution/random-008
=== CONT TestMeta/execution/random-008
=== RUN TestMeta/execution/random-014
=== PAUSE TestMeta/execution/random-014
=== CONT TestMeta/execution/random-014
```
<details><summary>Help</summary>
<p>
To reproduce, try:
```bash
go test -mod=vendor -tags 'invariants' -exec 'stress -p 1' -timeout 0 -test.v -run TestMeta$ ./internal/metamorphic -seed 1668072880985143374 -ops "uniform:5000-10000"
```
</p>
</details>
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestMeta.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
github.com/cockroachdb/pebble/internal/metamorphic: TestMeta failed - github.com/cockroachdb/pebble/internal/metamorphic.TestMeta [failed](https://buildConfiguration/Cockroach_Nightlies_Pebble_Metamorphic/7423668?buildTab=log) with [artifacts](https://buildConfiguration/Cockroach_Nightlies_Pebble_Metamorphic/7423668?buildTab=artifacts#meta) on refs/heads/master @ [7b30bd86ff65](https://github.com/cockroachdb/pebble/commits/7b30bd86ff65):
```
=== CONT TestMeta/execution/standard-023
=== RUN TestMeta/compare/standard-012
=== RUN TestMeta/execution
=== RUN TestMeta/execution/standard-010
=== PAUSE TestMeta/execution/standard-010
=== CONT TestMeta/execution/standard-010
=== RUN TestMeta/execution/random-024
=== PAUSE TestMeta/execution/random-024
=== CONT TestMeta/execution/random-024
=== RUN TestMeta/execution/random-013
=== PAUSE TestMeta/execution/random-013
=== CONT TestMeta/execution/random-013
=== RUN TestMeta/execution/random-023
=== PAUSE TestMeta/execution/random-023
=== CONT TestMeta/execution/random-023
=== RUN TestMeta/execution/random-015
=== PAUSE TestMeta/execution/random-015
=== CONT TestMeta/execution/random-015
=== RUN TestMeta/execution/random-020
=== PAUSE TestMeta/execution/random-020
=== CONT TestMeta/execution/random-020
=== RUN TestMeta/execution/standard-021
=== PAUSE TestMeta/execution/standard-021
=== CONT TestMeta/execution/standard-021
=== RUN TestMeta/execution/random-007
=== PAUSE TestMeta/execution/random-007
=== CONT TestMeta/execution/random-007
=== RUN TestMeta/execution/random-012
=== PAUSE TestMeta/execution/random-012
=== CONT TestMeta/execution/random-012
=== RUN TestMeta/execution/random-006
=== PAUSE TestMeta/execution/random-006
=== CONT TestMeta/execution/random-006
=== RUN TestMeta/execution/standard-011
=== PAUSE TestMeta/execution/standard-011
=== CONT TestMeta/execution/standard-011
=== RUN TestMeta/execution/standard-020
=== PAUSE TestMeta/execution/standard-020
=== CONT TestMeta/execution/standard-020
=== RUN TestMeta/compare/standard-007
=== RUN TestMeta/compare/standard-008
=== RUN TestMeta/execution/standard-002
=== PAUSE TestMeta/execution/standard-002
=== CONT TestMeta/execution/standard-002
=== RUN TestMeta/execution/random-008
=== PAUSE TestMeta/execution/random-008
=== CONT TestMeta/execution/random-008
=== RUN TestMeta/execution/random-014
=== PAUSE TestMeta/execution/random-014
=== CONT TestMeta/execution/random-014
```
<details><summary>Help</summary>
<p>
To reproduce, try:
```bash
go test -mod=vendor -tags 'invariants' -exec 'stress -p 1' -timeout 0 -test.v -run TestMeta$ ./internal/metamorphic -seed 1668072880985143374 -ops "uniform:5000-10000"
```
</p>
</details>
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestMeta.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
non_process
|
github com cockroachdb pebble internal metamorphic testmeta failed github com cockroachdb pebble internal metamorphic testmeta with on refs heads master cont testmeta execution standard run testmeta compare standard run testmeta execution run testmeta execution standard pause testmeta execution standard cont testmeta execution standard run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution standard pause testmeta execution standard cont testmeta execution standard run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution standard pause testmeta execution standard cont testmeta execution standard run testmeta execution standard pause testmeta execution standard cont testmeta execution standard run testmeta compare standard run testmeta compare standard run testmeta execution standard pause testmeta execution standard cont testmeta execution standard run testmeta execution random pause testmeta execution random cont testmeta execution random run testmeta execution random pause testmeta execution random cont testmeta execution random help to reproduce try bash go test mod vendor tags invariants exec stress p timeout test v run testmeta internal metamorphic seed ops uniform
| 0
|
390,566
| 26,866,379,752
|
IssuesEvent
|
2023-02-04 00:35:15
|
yugabyte/yugabyte-db
|
https://api.github.com/repos/yugabyte/yugabyte-db
|
closed
|
[docs] [infra] allow for unversioned content areas
|
area/documentation
|
Relates to #10790...
If we move version-independent sections (like cloud, faq, releases, and others) into a separate folder, we need to figure out to build a complete left nav that incorporates those version-independent sections.
The left nav should pull its content from the version you're currently in (latest, stable, v2.6, etc.), and default to pulling from latest if you're currently IN an unversioned section.
|
1.0
|
[docs] [infra] allow for unversioned content areas - Relates to #10790...
If we move version-independent sections (like cloud, faq, releases, and others) into a separate folder, we need to figure out to build a complete left nav that incorporates those version-independent sections.
The left nav should pull its content from the version you're currently in (latest, stable, v2.6, etc.), and default to pulling from latest if you're currently IN an unversioned section.
|
non_process
|
allow for unversioned content areas relates to if we move version independent sections like cloud faq releases and others into a separate folder we need to figure out to build a complete left nav that incorporates those version independent sections the left nav should pull its content from the version you re currently in latest stable etc and default to pulling from latest if you re currently in an unversioned section
| 0
|
15,615
| 19,753,063,200
|
IssuesEvent
|
2022-01-15 09:03:41
|
googleapis/java-pubsublite-spark
|
https://api.github.com/repos/googleapis/java-pubsublite-spark
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'pubsublite-spark' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'pubsublite-spark' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname pubsublite spark invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
| 1
|
22,066
| 30,591,680,572
|
IssuesEvent
|
2023-07-21 17:37:40
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
transform process won't load in metrics pipeline since v0.71.0+
|
bug processor/transform needs triage
|
### Component(s)
processor/transform
### What happened?
## Description
The [transform](https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/transformprocessor) processor won't load into the `metrics` pipeline since version 0.71.0.
## Steps to Reproduce
1. Copy config below into `agent.yaml`
2. Run `otelcol-contrib --config agent.yaml`
3. Check the logs and see that `transform` is not loaded
*NOTES*: The `filter` processor is loaded but not the `transform` processor.
## Expected Result
Transform processor should be loaded.
## Actual Result
Transform processor didn't load.
### Collector version
0.81.0
### Environment information
## Environment
OS: macOS Ventura 13.4
Also tried the container and Ubuntu, same issue.
### OpenTelemetry Collector configuration
```yaml
receivers:
hostmetrics:
collection_interval: 30s
scrapers:
load:
processors:
filter:
metrics:
include:
match_type: regexp
metric_names:
- jfrt_.*
transform:
metric_statements:
- context: resource
statements:
- delete_key(attributes, "service.instance.id")
- context: datapoint
statements:
- delete_key(attributes, "service.instance.id")
exporters:
logging:
service:
pipelines:
metrics:
receivers:
- hostmetrics
processors:
- filter
- transform
exporters:
- logging
```
### Log output
```shell
2023-07-20T14:03:05.008-0400 info service/telemetry.go:81 Setting up own telemetry...
2023-07-20T14:03:05.008-0400 info service/telemetry.go:104 Serving Prometheus metrics {"address": ":8888", "level": "Basic"}
2023-07-20T14:03:05.008-0400 info exporter@v0.81.0/exporter.go:275 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"}
2023-07-20T14:03:05.008-0400 info filterprocessor@v0.81.0/metrics.go:89 Metric filter configured {"kind": "processor", "name": "filter", "pipeline": "metrics", "include match_type": "regexp", "in
clude expressions": [], "include metric names": ["jfrt_.*"], "include metrics with resource attributes": null, "exclude match_type": "", "exclude expressions": [], "exclude metric names": [], "exclude metrics w
ith resource attributes": null}
2023-07-20T14:03:05.008-0400 info service/service.go:131 Starting otelcol-contrib... {"Version": "0.81.0", "NumCPU": 8}
2023-07-20T14:03:05.008-0400 info extensions/extensions.go:30 Starting extensions...
2023-07-20T14:03:05.008-0400 info service/service.go:148 Everything is ready. Begin running and processing data.
```
### Additional context
Below is the log for version 0.70.0 and `transform` processor is loaded.
```console
2023-07-20T14:09:38.895-0400 info service/telemetry.go:90 Setting up own telemetry...
2023-07-20T14:09:38.895-0400 info service/telemetry.go:116 Serving Prometheus metrics {"address": ":8888", "level": "Basic"}
2023-07-20T14:09:38.895-0400 info exporter/exporter.go:290 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"}
2023-07-20T14:09:38.896-0400 info filterprocessor@v0.70.0/metrics.go:97 Metric filter configured {"kind": "processor", "name": "filter", "pipeline": "metrics", "include match_type": "regexp", "in
clude expressions": [], "include metric names": ["jfrt_.*"], "include metrics with resource attributes": null, "exclude match_type": "", "exclude expressions": [], "exclude metric names": [], "exclude metrics w
ith resource attributes": null}
2023-07-20T14:09:38.897-0400 info service/service.go:128 Starting otelcol-contrib... {"Version": "0.70.0", "NumCPU": 8}
2023-07-20T14:09:38.897-0400 info extensions/extensions.go:41 Starting extensions...
2023-07-20T14:09:38.897-0400 info service/pipelines.go:86 Starting exporters...
2023-07-20T14:09:38.897-0400 info service/pipelines.go:90 Exporter is starting... {"kind": "exporter", "data_type": "metrics", "name": "logging"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:94 Exporter started. {"kind": "exporter", "data_type": "metrics", "name": "logging"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:98 Starting processors...
2023-07-20T14:09:38.897-0400 info service/pipelines.go:102 Processor is starting... {"kind": "processor", "name": "transform", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:106 Processor started. {"kind": "processor", "name": "transform", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:102 Processor is starting... {"kind": "processor", "name": "filter", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:106 Processor started. {"kind": "processor", "name": "filter", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:110 Starting receivers...
2023-07-20T14:09:38.897-0400 info service/pipelines.go:114 Receiver is starting... {"kind": "receiver", "name": "hostmetrics", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:118 Receiver started. {"kind": "receiver", "name": "hostmetrics", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/service.go:145 Everything is ready. Begin running and processing data.
```
|
1.0
|
transform process won't load in metrics pipeline since v0.71.0+ - ### Component(s)
processor/transform
### What happened?
## Description
The [transform](https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/transformprocessor) processor won't load into the `metrics` pipeline since version 0.71.0.
## Steps to Reproduce
1. Copy config below into `agent.yaml`
2. Run `otelcol-contrib --config agent.yaml`
3. Check the logs and see that `transform` is not loaded
*NOTES*: The `filter` processor is loaded but not the `transform` processor.
## Expected Result
Transform processor should be loaded.
## Actual Result
Transform processor didn't load.
### Collector version
0.81.0
### Environment information
## Environment
OS: macOS Ventura 13.4
Also tried the container and Ubuntu, same issue.
### OpenTelemetry Collector configuration
```yaml
receivers:
hostmetrics:
collection_interval: 30s
scrapers:
load:
processors:
filter:
metrics:
include:
match_type: regexp
metric_names:
- jfrt_.*
transform:
metric_statements:
- context: resource
statements:
- delete_key(attributes, "service.instance.id")
- context: datapoint
statements:
- delete_key(attributes, "service.instance.id")
exporters:
logging:
service:
pipelines:
metrics:
receivers:
- hostmetrics
processors:
- filter
- transform
exporters:
- logging
```
### Log output
```shell
2023-07-20T14:03:05.008-0400 info service/telemetry.go:81 Setting up own telemetry...
2023-07-20T14:03:05.008-0400 info service/telemetry.go:104 Serving Prometheus metrics {"address": ":8888", "level": "Basic"}
2023-07-20T14:03:05.008-0400 info exporter@v0.81.0/exporter.go:275 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"}
2023-07-20T14:03:05.008-0400 info filterprocessor@v0.81.0/metrics.go:89 Metric filter configured {"kind": "processor", "name": "filter", "pipeline": "metrics", "include match_type": "regexp", "in
clude expressions": [], "include metric names": ["jfrt_.*"], "include metrics with resource attributes": null, "exclude match_type": "", "exclude expressions": [], "exclude metric names": [], "exclude metrics w
ith resource attributes": null}
2023-07-20T14:03:05.008-0400 info service/service.go:131 Starting otelcol-contrib... {"Version": "0.81.0", "NumCPU": 8}
2023-07-20T14:03:05.008-0400 info extensions/extensions.go:30 Starting extensions...
2023-07-20T14:03:05.008-0400 info service/service.go:148 Everything is ready. Begin running and processing data.
```
### Additional context
Below is the log for version 0.70.0 and `transform` processor is loaded.
```console
2023-07-20T14:09:38.895-0400 info service/telemetry.go:90 Setting up own telemetry...
2023-07-20T14:09:38.895-0400 info service/telemetry.go:116 Serving Prometheus metrics {"address": ":8888", "level": "Basic"}
2023-07-20T14:09:38.895-0400 info exporter/exporter.go:290 Development component. May change in the future. {"kind": "exporter", "data_type": "metrics", "name": "logging"}
2023-07-20T14:09:38.896-0400 info filterprocessor@v0.70.0/metrics.go:97 Metric filter configured {"kind": "processor", "name": "filter", "pipeline": "metrics", "include match_type": "regexp", "in
clude expressions": [], "include metric names": ["jfrt_.*"], "include metrics with resource attributes": null, "exclude match_type": "", "exclude expressions": [], "exclude metric names": [], "exclude metrics w
ith resource attributes": null}
2023-07-20T14:09:38.897-0400 info service/service.go:128 Starting otelcol-contrib... {"Version": "0.70.0", "NumCPU": 8}
2023-07-20T14:09:38.897-0400 info extensions/extensions.go:41 Starting extensions...
2023-07-20T14:09:38.897-0400 info service/pipelines.go:86 Starting exporters...
2023-07-20T14:09:38.897-0400 info service/pipelines.go:90 Exporter is starting... {"kind": "exporter", "data_type": "metrics", "name": "logging"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:94 Exporter started. {"kind": "exporter", "data_type": "metrics", "name": "logging"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:98 Starting processors...
2023-07-20T14:09:38.897-0400 info service/pipelines.go:102 Processor is starting... {"kind": "processor", "name": "transform", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:106 Processor started. {"kind": "processor", "name": "transform", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:102 Processor is starting... {"kind": "processor", "name": "filter", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:106 Processor started. {"kind": "processor", "name": "filter", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:110 Starting receivers...
2023-07-20T14:09:38.897-0400 info service/pipelines.go:114 Receiver is starting... {"kind": "receiver", "name": "hostmetrics", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/pipelines.go:118 Receiver started. {"kind": "receiver", "name": "hostmetrics", "pipeline": "metrics"}
2023-07-20T14:09:38.897-0400 info service/service.go:145 Everything is ready. Begin running and processing data.
```
|
process
|
transform process won t load in metrics pipeline since component s processor transform what happened description the processor won t load into the metrics pipeline since version steps to reproduce copy config below into agent yaml run otelcol contrib config agent yaml check the logs and see that transform is not loaded notes the filter processor is loaded but not the transform processor expected result transform processor should be loaded actual result transform processor didn t load collector version environment information environment os macos ventura also tried the container and ubuntu same issue opentelemetry collector configuration yaml receivers hostmetrics collection interval scrapers load processors filter metrics include match type regexp metric names jfrt transform metric statements context resource statements delete key attributes service instance id context datapoint statements delete key attributes service instance id exporters logging service pipelines metrics receivers hostmetrics processors filter transform exporters logging log output shell info service telemetry go setting up own telemetry info service telemetry go serving prometheus metrics address level basic info exporter exporter go development component may change in the future kind exporter data type metrics name logging info filterprocessor metrics go metric filter configured kind processor name filter pipeline metrics include match type regexp in clude expressions include metric names include metrics with resource attributes null exclude match type exclude expressions exclude metric names exclude metrics w ith resource attributes null info service service go starting otelcol contrib version numcpu info extensions extensions go starting extensions info service service go everything is ready begin running and processing data additional context below is the log for version and transform processor is loaded console info service telemetry go setting up own telemetry info service telemetry go serving prometheus metrics address level basic info exporter exporter go development component may change in the future kind exporter data type metrics name logging info filterprocessor metrics go metric filter configured kind processor name filter pipeline metrics include match type regexp in clude expressions include metric names include metrics with resource attributes null exclude match type exclude expressions exclude metric names exclude metrics w ith resource attributes null info service service go starting otelcol contrib version numcpu info extensions extensions go starting extensions info service pipelines go starting exporters info service pipelines go exporter is starting kind exporter data type metrics name logging info service pipelines go exporter started kind exporter data type metrics name logging info service pipelines go starting processors info service pipelines go processor is starting kind processor name transform pipeline metrics info service pipelines go processor started kind processor name transform pipeline metrics info service pipelines go processor is starting kind processor name filter pipeline metrics info service pipelines go processor started kind processor name filter pipeline metrics info service pipelines go starting receivers info service pipelines go receiver is starting kind receiver name hostmetrics pipeline metrics info service pipelines go receiver started kind receiver name hostmetrics pipeline metrics info service service go everything is ready begin running and processing data
| 1
|
14,462
| 17,568,534,799
|
IssuesEvent
|
2021-08-14 07:19:31
|
Geoxor/Sakuria
|
https://api.github.com/repos/Geoxor/Sakuria
|
opened
|
Improve the GIF Encoder's performance
|
bug image processors
|
This shit is really bottlenecking the render speed

|
1.0
|
Improve the GIF Encoder's performance - This shit is really bottlenecking the render speed

|
process
|
improve the gif encoder s performance this shit is really bottlenecking the render speed
| 1
|
16,942
| 22,294,718,295
|
IssuesEvent
|
2022-06-12 21:56:48
|
0xffset/rOSt
|
https://api.github.com/repos/0xffset/rOSt
|
closed
|
Memory freeing on process termination
|
memory processes
|
Currently when a process is created the frames allocated to it are never freed back. We need to add freeing the process's memory after it's terminated.
To do that we would probably need to rewrite the frame allocator to support deallocation.
|
1.0
|
Memory freeing on process termination - Currently when a process is created the frames allocated to it are never freed back. We need to add freeing the process's memory after it's terminated.
To do that we would probably need to rewrite the frame allocator to support deallocation.
|
process
|
memory freeing on process termination currently when a process is created the frames allocated to it are never freed back we need to add freeing the process s memory after it s terminated to do that we would probably need to rewrite the frame allocator to support deallocation
| 1
|
3,720
| 6,732,884,771
|
IssuesEvent
|
2017-10-18 13:11:09
|
lockedata/rcms
|
https://api.github.com/repos/lockedata/rcms
|
opened
|
Build agenda
|
conference team odoo processes
|
## Detailed task
- Create a schedule over multiple rooms (and days if required)
- Publish agenda
## Assessing the task
Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks.
Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback.
## Extra Info
- Site: [odoo](//http://188.166.159.192:8069)
- System documentation: [odoo docs](https://www.odoo.com/page/docs)
- Role: Conference team
- Area: Processes
|
1.0
|
Build agenda - ## Detailed task
- Create a schedule over multiple rooms (and days if required)
- Publish agenda
## Assessing the task
Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks.
Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback.
## Extra Info
- Site: [odoo](//http://188.166.159.192:8069)
- System documentation: [odoo docs](https://www.odoo.com/page/docs)
- Role: Conference team
- Area: Processes
|
process
|
build agenda detailed task create a schedule over multiple rooms and days if required publish agenda assessing the task try to perform the task use google and the system documentation to help part of what we re trying to assess how easy it is for people to work out how to do tasks use a 👍 reaction to this task if you were able to perform the task use a 👎 reaction to the task if you could not complete it add a reply with any comments or feedback extra info site system documentation role conference team area processes
| 1
|
171,174
| 20,948,546,491
|
IssuesEvent
|
2022-03-26 08:18:02
|
npenin/thewheel-dotnet
|
https://api.github.com/repos/npenin/thewheel-dotnet
|
opened
|
CVE-2017-0249 (High) detected in system.net.http.4.3.0.nupkg
|
security vulnerability
|
## CVE-2017-0249 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- mstest.testframework.2.2.3.nupkg (Root Library)
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/682d2be920def33815c7fdb6916d426a96c8df76">682d2be920def33815c7fdb6916d426a96c8df76</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An elevation of privilege vulnerability exists when the ASP.NET Core fails to properly sanitize web requests.
<p>Publish Date: 2017-05-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-0249>CVE-2017-0249</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/aspnet/Announcements/issues/239">https://github.com/aspnet/Announcements/issues/239</a></p>
<p>Release Date: 2017-05-12</p>
<p>Fix Resolution: System.Text.Encodings.Web - 4.0.1,4.3.1;System.Net.Http - 4.1.2,4.3.2;System.Net.Http.WinHttpHandler - 4.0.2,4.3.1;System.Net.Security - 4.0.1,4.3.1;System.Net.WebSockets.Client - 4.0.1,4.3.1;Microsoft.AspNetCore.Mvc - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Core - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Abstractions - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ApiExplorer - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Cors - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.DataAnnotations - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Json - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Xml - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Localization - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor.Host - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.TagHelpers - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ViewFeatures - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.WebApiCompatShim - 1.0.4,1.1.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-0249 (High) detected in system.net.http.4.3.0.nupkg - ## CVE-2017-0249 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /TheWheel.Tests/TheWheel.Tests.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- mstest.testframework.2.2.3.nupkg (Root Library)
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/npenin/thewheel-dotnet/commit/682d2be920def33815c7fdb6916d426a96c8df76">682d2be920def33815c7fdb6916d426a96c8df76</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An elevation of privilege vulnerability exists when the ASP.NET Core fails to properly sanitize web requests.
<p>Publish Date: 2017-05-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-0249>CVE-2017-0249</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/aspnet/Announcements/issues/239">https://github.com/aspnet/Announcements/issues/239</a></p>
<p>Release Date: 2017-05-12</p>
<p>Fix Resolution: System.Text.Encodings.Web - 4.0.1,4.3.1;System.Net.Http - 4.1.2,4.3.2;System.Net.Http.WinHttpHandler - 4.0.2,4.3.1;System.Net.Security - 4.0.1,4.3.1;System.Net.WebSockets.Client - 4.0.1,4.3.1;Microsoft.AspNetCore.Mvc - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Core - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Abstractions - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ApiExplorer - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Cors - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.DataAnnotations - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Json - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Xml - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Localization - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor.Host - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.TagHelpers - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ViewFeatures - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.WebApiCompatShim - 1.0.4,1.1.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in system net http nupkg cve high severity vulnerability vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that library home page a href path to dependency file thewheel tests thewheel tests csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg dependency hierarchy mstest testframework nupkg root library netstandard library nupkg x system net http nupkg vulnerable library found in head commit a href found in base branch master vulnerability details an elevation of privilege vulnerability exists when the asp net core fails to properly sanitize web requests publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution system text encodings web system net http system net http winhttphandler system net security system net websockets client microsoft aspnetcore mvc microsoft aspnetcore mvc core microsoft aspnetcore mvc abstractions microsoft aspnetcore mvc apiexplorer microsoft aspnetcore mvc cors microsoft aspnetcore mvc dataannotations microsoft aspnetcore mvc formatters json microsoft aspnetcore mvc formatters xml microsoft aspnetcore mvc localization microsoft aspnetcore mvc razor host microsoft aspnetcore mvc razor microsoft aspnetcore mvc taghelpers microsoft aspnetcore mvc viewfeatures microsoft aspnetcore mvc webapicompatshim step up your open source security game with whitesource
| 0
|
16,859
| 22,139,931,280
|
IssuesEvent
|
2022-06-03 05:21:57
|
deepset-ai/haystack
|
https://api.github.com/repos/deepset-ai/haystack
|
closed
|
split_length not working as expected
|
type:question topic:preprocessing journey:intermediate
|
**Question**
I am trying to split my documents such that I have multiple sentences with 500 words each. But after executing the `PreProcessor` function, I get the warning:
```
> 0%| | 0/1052 [00:00<?, ?docs/s]WARNING - haystack.nodes.preprocessor.preprocessor - One or more sentence found with word count higher than the split length.
```
Upon checking the count of sentences, I see that nothing was split. I fed in 1052 documents and the resulting list also has 1052 elements. What am I doing wrong here?
This is the code I am using:
```
def pre_process(source):
all_docs = convert_files_to_docs(dir_path=source)
preprocessor = PreProcessor(
clean_empty_lines=True,
clean_whitespace=True,
clean_header_footer=False,
split_by="word",
split_length=500,
split_respect_sentence_boundary=True,
)
processed_docs = preprocessor.process(all_docs)
print(f"Number of input files: {len(all_docs)}\nNumber of output files: {len(processed_docs)}")
return [item.content for item in processed_docs]
```
**FAQ Check**
- [x] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)?
|
1.0
|
split_length not working as expected - **Question**
I am trying to split my documents such that I have multiple sentences with 500 words each. But after executing the `PreProcessor` function, I get the warning:
```
> 0%| | 0/1052 [00:00<?, ?docs/s]WARNING - haystack.nodes.preprocessor.preprocessor - One or more sentence found with word count higher than the split length.
```
Upon checking the count of sentences, I see that nothing was split. I fed in 1052 documents and the resulting list also has 1052 elements. What am I doing wrong here?
This is the code I am using:
```
def pre_process(source):
all_docs = convert_files_to_docs(dir_path=source)
preprocessor = PreProcessor(
clean_empty_lines=True,
clean_whitespace=True,
clean_header_footer=False,
split_by="word",
split_length=500,
split_respect_sentence_boundary=True,
)
processed_docs = preprocessor.process(all_docs)
print(f"Number of input files: {len(all_docs)}\nNumber of output files: {len(processed_docs)}")
return [item.content for item in processed_docs]
```
**FAQ Check**
- [x] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)?
|
process
|
split length not working as expected question i am trying to split my documents such that i have multiple sentences with words each but after executing the preprocessor function i get the warning warning haystack nodes preprocessor preprocessor one or more sentence found with word count higher than the split length upon checking the count of sentences i see that nothing was split i fed in documents and the resulting list also has elements what am i doing wrong here this is the code i am using def pre process source all docs convert files to docs dir path source preprocessor preprocessor clean empty lines true clean whitespace true clean header footer false split by word split length split respect sentence boundary true processed docs preprocessor process all docs print f number of input files len all docs nnumber of output files len processed docs return faq check have you had a look at
| 1
|
17,796
| 23,723,833,162
|
IssuesEvent
|
2022-08-30 17:38:59
|
hashgraph/hedera-mirror-node
|
https://api.github.com/repos/hashgraph/hedera-mirror-node
|
opened
|
Release checklist 0.64
|
enhancement process
|
### Problem
We need a checklist to verify the release is rolled out successfully.
### Solution
## Preparation
- [x] Milestone field populated on relevant [issues](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aclosed+no%3Amilestone+sort%3Aupdated-desc)
- [x] Nothing open for [milestone](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aopen+sort%3Aupdated-desc+milestone%3A0.64.0)
- [x] GitHub checks for branch are passing
- [x] Automated Kubernetes deployment successful
- [x] Tag release
- [x] Upload release artifacts
- [x] Publish release
## Integration
- [ ] Deploy to VM
## Performance
- [ ] Deploy to Kubernetes
- [ ] Deploy to VM
- [ ] gRPC API performance tests
- [ ] Importer performance tests
- [ ] REST API performance tests
- [ ] Migrations tested against mainnet clone
## Previewnet
- [ ] Deploy to VM
## Staging
- [ ] Deploy to Kubernetes EU
- [ ] Deploy to Kubernetes NA
## Testnet
- [ ] Deploy to VM
## Mainnet
- [ ] Deploy to Kubernetes EU
- [ ] Deploy to Kubernetes NA
- [ ] Deploy to VM
- [ ] Deploy to ETL
### Alternatives
_No response_
|
1.0
|
Release checklist 0.64 - ### Problem
We need a checklist to verify the release is rolled out successfully.
### Solution
## Preparation
- [x] Milestone field populated on relevant [issues](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aclosed+no%3Amilestone+sort%3Aupdated-desc)
- [x] Nothing open for [milestone](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aopen+sort%3Aupdated-desc+milestone%3A0.64.0)
- [x] GitHub checks for branch are passing
- [x] Automated Kubernetes deployment successful
- [x] Tag release
- [x] Upload release artifacts
- [x] Publish release
## Integration
- [ ] Deploy to VM
## Performance
- [ ] Deploy to Kubernetes
- [ ] Deploy to VM
- [ ] gRPC API performance tests
- [ ] Importer performance tests
- [ ] REST API performance tests
- [ ] Migrations tested against mainnet clone
## Previewnet
- [ ] Deploy to VM
## Staging
- [ ] Deploy to Kubernetes EU
- [ ] Deploy to Kubernetes NA
## Testnet
- [ ] Deploy to VM
## Mainnet
- [ ] Deploy to Kubernetes EU
- [ ] Deploy to Kubernetes NA
- [ ] Deploy to VM
- [ ] Deploy to ETL
### Alternatives
_No response_
|
process
|
release checklist problem we need a checklist to verify the release is rolled out successfully solution preparation milestone field populated on relevant nothing open for github checks for branch are passing automated kubernetes deployment successful tag release upload release artifacts publish release integration deploy to vm performance deploy to kubernetes deploy to vm grpc api performance tests importer performance tests rest api performance tests migrations tested against mainnet clone previewnet deploy to vm staging deploy to kubernetes eu deploy to kubernetes na testnet deploy to vm mainnet deploy to kubernetes eu deploy to kubernetes na deploy to vm deploy to etl alternatives no response
| 1
|
725,085
| 24,950,885,612
|
IssuesEvent
|
2022-11-01 07:06:35
|
AY2223S1-CS2103T-T12-3/tp
|
https://api.github.com/repos/AY2223S1-CS2103T-T12-3/tp
|
closed
|
[PE-D][Tester B] Difference in date behavior in UG and in programme
|
bug priority.High
|
In the user guide, it is stated that only one date can be added, while multiple tags can be added:

However, if multiple dates are added, the hobby will adopt multiple dates:

Select also shows the first date added, as opposed to the last date (in this case feb 2 2022 is shown as opposed to feb 1 2022:

<!--session: 1666944900526-b54b72e7-f84e-408d-9338-6c15585ac9ca-->
<!--Version: Web v3.4.4-->
-------------
Labels: `type.FunctionalityBug` `severity.High`
original: farisanadia/ped#1
|
1.0
|
[PE-D][Tester B] Difference in date behavior in UG and in programme - In the user guide, it is stated that only one date can be added, while multiple tags can be added:

However, if multiple dates are added, the hobby will adopt multiple dates:

Select also shows the first date added, as opposed to the last date (in this case feb 2 2022 is shown as opposed to feb 1 2022:

<!--session: 1666944900526-b54b72e7-f84e-408d-9338-6c15585ac9ca-->
<!--Version: Web v3.4.4-->
-------------
Labels: `type.FunctionalityBug` `severity.High`
original: farisanadia/ped#1
|
non_process
|
difference in date behavior in ug and in programme in the user guide it is stated that only one date can be added while multiple tags can be added however if multiple dates are added the hobby will adopt multiple dates select also shows the first date added as opposed to the last date in this case feb is shown as opposed to feb labels type functionalitybug severity high original farisanadia ped
| 0
|
747,953
| 26,102,415,301
|
IssuesEvent
|
2022-12-27 08:55:46
|
bounswe/bounswe2022group1
|
https://api.github.com/repos/bounswe/bounswe2022group1
|
opened
|
Checking PR and Merging Frontend Branch to Master
|
Priority: Critical Type: Task Status: In Progress Frontend
|
**Issue Description:**
We are going to check PR from `frontend` to `master` for latest version of our frontend app, resolve conflicts and merge this branch. Thanks all of my teaammates for their efforts, health to our hands 🤝
**Tasks to Do:**
- [ ] add issue labels
- [ ] add related links
- [ ] check PR
- [ ] resolve conflicts
- [ ] merge `frontend` to `master`
*Task Deadline: 27/12/2022 12:15 pm*
*Final Situation:*
|
1.0
|
Checking PR and Merging Frontend Branch to Master - **Issue Description:**
We are going to check PR from `frontend` to `master` for latest version of our frontend app, resolve conflicts and merge this branch. Thanks all of my teaammates for their efforts, health to our hands 🤝
**Tasks to Do:**
- [ ] add issue labels
- [ ] add related links
- [ ] check PR
- [ ] resolve conflicts
- [ ] merge `frontend` to `master`
*Task Deadline: 27/12/2022 12:15 pm*
*Final Situation:*
|
non_process
|
checking pr and merging frontend branch to master issue description we are going to check pr from frontend to master for latest version of our frontend app resolve conflicts and merge this branch thanks all of my teaammates for their efforts health to our hands 🤝 tasks to do add issue labels add related links check pr resolve conflicts merge frontend to master task deadline pm final situation
| 0
|
511,368
| 14,859,018,510
|
IssuesEvent
|
2021-01-18 17:43:33
|
Sphereserver/Source-X
|
https://api.github.com/repos/Sphereserver/Source-X
|
closed
|
Server Loop if wrong .set <SKILL NUMBER> Value
|
Priority: Medium Status-Bug: Confirmed Status: Done - waiting feedback
|
If for mistake try to use command .set <SKILLID> whit invalid skill ID, server enter in loop error.
**How reproduce:**
- .set 100
- Target self
**Console error:**
```
17:27:1:'a' commands 'set 100'=1
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CNetworkManager::tick | +0
17:27:DEBUG:(03172) T_Main | 1 | CNetworkThread::tick | +9
17:27:DEBUG:(03172) T_Main | 2 | CNetworkInput::processInput | +0
17:27:DEBUG:(03172) T_Main | 3 | CNetworkInput::processData | +1487
17:27:DEBUG:(03172) T_Main | 4 | CNetworkInput::processData | +0
17:27:DEBUG:(03172) T_Main | 5 | CNetworkInput::processGameClientData | +1
17:27:DEBUG:(03172) T_Main | 6 | PacketTarget::onReceive | +1
17:27:DEBUG:(03172) T_Main | 7 | CClient::Event_Target | +1
17:27:DEBUG:(03172) T_Main | 8 | CClient::OnTarg_Obj_Set | +1
17:27:DEBUG:(03172) T_Main | 9 | CChar::r_Verb | +7
17:27:DEBUG:(03172) T_Main | 10 | CObjBase::r_Verb | +3
17:27:DEBUG:(03172) T_Main | 11 | CScriptObj::r_Verb | +3
17:27:DEBUG:(03172) T_Main | 12 | CChar::r_LoadVal | +3
17:27:DEBUG:(03172) T_Main | 13 | CChar::Skill_SetBase | +5
17:27:DEBUG:(03172) T_Main | 14 | CClient::addSkillWindow | +0
17:27:DEBUG:(03172) T_Main | 15 | PacketSkills::PacketSkills | +1 <-- last function call (stack unwinding began here)
17:27:CRITICAL:"Access Violation" (0x1b2074), in ExcType=CSError catched in CChar::LoadVal() #3 "Keyword"
17:27:DEBUG:command '100' args ''
17:27:'a' commands uid=0307 (B) to '100'=0
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +9
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +1
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +2
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +8
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +1
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +12
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:CRITICAL:"Access Violation" (0x1ae76b), in ExcType=CSError catched in CChar::OnTickPeriodic() #1 "update stats"
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::CheckLocation | +10
17:27:DEBUG:(03172) T_Main | 3 | CItemMultiCustom::GetDesignArea | +3268 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +15
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +20
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +3 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +16
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:CRITICAL:"Access Violation" (0x1ae76b), in ExcType=CSError catched in CChar::OnTickPeriodic() #1 "update stats"
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +16
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +2
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
```
**Extra info:**
- Tested on build 3381
- Tested whit clear script pack
|
1.0
|
Server Loop if wrong .set <SKILL NUMBER> Value - If for mistake try to use command .set <SKILLID> whit invalid skill ID, server enter in loop error.
**How reproduce:**
- .set 100
- Target self
**Console error:**
```
17:27:1:'a' commands 'set 100'=1
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CNetworkManager::tick | +0
17:27:DEBUG:(03172) T_Main | 1 | CNetworkThread::tick | +9
17:27:DEBUG:(03172) T_Main | 2 | CNetworkInput::processInput | +0
17:27:DEBUG:(03172) T_Main | 3 | CNetworkInput::processData | +1487
17:27:DEBUG:(03172) T_Main | 4 | CNetworkInput::processData | +0
17:27:DEBUG:(03172) T_Main | 5 | CNetworkInput::processGameClientData | +1
17:27:DEBUG:(03172) T_Main | 6 | PacketTarget::onReceive | +1
17:27:DEBUG:(03172) T_Main | 7 | CClient::Event_Target | +1
17:27:DEBUG:(03172) T_Main | 8 | CClient::OnTarg_Obj_Set | +1
17:27:DEBUG:(03172) T_Main | 9 | CChar::r_Verb | +7
17:27:DEBUG:(03172) T_Main | 10 | CObjBase::r_Verb | +3
17:27:DEBUG:(03172) T_Main | 11 | CScriptObj::r_Verb | +3
17:27:DEBUG:(03172) T_Main | 12 | CChar::r_LoadVal | +3
17:27:DEBUG:(03172) T_Main | 13 | CChar::Skill_SetBase | +5
17:27:DEBUG:(03172) T_Main | 14 | CClient::addSkillWindow | +0
17:27:DEBUG:(03172) T_Main | 15 | PacketSkills::PacketSkills | +1 <-- last function call (stack unwinding began here)
17:27:CRITICAL:"Access Violation" (0x1b2074), in ExcType=CSError catched in CChar::LoadVal() #3 "Keyword"
17:27:DEBUG:command '100' args ''
17:27:'a' commands uid=0307 (B) to '100'=0
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +9
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +1
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +2
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +8
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +1
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +12
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:CRITICAL:"Access Violation" (0x1ae76b), in ExcType=CSError catched in CChar::OnTickPeriodic() #1 "update stats"
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::CheckLocation | +10
17:27:DEBUG:(03172) T_Main | 3 | CItemMultiCustom::GetDesignArea | +3268 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +15
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +20
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +3 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +16
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +14
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:CRITICAL:"Access Violation" (0x1ae76b), in ExcType=CSError catched in CChar::OnTickPeriodic() #1 "update stats"
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +1
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +16
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +1
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +2 <-- last function call (stack unwinding began here)
17:27:DEBUG:Printing STACK TRACE for debugging purposes.
17:27:DEBUG: __ thread (id) name __ | # | _____________ function _____________ | microseconds passed from previous function start
17:27:DEBUG:(03172) T_Main | 0 | CWorld::OnTick | +0
17:27:DEBUG:(03172) T_Main | 1 | CWorldTicker::Tick | +2
17:27:DEBUG:(03172) T_Main | 2 | CChar::OnTickStatusUpdate | +13
17:27:DEBUG:(03172) T_Main | 3 | CClient::UpdateStats | +0
17:27:DEBUG:(03172) T_Main | 4 | CClient::addStatusWindow | +0
17:27:DEBUG:(03172) T_Main | 5 | PacketObjectStatus::PacketObjectStat | +1 <-- last function call (stack unwinding began here)
```
**Extra info:**
- Tested on build 3381
- Tested whit clear script pack
|
non_process
|
server loop if wrong set value if for mistake try to use command set whit invalid skill id server enter in loop error how reproduce set target self console error a commands set debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cnetworkmanager tick debug t main cnetworkthread tick debug t main cnetworkinput processinput debug t main cnetworkinput processdata debug t main cnetworkinput processdata debug t main cnetworkinput processgameclientdata debug t main packettarget onreceive debug t main cclient event target debug t main cclient ontarg obj set debug t main cchar r verb debug t main cobjbase r verb debug t main cscriptobj r verb debug t main cchar r loadval debug t main cchar skill setbase debug t main cclient addskillwindow debug t main packetskills packetskills last function call stack unwinding began here critical access violation in exctype cserror catched in cchar loadval keyword debug command args a commands uid b to debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here critical access violation in exctype cserror catched in cchar ontickperiodic update stats debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar checklocation debug t main citemmulticustom getdesignarea last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here critical access violation in exctype cserror catched in cchar ontickperiodic update stats debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here debug printing stack trace for debugging purposes debug thread id name function microseconds passed from previous function start debug t main cworld ontick debug t main cworldticker tick debug t main cchar ontickstatusupdate debug t main cclient updatestats debug t main cclient addstatuswindow debug t main packetobjectstatus packetobjectstat last function call stack unwinding began here extra info tested on build tested whit clear script pack
| 0
|
112,328
| 14,238,928,303
|
IssuesEvent
|
2020-11-18 19:21:50
|
Subscribie/subscribie
|
https://api.github.com/repos/Subscribie/subscribie
|
opened
|
Stripe payment confirmation for one off item
|
needs-design needs-marketing needs-qa needs-user-story
|
Just tested a one off product, the confirmation page is blank and the auto confirmation email is the same as it would be for an ongoing subscription product.
|
1.0
|
Stripe payment confirmation for one off item - Just tested a one off product, the confirmation page is blank and the auto confirmation email is the same as it would be for an ongoing subscription product.
|
non_process
|
stripe payment confirmation for one off item just tested a one off product the confirmation page is blank and the auto confirmation email is the same as it would be for an ongoing subscription product
| 0
|
5,501
| 8,366,664,110
|
IssuesEvent
|
2018-10-04 09:46:50
|
linnovate/root
|
https://api.github.com/repos/linnovate/root
|
closed
|
in every entity, after selecting one of the entites ( a task, project, discussion, etc...) using multiple choice mode and trying to add a new tag, every time you hover over the update button with your mouse it adds the same tag over and over again
|
Process bug bug
|
in every entity, after selecting one of the entites ( a task, project, discussion, etc...) using multiple choice mode and trying to add a new tag, every time you hover over the update button with your mouse it adds the same tag over and over again
|
1.0
|
in every entity, after selecting one of the entites ( a task, project, discussion, etc...) using multiple choice mode and trying to add a new tag, every time you hover over the update button with your mouse it adds the same tag over and over again - in every entity, after selecting one of the entites ( a task, project, discussion, etc...) using multiple choice mode and trying to add a new tag, every time you hover over the update button with your mouse it adds the same tag over and over again
|
process
|
in every entity after selecting one of the entites a task project discussion etc using multiple choice mode and trying to add a new tag every time you hover over the update button with your mouse it adds the same tag over and over again in every entity after selecting one of the entites a task project discussion etc using multiple choice mode and trying to add a new tag every time you hover over the update button with your mouse it adds the same tag over and over again
| 1
|
17,814
| 23,741,281,459
|
IssuesEvent
|
2022-08-31 12:39:45
|
km4ack/patmenu2
|
https://api.github.com/repos/km4ack/patmenu2
|
closed
|
VARA modem isn't closed
|
bug in process
|
The VARA modem doesn't close when using the "Stop Modem" button in Pat
|
1.0
|
VARA modem isn't closed - The VARA modem doesn't close when using the "Stop Modem" button in Pat
|
process
|
vara modem isn t closed the vara modem doesn t close when using the stop modem button in pat
| 1
|
22,232
| 30,782,372,881
|
IssuesEvent
|
2023-07-31 10:55:19
|
ESMValGroup/ESMValCore
|
https://api.github.com/repos/ESMValGroup/ESMValCore
|
opened
|
Kalman filter
|
enhancement preprocessor
|
Here's a nice paper on how a Kalman filter is used as a post-processing (in their case) tool, in our case would be a nice preprocessor. I vaguely remember someone asking about a Kalman filter but I can't find an issue about it, so am plopping it here. I've not done any preprocs dev in a while, if peeps think it's useful I might do it meself, am getting a bit rusty these days, too much admin and maintenance stuff :grin: https://gmd.copernicus.org/articles/16/4233/2023/
|
1.0
|
Kalman filter - Here's a nice paper on how a Kalman filter is used as a post-processing (in their case) tool, in our case would be a nice preprocessor. I vaguely remember someone asking about a Kalman filter but I can't find an issue about it, so am plopping it here. I've not done any preprocs dev in a while, if peeps think it's useful I might do it meself, am getting a bit rusty these days, too much admin and maintenance stuff :grin: https://gmd.copernicus.org/articles/16/4233/2023/
|
process
|
kalman filter here s a nice paper on how a kalman filter is used as a post processing in their case tool in our case would be a nice preprocessor i vaguely remember someone asking about a kalman filter but i can t find an issue about it so am plopping it here i ve not done any preprocs dev in a while if peeps think it s useful i might do it meself am getting a bit rusty these days too much admin and maintenance stuff grin
| 1
|
694
| 3,184,492,529
|
IssuesEvent
|
2015-09-27 12:31:18
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
opened
|
process.versions.icu
|
feature request intl process
|
This would be useful to test for the presence of ICU and to implement version-specific behaviour. I've been trying to add [`U_ICU_VERSION`](https://ssl.icu-project.org/apiref/icu4c/uvernum_8h.html), but haven't gotten it working yet, If someone wants to take a shot.
Related: #3007
|
1.0
|
process.versions.icu - This would be useful to test for the presence of ICU and to implement version-specific behaviour. I've been trying to add [`U_ICU_VERSION`](https://ssl.icu-project.org/apiref/icu4c/uvernum_8h.html), but haven't gotten it working yet, If someone wants to take a shot.
Related: #3007
|
process
|
process versions icu this would be useful to test for the presence of icu and to implement version specific behaviour i ve been trying to add but haven t gotten it working yet if someone wants to take a shot related
| 1
|
676,916
| 23,142,559,762
|
IssuesEvent
|
2022-07-28 20:03:42
|
Cheos137/ArmorpointsPlusplus
|
https://api.github.com/repos/Cheos137/ArmorpointsPlusplus
|
closed
|
[Bug]: max health below 20 still draws heart borders
|
bug minecraft: all supported future update priority: medium
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Description
When having a maximum health below 20, all heart borders are drawn. However, vanilla doesn't draw unnecessary heart borders - and as this mod tries to follow vanilla behavior as closely as possible, this should be corrected.
Though, a config option should be added to re-enable the current behavior.
### Version
Other (this list is not up to date - again...)
### Version
3.0.0 for all supported minecraft versions
### To Reproduce
reduce max-health to a value below 20 (19 is not really useful for testing, too)
### Expexted Behavior
Unnecessary heart borders are not drawn.
### Additional Context
_No response_
### Screenshots
_No response_
### Relevant Log / Crash-Report Output
_No response_
|
1.0
|
[Bug]: max health below 20 still draws heart borders - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Description
When having a maximum health below 20, all heart borders are drawn. However, vanilla doesn't draw unnecessary heart borders - and as this mod tries to follow vanilla behavior as closely as possible, this should be corrected.
Though, a config option should be added to re-enable the current behavior.
### Version
Other (this list is not up to date - again...)
### Version
3.0.0 for all supported minecraft versions
### To Reproduce
reduce max-health to a value below 20 (19 is not really useful for testing, too)
### Expexted Behavior
Unnecessary heart borders are not drawn.
### Additional Context
_No response_
### Screenshots
_No response_
### Relevant Log / Crash-Report Output
_No response_
|
non_process
|
max health below still draws heart borders is there an existing issue for this i have searched the existing issues description when having a maximum health below all heart borders are drawn however vanilla doesn t draw unnecessary heart borders and as this mod tries to follow vanilla behavior as closely as possible this should be corrected though a config option should be added to re enable the current behavior version other this list is not up to date again version for all supported minecraft versions to reproduce reduce max health to a value below is not really useful for testing too expexted behavior unnecessary heart borders are not drawn additional context no response screenshots no response relevant log crash report output no response
| 0
|
196,097
| 14,799,050,401
|
IssuesEvent
|
2021-01-13 01:19:05
|
flutter/flutter
|
https://api.github.com/repos/flutter/flutter
|
closed
|
Make sure Scrollable.ensureVisible works with children before the center
|
P4 a: tests f: scrolling found in release: 1.22 framework has reproducible steps
|
See "ListView ensureVisible negative child" in `ensure_visible_test.dart`.
|
1.0
|
Make sure Scrollable.ensureVisible works with children before the center - See "ListView ensureVisible negative child" in `ensure_visible_test.dart`.
|
non_process
|
make sure scrollable ensurevisible works with children before the center see listview ensurevisible negative child in ensure visible test dart
| 0
|
16,348
| 21,006,561,479
|
IssuesEvent
|
2022-03-29 23:26:46
|
grpc/grpc-ios
|
https://api.github.com/repos/grpc/grpc-ios
|
opened
|
gRPC C++ build instructions for iOS
|
area/c-core kind/process
|
Add goto doc for how to build gRPC native c-core for iOS.
|
1.0
|
gRPC C++ build instructions for iOS - Add goto doc for how to build gRPC native c-core for iOS.
|
process
|
grpc c build instructions for ios add goto doc for how to build grpc native c core for ios
| 1
|
11,991
| 9,546,237,724
|
IssuesEvent
|
2019-05-01 19:20:49
|
trilinos/Trilinos
|
https://api.github.com/repos/trilinos/Trilinos
|
closed
|
KokkosKernels_graph tests timing out in ATDM builds
|
ATDM Sev: Nonblocker PA: Data Services client: ATDM pkg: KokkosKernels type: bug
|
CC: @trilinos/kokkoskernels, @kddevin (Trilinos Data Services Product Lead), @bartlettroscoe, @fryeguy52
<Checklist>
<???: Add label "ATDM">
<???: Add label "bug"?>
<???: Add label for affected packages (e.g. "MueLu", "Tpetra", "Kokkos", etc.)>
<???: Add milestone "Initial cleanup of new ATDM builds of Trilinos" or "Keep promoted ATDM builds of Trilinos clean">
<???: Once GitHub Issue is created, add entries for tests to TrilinosATDMStatus/*.csv files>
<???: Add label "PA: ???Project Area???" (e.g. "PA: Linear Solvers", "PA: Data Services")>
## Next Action Status
<status-and-or-first-action>
## Description
As shown in [this query](https://testing.sandia.gov/cdash/queryTests.php?project=Trilinos&filtercombine=and&filtercombine=&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercount=6&showfilters=1&filtercombine=and&field1=buildname&compare1=65&value1=Trilinos-atdm-&field2=testname&compare2=65&value2=KokkosKernels_graph_&field3=testname&compare3=66&value3=_MPI_1&field4=status&compare4=62&value4=Passed&field5=buildstarttime&compare5=83&value5=2019-04-07&field6=buildstarttime&compare6=84&value6=2019-04-10) the tests:
* KokkosKernels_graph_openmp_MPI_1
* KokkosKernels_graph_serial_MPI_1
started timing out on 2019-04-07 in the builds:
* Trilinos-atdm-hansen-shiller-gnu-debug-serial
* Trilinos-atdm-waterman-cuda-9.2-debug
* Trilinos-atdm-white-ride-cuda-9.2-gnu-7.2.0-debug
* Trilinos-atdm-white-ride-gnu-7.2.0-openmp-debug
## Current Status on CDash
Failures for the current testing day can be found [here](https://testing.sandia.gov/cdash/queryTests.php?project=Trilinos&filtercombine=and&filtercombine=&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercount=6&showfilters=1&filtercombine=and&field1=buildname&compare1=65&value1=Trilinos-atdm-&field2=testname&compare2=65&value2=KokkosKernels_graph_&field3=testname&compare3=66&value3=_MPI_1&field4=status&compare4=62&value4=Passed&field5=buildstarttime&compare5=83&value5=yesterday&field6=buildstarttime&compare6=84&value6=today)
## Steps to Reproduce
One should be able to reproduce this failure on ride or white as described in:
* https://github.com/trilinos/Trilinos/blob/develop/cmake/std/atdm/README.md
More specifically, the commands given for ride or white are provided at:
* https://github.com/trilinos/Trilinos/blob/develop/cmake/std/atdm/README.md#ridewhite
The exact commands to reproduce this issue should be:
```
$ cd <some_build_dir>/
$ source $TRILINOS_DIR/cmake/std/atdm/load-env.sh Trilinos-atdm-white-ride-gnu-7.2.0-openmp-debug
$ cmake \
-GNinja \
-DTrilinos_CONFIGURE_OPTIONS_FILE:STRING=cmake/std/atdm/ATDMDevEnv.cmake \
-DTrilinos_ENABLE_TESTS=ON -DTrilinos_ENABLE_KokkosKernels=ON \
$TRILINOS_DIR
$ make NP=16
$ bsub -x -Is -q rhel7F -n 16 ctest -j16
```
|
1.0
|
KokkosKernels_graph tests timing out in ATDM builds - CC: @trilinos/kokkoskernels, @kddevin (Trilinos Data Services Product Lead), @bartlettroscoe, @fryeguy52
<Checklist>
<???: Add label "ATDM">
<???: Add label "bug"?>
<???: Add label for affected packages (e.g. "MueLu", "Tpetra", "Kokkos", etc.)>
<???: Add milestone "Initial cleanup of new ATDM builds of Trilinos" or "Keep promoted ATDM builds of Trilinos clean">
<???: Once GitHub Issue is created, add entries for tests to TrilinosATDMStatus/*.csv files>
<???: Add label "PA: ???Project Area???" (e.g. "PA: Linear Solvers", "PA: Data Services")>
## Next Action Status
<status-and-or-first-action>
## Description
As shown in [this query](https://testing.sandia.gov/cdash/queryTests.php?project=Trilinos&filtercombine=and&filtercombine=&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercount=6&showfilters=1&filtercombine=and&field1=buildname&compare1=65&value1=Trilinos-atdm-&field2=testname&compare2=65&value2=KokkosKernels_graph_&field3=testname&compare3=66&value3=_MPI_1&field4=status&compare4=62&value4=Passed&field5=buildstarttime&compare5=83&value5=2019-04-07&field6=buildstarttime&compare6=84&value6=2019-04-10) the tests:
* KokkosKernels_graph_openmp_MPI_1
* KokkosKernels_graph_serial_MPI_1
started timing out on 2019-04-07 in the builds:
* Trilinos-atdm-hansen-shiller-gnu-debug-serial
* Trilinos-atdm-waterman-cuda-9.2-debug
* Trilinos-atdm-white-ride-cuda-9.2-gnu-7.2.0-debug
* Trilinos-atdm-white-ride-gnu-7.2.0-openmp-debug
## Current Status on CDash
Failures for the current testing day can be found [here](https://testing.sandia.gov/cdash/queryTests.php?project=Trilinos&filtercombine=and&filtercombine=&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercombine=and&filtercount=6&showfilters=1&filtercombine=and&field1=buildname&compare1=65&value1=Trilinos-atdm-&field2=testname&compare2=65&value2=KokkosKernels_graph_&field3=testname&compare3=66&value3=_MPI_1&field4=status&compare4=62&value4=Passed&field5=buildstarttime&compare5=83&value5=yesterday&field6=buildstarttime&compare6=84&value6=today)
## Steps to Reproduce
One should be able to reproduce this failure on ride or white as described in:
* https://github.com/trilinos/Trilinos/blob/develop/cmake/std/atdm/README.md
More specifically, the commands given for ride or white are provided at:
* https://github.com/trilinos/Trilinos/blob/develop/cmake/std/atdm/README.md#ridewhite
The exact commands to reproduce this issue should be:
```
$ cd <some_build_dir>/
$ source $TRILINOS_DIR/cmake/std/atdm/load-env.sh Trilinos-atdm-white-ride-gnu-7.2.0-openmp-debug
$ cmake \
-GNinja \
-DTrilinos_CONFIGURE_OPTIONS_FILE:STRING=cmake/std/atdm/ATDMDevEnv.cmake \
-DTrilinos_ENABLE_TESTS=ON -DTrilinos_ENABLE_KokkosKernels=ON \
$TRILINOS_DIR
$ make NP=16
$ bsub -x -Is -q rhel7F -n 16 ctest -j16
```
|
non_process
|
kokkoskernels graph tests timing out in atdm builds cc trilinos kokkoskernels kddevin trilinos data services product lead bartlettroscoe next action status description as shown in the tests kokkoskernels graph openmp mpi kokkoskernels graph serial mpi started timing out on in the builds trilinos atdm hansen shiller gnu debug serial trilinos atdm waterman cuda debug trilinos atdm white ride cuda gnu debug trilinos atdm white ride gnu openmp debug current status on cdash failures for the current testing day can be found steps to reproduce one should be able to reproduce this failure on ride or white as described in more specifically the commands given for ride or white are provided at the exact commands to reproduce this issue should be cd source trilinos dir cmake std atdm load env sh trilinos atdm white ride gnu openmp debug cmake gninja dtrilinos configure options file string cmake std atdm atdmdevenv cmake dtrilinos enable tests on dtrilinos enable kokkoskernels on trilinos dir make np bsub x is q n ctest
| 0
|
883
| 3,348,219,913
|
IssuesEvent
|
2015-11-17 00:25:16
|
beesmart-it/trend-hrm
|
https://api.github.com/repos/beesmart-it/trend-hrm
|
opened
|
Can't delete team member if owning processes
|
company requirement selection process
|
Can't delete team member if owning processes.
One option is to ask user to re-assign all processes to other user.
|
1.0
|
Can't delete team member if owning processes - Can't delete team member if owning processes.
One option is to ask user to re-assign all processes to other user.
|
process
|
can t delete team member if owning processes can t delete team member if owning processes one option is to ask user to re assign all processes to other user
| 1
|
422,290
| 12,269,443,616
|
IssuesEvent
|
2020-05-07 14:05:08
|
googleapis/java-iam
|
https://api.github.com/repos/googleapis/java-iam
|
opened
|
Synthesis failed for java-iam
|
autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate java-iam. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to a new branch 'autosynth'
Cloning into '/tmpfs/tmp/tmpd1dezaou/synthtool'...
Switched to branch 'autosynth-self'
Note: checking out '81a7470ee9633ff052ccb679cbe37b0079ee51ef'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2)
Note: checking out '04cb397eb7590ea1e6c10a39fca3d8fe0fb3d256'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 04cb397 chore: add labels with context info (#518)
Switched to a new branch 'autosynth-self-1'
2020-05-07 07:00:56 [INFO] Running synthtool
2020-05-07 07:00:56 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-05-07 07:00:56,457 synthtool > Executing /tmpfs/src/github/synthtool/working_repo/synth.py.
On branch autosynth-self-1
nothing to commit, working tree clean
2020-05-07 07:00:56,589 synthtool > Ensuring dependencies.
2020-05-07 07:00:56,601 synthtool > Cloning googleapis.
2020-05-07 07:00:58,052 synthtool > Generating code for: //google/iam/v1:google-iam-v1-java.
2020-05-07 07:00:58,168 synthtool > Failed executing bazel build //google/iam/v1:google-iam-v1-java:
Starting local Bazel server and connecting to it...
Server crashed during startup. Now printing /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/server/jvm.out
OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x0000000715200000, 247463936, 0) failed; error='Not enough space' (errno=12)
#
# There is insufficient memory for the Java Runtime Environment to continue.
# Native memory allocation (mmap) failed to map 247463936 bytes for committing reserved memory.
# An error report file with more information is saved as:
# /home/kbuilder/.cache/synthtool/googleapis/hs_err_pid15482.log
2020-05-07 07:00:58,168 synthtool > Wrote metadata to synth.metadata.
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/tmpfs/src/github/synthtool/working_repo/synth.py", line 31, in <module>
bazel_target=f'//google/iam/{version}:google-iam-{version}-java',
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 63, in java_library
return self._generate_code(service, version, "java", **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 177, in _generate_code
shell.run(bazel_run_args)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['bazel', 'build', '//google/iam/v1:google-iam-v1-java']' returned non-zero exit status 37.
2020-05-07 07:00:58 [ERROR] Synthesis failed
HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2)
Switched to branch 'autosynth-self'
Note: checking out '81a7470ee9633ff052ccb679cbe37b0079ee51ef'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2)
Previous HEAD position was 04cb397 chore: add labels with context info (#518)
HEAD is now at be74d3e build: do not fail builds on codecov errors (#528)
Switched to a new branch 'autosynth-8'
2020-05-07 07:00:58 [INFO] Running synthtool
2020-05-07 07:00:58 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-05-07 07:00:58,447 synthtool > Executing /tmpfs/src/github/synthtool/working_repo/synth.py.
On branch autosynth-8
nothing to commit, working tree clean
2020-05-07 07:00:58,581 synthtool > Ensuring dependencies.
2020-05-07 07:00:58,593 synthtool > Cloning googleapis.
2020-05-07 07:00:58,986 synthtool > Generating code for: //google/iam/v1:google-iam-v1-java.
2020-05-07 07:00:59,149 synthtool > Failed executing bazel build //google/iam/v1:google-iam-v1-java:
Starting local Bazel server and connecting to it...
Server crashed during startup. Now printing /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/server/jvm.out
OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x0000000715200000, 247463936, 0) failed; error='Not enough space' (errno=12)
#
# There is insufficient memory for the Java Runtime Environment to continue.
# Native memory allocation (mmap) failed to map 247463936 bytes for committing reserved memory.
# An error report file with more information is saved as:
# /home/kbuilder/.cache/synthtool/googleapis/hs_err_pid15513.log
2020-05-07 07:00:59,150 synthtool > Wrote metadata to synth.metadata.
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/tmpfs/src/github/synthtool/working_repo/synth.py", line 31, in <module>
bazel_target=f'//google/iam/{version}:google-iam-{version}-java',
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 63, in java_library
return self._generate_code(service, version, "java", **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 177, in _generate_code
shell.run(bazel_run_args)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['bazel', 'build', '//google/iam/v1:google-iam-v1-java']' returned non-zero exit status 37.
2020-05-07 07:00:59 [ERROR] Synthesis failed
HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2)
Switched to branch 'autosynth'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 584, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 465, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 574, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 360, in synthesize_loop
synthesize_inner_loop(toolbox, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 370, in synthesize_inner_loop
synthesizer, len(toolbox.versions) - 1
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 259, in synthesize_version_in_new_branch
synthesizer.synthesize(self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 115, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/f9c87743-f01c-4b38-93fb-8cba46610304).
|
1.0
|
Synthesis failed for java-iam - Hello! Autosynth couldn't regenerate java-iam. :broken_heart:
Here's the output from running `synth.py`:
```
Cloning into 'working_repo'...
Switched to a new branch 'autosynth'
Cloning into '/tmpfs/tmp/tmpd1dezaou/synthtool'...
Switched to branch 'autosynth-self'
Note: checking out '81a7470ee9633ff052ccb679cbe37b0079ee51ef'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2)
Note: checking out '04cb397eb7590ea1e6c10a39fca3d8fe0fb3d256'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 04cb397 chore: add labels with context info (#518)
Switched to a new branch 'autosynth-self-1'
2020-05-07 07:00:56 [INFO] Running synthtool
2020-05-07 07:00:56 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-05-07 07:00:56,457 synthtool > Executing /tmpfs/src/github/synthtool/working_repo/synth.py.
On branch autosynth-self-1
nothing to commit, working tree clean
2020-05-07 07:00:56,589 synthtool > Ensuring dependencies.
2020-05-07 07:00:56,601 synthtool > Cloning googleapis.
2020-05-07 07:00:58,052 synthtool > Generating code for: //google/iam/v1:google-iam-v1-java.
2020-05-07 07:00:58,168 synthtool > Failed executing bazel build //google/iam/v1:google-iam-v1-java:
Starting local Bazel server and connecting to it...
Server crashed during startup. Now printing /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/server/jvm.out
OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x0000000715200000, 247463936, 0) failed; error='Not enough space' (errno=12)
#
# There is insufficient memory for the Java Runtime Environment to continue.
# Native memory allocation (mmap) failed to map 247463936 bytes for committing reserved memory.
# An error report file with more information is saved as:
# /home/kbuilder/.cache/synthtool/googleapis/hs_err_pid15482.log
2020-05-07 07:00:58,168 synthtool > Wrote metadata to synth.metadata.
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/tmpfs/src/github/synthtool/working_repo/synth.py", line 31, in <module>
bazel_target=f'//google/iam/{version}:google-iam-{version}-java',
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 63, in java_library
return self._generate_code(service, version, "java", **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 177, in _generate_code
shell.run(bazel_run_args)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['bazel', 'build', '//google/iam/v1:google-iam-v1-java']' returned non-zero exit status 37.
2020-05-07 07:00:58 [ERROR] Synthesis failed
HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2)
Switched to branch 'autosynth-self'
Note: checking out '81a7470ee9633ff052ccb679cbe37b0079ee51ef'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2)
Previous HEAD position was 04cb397 chore: add labels with context info (#518)
HEAD is now at be74d3e build: do not fail builds on codecov errors (#528)
Switched to a new branch 'autosynth-8'
2020-05-07 07:00:58 [INFO] Running synthtool
2020-05-07 07:00:58 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-05-07 07:00:58,447 synthtool > Executing /tmpfs/src/github/synthtool/working_repo/synth.py.
On branch autosynth-8
nothing to commit, working tree clean
2020-05-07 07:00:58,581 synthtool > Ensuring dependencies.
2020-05-07 07:00:58,593 synthtool > Cloning googleapis.
2020-05-07 07:00:58,986 synthtool > Generating code for: //google/iam/v1:google-iam-v1-java.
2020-05-07 07:00:59,149 synthtool > Failed executing bazel build //google/iam/v1:google-iam-v1-java:
Starting local Bazel server and connecting to it...
Server crashed during startup. Now printing /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/server/jvm.out
OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x0000000715200000, 247463936, 0) failed; error='Not enough space' (errno=12)
#
# There is insufficient memory for the Java Runtime Environment to continue.
# Native memory allocation (mmap) failed to map 247463936 bytes for committing reserved memory.
# An error report file with more information is saved as:
# /home/kbuilder/.cache/synthtool/googleapis/hs_err_pid15513.log
2020-05-07 07:00:59,150 synthtool > Wrote metadata to synth.metadata.
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/tmpfs/src/github/synthtool/working_repo/synth.py", line 31, in <module>
bazel_target=f'//google/iam/{version}:google-iam-{version}-java',
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 63, in java_library
return self._generate_code(service, version, "java", **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 177, in _generate_code
shell.run(bazel_run_args)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['bazel', 'build', '//google/iam/v1:google-iam-v1-java']' returned non-zero exit status 37.
2020-05-07 07:00:59 [ERROR] Synthesis failed
HEAD is now at 81a7470 deps: update dependency com.google.api.grpc:proto-google-common-protos to v1.18.0 (#2)
Switched to branch 'autosynth'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 584, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 465, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 574, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 360, in synthesize_loop
synthesize_inner_loop(toolbox, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 370, in synthesize_inner_loop
synthesizer, len(toolbox.versions) - 1
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 259, in synthesize_version_in_new_branch
synthesizer.synthesize(self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 115, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/f9c87743-f01c-4b38-93fb-8cba46610304).
|
non_process
|
synthesis failed for java iam hello autosynth couldn t regenerate java iam broken heart here s the output from running synth py cloning into working repo switched to a new branch autosynth cloning into tmpfs tmp synthtool switched to branch autosynth self note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at deps update dependency com google api grpc proto google common protos to note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at chore add labels with context info switched to a new branch autosynth self running synthtool synthtool executing tmpfs src github synthtool working repo synth py on branch autosynth self nothing to commit working tree clean synthtool ensuring dependencies synthtool cloning googleapis synthtool generating code for google iam google iam java synthtool failed executing bazel build google iam google iam java starting local bazel server and connecting to it server crashed during startup now printing home kbuilder cache bazel bazel kbuilder server jvm out openjdk bit server vm warning info os commit memory failed error not enough space errno there is insufficient memory for the java runtime environment to continue native memory allocation mmap failed to map bytes for committing reserved memory an error report file with more information is saved as home kbuilder cache synthtool googleapis hs err log synthtool wrote metadata to synth metadata traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file tmpfs src github synthtool working repo synth py line in bazel target f google iam version google iam version java file tmpfs src github synthtool synthtool gcp gapic bazel py line in java library return self generate code service version java kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in generate code shell run bazel run args file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status synthesis failed head is now at deps update dependency com google api grpc proto google common protos to switched to branch autosynth self note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at deps update dependency com google api grpc proto google common protos to previous head position was chore add labels with context info head is now at build do not fail builds on codecov errors switched to a new branch autosynth running synthtool synthtool executing tmpfs src github synthtool working repo synth py on branch autosynth nothing to commit working tree clean synthtool ensuring dependencies synthtool cloning googleapis synthtool generating code for google iam google iam java synthtool failed executing bazel build google iam google iam java starting local bazel server and connecting to it server crashed during startup now printing home kbuilder cache bazel bazel kbuilder server jvm out openjdk bit server vm warning info os commit memory failed error not enough space errno there is insufficient memory for the java runtime environment to continue native memory allocation mmap failed to map bytes for committing reserved memory an error report file with more information is saved as home kbuilder cache synthtool googleapis hs err log synthtool wrote metadata to synth metadata traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file tmpfs src github synthtool working repo synth py line in bazel target f google iam version google iam version java file tmpfs src github synthtool synthtool gcp gapic bazel py line in java library return self generate code service version java kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in generate code shell run bazel run args file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status synthesis failed head is now at deps update dependency com google api grpc proto google common protos to switched to branch autosynth traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop synthesize inner loop toolbox synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize inner loop synthesizer len toolbox versions file tmpfs src github synthtool autosynth synth py line in synthesize version in new branch synthesizer synthesize self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
| 0
|
1,798
| 4,539,979,563
|
IssuesEvent
|
2016-09-09 13:14:01
|
openvstorage/alba
|
https://api.github.com/repos/openvstorage/alba
|
closed
|
ASDs can refuse to start up after power loss
|
process_wontfix type_bug
|
```
2016/05/12 14:39:55 259633 +0200 - ovs03 - 8878/0 - alba/asd - 0 - info - ETCD: etcdctl --peers=127.0.0.1:2379 get ovs/alba/asds/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX/config
2016/05/12 14:39:55 265048 +0200 - ovs03 - 8878/0 - alba/asd - 1 - info - Found the following config: {"asd_id": "HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX", "log_level": "info", "ips": ["172.19.197.3"], "node_id": "LxQuz79r7xe7k4gFhPUqJMGs9Cs7PxF6", "home": "/mnt/alba-asd/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX", "port": 8602}
2016/05/12 14:39:55 265150 +0200 - ovs03 - 8878/0 - alba/asd - 2 - info - Interpreted the config as: { Asd_config.Config.ips = ["172.19.197.3"]; port = (Some 8602);; node_id = "LxQuz79r7xe7k4gFhPUqJMGs9Cs7PxF6";; home = "/mnt/alba-asd/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX";; log_level = "info"; asd_id = (Some "HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX");; __sync_dont_use = true; limit = 99L; buffer_size = 786432;; multicast = (Some 10.); tls = None;; tcp_keepalive = { Tcp_keepalive.enable_tcp_keepalive = true;; tcp_keepalive_time = 20; tcp_keepalive_intvl = 20;; tcp_keepalive_probes = 3 };; __warranty_void__write_blobs = true; use_fadvise = true;; use_fallocate = true; rocksdb_block_cache_size = None }
2016/05/12 14:39:55 265172 +0200 - ovs03 - 8878/0 - alba/asd - 3 - info - asd_server version:0.9.3-0-gd6d4a3d
alba: internal error, uncaught exception:
(Failure "Corruption: checksum mismatch")
Raised at file "src/core/lwt.ml", line 789, characters 22-23
Called from file "src/unix/lwt_main.ml", line 34, characters 8-18
```
The rocksdb log shows this:
```
2016/05/17-17:51:34.132241 7f2446538940 Recovered from manifest file:/mnt/alba-asd/7aSGYFzBuY7uFODPysPntsDijG7ZxtPm/db/MANIFEST-004971 succeeded,manifest_file_number is 4971, next_file_number is 5136, last_sequence is 47018602, log_number is 0,prev_log_number is 0,max_column_family is 0
2016/05/17-17:51:34.132256 7f2446538940 Column family [default] (ID 0), log number is 5133
2016/05/17-17:51:34.132418 7f2446538940 EVENT_LOG_v1 {"time_micros": 1463487694132404, "job": 1, "event": "recovery_started", "log_files": [5133]}
2016/05/17-17:51:34.132433 7f2446538940 Recovering log #5133 mode 0 skip-recovery 0
2016/05/17-17:51:34.163110 7f2446538940 [WARN] /mnt/alba-asd/7aSGYFzBuY7uFODPysPntsDijG7ZxtPm/db/005133.log: dropping 5792 bytes; Corruption: checksum mismatch
```
We started seeing this after we started using the rocksdb recycle_log_file_name option, so it will probably be solved by https://github.com/facebook/rocksdb/pull/881.
|
1.0
|
ASDs can refuse to start up after power loss - ```
2016/05/12 14:39:55 259633 +0200 - ovs03 - 8878/0 - alba/asd - 0 - info - ETCD: etcdctl --peers=127.0.0.1:2379 get ovs/alba/asds/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX/config
2016/05/12 14:39:55 265048 +0200 - ovs03 - 8878/0 - alba/asd - 1 - info - Found the following config: {"asd_id": "HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX", "log_level": "info", "ips": ["172.19.197.3"], "node_id": "LxQuz79r7xe7k4gFhPUqJMGs9Cs7PxF6", "home": "/mnt/alba-asd/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX", "port": 8602}
2016/05/12 14:39:55 265150 +0200 - ovs03 - 8878/0 - alba/asd - 2 - info - Interpreted the config as: { Asd_config.Config.ips = ["172.19.197.3"]; port = (Some 8602);; node_id = "LxQuz79r7xe7k4gFhPUqJMGs9Cs7PxF6";; home = "/mnt/alba-asd/HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX";; log_level = "info"; asd_id = (Some "HFvP96Pjb10yf9LS7NFQaAg0oD26xrMX");; __sync_dont_use = true; limit = 99L; buffer_size = 786432;; multicast = (Some 10.); tls = None;; tcp_keepalive = { Tcp_keepalive.enable_tcp_keepalive = true;; tcp_keepalive_time = 20; tcp_keepalive_intvl = 20;; tcp_keepalive_probes = 3 };; __warranty_void__write_blobs = true; use_fadvise = true;; use_fallocate = true; rocksdb_block_cache_size = None }
2016/05/12 14:39:55 265172 +0200 - ovs03 - 8878/0 - alba/asd - 3 - info - asd_server version:0.9.3-0-gd6d4a3d
alba: internal error, uncaught exception:
(Failure "Corruption: checksum mismatch")
Raised at file "src/core/lwt.ml", line 789, characters 22-23
Called from file "src/unix/lwt_main.ml", line 34, characters 8-18
```
The rocksdb log shows this:
```
2016/05/17-17:51:34.132241 7f2446538940 Recovered from manifest file:/mnt/alba-asd/7aSGYFzBuY7uFODPysPntsDijG7ZxtPm/db/MANIFEST-004971 succeeded,manifest_file_number is 4971, next_file_number is 5136, last_sequence is 47018602, log_number is 0,prev_log_number is 0,max_column_family is 0
2016/05/17-17:51:34.132256 7f2446538940 Column family [default] (ID 0), log number is 5133
2016/05/17-17:51:34.132418 7f2446538940 EVENT_LOG_v1 {"time_micros": 1463487694132404, "job": 1, "event": "recovery_started", "log_files": [5133]}
2016/05/17-17:51:34.132433 7f2446538940 Recovering log #5133 mode 0 skip-recovery 0
2016/05/17-17:51:34.163110 7f2446538940 [WARN] /mnt/alba-asd/7aSGYFzBuY7uFODPysPntsDijG7ZxtPm/db/005133.log: dropping 5792 bytes; Corruption: checksum mismatch
```
We started seeing this after we started using the rocksdb recycle_log_file_name option, so it will probably be solved by https://github.com/facebook/rocksdb/pull/881.
|
process
|
asds can refuse to start up after power loss alba asd info etcd etcdctl peers get ovs alba asds config alba asd info found the following config asd id log level info ips node id home mnt alba asd port alba asd info interpreted the config as asd config config ips port some node id home mnt alba asd log level info asd id some sync dont use true limit buffer size multicast some tls none tcp keepalive tcp keepalive enable tcp keepalive true tcp keepalive time tcp keepalive intvl tcp keepalive probes warranty void write blobs true use fadvise true use fallocate true rocksdb block cache size none alba asd info asd server version alba internal error uncaught exception failure corruption checksum mismatch raised at file src core lwt ml line characters called from file src unix lwt main ml line characters the rocksdb log shows this recovered from manifest file mnt alba asd db manifest succeeded manifest file number is next file number is last sequence is log number is prev log number is max column family is column family id log number is event log time micros job event recovery started log files recovering log mode skip recovery mnt alba asd db log dropping bytes corruption checksum mismatch we started seeing this after we started using the rocksdb recycle log file name option so it will probably be solved by
| 1
|
132,257
| 28,128,141,000
|
IssuesEvent
|
2023-03-31 19:43:26
|
creativecommons/cc-resource-archive
|
https://api.github.com/repos/creativecommons/cc-resource-archive
|
closed
|
[Feature] Adding Footer
|
🟩 priority: low ⛔️ status: discarded 🚦 status: awaiting triage ✨ goal: improvement 💻 aspect: code
|
## Problem
There is No footer Section
## Description
Adding footer for showing contact information about Creative Commons.
## Implementation

<!-- Replace the [ ] with [x] to check the box. -->
- [x] I would be interested in implementing this feature.
|
1.0
|
[Feature] Adding Footer - ## Problem
There is No footer Section
## Description
Adding footer for showing contact information about Creative Commons.
## Implementation

<!-- Replace the [ ] with [x] to check the box. -->
- [x] I would be interested in implementing this feature.
|
non_process
|
adding footer problem there is no footer section description adding footer for showing contact information about creative commons implementation i would be interested in implementing this feature
| 0
|
11,114
| 13,957,681,476
|
IssuesEvent
|
2020-10-24 08:07:26
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
DE: request for a new harvesting
|
DE - Germany Geoportal Harvesting process
|
Dear Geoportal Helpdesk,
As mentioned in Roberts Mail from 2020/03/02 we would like to initiate a new push of our metadata records to the EU Geoportal. For this reason we kindly ask you to start a new harvesting of our catalogue instance and publish them for us in the Geoportal harvesting "sandbox", please.
Also we kindly ask you, if you could provide us two or three original csw-requests (for an internal validation/review on our side), which you are using to get the metadata records from our catalogue instance.
Thanks in advance and best regards,
Anja (on behalf of Coordination Office SDI Germany)
|
1.0
|
DE: request for a new harvesting - Dear Geoportal Helpdesk,
As mentioned in Roberts Mail from 2020/03/02 we would like to initiate a new push of our metadata records to the EU Geoportal. For this reason we kindly ask you to start a new harvesting of our catalogue instance and publish them for us in the Geoportal harvesting "sandbox", please.
Also we kindly ask you, if you could provide us two or three original csw-requests (for an internal validation/review on our side), which you are using to get the metadata records from our catalogue instance.
Thanks in advance and best regards,
Anja (on behalf of Coordination Office SDI Germany)
|
process
|
de request for a new harvesting dear geoportal helpdesk as mentioned in roberts mail from we would like to initiate a new push of our metadata records to the eu geoportal for this reason we kindly ask you to start a new harvesting of our catalogue instance and publish them for us in the geoportal harvesting quot sandbox quot please also we kindly ask you if you could provide us two or three original csw requests for an internal validation review on our side which you are using to get the metadata records from our catalogue instance thanks in advance and best regards anja on behalf of coordination office sdi germany
| 1
|
18,344
| 24,467,515,861
|
IssuesEvent
|
2022-10-07 16:19:43
|
bondaleksey/credit-card-fraud-detection
|
https://api.github.com/repos/bondaleksey/credit-card-fraud-detection
|
opened
|
daily data generation
|
work plan data preprocessing
|
- Write a script that generates a new chunk of data and saves it to HDFS.
- Automate your regular launch with AirFlow.
|
1.0
|
daily data generation - - Write a script that generates a new chunk of data and saves it to HDFS.
- Automate your regular launch with AirFlow.
|
process
|
daily data generation write a script that generates a new chunk of data and saves it to hdfs automate your regular launch with airflow
| 1
|
277,176
| 24,054,501,787
|
IssuesEvent
|
2022-09-16 15:32:31
|
wazuh/wazuh
|
https://api.github.com/repos/wazuh/wazuh
|
closed
|
Release 4.3.8 - Release Candidate 1
|
release test/4.3.8
|
The following issue will gather all the info regarding testing and fixing in order to validate this release candidate.
The definition of done for this one is the validation from the product owner of each QA analysis and the acceptance of the implemented fixes implemented, all the below issues must be closed in order to close this one.
## Release candidate info
|Project|Main issue|Version|Release candidate #|Tag|Previous RC issue|Next RC issue|
|---|---|---|---|---|---|---|
|[v4.3.8](https://github.com/orgs/wazuh/projects/20/views/1)|[#14826](https://github.com/wazuh/wazuh/issues/14826)|4.3.8|RC1|[v4.3.8-rc1](https://github.com/wazuh/wazuh/tree/v4.3.8-rc1)|-|-|
## QA testing issues
In order to move to a new release candidate or the GA version, all tests and metrics analyses below must be in Closed status.
| Name | Issue | Status |DRI|
|-----------------------------|-----------------------------------------------|-------------|---|
| C unit | https://github.com/wazuh/wazuh/issues/14879 | 🟣 Completed |@wazuh/core|
| Python unit | https://github.com/wazuh/wazuh/issues/14870 | 🟣 Completed |@wazuh/framework|
| Footprint metrics | https://github.com/wazuh/wazuh/issues/14859 | 🟣 Completed |@wazuh/cicd|
| Workload benchmarks metrics | https://github.com/wazuh/wazuh/issues/14860 | 🟣 Completed |@wazuh/framework|
| Integration | https://github.com/wazuh/wazuh/issues/14877 | 🟣 Completed |@wazuh/qa|
| API integration | https://github.com/wazuh/wazuh/issues/14871 | 🟣 Completed |@wazuh/framework|
| System | https://github.com/wazuh/wazuh/issues/14887 | 🟣 Completed | @wazuh/framework|
| External integrations modules | - | ⚪ Skipped | @wazuh/framework|
| Demo uses cases | https://github.com/wazuh/wazuh/issues/14921 | 🟣 Completed |@wazuh/cicd|
| Packages | https://github.com/wazuh/wazuh/issues/14862 | 🟣 Completed |@wazuh/cicd|
| Coverity scan | https://github.com/wazuh/wazuh/issues/14881 | 🟣 Completed |@wazuh/core|
| Ruleset | https://github.com/wazuh/wazuh/issues/14872 | 🟣 Completed |@wazuh/threat-intel|
| Kibana UI regression | https://github.com/wazuh/wazuh-kibana-app/issues/4535 | 🟣 Completed |@wazuh/frontend|
| Splunk UI regression | https://github.com/wazuh/wazuh-splunk/issues/1366 | 🟣 Completed |@wazuh/frontend|
| WPK Upgrade |https://github.com/wazuh/wazuh/issues/14861 | 🟣 Completed |@wazuh/core|
| E2E UX | https://github.com/wazuh/wazuh/issues/14828 | 🟣 Completed |@wazuh|
⚫ _Not started: The tasks didn't start yet._
🟡 _In progress: The team is already working on it._
🟢 _Ready to review: The product owner must audit and validate the results._
⚪ _Skipped: The task has been deemed not necessary for this RC._
🟣 _Completed: Task finished. Nothing to do here._
🔴 _Completed with failures: Some issues raised here._
## Auditors' validation
In order to close and proceed with the release or the next candidate version, the following auditors must give the green light to this RC.
- [x] @davidjiglesias
|
1.0
|
Release 4.3.8 - Release Candidate 1 - The following issue will gather all the info regarding testing and fixing in order to validate this release candidate.
The definition of done for this one is the validation from the product owner of each QA analysis and the acceptance of the implemented fixes implemented, all the below issues must be closed in order to close this one.
## Release candidate info
|Project|Main issue|Version|Release candidate #|Tag|Previous RC issue|Next RC issue|
|---|---|---|---|---|---|---|
|[v4.3.8](https://github.com/orgs/wazuh/projects/20/views/1)|[#14826](https://github.com/wazuh/wazuh/issues/14826)|4.3.8|RC1|[v4.3.8-rc1](https://github.com/wazuh/wazuh/tree/v4.3.8-rc1)|-|-|
## QA testing issues
In order to move to a new release candidate or the GA version, all tests and metrics analyses below must be in Closed status.
| Name | Issue | Status |DRI|
|-----------------------------|-----------------------------------------------|-------------|---|
| C unit | https://github.com/wazuh/wazuh/issues/14879 | 🟣 Completed |@wazuh/core|
| Python unit | https://github.com/wazuh/wazuh/issues/14870 | 🟣 Completed |@wazuh/framework|
| Footprint metrics | https://github.com/wazuh/wazuh/issues/14859 | 🟣 Completed |@wazuh/cicd|
| Workload benchmarks metrics | https://github.com/wazuh/wazuh/issues/14860 | 🟣 Completed |@wazuh/framework|
| Integration | https://github.com/wazuh/wazuh/issues/14877 | 🟣 Completed |@wazuh/qa|
| API integration | https://github.com/wazuh/wazuh/issues/14871 | 🟣 Completed |@wazuh/framework|
| System | https://github.com/wazuh/wazuh/issues/14887 | 🟣 Completed | @wazuh/framework|
| External integrations modules | - | ⚪ Skipped | @wazuh/framework|
| Demo uses cases | https://github.com/wazuh/wazuh/issues/14921 | 🟣 Completed |@wazuh/cicd|
| Packages | https://github.com/wazuh/wazuh/issues/14862 | 🟣 Completed |@wazuh/cicd|
| Coverity scan | https://github.com/wazuh/wazuh/issues/14881 | 🟣 Completed |@wazuh/core|
| Ruleset | https://github.com/wazuh/wazuh/issues/14872 | 🟣 Completed |@wazuh/threat-intel|
| Kibana UI regression | https://github.com/wazuh/wazuh-kibana-app/issues/4535 | 🟣 Completed |@wazuh/frontend|
| Splunk UI regression | https://github.com/wazuh/wazuh-splunk/issues/1366 | 🟣 Completed |@wazuh/frontend|
| WPK Upgrade |https://github.com/wazuh/wazuh/issues/14861 | 🟣 Completed |@wazuh/core|
| E2E UX | https://github.com/wazuh/wazuh/issues/14828 | 🟣 Completed |@wazuh|
⚫ _Not started: The tasks didn't start yet._
🟡 _In progress: The team is already working on it._
🟢 _Ready to review: The product owner must audit and validate the results._
⚪ _Skipped: The task has been deemed not necessary for this RC._
🟣 _Completed: Task finished. Nothing to do here._
🔴 _Completed with failures: Some issues raised here._
## Auditors' validation
In order to close and proceed with the release or the next candidate version, the following auditors must give the green light to this RC.
- [x] @davidjiglesias
|
non_process
|
release release candidate the following issue will gather all the info regarding testing and fixing in order to validate this release candidate the definition of done for this one is the validation from the product owner of each qa analysis and the acceptance of the implemented fixes implemented all the below issues must be closed in order to close this one release candidate info project main issue version release candidate tag previous rc issue next rc issue qa testing issues in order to move to a new release candidate or the ga version all tests and metrics analyses below must be in closed status name issue status dri c unit 🟣 completed wazuh core python unit 🟣 completed wazuh framework footprint metrics 🟣 completed wazuh cicd workload benchmarks metrics 🟣 completed wazuh framework integration 🟣 completed wazuh qa api integration 🟣 completed wazuh framework system 🟣 completed wazuh framework external integrations modules ⚪ skipped wazuh framework demo uses cases 🟣 completed wazuh cicd packages 🟣 completed wazuh cicd coverity scan 🟣 completed wazuh core ruleset 🟣 completed wazuh threat intel kibana ui regression 🟣 completed wazuh frontend splunk ui regression 🟣 completed wazuh frontend wpk upgrade 🟣 completed wazuh core ux 🟣 completed wazuh ⚫ not started the tasks didn t start yet 🟡 in progress the team is already working on it 🟢 ready to review the product owner must audit and validate the results ⚪ skipped the task has been deemed not necessary for this rc 🟣 completed task finished nothing to do here 🔴 completed with failures some issues raised here auditors validation in order to close and proceed with the release or the next candidate version the following auditors must give the green light to this rc davidjiglesias
| 0
|
5,027
| 7,849,611,670
|
IssuesEvent
|
2018-06-20 04:51:03
|
kerubistan/kerub
|
https://api.github.com/repos/kerubistan/kerub
|
opened
|
planner: filter out repeating inverses
|
component:data processing performance
|
> filter out repeating inverses: when the plan has already steps "power host-1 up" and then later "power host-1 down" then any new "power host-1 up" steps could be filtered out - and let's see how this effects performance
|
1.0
|
planner: filter out repeating inverses - > filter out repeating inverses: when the plan has already steps "power host-1 up" and then later "power host-1 down" then any new "power host-1 up" steps could be filtered out - and let's see how this effects performance
|
process
|
planner filter out repeating inverses filter out repeating inverses when the plan has already steps power host up and then later power host down then any new power host up steps could be filtered out and let s see how this effects performance
| 1
|
368,600
| 25,799,866,232
|
IssuesEvent
|
2022-12-10 22:51:58
|
EliasSchaut/Movie-Monday-Manager
|
https://api.github.com/repos/EliasSchaut/Movie-Monday-Manager
|
closed
|
The Dev Update
|
documentation enhancement
|
- Dev Documentation
- Api Documentation
- Type definitions for data transmision objects
- Improve set up and installation
|
1.0
|
The Dev Update - - Dev Documentation
- Api Documentation
- Type definitions for data transmision objects
- Improve set up and installation
|
non_process
|
the dev update dev documentation api documentation type definitions for data transmision objects improve set up and installation
| 0
|
69,806
| 7,164,635,682
|
IssuesEvent
|
2018-01-29 11:56:38
|
smap-consulting/smapserver
|
https://api.github.com/repos/smap-consulting/smapserver
|
closed
|
BUG - 17.12 - Help link in sessions goes to wrong help folder
|
In Test bug
|
### SmapServer
### Version 17.12
### Expected Behaviour: Link must goes to: [Here](https://drive.google.com/drive/folders/0B966-DLxVMApbDJzd1NVVUMyNms)
### Actual Behaviour link goes to: [SMAP](https://drive.google.com/drive/folders/0B5_SmpWlQYxvNExwSFV5X051Yjg)

|
1.0
|
BUG - 17.12 - Help link in sessions goes to wrong help folder - ### SmapServer
### Version 17.12
### Expected Behaviour: Link must goes to: [Here](https://drive.google.com/drive/folders/0B966-DLxVMApbDJzd1NVVUMyNms)
### Actual Behaviour link goes to: [SMAP](https://drive.google.com/drive/folders/0B5_SmpWlQYxvNExwSFV5X051Yjg)

|
non_process
|
bug help link in sessions goes to wrong help folder smapserver version expected behaviour link must goes to actual behaviour link goes to
| 0
|
1,963
| 4,779,096,737
|
IssuesEvent
|
2016-10-27 21:19:45
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
doc: need better link to signals list in process documentation
|
doc good first contribution process
|
* **Version**: all
* **Platform**: n/a
* **Subsystem**: doc
In the 'Signal Events' section of the `process` documentation, there is a link to a man page that supposedly lists the names of the various signals that can be listened for. However, the page that is currently linked ([sigaction(2)](http://man7.org/linux/man-pages/man2/sigaction.2.html)) doesn't show all of the valid signals and the ones that it does show are not formatted very well. I think a better link would be to [signal(7)](http://man7.org/linux/man-pages/man7/signal.7.html), which *does* list all of the signal names and in an easy-to-spot/read format.
|
1.0
|
doc: need better link to signals list in process documentation - * **Version**: all
* **Platform**: n/a
* **Subsystem**: doc
In the 'Signal Events' section of the `process` documentation, there is a link to a man page that supposedly lists the names of the various signals that can be listened for. However, the page that is currently linked ([sigaction(2)](http://man7.org/linux/man-pages/man2/sigaction.2.html)) doesn't show all of the valid signals and the ones that it does show are not formatted very well. I think a better link would be to [signal(7)](http://man7.org/linux/man-pages/man7/signal.7.html), which *does* list all of the signal names and in an easy-to-spot/read format.
|
process
|
doc need better link to signals list in process documentation version all platform n a subsystem doc in the signal events section of the process documentation there is a link to a man page that supposedly lists the names of the various signals that can be listened for however the page that is currently linked doesn t show all of the valid signals and the ones that it does show are not formatted very well i think a better link would be to which does list all of the signal names and in an easy to spot read format
| 1
|
26,457
| 2,684,552,301
|
IssuesEvent
|
2015-03-29 03:23:59
|
cs2103jan2015-w13-1j/main
|
https://api.github.com/repos/cs2103jan2015-w13-1j/main
|
closed
|
A user can remove multiple tasks by tags
|
priority.medium
|
... so that I do not need to remove them individually
|
1.0
|
A user can remove multiple tasks by tags - ... so that I do not need to remove them individually
|
non_process
|
a user can remove multiple tasks by tags so that i do not need to remove them individually
| 0
|
21,920
| 30,446,458,712
|
IssuesEvent
|
2023-07-15 18:31:11
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
pyutils 0.0.1a1 has 2 GuardDog issues
|
guarddog typosquatting silent-process-execution
|
https://pypi.org/project/pyutils
https://inspector.pypi.io/project/pyutils
```{
"dependency": "pyutils",
"version": "0.0.1a1",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils",
"silent-process-execution": [
{
"location": "pyutils-0.0.1a1/src/pyutils/exec_utils.py:200",
"code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpxi7_due4/pyutils"
}
}```
|
1.0
|
pyutils 0.0.1a1 has 2 GuardDog issues - https://pypi.org/project/pyutils
https://inspector.pypi.io/project/pyutils
```{
"dependency": "pyutils",
"version": "0.0.1a1",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils",
"silent-process-execution": [
{
"location": "pyutils-0.0.1a1/src/pyutils/exec_utils.py:200",
"code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpxi7_due4/pyutils"
}
}```
|
process
|
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt pytils python utils silent process execution location pyutils src pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp pyutils
| 1
|
4,499
| 7,348,662,047
|
IssuesEvent
|
2018-03-08 07:40:11
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
Missed help for all Processing Graphics algorithms
|
Processing
|
Moved from QGIS redmine https://hub.qgis.org/issues/12702
There are no descriptions for Bar plot, Mean and std.dev plot, ~~Polar plot~~ and Vector layer scatterplot algorithms
|
1.0
|
Missed help for all Processing Graphics algorithms - Moved from QGIS redmine https://hub.qgis.org/issues/12702
There are no descriptions for Bar plot, Mean and std.dev plot, ~~Polar plot~~ and Vector layer scatterplot algorithms
|
process
|
missed help for all processing graphics algorithms moved from qgis redmine there are no descriptions for bar plot mean and std dev plot polar plot and vector layer scatterplot algorithms
| 1
|
38,358
| 2,846,359,873
|
IssuesEvent
|
2015-05-29 10:55:55
|
WarEmu/WarBugs
|
https://api.github.com/repos/WarEmu/WarBugs
|
closed
|
Relogging onto another character makes the previous characters abilities appear on the action bar
|
Ability Emulator Interface Low Priority
|
When you relog a character to another one. The abilities from the previous character are shown on the second character's hot bar.
For example: Relogging from an Archmage to a Sword master. You get the abilities of the Archmage on the Sword master's hot bar.

|
1.0
|
Relogging onto another character makes the previous characters abilities appear on the action bar - When you relog a character to another one. The abilities from the previous character are shown on the second character's hot bar.
For example: Relogging from an Archmage to a Sword master. You get the abilities of the Archmage on the Sword master's hot bar.

|
non_process
|
relogging onto another character makes the previous characters abilities appear on the action bar when you relog a character to another one the abilities from the previous character are shown on the second character s hot bar for example relogging from an archmage to a sword master you get the abilities of the archmage on the sword master s hot bar
| 0
|
9,552
| 12,514,450,665
|
IssuesEvent
|
2020-06-03 05:17:58
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Aggregate tool not working
|
Bug Processing Regression
|
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
Aggregate tool doesn't work
**How to Reproduce**
1. Go to Toolbox
2. Click on Aggregate tool
3. Python error message pops and the tools fails to "start"
4. See error:
`Traceback (most recent call last):
File "/usr/share/qgis/python/plugins/processing/gui/ProcessingToolbox.py", line 262, in executeAlgorithm
dlg = AlgorithmDialog(alg, self.in_place_mode, iface.mainWindow())
File "/usr/share/qgis/python/plugins/processing/gui/AlgorithmDialog.py", line 67, in __init__
self.setMainWidget(self.getParametersPanel(alg, self))
File "/usr/share/qgis/python/plugins/processing/gui/AlgorithmDialog.py", line 87, in getParametersPanel
panel = ParametersPanel(parent, alg, self.in_place)
File "/usr/share/qgis/python/plugins/processing/gui/ParametersPanel.py", line 70, in __init__
self.initWidgets()
File "/usr/share/qgis/python/plugins/processing/gui/ParametersPanel.py", line 104, in initWidgets
wrapper = WidgetWrapperFactory.create_wrapper(param, self.parent())
File "/usr/share/qgis/python/plugins/processing/gui/wrappers.py", line 1845, in create_wrapper
return WidgetWrapperFactory.create_wrapper_from_metadata(param, dialog, row, col)
File "/usr/share/qgis/python/plugins/processing/gui/wrappers.py", line 1876, in create_wrapper_from_metadata
mod = __import__('.'.join(tokens[:-1]), fromlist=[tokens[-1]])
File "/usr/lib/python3/dist-packages/qgis/utils.py", line 743, in _import
mod = _builtin_import(name, globals, locals, fromlist, level)
File "/usr/share/qgis/python/plugins/processing/algs/qgis/ui/AggregatesPanel.py", line 45, in
from processing.algs.qgis.ui.FieldsMappingPanel import (
ImportError: cannot import name 'ExpressionDelegate' from 'processing.algs.qgis.ui.FieldsMappingPanel' (/usr/share/qgis/python/plugins/processing/algs/qgis/ui/FieldsMappingPanel.py)
`
**QGIS and OS versions**
QGIS 3.13 Master
PopOS 20.04 / Ubuntu 20.04
|
1.0
|
Aggregate tool not working - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
Aggregate tool doesn't work
**How to Reproduce**
1. Go to Toolbox
2. Click on Aggregate tool
3. Python error message pops and the tools fails to "start"
4. See error:
`Traceback (most recent call last):
File "/usr/share/qgis/python/plugins/processing/gui/ProcessingToolbox.py", line 262, in executeAlgorithm
dlg = AlgorithmDialog(alg, self.in_place_mode, iface.mainWindow())
File "/usr/share/qgis/python/plugins/processing/gui/AlgorithmDialog.py", line 67, in __init__
self.setMainWidget(self.getParametersPanel(alg, self))
File "/usr/share/qgis/python/plugins/processing/gui/AlgorithmDialog.py", line 87, in getParametersPanel
panel = ParametersPanel(parent, alg, self.in_place)
File "/usr/share/qgis/python/plugins/processing/gui/ParametersPanel.py", line 70, in __init__
self.initWidgets()
File "/usr/share/qgis/python/plugins/processing/gui/ParametersPanel.py", line 104, in initWidgets
wrapper = WidgetWrapperFactory.create_wrapper(param, self.parent())
File "/usr/share/qgis/python/plugins/processing/gui/wrappers.py", line 1845, in create_wrapper
return WidgetWrapperFactory.create_wrapper_from_metadata(param, dialog, row, col)
File "/usr/share/qgis/python/plugins/processing/gui/wrappers.py", line 1876, in create_wrapper_from_metadata
mod = __import__('.'.join(tokens[:-1]), fromlist=[tokens[-1]])
File "/usr/lib/python3/dist-packages/qgis/utils.py", line 743, in _import
mod = _builtin_import(name, globals, locals, fromlist, level)
File "/usr/share/qgis/python/plugins/processing/algs/qgis/ui/AggregatesPanel.py", line 45, in
from processing.algs.qgis.ui.FieldsMappingPanel import (
ImportError: cannot import name 'ExpressionDelegate' from 'processing.algs.qgis.ui.FieldsMappingPanel' (/usr/share/qgis/python/plugins/processing/algs/qgis/ui/FieldsMappingPanel.py)
`
**QGIS and OS versions**
QGIS 3.13 Master
PopOS 20.04 / Ubuntu 20.04
|
process
|
aggregate tool not working bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug aggregate tool doesn t work how to reproduce go to toolbox click on aggregate tool python error message pops and the tools fails to start see error traceback most recent call last file usr share qgis python plugins processing gui processingtoolbox py line in executealgorithm dlg algorithmdialog alg self in place mode iface mainwindow file usr share qgis python plugins processing gui algorithmdialog py line in init self setmainwidget self getparameterspanel alg self file usr share qgis python plugins processing gui algorithmdialog py line in getparameterspanel panel parameterspanel parent alg self in place file usr share qgis python plugins processing gui parameterspanel py line in init self initwidgets file usr share qgis python plugins processing gui parameterspanel py line in initwidgets wrapper widgetwrapperfactory create wrapper param self parent file usr share qgis python plugins processing gui wrappers py line in create wrapper return widgetwrapperfactory create wrapper from metadata param dialog row col file usr share qgis python plugins processing gui wrappers py line in create wrapper from metadata mod import join tokens fromlist file usr lib dist packages qgis utils py line in import mod builtin import name globals locals fromlist level file usr share qgis python plugins processing algs qgis ui aggregatespanel py line in from processing algs qgis ui fieldsmappingpanel import importerror cannot import name expressiondelegate from processing algs qgis ui fieldsmappingpanel usr share qgis python plugins processing algs qgis ui fieldsmappingpanel py qgis and os versions qgis master popos ubuntu
| 1
|
619
| 3,086,178,612
|
IssuesEvent
|
2015-08-25 00:39:48
|
e107inc/e107
|
https://api.github.com/repos/e107inc/e107
|
closed
|
alpha 2 upgrade database update
|
bug upgrade process 1.x to 2.x
|
How many times will the database update run? Is there any way I can check progress. I have clicked the update tab about 50 times now and it still says it is functioning correctly and the query stats vary slightly each time suggesting they are doing different work, but there is no end in site. It will not let me continue on to add additional/missing core files until this step completes.
|
1.0
|
alpha 2 upgrade database update - How many times will the database update run? Is there any way I can check progress. I have clicked the update tab about 50 times now and it still says it is functioning correctly and the query stats vary slightly each time suggesting they are doing different work, but there is no end in site. It will not let me continue on to add additional/missing core files until this step completes.
|
process
|
alpha upgrade database update how many times will the database update run is there any way i can check progress i have clicked the update tab about times now and it still says it is functioning correctly and the query stats vary slightly each time suggesting they are doing different work but there is no end in site it will not let me continue on to add additional missing core files until this step completes
| 1
|
20,988
| 27,853,626,269
|
IssuesEvent
|
2023-03-20 20:46:01
|
dDevTech/tapas-top-frontend
|
https://api.github.com/repos/dDevTech/tapas-top-frontend
|
opened
|
Panel usuario
|
in process
|
Diseño de panel con información resumida del perfil
También incluir estadísticas
- Número de degustaciones
- Locales nuevos ult.7 días con opción de ver todo
|
1.0
|
Panel usuario - Diseño de panel con información resumida del perfil
También incluir estadísticas
- Número de degustaciones
- Locales nuevos ult.7 días con opción de ver todo
|
process
|
panel usuario diseño de panel con información resumida del perfil también incluir estadísticas número de degustaciones locales nuevos ult días con opción de ver todo
| 1
|
15,052
| 18,762,899,670
|
IssuesEvent
|
2021-11-05 18:46:39
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
closed
|
Moisture in combustion air
|
Process Heating
|
Field in Flue Gas calculator
This is implemented for Solid/Liquid Fuels but not gas fuels. I think it is in the math for gas fuels and will be needed for the ESC 31 calc #4060
Hopefully just need the field visible with help, etc.
|
1.0
|
Moisture in combustion air - Field in Flue Gas calculator
This is implemented for Solid/Liquid Fuels but not gas fuels. I think it is in the math for gas fuels and will be needed for the ESC 31 calc #4060
Hopefully just need the field visible with help, etc.
|
process
|
moisture in combustion air field in flue gas calculator this is implemented for solid liquid fuels but not gas fuels i think it is in the math for gas fuels and will be needed for the esc calc hopefully just need the field visible with help etc
| 1
|
437,307
| 12,576,526,931
|
IssuesEvent
|
2020-06-09 08:02:54
|
kubeflow/manifests
|
https://api.github.com/repos/kubeflow/manifests
|
closed
|
application CR's istio-install and knative-serving-install do not deploy
|
area/kfctl kind/bug lifecycle/stale priority/p1
|
both fail for different reasons
knative-serving-install deployment seems to wait forever
istio-install seems to not resolve all $(generateName)'s
|
1.0
|
application CR's istio-install and knative-serving-install do not deploy - both fail for different reasons
knative-serving-install deployment seems to wait forever
istio-install seems to not resolve all $(generateName)'s
|
non_process
|
application cr s istio install and knative serving install do not deploy both fail for different reasons knative serving install deployment seems to wait forever istio install seems to not resolve all generatename s
| 0
|
1,895
| 4,724,814,764
|
IssuesEvent
|
2016-10-18 02:36:35
|
webshell/materia-designer
|
https://api.github.com/repos/webshell/materia-designer
|
closed
|
Ghost created when Materia designer is restarted from the Chrome dev tools
|
area/materia-server-process type/bug
|
### Expected behavior
Clear the process before the restart
### Actual behavior
Restart without deleting internal process.
### Step to reproduce the behavior
* Navigate in Materia designer a while
* Open the Chrome dev tools
* Go in the **Network** tab => it asks you to refresh with `Command+R`
* Hit `Command+R` => Materia designer restart => a ghost of the latest Materia-server process is still in memory.
---
* Materia Server Version: 0.1.1
* Materia Designer Version: 0.1.0-rc.3
|
1.0
|
Ghost created when Materia designer is restarted from the Chrome dev tools - ### Expected behavior
Clear the process before the restart
### Actual behavior
Restart without deleting internal process.
### Step to reproduce the behavior
* Navigate in Materia designer a while
* Open the Chrome dev tools
* Go in the **Network** tab => it asks you to refresh with `Command+R`
* Hit `Command+R` => Materia designer restart => a ghost of the latest Materia-server process is still in memory.
---
* Materia Server Version: 0.1.1
* Materia Designer Version: 0.1.0-rc.3
|
process
|
ghost created when materia designer is restarted from the chrome dev tools expected behavior clear the process before the restart actual behavior restart without deleting internal process step to reproduce the behavior navigate in materia designer a while open the chrome dev tools go in the network tab it asks you to refresh with command r hit command r materia designer restart a ghost of the latest materia server process is still in memory materia server version materia designer version rc
| 1
|
2,109
| 4,944,604,554
|
IssuesEvent
|
2016-11-30 08:32:59
|
mathiasbynens/es-regexp-dotall-flag
|
https://api.github.com/repos/mathiasbynens/es-regexp-dotall-flag
|
closed
|
Advance to stage 1
|
process
|
Criteria taken from [the TC39 process document](https://tc39.github.io/process-document/):
> - [x] Identified “champion” who will advance the addition
Done: @bterlson.
> - [x] Prose outlining the problem or need and the general shape of a solution
https://github.com/mathiasbynens/es-regexp-singleline-flag#motivation
> - [x] Illustrative examples of usage
https://github.com/mathiasbynens/es-regexp-dotall-flag#high-level-api
> - [x] High-level API
https://github.com/mathiasbynens/es-regexp-singleline-flag#high-level-api
> - [x] Discussion of key algorithms, abstractions and semantics
https://github.com/mathiasbynens/es-regexp-singleline-flag#faq
> - [x] Identification of potential “cross-cutting” concerns and implementation challenges/complexity
There should be none.
|
1.0
|
Advance to stage 1 - Criteria taken from [the TC39 process document](https://tc39.github.io/process-document/):
> - [x] Identified “champion” who will advance the addition
Done: @bterlson.
> - [x] Prose outlining the problem or need and the general shape of a solution
https://github.com/mathiasbynens/es-regexp-singleline-flag#motivation
> - [x] Illustrative examples of usage
https://github.com/mathiasbynens/es-regexp-dotall-flag#high-level-api
> - [x] High-level API
https://github.com/mathiasbynens/es-regexp-singleline-flag#high-level-api
> - [x] Discussion of key algorithms, abstractions and semantics
https://github.com/mathiasbynens/es-regexp-singleline-flag#faq
> - [x] Identification of potential “cross-cutting” concerns and implementation challenges/complexity
There should be none.
|
process
|
advance to stage criteria taken from identified “champion” who will advance the addition done bterlson prose outlining the problem or need and the general shape of a solution illustrative examples of usage high level api discussion of key algorithms abstractions and semantics identification of potential “cross cutting” concerns and implementation challenges complexity there should be none
| 1
|
20,036
| 26,520,442,149
|
IssuesEvent
|
2023-01-19 01:47:18
|
DIGGSml/Geophysics
|
https://api.github.com/repos/DIGGSml/Geophysics
|
opened
|
General Topic: Spatial-temporal domain data
|
Data acquisition Processed data
|
The Test measurement structure currently only supports data in a spatial domain, where all coordinate axes are spatial. To support spatial-temporal results, where the "locatioN" of a data value has a temporal axis, likely can be handled by providing reference compound spatial reference systems and/or the ability for the user to define a custom spatial-temporal compound coordinate
For reference material on compound CRS's: https://docs.opengeospatial.org/is/18-010r7/18-010r7.html.
|
1.0
|
General Topic: Spatial-temporal domain data - The Test measurement structure currently only supports data in a spatial domain, where all coordinate axes are spatial. To support spatial-temporal results, where the "locatioN" of a data value has a temporal axis, likely can be handled by providing reference compound spatial reference systems and/or the ability for the user to define a custom spatial-temporal compound coordinate
For reference material on compound CRS's: https://docs.opengeospatial.org/is/18-010r7/18-010r7.html.
|
process
|
general topic spatial temporal domain data the test measurement structure currently only supports data in a spatial domain where all coordinate axes are spatial to support spatial temporal results where the location of a data value has a temporal axis likely can be handled by providing reference compound spatial reference systems and or the ability for the user to define a custom spatial temporal compound coordinate for reference material on compound crs s
| 1
|
211,373
| 23,808,128,890
|
IssuesEvent
|
2022-09-04 11:03:09
|
neptune-mutual-blue/protocol
|
https://api.github.com/repos/neptune-mutual-blue/protocol
|
closed
|
Decouple pause and unpause ACL logic in the store contract
|
enhancement security rc2
|
Unlike the `unpause` feature, which is solely available to the owner, the `pause` feature of the store should be accessible to a separate account in order to programmatically pause the contract. In a hosted environment, this helps to avoid using the owner's private key.
|
True
|
Decouple pause and unpause ACL logic in the store contract - Unlike the `unpause` feature, which is solely available to the owner, the `pause` feature of the store should be accessible to a separate account in order to programmatically pause the contract. In a hosted environment, this helps to avoid using the owner's private key.
|
non_process
|
decouple pause and unpause acl logic in the store contract unlike the unpause feature which is solely available to the owner the pause feature of the store should be accessible to a separate account in order to programmatically pause the contract in a hosted environment this helps to avoid using the owner s private key
| 0
|
7,321
| 10,454,860,287
|
IssuesEvent
|
2019-09-19 19:53:16
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
closed
|
Inconsistencies between Efficiency Improvement Calculator and O2 Enrichment calculator
|
Process Heating bug
|
User reported differences via email:
Consider the basic problem below:
Natural Gas Furnace
Current Flue Gas Oxygen = New Flue Gas Oxygen = 4%
Current Flue Gas Temp = New Flue Gas Temp = 500 F
Current Combustion Air = 80 F
New Combustion Air = 300 F
The Fuel Savings (%) calculated by the "Efficiency Improvement Calculator" does not match with the Fuel Savings calculated by the "Oxygen Enrichment Calculator".
The latter is closer to PHAST's calculations. Can you look into why the two calculators give different answers?
|
1.0
|
Inconsistencies between Efficiency Improvement Calculator and O2 Enrichment calculator - User reported differences via email:
Consider the basic problem below:
Natural Gas Furnace
Current Flue Gas Oxygen = New Flue Gas Oxygen = 4%
Current Flue Gas Temp = New Flue Gas Temp = 500 F
Current Combustion Air = 80 F
New Combustion Air = 300 F
The Fuel Savings (%) calculated by the "Efficiency Improvement Calculator" does not match with the Fuel Savings calculated by the "Oxygen Enrichment Calculator".
The latter is closer to PHAST's calculations. Can you look into why the two calculators give different answers?
|
process
|
inconsistencies between efficiency improvement calculator and enrichment calculator user reported differences via email consider the basic problem below natural gas furnace current flue gas oxygen new flue gas oxygen current flue gas temp new flue gas temp f current combustion air f new combustion air f the fuel savings calculated by the efficiency improvement calculator does not match with the fuel savings calculated by the oxygen enrichment calculator the latter is closer to phast s calculations can you look into why the two calculators give different answers
| 1
|
296,215
| 22,292,681,854
|
IssuesEvent
|
2022-06-12 15:45:15
|
noradream/SwimSetGenerator
|
https://api.github.com/repos/noradream/SwimSetGenerator
|
opened
|
Improve Code Readibility
|
documentation
|
JavaDoc & Other comments are very clunky right now, and need an overhaul.
|
1.0
|
Improve Code Readibility - JavaDoc & Other comments are very clunky right now, and need an overhaul.
|
non_process
|
improve code readibility javadoc other comments are very clunky right now and need an overhaul
| 0
|
17,290
| 23,098,664,137
|
IssuesEvent
|
2022-07-26 22:41:33
|
maticnetwork/miden
|
https://api.github.com/repos/maticnetwork/miden
|
closed
|
MPVERIFY instruction potential refactoring
|
assembly instruction set processor
|
`MPVERIFY` computes the Merkle root of a path for a given specified node in the stack. The stack is arranged as follows (from the top):
- depth of the path, 1 element.
- index of the node, 1 element.
- value of the node, 4 elements.
- root of the tree, 4 elements.
Currently, post this operation, we are replacing the node value with the newly computed root calculated by the `hasher` co-processor. The stack will look like this post this operation:
- depth of the path, 1 element.
- index of the node, 1 element.
- computed root of the path, 4 elements.
- root of the tree, 4 elements.
We are already asserting the computed root of the path with root of the tree and repeating this in assembly(`mtree.get`) wastes couple of clock cycles. As noted by @bobbinth:
> "Instead of replacing the node with the newly computed root, we should just copy over the stack to the next cycle. The instruction should fail if the roots are not the same".
Operations relying on `MPVERIFY` might get a bit simplified due to fewer VM cycles needed now.
edit:
The instruction also needs to be populated with six helper registers which are as follows:
- row address of the start of the computation in hasher chiplet - 1 element
- the least significant bit of the node index - 1 element
- sibling node of the leaf - 4 elements
The `mtree.get` assembly instruction also needs to be updated to be in line with the new changes made in `MPVERIFY`as there are some potential optimization possibilities..
|
1.0
|
MPVERIFY instruction potential refactoring - `MPVERIFY` computes the Merkle root of a path for a given specified node in the stack. The stack is arranged as follows (from the top):
- depth of the path, 1 element.
- index of the node, 1 element.
- value of the node, 4 elements.
- root of the tree, 4 elements.
Currently, post this operation, we are replacing the node value with the newly computed root calculated by the `hasher` co-processor. The stack will look like this post this operation:
- depth of the path, 1 element.
- index of the node, 1 element.
- computed root of the path, 4 elements.
- root of the tree, 4 elements.
We are already asserting the computed root of the path with root of the tree and repeating this in assembly(`mtree.get`) wastes couple of clock cycles. As noted by @bobbinth:
> "Instead of replacing the node with the newly computed root, we should just copy over the stack to the next cycle. The instruction should fail if the roots are not the same".
Operations relying on `MPVERIFY` might get a bit simplified due to fewer VM cycles needed now.
edit:
The instruction also needs to be populated with six helper registers which are as follows:
- row address of the start of the computation in hasher chiplet - 1 element
- the least significant bit of the node index - 1 element
- sibling node of the leaf - 4 elements
The `mtree.get` assembly instruction also needs to be updated to be in line with the new changes made in `MPVERIFY`as there are some potential optimization possibilities..
|
process
|
mpverify instruction potential refactoring mpverify computes the merkle root of a path for a given specified node in the stack the stack is arranged as follows from the top depth of the path element index of the node element value of the node elements root of the tree elements currently post this operation we are replacing the node value with the newly computed root calculated by the hasher co processor the stack will look like this post this operation depth of the path element index of the node element computed root of the path elements root of the tree elements we are already asserting the computed root of the path with root of the tree and repeating this in assembly mtree get wastes couple of clock cycles as noted by bobbinth instead of replacing the node with the newly computed root we should just copy over the stack to the next cycle the instruction should fail if the roots are not the same operations relying on mpverify might get a bit simplified due to fewer vm cycles needed now edit the instruction also needs to be populated with six helper registers which are as follows row address of the start of the computation in hasher chiplet element the least significant bit of the node index element sibling node of the leaf elements the mtree get assembly instruction also needs to be updated to be in line with the new changes made in mpverify as there are some potential optimization possibilities
| 1
|
80,880
| 15,589,336,045
|
IssuesEvent
|
2021-03-18 07:54:04
|
AlchTech/JeremySalettes-Wozniak_4_05032021
|
https://api.github.com/repos/AlchTech/JeremySalettes-Wozniak_4_05032021
|
closed
|
CVE-2016-10735 (Medium) detected in bootstrap-3.3.5.min.js
|
security vulnerability
|
## CVE-2016-10735 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.5.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js</a></p>
<p>Path to dependency file: JeremySalettes-Wozniak_4_05032021/index.html</p>
<p>Path to vulnerable library: JeremySalettes-Wozniak_4_05032021/js/bootstrap.js,JeremySalettes-Wozniak_4_05032021/./js/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.5.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlchTech/JeremySalettes-Wozniak_4_05032021/commit/6596720eb100c60b452ca547f2a47690766344f7">6596720eb100c60b452ca547f2a47690766344f7</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-10735 (Medium) detected in bootstrap-3.3.5.min.js - ## CVE-2016-10735 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.5.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js</a></p>
<p>Path to dependency file: JeremySalettes-Wozniak_4_05032021/index.html</p>
<p>Path to vulnerable library: JeremySalettes-Wozniak_4_05032021/js/bootstrap.js,JeremySalettes-Wozniak_4_05032021/./js/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.5.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlchTech/JeremySalettes-Wozniak_4_05032021/commit/6596720eb100c60b452ca547f2a47690766344f7">6596720eb100c60b452ca547f2a47690766344f7</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file jeremysalettes wozniak index html path to vulnerable library jeremysalettes wozniak js bootstrap js jeremysalettes wozniak js bootstrap js dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch main vulnerability details in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
16,097
| 20,268,502,150
|
IssuesEvent
|
2022-02-15 14:17:24
|
streamnative/flink
|
https://api.github.com/repos/streamnative/flink
|
closed
|
[FLINK-26161][BUG] PulsarSource: Split should be added to finishedSplit if the the split is empty from beginning.
|
compute/data-processing type/bug
|
Currently we add a split to finishedSplit in the read for loop, but if the split is empty and never enters for loop it should be checked whether to add to finishedSplit as well. This is an edge case and priority is low.
```
// current
if (stopCursor.shouldStop(message)) {
builder.addFinishedSplit(splitId);
break;
}
```
|
1.0
|
[FLINK-26161][BUG] PulsarSource: Split should be added to finishedSplit if the the split is empty from beginning. - Currently we add a split to finishedSplit in the read for loop, but if the split is empty and never enters for loop it should be checked whether to add to finishedSplit as well. This is an edge case and priority is low.
```
// current
if (stopCursor.shouldStop(message)) {
builder.addFinishedSplit(splitId);
break;
}
```
|
process
|
pulsarsource split should be added to finishedsplit if the the split is empty from beginning currently we add a split to finishedsplit in the read for loop but if the split is empty and never enters for loop it should be checked whether to add to finishedsplit as well this is an edge case and priority is low current if stopcursor shouldstop message builder addfinishedsplit splitid break
| 1
|
9,589
| 12,540,472,181
|
IssuesEvent
|
2020-06-05 10:25:56
|
ESMValGroup/ESMValCore
|
https://api.github.com/repos/ESMValGroup/ESMValCore
|
closed
|
Anomalies preprocessor with period 'full' removes important metadata from cube
|
bug preprocessor
|
**Describe the bug**
When the `anomalies` preprocessor function is applied with `period=full`, it removes important metadata such as the variable name, all attributes, etc. This doesn't happen for other periods like `month` or `day`. Consequently, subsequent preprocessors or diagnostics fail.
I think it's happening [here](https://github.com/ESMValGroup/ESMValCore/blob/8306a6561a011f61dec949dc56889568b1af815e/esmvalcore/preprocessor/_time.py#L463):
```
cube = cube - reference
```
and this seems to be the intended behaviour in Iris, as it is documented [here](https://scitools.org.uk/iris/docs/latest/userguide/cube_maths.html#calculating-a-cube-anomaly).
Perhaps EMSValTool should only subtract the data instead of the full cube?
**Example**
`print(cube)`:
```
air_temperature / (K) (time: 1668)
Dimension coordinates:
time x
Auxiliary coordinates:
day_of_year x
month_number x
Scalar coordinates:
height: 1.5 m
latitude: 0.0 degrees, bound=(-90.0, 90.0) degrees
longitude: 179.0625 degrees, bound=(-0.9375, 359.0625) degrees
Attributes:
Conventions: CF-1.7
associated_files: baseURL: http://cmip-pcmdi.llnl.gov/CMIP5/dataLocation gridspecFile: gridspec_atmos_fx_ACCESS1-0_rcp45_r0i0p0.nc...
branch_time: 732311.0;109207.0;732311.0
cmor_version: 2.8.0
contact: The ACCESS wiki: http://wiki.csiro.au/confluence/display/ACCESS/Home. Contact...
experiment: RCP4.5;historical;RCP4.5
experiment_id: rcp45;historical;rcp45
forcing: GHG, Oz, SA, Sl, Vl, BC, OC, (GHG = CO2, N2O, CH4, CFC11, CFC12, CFC113,...
frequency: mon
initialization_method: 1
institute_id: CSIRO-BOM
institution: CSIRO (Commonwealth Scientific and Industrial Research Organisation, Australia),...
model_id: ACCESS1-0
modeling_realm: atmos
parent_experiment: historical;pre-industrial control;historical
parent_experiment_id: historical;piControl;historical
parent_experiment_rip: r1i1p1
physics_version: 1
product: output
project_id: CMIP5
realization: 1
references: See http://wiki.csiro.au/confluence/display/ACCESS/ACCESS+Publications
source: ACCESS1-0 2011. Atmosphere: AGCM v1.0 (N96 grid-point, 1.875 degrees EW...
table_id: Table Amon (27 April 2011) 9c851218e3842df9a62ef38b1e2575bb
title: ACCESS1-0 model output prepared for CMIP5 RCP4.5;ACCESS1-0 model output...
version_number: v20120115
Cell methods:
mean: time
mean: longitude, latitude
```
```
anoms = anomalies(cube, period='full', reference=dict(start_year=1980, start_month=1, start_day=1, end_year=2009, end_month=12, end_day=31))
print(anoms)
```
gives
```
unknown / (K) (time: 1668)
Dimension coordinates:
time x
Auxiliary coordinates:
month_number x
Scalar coordinates:
height: 1.5 m
latitude: 0.0 degrees, bound=(-90.0, 90.0) degrees
longitude: 179.0625 degrees, bound=(-0.9375, 359.0625) degrees
```
|
1.0
|
Anomalies preprocessor with period 'full' removes important metadata from cube - **Describe the bug**
When the `anomalies` preprocessor function is applied with `period=full`, it removes important metadata such as the variable name, all attributes, etc. This doesn't happen for other periods like `month` or `day`. Consequently, subsequent preprocessors or diagnostics fail.
I think it's happening [here](https://github.com/ESMValGroup/ESMValCore/blob/8306a6561a011f61dec949dc56889568b1af815e/esmvalcore/preprocessor/_time.py#L463):
```
cube = cube - reference
```
and this seems to be the intended behaviour in Iris, as it is documented [here](https://scitools.org.uk/iris/docs/latest/userguide/cube_maths.html#calculating-a-cube-anomaly).
Perhaps EMSValTool should only subtract the data instead of the full cube?
**Example**
`print(cube)`:
```
air_temperature / (K) (time: 1668)
Dimension coordinates:
time x
Auxiliary coordinates:
day_of_year x
month_number x
Scalar coordinates:
height: 1.5 m
latitude: 0.0 degrees, bound=(-90.0, 90.0) degrees
longitude: 179.0625 degrees, bound=(-0.9375, 359.0625) degrees
Attributes:
Conventions: CF-1.7
associated_files: baseURL: http://cmip-pcmdi.llnl.gov/CMIP5/dataLocation gridspecFile: gridspec_atmos_fx_ACCESS1-0_rcp45_r0i0p0.nc...
branch_time: 732311.0;109207.0;732311.0
cmor_version: 2.8.0
contact: The ACCESS wiki: http://wiki.csiro.au/confluence/display/ACCESS/Home. Contact...
experiment: RCP4.5;historical;RCP4.5
experiment_id: rcp45;historical;rcp45
forcing: GHG, Oz, SA, Sl, Vl, BC, OC, (GHG = CO2, N2O, CH4, CFC11, CFC12, CFC113,...
frequency: mon
initialization_method: 1
institute_id: CSIRO-BOM
institution: CSIRO (Commonwealth Scientific and Industrial Research Organisation, Australia),...
model_id: ACCESS1-0
modeling_realm: atmos
parent_experiment: historical;pre-industrial control;historical
parent_experiment_id: historical;piControl;historical
parent_experiment_rip: r1i1p1
physics_version: 1
product: output
project_id: CMIP5
realization: 1
references: See http://wiki.csiro.au/confluence/display/ACCESS/ACCESS+Publications
source: ACCESS1-0 2011. Atmosphere: AGCM v1.0 (N96 grid-point, 1.875 degrees EW...
table_id: Table Amon (27 April 2011) 9c851218e3842df9a62ef38b1e2575bb
title: ACCESS1-0 model output prepared for CMIP5 RCP4.5;ACCESS1-0 model output...
version_number: v20120115
Cell methods:
mean: time
mean: longitude, latitude
```
```
anoms = anomalies(cube, period='full', reference=dict(start_year=1980, start_month=1, start_day=1, end_year=2009, end_month=12, end_day=31))
print(anoms)
```
gives
```
unknown / (K) (time: 1668)
Dimension coordinates:
time x
Auxiliary coordinates:
month_number x
Scalar coordinates:
height: 1.5 m
latitude: 0.0 degrees, bound=(-90.0, 90.0) degrees
longitude: 179.0625 degrees, bound=(-0.9375, 359.0625) degrees
```
|
process
|
anomalies preprocessor with period full removes important metadata from cube describe the bug when the anomalies preprocessor function is applied with period full it removes important metadata such as the variable name all attributes etc this doesn t happen for other periods like month or day consequently subsequent preprocessors or diagnostics fail i think it s happening cube cube reference and this seems to be the intended behaviour in iris as it is documented perhaps emsvaltool should only subtract the data instead of the full cube example print cube air temperature k time dimension coordinates time x auxiliary coordinates day of year x month number x scalar coordinates height m latitude degrees bound degrees longitude degrees bound degrees attributes conventions cf associated files baseurl gridspecfile gridspec atmos fx nc branch time cmor version contact the access wiki contact experiment historical experiment id historical forcing ghg oz sa sl vl bc oc ghg frequency mon initialization method institute id csiro bom institution csiro commonwealth scientific and industrial research organisation australia model id modeling realm atmos parent experiment historical pre industrial control historical parent experiment id historical picontrol historical parent experiment rip physics version product output project id realization references see source atmosphere agcm grid point degrees ew table id table amon april title model output prepared for model output version number cell methods mean time mean longitude latitude anoms anomalies cube period full reference dict start year start month start day end year end month end day print anoms gives unknown k time dimension coordinates time x auxiliary coordinates month number x scalar coordinates height m latitude degrees bound degrees longitude degrees bound degrees
| 1
|
8,628
| 11,782,266,773
|
IssuesEvent
|
2020-03-17 01:17:55
|
googleapis/java-storage-nio
|
https://api.github.com/repos/googleapis/java-storage-nio
|
closed
|
org.mockito.matcher + org.junit.Assert + org.junit.rules.ExpectedException has been deprecated
|
api: storage type: cleanup type: process
|
```
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileSystemTest.java:[61,60] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[106,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[125,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[126,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[127,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[128,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[135,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[136,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[137,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[138,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[139,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[140,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[145,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[146,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[147,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[148,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[149,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[150,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[151,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[156,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[162,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[163,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[167,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[168,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[169,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[175,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[176,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[177,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[178,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[184,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[186,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadTest.java:[58,60] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[50,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[177,5] verifyZeroInteractions(java.lang.Object...) in org.mockito.Mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStoragePathTest.java:[43,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[106,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[126,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[127,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[128,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[129,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[139,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[140,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[141,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[143,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[144,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[145,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[150,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[151,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[152,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[153,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[154,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[155,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[156,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[161,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[163,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[165,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[166,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[175,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[176,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[178,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[179,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[180,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[187,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[188,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[189,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[196,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[197,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[199,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[200,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[201,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[45,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[98,5] verifyZeroInteractions(java.lang.Object...) in org.mockito.Mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/UnixPathTest.java:[34,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageLateInitializationTest.java:[39,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileSystemProviderTest.java:[88,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileAttributeViewTest.java:[45,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageConfigurationTest.java:[34,66] none() in org.junit.rules.ExpectedException has been deprecated
```
|
1.0
|
org.mockito.matcher + org.junit.Assert + org.junit.rules.ExpectedException has been deprecated - ```
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[20,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[21,26] org.mockito.Matchers in org.mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileSystemTest.java:[61,60] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[106,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[125,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[126,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[127,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[128,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[135,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[136,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[137,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[138,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[139,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[140,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[145,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[146,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[147,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[148,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[149,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[150,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[151,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[156,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[162,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[163,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[167,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[168,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[169,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[175,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[176,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[177,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[178,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[184,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadFileChannelTest.java:[186,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadTest.java:[58,60] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[50,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageReadChannelTest.java:[177,5] verifyZeroInteractions(java.lang.Object...) in org.mockito.Mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStoragePathTest.java:[43,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[106,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[126,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[127,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[128,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[129,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[139,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[140,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[141,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[143,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[144,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[145,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[150,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[151,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[152,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[153,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[154,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[155,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[156,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[161,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[163,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[165,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[166,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[175,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[176,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[178,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[179,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[180,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[187,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[188,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[189,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[196,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[197,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[199,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[200,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteFileChannelTest.java:[201,5] <T>assertThat(T,org.hamcrest.Matcher<? super T>) in org.junit.Assert has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[45,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageWriteChannelTest.java:[98,5] verifyZeroInteractions(java.lang.Object...) in org.mockito.Mockito has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/UnixPathTest.java:[34,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageLateInitializationTest.java:[39,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileSystemProviderTest.java:[88,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageFileAttributeViewTest.java:[45,66] none() in org.junit.rules.ExpectedException has been deprecated
[WARNING] /home/ajit/Documents/projects/github/java-storage-nio/google-cloud-nio/src/test/java/com/google/cloud/storage/contrib/nio/CloudStorageConfigurationTest.java:[34,66] none() in org.junit.rules.ExpectedException has been deprecated
```
|
process
|
org mockito matcher org junit assert org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java org mockito matchers in org mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragefilesystemtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadfilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragereadchanneltest java verifyzerointeractions java lang object in org mockito mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragepathtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritefilechanneltest java assertthat t org hamcrest matcher in org junit assert has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragewritechanneltest java verifyzerointeractions java lang object in org mockito mockito has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio unixpathtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragelateinitializationtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragefilesystemprovidertest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstoragefileattributeviewtest java none in org junit rules expectedexception has been deprecated home ajit documents projects github java storage nio google cloud nio src test java com google cloud storage contrib nio cloudstorageconfigurationtest java none in org junit rules expectedexception has been deprecated
| 1
|
13,419
| 15,880,355,727
|
IssuesEvent
|
2021-04-09 13:36:34
|
department-of-veterans-affairs/notification-api
|
https://api.github.com/repos/department-of-veterans-affairs/notification-api
|
closed
|
Get Amazon SES into production
|
Process Task
|
# Value Statement
**As** VANotify platform developer
**I want to** onboard Amazon SES as a provider into production
**So that**I can use it to send emails and begin to "warm up"
# Acceptance Criteria
**GIVEN** a VANotify platform developer
**WHEN** is viewing email providers in the VANotify platform
**THEN** Amazon SES is an option with which to send emails
# Checklist
- [x] AWS Support request to get dedicated IP addresses
- [x] AWS Support request for sending limits?
- [x] AWS Support request to take us out of sandbox
- [x] Submit forms with main ISPs to clear IP history
- [x] Make sure account suppression list is on
# Assumptions
-
# Additional Info/Resources
-
# Out of Scope
-
# Open Questions
-
|
1.0
|
Get Amazon SES into production - # Value Statement
**As** VANotify platform developer
**I want to** onboard Amazon SES as a provider into production
**So that**I can use it to send emails and begin to "warm up"
# Acceptance Criteria
**GIVEN** a VANotify platform developer
**WHEN** is viewing email providers in the VANotify platform
**THEN** Amazon SES is an option with which to send emails
# Checklist
- [x] AWS Support request to get dedicated IP addresses
- [x] AWS Support request for sending limits?
- [x] AWS Support request to take us out of sandbox
- [x] Submit forms with main ISPs to clear IP history
- [x] Make sure account suppression list is on
# Assumptions
-
# Additional Info/Resources
-
# Out of Scope
-
# Open Questions
-
|
process
|
get amazon ses into production value statement as vanotify platform developer i want to onboard amazon ses as a provider into production so that i can use it to send emails and begin to warm up acceptance criteria given a vanotify platform developer when is viewing email providers in the vanotify platform then amazon ses is an option with which to send emails checklist aws support request to get dedicated ip addresses aws support request for sending limits aws support request to take us out of sandbox submit forms with main isps to clear ip history make sure account suppression list is on assumptions additional info resources out of scope open questions
| 1
|
12,552
| 14,976,925,172
|
IssuesEvent
|
2021-01-28 08:48:01
|
threefoldtech/js-sdk
|
https://api.github.com/repos/threefoldtech/js-sdk
|
closed
|
failed to deploy mattermost
|
process_wontfix
|

failed to deploy mattermost after providing my MySQL username and password
|
1.0
|
failed to deploy mattermost - 
failed to deploy mattermost after providing my MySQL username and password
|
process
|
failed to deploy mattermost failed to deploy mattermost after providing my mysql username and password
| 1
|
809,158
| 30,177,465,968
|
IssuesEvent
|
2023-07-04 06:23:40
|
channel-io/bezier-react
|
https://api.github.com/repos/channel-io/bezier-react
|
closed
|
Improve `Tooltip`'s ghosting
|
enhancement priority:A
|
### Description
툴팁의 잔상 현상을 개선합니다.
### Reasons for suggestion
다수의 툴팁이 있는 경우, 툴팁의 딜레이가 현재 `0` 으로 설정되어 있기 때문에 마우스 포인터 이동에 불필요하게 많은 툴팁이 보이는 경우가 있습니다.
### Proposed solution
1. `delayShow`, `skipDelayShow` 옵션을 적절히 조절합니다.
2. 애니메이션의 영향일 수 있습니다. 1번 안에 더해, 애니메이션을 제거한 상태에서 동작을 점검해보면 좋겠습니다. 필요에 따라 애니메이션을 제거합니다.
### References
- [Feedback](https://desk.channel.io/root/groups/WebDesk-113051/649e88b11b0ee9ec8965)
|
1.0
|
Improve `Tooltip`'s ghosting - ### Description
툴팁의 잔상 현상을 개선합니다.
### Reasons for suggestion
다수의 툴팁이 있는 경우, 툴팁의 딜레이가 현재 `0` 으로 설정되어 있기 때문에 마우스 포인터 이동에 불필요하게 많은 툴팁이 보이는 경우가 있습니다.
### Proposed solution
1. `delayShow`, `skipDelayShow` 옵션을 적절히 조절합니다.
2. 애니메이션의 영향일 수 있습니다. 1번 안에 더해, 애니메이션을 제거한 상태에서 동작을 점검해보면 좋겠습니다. 필요에 따라 애니메이션을 제거합니다.
### References
- [Feedback](https://desk.channel.io/root/groups/WebDesk-113051/649e88b11b0ee9ec8965)
|
non_process
|
improve tooltip s ghosting description 툴팁의 잔상 현상을 개선합니다 reasons for suggestion 다수의 툴팁이 있는 경우 툴팁의 딜레이가 현재 으로 설정되어 있기 때문에 마우스 포인터 이동에 불필요하게 많은 툴팁이 보이는 경우가 있습니다 proposed solution delayshow skipdelayshow 옵션을 적절히 조절합니다 애니메이션의 영향일 수 있습니다 안에 더해 애니메이션을 제거한 상태에서 동작을 점검해보면 좋겠습니다 필요에 따라 애니메이션을 제거합니다 references
| 0
|
26,505
| 2,684,633,813
|
IssuesEvent
|
2015-03-29 06:08:13
|
gtcasl/gpuocelot
|
https://api.github.com/repos/gtcasl/gpuocelot
|
opened
|
Build Error: llvm/Assembly/Parser.h: No such file or directory
|
bug imported Priority-Medium
|
_From [RichardH...@gmail.com](https://code.google.com/u/107171517134746479087/) on November 11, 2014 16:57:28_
What steps will reproduce the problem? 1. Build Ocelot using the build.py script: What is the expected output? What do you see instead? expect: build successfully
build error: ocelot/ocelot/ir/implementation/ExternalFunctionSet.cpp:28:34: fatal error: llvm/Assembly/Parser.h: No such file or directory What version of the product are you using? On what operating system? I checkout the ocelot from svn.
Also, checkout the llvm from svn. Please provide any additional information below.
_Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=100_
|
1.0
|
Build Error: llvm/Assembly/Parser.h: No such file or directory - _From [RichardH...@gmail.com](https://code.google.com/u/107171517134746479087/) on November 11, 2014 16:57:28_
What steps will reproduce the problem? 1. Build Ocelot using the build.py script: What is the expected output? What do you see instead? expect: build successfully
build error: ocelot/ocelot/ir/implementation/ExternalFunctionSet.cpp:28:34: fatal error: llvm/Assembly/Parser.h: No such file or directory What version of the product are you using? On what operating system? I checkout the ocelot from svn.
Also, checkout the llvm from svn. Please provide any additional information below.
_Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=100_
|
non_process
|
build error llvm assembly parser h no such file or directory from on november what steps will reproduce the problem build ocelot using the build py script what is the expected output what do you see instead expect build successfully build error ocelot ocelot ir implementation externalfunctionset cpp fatal error llvm assembly parser h no such file or directory what version of the product are you using on what operating system i checkout the ocelot from svn also checkout the llvm from svn please provide any additional information below original issue
| 0
|
238,405
| 7,779,261,895
|
IssuesEvent
|
2018-06-05 16:16:32
|
emory-libraries/ezpaarse-platforms
|
https://api.github.com/repos/emory-libraries/ezpaarse-platforms
|
opened
|
Cambridge Companions Online
|
High Priority
|
### Example:star::star: :
https://proxy.library.emory.edu/login?url=http://cco.cambridge.org/uid=2892/private_home
### Priority:
High
### Subscriber (Library):
Theology
|
1.0
|
Cambridge Companions Online - ### Example:star::star: :
https://proxy.library.emory.edu/login?url=http://cco.cambridge.org/uid=2892/private_home
### Priority:
High
### Subscriber (Library):
Theology
|
non_process
|
cambridge companions online example star star priority high subscriber library theology
| 0
|
272,791
| 23,703,651,805
|
IssuesEvent
|
2022-08-29 21:39:44
|
KaTeX/KaTeX
|
https://api.github.com/repos/KaTeX/KaTeX
|
closed
|
Update `toSplitInto` matcher to accept the same delimiter object type that's used in renderMathInElement
|
testing good first issue
|
Right now `toSplitInto` uses `display: false` for all tests.
|
1.0
|
Update `toSplitInto` matcher to accept the same delimiter object type that's used in renderMathInElement - Right now `toSplitInto` uses `display: false` for all tests.
|
non_process
|
update tosplitinto matcher to accept the same delimiter object type that s used in rendermathinelement right now tosplitinto uses display false for all tests
| 0
|
239,157
| 18,264,849,594
|
IssuesEvent
|
2021-10-04 07:09:40
|
ayushs1ngh/Fun-and-Tricky-Programs
|
https://api.github.com/repos/ayushs1ngh/Fun-and-Tricky-Programs
|
closed
|
Add fun/tricky/ uncommon programs
|
documentation help wanted good first issue question hacktoberfest up-for-grabs
|
Add a PR with any tricky or uncommon coding problem.
Add coding problems and solutions in the respective language folder. if that language folder doesn't exist then create it.
Check other folders in Java or C/CPP how it is done in the rest of the file.
E.g
Add a folder in the Java folder with the Name of Problem Title.
Inside that folder add the solution ProblemTitle.java and a README.md
Make sure the class name is ProblemTitle that is the name of the file.
README.md must contain a detailed description of the problem with the test cases or example input/output.
|
1.0
|
Add fun/tricky/ uncommon programs - Add a PR with any tricky or uncommon coding problem.
Add coding problems and solutions in the respective language folder. if that language folder doesn't exist then create it.
Check other folders in Java or C/CPP how it is done in the rest of the file.
E.g
Add a folder in the Java folder with the Name of Problem Title.
Inside that folder add the solution ProblemTitle.java and a README.md
Make sure the class name is ProblemTitle that is the name of the file.
README.md must contain a detailed description of the problem with the test cases or example input/output.
|
non_process
|
add fun tricky uncommon programs add a pr with any tricky or uncommon coding problem add coding problems and solutions in the respective language folder if that language folder doesn t exist then create it check other folders in java or c cpp how it is done in the rest of the file e g add a folder in the java folder with the name of problem title inside that folder add the solution problemtitle java and a readme md make sure the class name is problemtitle that is the name of the file readme md must contain a detailed description of the problem with the test cases or example input output
| 0
|
967
| 2,522,132,609
|
IssuesEvent
|
2015-01-19 19:38:07
|
YaleSTC/shifts
|
https://api.github.com/repos/YaleSTC/shifts
|
closed
|
Fix fixtures
|
testing
|
Currently, the RSpec suite (#281) runs on fixtures. [Whether or not this is an antipattern remains a matter of taste.](https://whatdoitest.com/getting-friendly-with-fixtures) Since we have fixtures, not factories, fixing them is worth investigating.
(And if fixing fixtures is too taxing, migrating to factories will make #325 easier.)
|
1.0
|
Fix fixtures - Currently, the RSpec suite (#281) runs on fixtures. [Whether or not this is an antipattern remains a matter of taste.](https://whatdoitest.com/getting-friendly-with-fixtures) Since we have fixtures, not factories, fixing them is worth investigating.
(And if fixing fixtures is too taxing, migrating to factories will make #325 easier.)
|
non_process
|
fix fixtures currently the rspec suite runs on fixtures since we have fixtures not factories fixing them is worth investigating and if fixing fixtures is too taxing migrating to factories will make easier
| 0
|
45,793
| 13,133,247,902
|
IssuesEvent
|
2020-08-06 20:31:48
|
shaundmorris/ddf
|
https://api.github.com/repos/shaundmorris/ddf
|
closed
|
CVE-2016-1000343 High Severity Vulnerability detected by WhiteSource
|
security vulnerability wontfix
|
## CVE-2016-1000343 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.54.jar</b></p></summary>
<p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p>
<p>path: /root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar</p>
<p>
<p>Library home page: <a href=http://www.bouncycastle.org/java.html>http://www.bouncycastle.org/java.html</a></p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **bcprov-jdk15on-1.54.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Bouncy Castle JCE Provider version 1.55 and earlier the DSA key pair generator generates a weak private key if used with default values. If the JCA key pair generator is not explicitly initialised with DSA parameters, 1.55 and earlier generates a private value assuming a 1024 bit key size. In earlier releases this can be dealt with by explicitly passing parameters to the key pair generator.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-1000343>CVE-2016-1000343</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/bcgit/bc-java/commit/50a53068c094d6cff37659da33c9b4505becd389#diff-5578e61500abb2b87b300d3114bdfd7d">https://github.com/bcgit/bc-java/commit/50a53068c094d6cff37659da33c9b4505becd389#diff-5578e61500abb2b87b300d3114bdfd7d</a></p>
<p>Release Date: 2016-11-03</p>
<p>Fix Resolution: Replace or update the following files: KeyPairGeneratorSpi.java, DSATest.java</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-1000343 High Severity Vulnerability detected by WhiteSource - ## CVE-2016-1000343 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.54.jar</b></p></summary>
<p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p>
<p>path: /root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar,/root/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.54/bcprov-jdk15on-1.54.jar</p>
<p>
<p>Library home page: <a href=http://www.bouncycastle.org/java.html>http://www.bouncycastle.org/java.html</a></p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **bcprov-jdk15on-1.54.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Bouncy Castle JCE Provider version 1.55 and earlier the DSA key pair generator generates a weak private key if used with default values. If the JCA key pair generator is not explicitly initialised with DSA parameters, 1.55 and earlier generates a private value assuming a 1024 bit key size. In earlier releases this can be dealt with by explicitly passing parameters to the key pair generator.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-1000343>CVE-2016-1000343</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/bcgit/bc-java/commit/50a53068c094d6cff37659da33c9b4505becd389#diff-5578e61500abb2b87b300d3114bdfd7d">https://github.com/bcgit/bc-java/commit/50a53068c094d6cff37659da33c9b4505becd389#diff-5578e61500abb2b87b300d3114bdfd7d</a></p>
<p>Release Date: 2016-11-03</p>
<p>Fix Resolution: Replace or update the following files: KeyPairGeneratorSpi.java, DSATest.java</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high severity vulnerability detected by whitesource cve high severity vulnerability vulnerable library bcprov jar the bouncy castle crypto package is a java implementation of cryptographic algorithms this jar contains jce provider and lightweight api for the bouncy castle cryptography apis for jdk to jdk path root repository org bouncycastle bcprov bcprov jar root repository org bouncycastle bcprov bcprov jar root repository org bouncycastle bcprov bcprov jar root repository org bouncycastle bcprov bcprov jar root repository org bouncycastle bcprov bcprov jar library home page a href dependency hierarchy tika parsers jar root library x bcprov jar vulnerable library vulnerability details in the bouncy castle jce provider version and earlier the dsa key pair generator generates a weak private key if used with default values if the jca key pair generator is not explicitly initialised with dsa parameters and earlier generates a private value assuming a bit key size in earlier releases this can be dealt with by explicitly passing parameters to the key pair generator publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following files keypairgeneratorspi java dsatest java step up your open source security game with whitesource
| 0
|
507,406
| 14,679,979,421
|
IssuesEvent
|
2020-12-31 08:40:45
|
k8smeetup/website-tasks
|
https://api.github.com/repos/k8smeetup/website-tasks
|
opened
|
/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md
|
lang/zh priority/P0 sync/update version/master welcome
|
Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md](https://github.com/kubernetes/website/blob/master/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md)
Diff 命令参考:
```bash
# 查看原始文档与翻译文档更新差异
git diff --no-index -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md content/zh/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md
# 跨分支持查看原始文档更新差异
git diff release-1.19 master -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md
```
|
1.0
|
/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md - Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md](https://github.com/kubernetes/website/blob/master/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md)
Diff 命令参考:
```bash
# 查看原始文档与翻译文档更新差异
git diff --no-index -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md content/zh/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md
# 跨分支持查看原始文档更新差异
git diff release-1.19 master -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha_selfhosting.md
```
|
non_process
|
docs reference setup tools kubeadm generated kubeadm alpha selfhosting md source file diff 命令参考 bash 查看原始文档与翻译文档更新差异 git diff no index content en docs reference setup tools kubeadm generated kubeadm alpha selfhosting md content zh docs reference setup tools kubeadm generated kubeadm alpha selfhosting md 跨分支持查看原始文档更新差异 git diff release master content en docs reference setup tools kubeadm generated kubeadm alpha selfhosting md
| 0
|
57,785
| 3,083,786,505
|
IssuesEvent
|
2015-08-24 11:18:54
|
MinetestForFun/minetest-minetestforfun-server
|
https://api.github.com/repos/MinetestForFun/minetest-minetestforfun-server
|
opened
|
Some Maize node are not animated
|
Modding ➤ BugFix Priority: Low
|

Very small maize is animated, and the biggest one too. But not some intermediaries.
Very important. Or not. But it should be investigated once.
|
1.0
|
Some Maize node are not animated - 
Very small maize is animated, and the biggest one too. But not some intermediaries.
Very important. Or not. But it should be investigated once.
|
non_process
|
some maize node are not animated very small maize is animated and the biggest one too but not some intermediaries very important or not but it should be investigated once
| 0
|
3,178
| 6,254,039,060
|
IssuesEvent
|
2017-07-14 00:17:20
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
opened
|
UAPAOT: System.Diagnostics.Tests.ProcessStartInfoTests/StartInfo_NotepadWithContent(useShellExecute: False)
|
area-System.Diagnostics.Process os-windows-uwp test-run-uwp-coreclr
|
Failed in Outerloop during validation for #22237:
https://mc.dot.net/#/user/CIPop/pr~2Fjenkins~2Fdotnet~2Fcorefx~2Fmaster~2F/test~2Ffunctional~2Fcli~2F/343581b48553b789c3a92222e109b9d4b07927c1/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessStartInfoTests~2FStartInfo_NotepadWithContent(useShellExecute:%20False)
```
Windows.10.Amd64.ClientRS2.Open-Debug-x64
Unhandled Exception of Type System.InvalidOperationException
Message :
System.InvalidOperationException : WaitForInputIdle failed. This could be because the process does not have a graphical interface.
Stack Trace :
at System.Diagnostics.Process.WaitForInputIdleCore(Int32 milliseconds)
at System.Diagnostics.Process.WaitForInputIdle(Int32 milliseconds)
at System.Diagnostics.Process.WaitForInputIdle()
at System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_NotepadWithContent(Boolean useShellExecute)
```
|
1.0
|
UAPAOT: System.Diagnostics.Tests.ProcessStartInfoTests/StartInfo_NotepadWithContent(useShellExecute: False) - Failed in Outerloop during validation for #22237:
https://mc.dot.net/#/user/CIPop/pr~2Fjenkins~2Fdotnet~2Fcorefx~2Fmaster~2F/test~2Ffunctional~2Fcli~2F/343581b48553b789c3a92222e109b9d4b07927c1/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessStartInfoTests~2FStartInfo_NotepadWithContent(useShellExecute:%20False)
```
Windows.10.Amd64.ClientRS2.Open-Debug-x64
Unhandled Exception of Type System.InvalidOperationException
Message :
System.InvalidOperationException : WaitForInputIdle failed. This could be because the process does not have a graphical interface.
Stack Trace :
at System.Diagnostics.Process.WaitForInputIdleCore(Int32 milliseconds)
at System.Diagnostics.Process.WaitForInputIdle(Int32 milliseconds)
at System.Diagnostics.Process.WaitForInputIdle()
at System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_NotepadWithContent(Boolean useShellExecute)
```
|
process
|
uapaot system diagnostics tests processstartinfotests startinfo notepadwithcontent useshellexecute false failed in outerloop during validation for windows open debug unhandled exception of type system invalidoperationexception message system invalidoperationexception waitforinputidle failed this could be because the process does not have a graphical interface stack trace at system diagnostics process waitforinputidlecore milliseconds at system diagnostics process waitforinputidle milliseconds at system diagnostics process waitforinputidle at system diagnostics tests processstartinfotests startinfo notepadwithcontent boolean useshellexecute
| 1
|
19,676
| 26,031,724,271
|
IssuesEvent
|
2022-12-21 22:06:36
|
biocodellc/localcontexts_db
|
https://api.github.com/repos/biocodellc/localcontexts_db
|
closed
|
Registration: register button disable on submit
|
registration process
|
Disable button on registration page on submit while the page loads so that duplicate user profiles are not created.
|
1.0
|
Registration: register button disable on submit - Disable button on registration page on submit while the page loads so that duplicate user profiles are not created.
|
process
|
registration register button disable on submit disable button on registration page on submit while the page loads so that duplicate user profiles are not created
| 1
|
85,336
| 3,689,595,090
|
IssuesEvent
|
2016-02-25 16:57:22
|
icatproject/topcat
|
https://api.github.com/repos/icatproject/topcat
|
closed
|
Availability wording needs improving
|
Priority-Low TopCAT v2
|
The use of ONLINE, ARCHIVED etc may not be obvious to users. It is described in the help but many users will not use that. These values should probably be configurable with maybe a description given as a tooltip.
|
1.0
|
Availability wording needs improving - The use of ONLINE, ARCHIVED etc may not be obvious to users. It is described in the help but many users will not use that. These values should probably be configurable with maybe a description given as a tooltip.
|
non_process
|
availability wording needs improving the use of online archived etc may not be obvious to users it is described in the help but many users will not use that these values should probably be configurable with maybe a description given as a tooltip
| 0
|
110,229
| 16,977,084,746
|
IssuesEvent
|
2021-06-30 01:42:13
|
SmartBear/readyapi-swagger-assertion-plugin
|
https://api.github.com/repos/SmartBear/readyapi-swagger-assertion-plugin
|
closed
|
CVE-2014-0114 (High) detected in commons-beanutils-1.8.0.jar - autoclosed
|
security vulnerability
|
## CVE-2014-0114 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.8.0.jar</b></p></summary>
<p>BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Path to dependency file: readyapi-swagger-assertion-plugin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.0/commons-beanutils-1.8.0.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-1.7.0.jar (Root Library)
- jasperreports-4.0.1.jar
- :x: **commons-beanutils-1.8.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SmartBear/readyapi-swagger-assertion-plugin/commit/dc701947b2244c9b5dff8c9a75808fab2d107d60">dc701947b2244c9b5dff8c9a75808fab2d107d60</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.
<p>Publish Date: 2014-04-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114>CVE-2014-0114</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p>
<p>Release Date: 2014-04-30</p>
<p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-beanutils","packageName":"commons-beanutils","packageVersion":"1.8.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;net.sf.jasperreports:jasperreports:4.0.1;commons-beanutils:commons-beanutils:1.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2014-0114","vulnerabilityDetails":"Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to \"manipulate\" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2014-0114 (High) detected in commons-beanutils-1.8.0.jar - autoclosed - ## CVE-2014-0114 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.8.0.jar</b></p></summary>
<p>BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Path to dependency file: readyapi-swagger-assertion-plugin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.8.0/commons-beanutils-1.8.0.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-1.7.0.jar (Root Library)
- jasperreports-4.0.1.jar
- :x: **commons-beanutils-1.8.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SmartBear/readyapi-swagger-assertion-plugin/commit/dc701947b2244c9b5dff8c9a75808fab2d107d60">dc701947b2244c9b5dff8c9a75808fab2d107d60</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.
<p>Publish Date: 2014-04-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114>CVE-2014-0114</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p>
<p>Release Date: 2014-04-30</p>
<p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-beanutils","packageName":"commons-beanutils","packageVersion":"1.8.0","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;net.sf.jasperreports:jasperreports:4.0.1;commons-beanutils:commons-beanutils:1.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2014-0114","vulnerabilityDetails":"Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to \"manipulate\" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-0114","cvss2Severity":"high","cvss2Score":"7.5","extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in commons beanutils jar autoclosed cve high severity vulnerability vulnerable library commons beanutils jar beanutils provides an easy to use but flexible wrapper around reflection and introspection path to dependency file readyapi swagger assertion plugin pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy ready api soapui pro jar root library jasperreports jar x commons beanutils jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons beanutils as distributed in lib commons beanutils jar in apache struts x through and in other products requiring commons beanutils through does not suppress the class property which allows remote attackers to manipulate the classloader and execute arbitrary code via the class parameter as demonstrated by the passing of this parameter to the getclass method of the actionform object in struts publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution commons beanutils commons beanutils org apache struts core isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com smartbear ready api soapui pro net sf jasperreports jasperreports commons beanutils commons beanutils isminimumfixversionavailable true minimumfixversion commons beanutils commons beanutils org apache struts core basebranches vulnerabilityidentifier cve vulnerabilitydetails apache commons beanutils as distributed in lib commons beanutils jar in apache struts x through and in other products requiring commons beanutils through does not suppress the class property which allows remote attackers to manipulate the classloader and execute arbitrary code via the class parameter as demonstrated by the passing of this parameter to the getclass method of the actionform object in struts vulnerabilityurl
| 0
|
15,305
| 19,347,233,234
|
IssuesEvent
|
2021-12-15 12:09:32
|
ESMValGroup/ESMValCore
|
https://api.github.com/repos/ESMValGroup/ESMValCore
|
opened
|
Explicitly **not** using any fx variable is not possible in preprocessor
|
preprocessor
|
Due to the implementation of the check for the `fx_variables` option in the preprocessors, i.e. basically something like
```py
if not step_settings.get('fx_variables'):
# add default fx_vars
else:
# use fx_variables that have been specified
```
it is currently not possible to explicitly **not** use any fx variable. Empty dictionaries, lists, or `None` will all trigger the first `if`, and the default values are automatically added.
<details>
https://github.com/ESMValGroup/ESMValCore/blob/5131ffc4cdf5aa5c46215aea659740ea442769dc/esmvalcore/_recipe.py#L506-L551
(specifically line 515)
</details>
I think by explicitly stating something like
```yml
area_statistics:
operator: mean
fx_variables: {} # or fx_variables: [] or fx_variables: null
```
**no** fx variables should be used. Only if `fx_variables` is not used at all, the default values should be used.
Any opinions on this @ESMValGroup/esmvaltool-coreteam ?
|
1.0
|
Explicitly **not** using any fx variable is not possible in preprocessor - Due to the implementation of the check for the `fx_variables` option in the preprocessors, i.e. basically something like
```py
if not step_settings.get('fx_variables'):
# add default fx_vars
else:
# use fx_variables that have been specified
```
it is currently not possible to explicitly **not** use any fx variable. Empty dictionaries, lists, or `None` will all trigger the first `if`, and the default values are automatically added.
<details>
https://github.com/ESMValGroup/ESMValCore/blob/5131ffc4cdf5aa5c46215aea659740ea442769dc/esmvalcore/_recipe.py#L506-L551
(specifically line 515)
</details>
I think by explicitly stating something like
```yml
area_statistics:
operator: mean
fx_variables: {} # or fx_variables: [] or fx_variables: null
```
**no** fx variables should be used. Only if `fx_variables` is not used at all, the default values should be used.
Any opinions on this @ESMValGroup/esmvaltool-coreteam ?
|
process
|
explicitly not using any fx variable is not possible in preprocessor due to the implementation of the check for the fx variables option in the preprocessors i e basically something like py if not step settings get fx variables add default fx vars else use fx variables that have been specified it is currently not possible to explicitly not use any fx variable empty dictionaries lists or none will all trigger the first if and the default values are automatically added specifically line i think by explicitly stating something like yml area statistics operator mean fx variables or fx variables or fx variables null no fx variables should be used only if fx variables is not used at all the default values should be used any opinions on this esmvalgroup esmvaltool coreteam
| 1
|
168
| 2,586,801,090
|
IssuesEvent
|
2015-02-17 14:39:51
|
tinkerpop/tinkerpop3
|
https://api.github.com/repos/tinkerpop/tinkerpop3
|
closed
|
We need a couple more Scope.local steps.
|
enhancement process
|
We have `order(local)`. But we also need:
* `dedup(local)` : if the current object is a collection, dedup it.
* `sample(local)`: if the current object is a collection, sample it.
I think thats it. Look at `OrderLocalStep` and `OrderGlobalStep` to see why its important that they are different classes --- one is a barrier, the other is not.
|
1.0
|
We need a couple more Scope.local steps. - We have `order(local)`. But we also need:
* `dedup(local)` : if the current object is a collection, dedup it.
* `sample(local)`: if the current object is a collection, sample it.
I think thats it. Look at `OrderLocalStep` and `OrderGlobalStep` to see why its important that they are different classes --- one is a barrier, the other is not.
|
process
|
we need a couple more scope local steps we have order local but we also need dedup local if the current object is a collection dedup it sample local if the current object is a collection sample it i think thats it look at orderlocalstep and orderglobalstep to see why its important that they are different classes one is a barrier the other is not
| 1
|
10,354
| 13,179,289,990
|
IssuesEvent
|
2020-08-12 10:38:44
|
shirou/gopsutil
|
https://api.github.com/repos/shirou/gopsutil
|
closed
|
func Processes() ([]*Process, error) can pass parameters
|
package:process
|
func Processes() ([]*Process, error) This function, whether it is possible to pass parameters, once called, will collect all the process information of the system, which consumes more CPU, if you can selectively collect process information, it will be more it is good. Hope to adopt
**Additional context**
[Add any other context or screenshots about the feature request here.]
|
1.0
|
func Processes() ([]*Process, error) can pass parameters - func Processes() ([]*Process, error) This function, whether it is possible to pass parameters, once called, will collect all the process information of the system, which consumes more CPU, if you can selectively collect process information, it will be more it is good. Hope to adopt
**Additional context**
[Add any other context or screenshots about the feature request here.]
|
process
|
func processes process error can pass parameters func processes process error this function whether it is possible to pass parameters once called will collect all the process information of the system which consumes more cpu if you can selectively collect process information it will be more it is good hope to adopt additional context
| 1
|
21,382
| 29,202,230,096
|
IssuesEvent
|
2023-05-21 00:37:14
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Hibrido / ] Fullstack Developer (Sênior - Híbrido - Belo Horizonte) na Coodesh
|
SALVADOR PJ JAVA MYSQL JAVASCRIPT FULL-STACK SENIOR PRIMEFACES JSF SPRING SQL GIT HIBERNATE MAVEN REST SOAP JSON ANGULAR REQUISITOS NGINX PROCESSOS INOVAÇÃO GITHUB APACHE UMA C DOCUMENTAÇÃO WILDFLY HTTP MANUTENÇÃO HIBRIDO ALOCADO Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/fullstack-developer-senior-hibrido-belo-horizonte-200652198?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>Prime Results </strong>está buscando <strong><ins>Fullstack Developer</ins></strong> para compor seu time!</p>
<p>Acreditamos no poder de transformação social realizado pelas empresas. Acreditamos no poder transformador das pessoas, aliado à gestão e tecnologia. Compartilhamos nosso conhecimento para solucionar problemas complexos e gerar valor para nossos clientes.</p>
<p><strong>Responsabilidades:</strong></p>
<ul>
<li>Desenvolvimento/implementação e manutenção de aplicações;</li>
<li>Participar da análise e execução dos projetos e execução dos tickets;</li>
<li>Definir as atividades necessárias para a realização de projetos, analisando os impactos em sistemas e processos através do entendimento da necessidade, conhecimento técnico e arquitetônico dos sistemas;</li>
<li>Desenvolver códigos para atendimento às áreas e empresas clientes, proporcionando o esclarecimento de dúvidas relacionados ao projeto, contribuindo para uma melhor análise de impactos de processos e sistemas sob sua responsabilidade; </li>
<li>Participar das atividades de planejamento para a liberação do produto para homologação e produção, por meio da validação de testes de aceite, assim como documentação de não conformidades avaliando e planejando a execução das correções reportadas;</li>
<li>Participar da rotina de SQUADs. </li>
</ul>
<p></p>
## Prime Results :
<p>O Best Seller Simon Sinek, diz que a maioria das empresas sabem o que fazem, porém não sabem por que o fazem. Não é o nosso caso. A Prime Results é uma empresa especializada em gestão organizacional que usa seu potencial de transformação em empresas que geram impacto positivo na sociedade. Nossos clientes hoje, fazem a diferença na vida de mais de 250.000 brasileiros, nas áreas de proteção patrimonial, saúde e assistência 24 horas. </p>
<p>Nosso objetivo central é criar um ambiente criativo, dinâmico e engajado, sempre aliados a métodos, processos inteligentes e muita inovação.</p><a href='https://coodesh.com/empresas/prime-results'>Veja mais no site</a>
## Habilidades:
- Java
- Hibernate
- Angular
- Javascript
- JSON
- Apache
- MySQL
- Microsoft SQL Server
- Spring
## Local:
undefined
## Requisitos:
- Experiência em Java: JSF, Spring, PrimeFaces, Hibernate, JasperReports;
- Conhecimentos em modelagem e desenvolvimento de Bancos de Dados relacionais: MySQL, SQL Server;
- Conhecimentos em Tecnologias Web: HTML5, CSS e Frameworks JavaScript, Angular;
- Conhecimento de Arquiteturas Web e Serviços (HTTP, SOAP, REST ou JSON);
- Conhecimentos nas ferramentas: GIT e Maven;
- Conhecimentos técnicos em servidores de aplicação (Wildfly - J2EE), servidores web (Apache e NGINX) e Spring Boot.
## Benefícios:
- GymPass;
- Assistência Médica após o período de experiência.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (Sênior - Híbrido - Belo Horizonte) na Prime Results ](https://coodesh.com/vagas/fullstack-developer-senior-hibrido-belo-horizonte-200652198?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Alocado
#### Regime
PJ
#### Categoria
Full-Stack
|
1.0
|
[Hibrido / ] Fullstack Developer (Sênior - Híbrido - Belo Horizonte) na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/fullstack-developer-senior-hibrido-belo-horizonte-200652198?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>Prime Results </strong>está buscando <strong><ins>Fullstack Developer</ins></strong> para compor seu time!</p>
<p>Acreditamos no poder de transformação social realizado pelas empresas. Acreditamos no poder transformador das pessoas, aliado à gestão e tecnologia. Compartilhamos nosso conhecimento para solucionar problemas complexos e gerar valor para nossos clientes.</p>
<p><strong>Responsabilidades:</strong></p>
<ul>
<li>Desenvolvimento/implementação e manutenção de aplicações;</li>
<li>Participar da análise e execução dos projetos e execução dos tickets;</li>
<li>Definir as atividades necessárias para a realização de projetos, analisando os impactos em sistemas e processos através do entendimento da necessidade, conhecimento técnico e arquitetônico dos sistemas;</li>
<li>Desenvolver códigos para atendimento às áreas e empresas clientes, proporcionando o esclarecimento de dúvidas relacionados ao projeto, contribuindo para uma melhor análise de impactos de processos e sistemas sob sua responsabilidade; </li>
<li>Participar das atividades de planejamento para a liberação do produto para homologação e produção, por meio da validação de testes de aceite, assim como documentação de não conformidades avaliando e planejando a execução das correções reportadas;</li>
<li>Participar da rotina de SQUADs. </li>
</ul>
<p></p>
## Prime Results :
<p>O Best Seller Simon Sinek, diz que a maioria das empresas sabem o que fazem, porém não sabem por que o fazem. Não é o nosso caso. A Prime Results é uma empresa especializada em gestão organizacional que usa seu potencial de transformação em empresas que geram impacto positivo na sociedade. Nossos clientes hoje, fazem a diferença na vida de mais de 250.000 brasileiros, nas áreas de proteção patrimonial, saúde e assistência 24 horas. </p>
<p>Nosso objetivo central é criar um ambiente criativo, dinâmico e engajado, sempre aliados a métodos, processos inteligentes e muita inovação.</p><a href='https://coodesh.com/empresas/prime-results'>Veja mais no site</a>
## Habilidades:
- Java
- Hibernate
- Angular
- Javascript
- JSON
- Apache
- MySQL
- Microsoft SQL Server
- Spring
## Local:
undefined
## Requisitos:
- Experiência em Java: JSF, Spring, PrimeFaces, Hibernate, JasperReports;
- Conhecimentos em modelagem e desenvolvimento de Bancos de Dados relacionais: MySQL, SQL Server;
- Conhecimentos em Tecnologias Web: HTML5, CSS e Frameworks JavaScript, Angular;
- Conhecimento de Arquiteturas Web e Serviços (HTTP, SOAP, REST ou JSON);
- Conhecimentos nas ferramentas: GIT e Maven;
- Conhecimentos técnicos em servidores de aplicação (Wildfly - J2EE), servidores web (Apache e NGINX) e Spring Boot.
## Benefícios:
- GymPass;
- Assistência Médica após o período de experiência.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (Sênior - Híbrido - Belo Horizonte) na Prime Results ](https://coodesh.com/vagas/fullstack-developer-senior-hibrido-belo-horizonte-200652198?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Alocado
#### Regime
PJ
#### Categoria
Full-Stack
|
process
|
fullstack developer sênior híbrido belo horizonte na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a prime results está buscando fullstack developer para compor seu time acreditamos no poder de transformação social realizado pelas empresas acreditamos no poder transformador das pessoas aliado à gestão e tecnologia compartilhamos nosso conhecimento para solucionar problemas complexos e gerar valor para nossos clientes responsabilidades desenvolvimento implementação e manutenção de aplicações participar da análise e execução dos projetos e execução dos tickets definir as atividades necessárias para a realização de projetos analisando os impactos em sistemas e processos através do entendimento da necessidade conhecimento técnico e arquitetônico dos sistemas desenvolver códigos para atendimento às áreas e empresas clientes proporcionando o esclarecimento de dúvidas relacionados ao projeto contribuindo para uma melhor análise de impactos de processos e sistemas sob sua responsabilidade nbsp participar das atividades de planejamento para a liberação do produto para homologação e produção por meio da validação de testes de aceite assim como documentação de não conformidades avaliando e planejando a execução das correções reportadas participar da rotina de squads nbsp prime results o best seller simon sinek diz que a maioria das empresas sabem o que fazem porém não sabem por que o fazem não é o nosso caso a prime results é uma empresa especializada em gestão organizacional que usa seu potencial de transformação em empresas que geram impacto positivo na sociedade nossos clientes hoje fazem a diferença na vida de mais de brasileiros nas áreas de proteção patrimonial saúde e assistência horas nbsp nosso objetivo central é criar um ambiente criativo dinâmico e engajado sempre aliados a métodos processos inteligentes e muita inovação habilidades java hibernate angular javascript json apache mysql microsoft sql server spring local undefined requisitos experiência em java jsf spring primefaces hibernate jasperreports conhecimentos em modelagem e desenvolvimento de bancos de dados relacionais mysql sql server conhecimentos em tecnologias web css e frameworks javascript angular conhecimento de arquiteturas web e serviços http soap rest ou json conhecimentos nas ferramentas git e maven conhecimentos técnicos em servidores de aplicação wildfly servidores web apache e nginx e spring boot benefícios gympass assistência médica após o período de experiência como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado regime pj categoria full stack
| 1
|
116,747
| 24,983,260,346
|
IssuesEvent
|
2022-11-02 13:22:48
|
arduino/arduino-cli
|
https://api.github.com/repos/arduino/arduino-cli
|
opened
|
Previous custom discovery tool process left running after update
|
topic: code type: imperfection
|
### Describe the problem
In addition to the always available [**serial-discovery**](https://github.com/arduino/serial-discovery) and [**mdns-discovery**](https://github.com/arduino/mdns-discovery) [pluggable discovery](https://arduino.github.io/arduino-cli/dev/pluggable-discovery-specification/) tools of the `builtin` package, dependencies on custom discovery tools [can be specified](https://arduino.github.io/arduino-cli/dev/package_index_json-specification/#tools-definitions:~:text=discoveryDependencies%3A%20the%20Pluggable%20Discoveries%20needed%20by%20this%20platform) by boards platforms.
[The latest version of the custom discovery dependency is installed](https://arduino.github.io/arduino-cli/dev/package_index_json-specification/#tools-definitions:~:text=the%20latest%20installed%20discovery%20tool%20will%20always%20be%20used) when installing or updating that platform.
🐛 The process for the previous version of a custom discovery tool is left running after a boards platform update that updates the discovery tool.
### To reproduce
#### Setup
```text
$ arduino-cli version
arduino-cli Version: git-snapshot Commit: 732bdc54 Date: 2022-09-07T09:00:01Z
$ export ARDUINO_DIRECTORIES_DATA="/tmp/arduino-cli-directories/data"
$ export ARDUINO_BOARD_MANAGER_ADDITIONAL_URLS="https://www.pjrc.com/teensy/package_teensy_index.json"
$ arduino-cli core update-index
$ arduino-cli core install teensy:avr@1.57.0 # Install outdated version of a platform with custom pluggable discovery tool dependency
[...]
Installing teensy:teensy-discovery@1.57.1...
teensy:teensy-discovery@1.57.1 installed
[...]
$ mv "${ARDUINO_DIRECTORIES_DATA}/packages/teensy/tools/teensy-discovery/1.57.1" "${ARDUINO_DIRECTORIES_DATA}/packages/teensy/tools/teensy-discovery/1.57.0" # Simulate outdated discovery tool
$ arduino-cli daemon --debug
```
#### Demo
run the following [`grpcurl`](https://github.com/fullstorydev/grpcurl) commands in another terminal:
```text
$ grpcurl \
-plaintext \
-import-path ./rpc \
-proto cc/arduino/cli/commands/v1/commands.proto \
127.0.0.1:50051 \
cc.arduino.cli.commands.v1.ArduinoCoreService.Create
{
"instance": {
"id": 1
}
}
$ grpcurl \
-plaintext \
-import-path ./rpc \
-proto cc/arduino/cli/commands/v1/commands.proto \
-d '{"instance": {"id": 1}}' \
127.0.0.1:50051 \
cc.arduino.cli.commands.v1.ArduinoCoreService.Init
$ pgrep -a -f 'teensy-discovery'
$ grpcurl \
-plaintext \
-import-path ./rpc \
-proto cc/arduino/cli/commands/v1/commands.proto \
-d '{"instance": {"id": 1}}' \
127.0.0.1:50051 \
cc.arduino.cli.commands.v1.ArduinoCoreService.BoardListWatch
[...]
$ pgrep -a -f 'teensy-discovery'
142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery
$ grpcurl \
-plaintext \
-import-path ./rpc \
-proto cc/arduino/cli/commands/v1/commands.proto \
-d '{"instance": {"id": 1}, "platform_package": "teensy", "architecture": "avr"}' \
127.0.0.1:50051 \
cc.arduino.cli.commands.v1.ArduinoCoreService.PlatformUpgrade
[...]
{
"taskProgress": {
"name": "Installing teensy:teensy-discovery@1.57.1"
}
}
[...]
$ pgrep -a -f 'teensy-discovery'
142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery
$ grpcurl \
-plaintext \
-import-path ./rpc \
-proto cc/arduino/cli/commands/v1/commands.proto \
-d '{"instance": {"id": 1}}' \
127.0.0.1:50051 \
cc.arduino.cli.commands.v1.ArduinoCoreService.BoardListWatch
[...]
$ pgrep -a -f 'teensy-discovery'
142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery
142363 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.1/teensy-discovery
```
🐛 There is a leftover `teensy:teensy-discovery@1.57.0` process still running in addition to `teensy:teensy-discovery@1.57.1`.
### Expected behavior
Discovery tool processes are cleaned up.
### Arduino CLI version
732bdc54
### Operating system
Windows, Ubuntu
### Operating system version
Windows 10, Ubuntu 20.04
### Additional context
Originally reported by @PaulStoffregen at https://forum.pjrc.com/threads/71370-Teensy4-1-reading-from-Serial-Monitor-of-Arduino-IDE?p=315184#post315184
---
I can also reproduce the issue using the equivalent command line interface commands. I chose to use the gRPC interface for the demo because the command line version seems unlikely to occur in real world usage, while the gRPC version is a simplification of an existing real world occurrence in Arduino IDE 2.x.
---
I suspected a similar problem might apply to the pluggable monitor tools (e.g., **teensy-monitor**). I did a quick check using Arduino IDE 2.x and found that only the previously running **teensy-monitor** process was running after the update. After I closed and reopened Serial Monitor, only a single **teensy-monitor** process for the updated version was running.
### Issue checklist
- [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-cli/issues?q=)
- [X] I verified the problem still occurs when using the [nightly build](https://arduino.github.io/arduino-cli/dev/installation/#nightly-builds)
- [X] My report contains all necessary details
|
1.0
|
Previous custom discovery tool process left running after update - ### Describe the problem
In addition to the always available [**serial-discovery**](https://github.com/arduino/serial-discovery) and [**mdns-discovery**](https://github.com/arduino/mdns-discovery) [pluggable discovery](https://arduino.github.io/arduino-cli/dev/pluggable-discovery-specification/) tools of the `builtin` package, dependencies on custom discovery tools [can be specified](https://arduino.github.io/arduino-cli/dev/package_index_json-specification/#tools-definitions:~:text=discoveryDependencies%3A%20the%20Pluggable%20Discoveries%20needed%20by%20this%20platform) by boards platforms.
[The latest version of the custom discovery dependency is installed](https://arduino.github.io/arduino-cli/dev/package_index_json-specification/#tools-definitions:~:text=the%20latest%20installed%20discovery%20tool%20will%20always%20be%20used) when installing or updating that platform.
🐛 The process for the previous version of a custom discovery tool is left running after a boards platform update that updates the discovery tool.
### To reproduce
#### Setup
```text
$ arduino-cli version
arduino-cli Version: git-snapshot Commit: 732bdc54 Date: 2022-09-07T09:00:01Z
$ export ARDUINO_DIRECTORIES_DATA="/tmp/arduino-cli-directories/data"
$ export ARDUINO_BOARD_MANAGER_ADDITIONAL_URLS="https://www.pjrc.com/teensy/package_teensy_index.json"
$ arduino-cli core update-index
$ arduino-cli core install teensy:avr@1.57.0 # Install outdated version of a platform with custom pluggable discovery tool dependency
[...]
Installing teensy:teensy-discovery@1.57.1...
teensy:teensy-discovery@1.57.1 installed
[...]
$ mv "${ARDUINO_DIRECTORIES_DATA}/packages/teensy/tools/teensy-discovery/1.57.1" "${ARDUINO_DIRECTORIES_DATA}/packages/teensy/tools/teensy-discovery/1.57.0" # Simulate outdated discovery tool
$ arduino-cli daemon --debug
```
#### Demo
run the following [`grpcurl`](https://github.com/fullstorydev/grpcurl) commands in another terminal:
```text
$ grpcurl \
-plaintext \
-import-path ./rpc \
-proto cc/arduino/cli/commands/v1/commands.proto \
127.0.0.1:50051 \
cc.arduino.cli.commands.v1.ArduinoCoreService.Create
{
"instance": {
"id": 1
}
}
$ grpcurl \
-plaintext \
-import-path ./rpc \
-proto cc/arduino/cli/commands/v1/commands.proto \
-d '{"instance": {"id": 1}}' \
127.0.0.1:50051 \
cc.arduino.cli.commands.v1.ArduinoCoreService.Init
$ pgrep -a -f 'teensy-discovery'
$ grpcurl \
-plaintext \
-import-path ./rpc \
-proto cc/arduino/cli/commands/v1/commands.proto \
-d '{"instance": {"id": 1}}' \
127.0.0.1:50051 \
cc.arduino.cli.commands.v1.ArduinoCoreService.BoardListWatch
[...]
$ pgrep -a -f 'teensy-discovery'
142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery
$ grpcurl \
-plaintext \
-import-path ./rpc \
-proto cc/arduino/cli/commands/v1/commands.proto \
-d '{"instance": {"id": 1}, "platform_package": "teensy", "architecture": "avr"}' \
127.0.0.1:50051 \
cc.arduino.cli.commands.v1.ArduinoCoreService.PlatformUpgrade
[...]
{
"taskProgress": {
"name": "Installing teensy:teensy-discovery@1.57.1"
}
}
[...]
$ pgrep -a -f 'teensy-discovery'
142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery
$ grpcurl \
-plaintext \
-import-path ./rpc \
-proto cc/arduino/cli/commands/v1/commands.proto \
-d '{"instance": {"id": 1}}' \
127.0.0.1:50051 \
cc.arduino.cli.commands.v1.ArduinoCoreService.BoardListWatch
[...]
$ pgrep -a -f 'teensy-discovery'
142276 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.0/teensy-discovery
142363 /tmp/arduino-cli-directories/data/packages/teensy/tools/teensy-discovery/1.57.1/teensy-discovery
```
🐛 There is a leftover `teensy:teensy-discovery@1.57.0` process still running in addition to `teensy:teensy-discovery@1.57.1`.
### Expected behavior
Discovery tool processes are cleaned up.
### Arduino CLI version
732bdc54
### Operating system
Windows, Ubuntu
### Operating system version
Windows 10, Ubuntu 20.04
### Additional context
Originally reported by @PaulStoffregen at https://forum.pjrc.com/threads/71370-Teensy4-1-reading-from-Serial-Monitor-of-Arduino-IDE?p=315184#post315184
---
I can also reproduce the issue using the equivalent command line interface commands. I chose to use the gRPC interface for the demo because the command line version seems unlikely to occur in real world usage, while the gRPC version is a simplification of an existing real world occurrence in Arduino IDE 2.x.
---
I suspected a similar problem might apply to the pluggable monitor tools (e.g., **teensy-monitor**). I did a quick check using Arduino IDE 2.x and found that only the previously running **teensy-monitor** process was running after the update. After I closed and reopened Serial Monitor, only a single **teensy-monitor** process for the updated version was running.
### Issue checklist
- [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-cli/issues?q=)
- [X] I verified the problem still occurs when using the [nightly build](https://arduino.github.io/arduino-cli/dev/installation/#nightly-builds)
- [X] My report contains all necessary details
|
non_process
|
previous custom discovery tool process left running after update describe the problem in addition to the always available and tools of the builtin package dependencies on custom discovery tools by boards platforms when installing or updating that platform 🐛 the process for the previous version of a custom discovery tool is left running after a boards platform update that updates the discovery tool to reproduce setup text arduino cli version arduino cli version git snapshot commit date export arduino directories data tmp arduino cli directories data export arduino board manager additional urls arduino cli core update index arduino cli core install teensy avr install outdated version of a platform with custom pluggable discovery tool dependency installing teensy teensy discovery teensy teensy discovery installed mv arduino directories data packages teensy tools teensy discovery arduino directories data packages teensy tools teensy discovery simulate outdated discovery tool arduino cli daemon debug demo run the following commands in another terminal text grpcurl plaintext import path rpc proto cc arduino cli commands commands proto cc arduino cli commands arduinocoreservice create instance id grpcurl plaintext import path rpc proto cc arduino cli commands commands proto d instance id cc arduino cli commands arduinocoreservice init pgrep a f teensy discovery grpcurl plaintext import path rpc proto cc arduino cli commands commands proto d instance id cc arduino cli commands arduinocoreservice boardlistwatch pgrep a f teensy discovery tmp arduino cli directories data packages teensy tools teensy discovery teensy discovery grpcurl plaintext import path rpc proto cc arduino cli commands commands proto d instance id platform package teensy architecture avr cc arduino cli commands arduinocoreservice platformupgrade taskprogress name installing teensy teensy discovery pgrep a f teensy discovery tmp arduino cli directories data packages teensy tools teensy discovery teensy discovery grpcurl plaintext import path rpc proto cc arduino cli commands commands proto d instance id cc arduino cli commands arduinocoreservice boardlistwatch pgrep a f teensy discovery tmp arduino cli directories data packages teensy tools teensy discovery teensy discovery tmp arduino cli directories data packages teensy tools teensy discovery teensy discovery 🐛 there is a leftover teensy teensy discovery process still running in addition to teensy teensy discovery expected behavior discovery tool processes are cleaned up arduino cli version operating system windows ubuntu operating system version windows ubuntu additional context originally reported by paulstoffregen at i can also reproduce the issue using the equivalent command line interface commands i chose to use the grpc interface for the demo because the command line version seems unlikely to occur in real world usage while the grpc version is a simplification of an existing real world occurrence in arduino ide x i suspected a similar problem might apply to the pluggable monitor tools e g teensy monitor i did a quick check using arduino ide x and found that only the previously running teensy monitor process was running after the update after i closed and reopened serial monitor only a single teensy monitor process for the updated version was running issue checklist i searched for previous reports in i verified the problem still occurs when using the my report contains all necessary details
| 0
|
6,316
| 9,329,627,188
|
IssuesEvent
|
2019-03-28 03:11:50
|
brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
|
https://api.github.com/repos/brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
|
closed
|
Реализовать часть программы
|
C++ Work in process
|
# Написать часть программы.
###
* В этой части должен содержаться хотя бы ввод исходной строки. Пример кода.
`#include <iostream>`
`int main ()`
`{`
`string ishodnik;`
`cin >> ishodnik;`
`return 0;`
`}`
|
1.0
|
Реализовать часть программы - # Написать часть программы.
###
* В этой части должен содержаться хотя бы ввод исходной строки. Пример кода.
`#include <iostream>`
`int main ()`
`{`
`string ishodnik;`
`cin >> ishodnik;`
`return 0;`
`}`
|
process
|
реализовать часть программы написать часть программы в этой части должен содержаться хотя бы ввод исходной строки пример кода include int main string ishodnik cin ishodnik return
| 1
|
91,000
| 11,456,395,416
|
IssuesEvent
|
2020-02-06 21:08:07
|
tokio-rs/tracing
|
https://api.github.com/repos/tokio-rs/tracing
|
closed
|
trace: Determine new syntax for uninitialized fields
|
crate/tracing kind/rfc needs/design
|
Followup from #1062 and https://github.com/tokio-rs/tokio/pull/1103#issuecomment-500019388.
Prior to #1103, the syntax `field,` was used by the `tokio-trace` macros to indicate span fields which have yet to have a value recorded. In order to support shorthand for using local variables as fields, we changed the meaning of this syntax, since local variable shorthand is likely to be a more common use-case (see #1103).
This leaves uninitialized fields without an appropriate macro syntax. There have been some suggestions, of which I think the strongest thus far are `let field,` and `field = _,`.
`let field` nicely mirrors the way let bindings with unset values are created in Rust; it does, however, seem somewhat inconsistent with the existing syntax, since other fields do not begin with `let`. On the other hand, `field = _` fits in with the existing field syntax, but @carllerche has pointed out that it uses the assignment syntax to indicate something which is _unassigned_, which seems confusing.
We should try to agree on a new syntax for uninitialized fields, and re-enable support for them in the macros.
|
1.0
|
trace: Determine new syntax for uninitialized fields - Followup from #1062 and https://github.com/tokio-rs/tokio/pull/1103#issuecomment-500019388.
Prior to #1103, the syntax `field,` was used by the `tokio-trace` macros to indicate span fields which have yet to have a value recorded. In order to support shorthand for using local variables as fields, we changed the meaning of this syntax, since local variable shorthand is likely to be a more common use-case (see #1103).
This leaves uninitialized fields without an appropriate macro syntax. There have been some suggestions, of which I think the strongest thus far are `let field,` and `field = _,`.
`let field` nicely mirrors the way let bindings with unset values are created in Rust; it does, however, seem somewhat inconsistent with the existing syntax, since other fields do not begin with `let`. On the other hand, `field = _` fits in with the existing field syntax, but @carllerche has pointed out that it uses the assignment syntax to indicate something which is _unassigned_, which seems confusing.
We should try to agree on a new syntax for uninitialized fields, and re-enable support for them in the macros.
|
non_process
|
trace determine new syntax for uninitialized fields followup from and prior to the syntax field was used by the tokio trace macros to indicate span fields which have yet to have a value recorded in order to support shorthand for using local variables as fields we changed the meaning of this syntax since local variable shorthand is likely to be a more common use case see this leaves uninitialized fields without an appropriate macro syntax there have been some suggestions of which i think the strongest thus far are let field and field let field nicely mirrors the way let bindings with unset values are created in rust it does however seem somewhat inconsistent with the existing syntax since other fields do not begin with let on the other hand field fits in with the existing field syntax but carllerche has pointed out that it uses the assignment syntax to indicate something which is unassigned which seems confusing we should try to agree on a new syntax for uninitialized fields and re enable support for them in the macros
| 0
|
11,573
| 14,442,128,623
|
IssuesEvent
|
2020-12-07 17:42:50
|
Arch666Angel/mods
|
https://api.github.com/repos/Arch666Angel/mods
|
opened
|
Module tech icons wrong size
|
Angels Bio Processing Impact: Bug
|
**Describe the bug**
Module tech icons have the wrong size
**Screenshots**
If applicable, add screenshots to help explain your problem.

**Additional context**
This is already fixed for the player crafting menu group, however the tech icons I overlooked:
https://github.com/Arch666Angel/mods/blob/b45c76fe1eb650ea5acf3a1230fbdb2c356c2743/angelsbioprocessing/prototypes/overrides/bio-processing-override-bob.lua#L160-L163
https://github.com/Arch666Angel/mods/blob/b45c76fe1eb650ea5acf3a1230fbdb2c356c2743/angelsbioprocessing/prototypes/overrides/bio-processing-override-bob.lua#L226-L229
|
1.0
|
Module tech icons wrong size - **Describe the bug**
Module tech icons have the wrong size
**Screenshots**
If applicable, add screenshots to help explain your problem.

**Additional context**
This is already fixed for the player crafting menu group, however the tech icons I overlooked:
https://github.com/Arch666Angel/mods/blob/b45c76fe1eb650ea5acf3a1230fbdb2c356c2743/angelsbioprocessing/prototypes/overrides/bio-processing-override-bob.lua#L160-L163
https://github.com/Arch666Angel/mods/blob/b45c76fe1eb650ea5acf3a1230fbdb2c356c2743/angelsbioprocessing/prototypes/overrides/bio-processing-override-bob.lua#L226-L229
|
process
|
module tech icons wrong size describe the bug module tech icons have the wrong size screenshots if applicable add screenshots to help explain your problem additional context this is already fixed for the player crafting menu group however the tech icons i overlooked
| 1
|
13,418
| 23,061,384,452
|
IssuesEvent
|
2022-07-25 10:14:48
|
renovatebot/renovate
|
https://api.github.com/repos/renovatebot/renovate
|
opened
|
npm package latest version not seen
|
type:bug status:requirements priority-5-triage
|
### How are you running Renovate?
Self-hosted
### If you're self-hosting Renovate, tell us what version of Renovate you run.
32.126.2
### Please select which platform you are using if self-hosting.
GitHub Enterprise Server
### If you're self-hosting Renovate, tell us what version of the platform you run.
3.3.8
### Was this something which used to work for you, and then stopped?
It used to work, and then stopped
### Describe the bug
Renovate doesn't create PRs for npm packages. For example, there is `3.8.2` version of chart.js available (we have `3.8.0`) but PR is not created.
When running renovate with debug logging, we can see it fetches information from our NPM source (Artifactory) and that the Artifactoy URL returns the expected JSON (so `3.8.2` is included in the versions).
Please note that we performed the update from 2.x to 3x manually but it shouldn't be any problem.

### Relevant debug logs
<details><summary>Logs</summary>
```
INFO: Dependency extraction complete (repository=project/ui)
"baseBranch": "master",
"stats": {
"managers": {
"kubernetes": {"fileCount": 1, "depCount": 4},
"npm": {"fileCount": 2, "depCount": 48}
},
"total": {"fileCount": 3, "depCount": 52}
}
{
"depType": "dependencies",
"depName": "chart.js",
"currentValue": "^3.8.0",
"datasource": "npm",
"prettyDepType": "dependency",
"lockedVersion": "3.8.0",
"depIndex": 4,
"updates": [],
"warnings": [],
"versioning": "npm",
"sourceUrl": "https://github.com/chartjs/Chart.js",
"homepage": "https://www.chartjs.org",
"currentVersion": "3.8.0",
"fixedVersion": "3.8.0"
},
DEBUG: http statistics (repository=project/ui)
"urls": {
"https://artifactory.our.company.com/artifactory/api/npm/registry.npmjs.org/chart.js (GET,200)": 1,
```
</details>
### Have you created a minimal reproduction repository?
No reproduction repository
|
1.0
|
npm package latest version not seen - ### How are you running Renovate?
Self-hosted
### If you're self-hosting Renovate, tell us what version of Renovate you run.
32.126.2
### Please select which platform you are using if self-hosting.
GitHub Enterprise Server
### If you're self-hosting Renovate, tell us what version of the platform you run.
3.3.8
### Was this something which used to work for you, and then stopped?
It used to work, and then stopped
### Describe the bug
Renovate doesn't create PRs for npm packages. For example, there is `3.8.2` version of chart.js available (we have `3.8.0`) but PR is not created.
When running renovate with debug logging, we can see it fetches information from our NPM source (Artifactory) and that the Artifactoy URL returns the expected JSON (so `3.8.2` is included in the versions).
Please note that we performed the update from 2.x to 3x manually but it shouldn't be any problem.

### Relevant debug logs
<details><summary>Logs</summary>
```
INFO: Dependency extraction complete (repository=project/ui)
"baseBranch": "master",
"stats": {
"managers": {
"kubernetes": {"fileCount": 1, "depCount": 4},
"npm": {"fileCount": 2, "depCount": 48}
},
"total": {"fileCount": 3, "depCount": 52}
}
{
"depType": "dependencies",
"depName": "chart.js",
"currentValue": "^3.8.0",
"datasource": "npm",
"prettyDepType": "dependency",
"lockedVersion": "3.8.0",
"depIndex": 4,
"updates": [],
"warnings": [],
"versioning": "npm",
"sourceUrl": "https://github.com/chartjs/Chart.js",
"homepage": "https://www.chartjs.org",
"currentVersion": "3.8.0",
"fixedVersion": "3.8.0"
},
DEBUG: http statistics (repository=project/ui)
"urls": {
"https://artifactory.our.company.com/artifactory/api/npm/registry.npmjs.org/chart.js (GET,200)": 1,
```
</details>
### Have you created a minimal reproduction repository?
No reproduction repository
|
non_process
|
npm package latest version not seen how are you running renovate self hosted if you re self hosting renovate tell us what version of renovate you run please select which platform you are using if self hosting github enterprise server if you re self hosting renovate tell us what version of the platform you run was this something which used to work for you and then stopped it used to work and then stopped describe the bug renovate doesn t create prs for npm packages for example there is version of chart js available we have but pr is not created when running renovate with debug logging we can see it fetches information from our npm source artifactory and that the artifactoy url returns the expected json so is included in the versions please note that we performed the update from x to manually but it shouldn t be any problem relevant debug logs logs info dependency extraction complete repository project ui basebranch master stats managers kubernetes filecount depcount npm filecount depcount total filecount depcount deptype dependencies depname chart js currentvalue datasource npm prettydeptype dependency lockedversion depindex updates warnings versioning npm sourceurl homepage currentversion fixedversion debug http statistics repository project ui urls get have you created a minimal reproduction repository no reproduction repository
| 0
|
53,975
| 6,777,427,927
|
IssuesEvent
|
2017-10-27 22:12:54
|
Automattic/jetpack
|
https://api.github.com/repos/Automattic/jetpack
|
opened
|
Connect Banner styles off
|
Admin Page General [Status] Design Polish [Type] Bug [Type] Good First Bug
|
#### Steps to reproduce the issue
Run the latest version of Jetpack
#### What I expected
Things to look more styled.
#### What happened instead
The banner looks brokern see "By connecting your site you agree to our fascinatin"

<!--
PLEASE NOTE
- These comments won't show up when you submit the issue.
- Everything is optional, but try to add as many details as possible.
- If requesting a new feature, explain why you'd like to see it added.
- This issue tracker is not for support. If you have questions about Jetpack, you can [start a new thread in the Jetpack support forums](https://wordpress.org/support/plugin/jetpack#new-post), or [send us an email](http://jetpack.com/contact-support/).
- Do not report potential security vulnerabilities here. For responsible disclosure of security issues and to be eligible for our bug bounty program, please submit your report via [the HackerOne portal](https://hackerone.com/automattic).
-->
|
1.0
|
Connect Banner styles off -
#### Steps to reproduce the issue
Run the latest version of Jetpack
#### What I expected
Things to look more styled.
#### What happened instead
The banner looks brokern see "By connecting your site you agree to our fascinatin"

<!--
PLEASE NOTE
- These comments won't show up when you submit the issue.
- Everything is optional, but try to add as many details as possible.
- If requesting a new feature, explain why you'd like to see it added.
- This issue tracker is not for support. If you have questions about Jetpack, you can [start a new thread in the Jetpack support forums](https://wordpress.org/support/plugin/jetpack#new-post), or [send us an email](http://jetpack.com/contact-support/).
- Do not report potential security vulnerabilities here. For responsible disclosure of security issues and to be eligible for our bug bounty program, please submit your report via [the HackerOne portal](https://hackerone.com/automattic).
-->
|
non_process
|
connect banner styles off steps to reproduce the issue run the latest version of jetpack what i expected things to look more styled what happened instead the banner looks brokern see by connecting your site you agree to our fascinatin please note these comments won t show up when you submit the issue everything is optional but try to add as many details as possible if requesting a new feature explain why you d like to see it added this issue tracker is not for support if you have questions about jetpack you can or do not report potential security vulnerabilities here for responsible disclosure of security issues and to be eligible for our bug bounty program please submit your report via
| 0
|
12,822
| 15,196,541,024
|
IssuesEvent
|
2021-02-16 08:26:31
|
threefoldtech/js-sdk
|
https://api.github.com/repos/threefoldtech/js-sdk
|
closed
|
Blog can't get verified certificate
|
process_wontfix type_bug
|
### Description
While testing, I found that Blog solution can't get SSL verified certificate.
### Steps to reproduce
- Deploy VDC
- Deploy Blog solution
- Go to URL domain


|
1.0
|
Blog can't get verified certificate - ### Description
While testing, I found that Blog solution can't get SSL verified certificate.
### Steps to reproduce
- Deploy VDC
- Deploy Blog solution
- Go to URL domain


|
process
|
blog can t get verified certificate description while testing i found that blog solution can t get ssl verified certificate steps to reproduce deploy vdc deploy blog solution go to url domain
| 1
|
5,404
| 27,115,681,186
|
IssuesEvent
|
2023-02-15 18:22:31
|
VA-Explorer/va_explorer
|
https://api.github.com/repos/VA-Explorer/va_explorer
|
closed
|
Calculate and highlight outlier data within VA trends
|
Type: Maintainance good first issue Domain: Frontend Status: Inactive
|
**What is the expected state?**
As a data analyst I expect to be able to tell at a glance if my trends data contains outliers so I can quickly identify them or further investigate them. I would like the trends tab chart to somehow highlight data that qualifies as an outlier.
**What is the actual state?**
The trends chart currently shows data without regard to outliers. It doesn't calculate them and it doesn't show them.
**Relevant context**
- `va_analytics/dash_apps/va_dashboard.py`
|
True
|
Calculate and highlight outlier data within VA trends - **What is the expected state?**
As a data analyst I expect to be able to tell at a glance if my trends data contains outliers so I can quickly identify them or further investigate them. I would like the trends tab chart to somehow highlight data that qualifies as an outlier.
**What is the actual state?**
The trends chart currently shows data without regard to outliers. It doesn't calculate them and it doesn't show them.
**Relevant context**
- `va_analytics/dash_apps/va_dashboard.py`
|
non_process
|
calculate and highlight outlier data within va trends what is the expected state as a data analyst i expect to be able to tell at a glance if my trends data contains outliers so i can quickly identify them or further investigate them i would like the trends tab chart to somehow highlight data that qualifies as an outlier what is the actual state the trends chart currently shows data without regard to outliers it doesn t calculate them and it doesn t show them relevant context va analytics dash apps va dashboard py
| 0
|
72,532
| 9,596,906,567
|
IssuesEvent
|
2019-05-09 19:52:21
|
biotaphy/analyses
|
https://api.github.com/repos/biotaphy/analyses
|
opened
|
Document how to use phylo beta diversity
|
documentation help wanted
|
This includes how to use the script for performing the computations as well as to what the actual computation is and why someone would want to do it.
|
1.0
|
Document how to use phylo beta diversity - This includes how to use the script for performing the computations as well as to what the actual computation is and why someone would want to do it.
|
non_process
|
document how to use phylo beta diversity this includes how to use the script for performing the computations as well as to what the actual computation is and why someone would want to do it
| 0
|
47,432
| 19,650,833,723
|
IssuesEvent
|
2022-01-10 06:49:01
|
IBM-Cloud/terraform-provider-ibm
|
https://api.github.com/repos/IBM-Cloud/terraform-provider-ibm
|
closed
|
Add Placement Group support for PowerVS
|
enhancement service/Power Systems
|
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
In June 2021 the PowerVS REST API added support for Placement Groups:
* https://cloud.ibm.com/apidocs/power-cloud#release-notes
Documented here:
* https://cloud.ibm.com/apidocs/power-cloud#pcloud-placementgroups-post
These would be a useful addition to the Terraform provider.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* ibm_pi_placement_group
* ibm_pi_instance
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
Proposed new resource:
```hcl
resource "ibm_pi_placement_group" "placement_group" {
pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60"
pi_placement_name = "some_group_name"
pi_placement_policy = "affinity"
}
```
Proposed new data source:
```hcl
data "ibm_pi_placement_group" "ds_placement_group" {
pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60"
pi_placement_group_name = "some_group_name"
}
```
Addition to existing ibm_pi_instance resource:
```hcl
resource "ibm_pi_instance" "test-instance" {
pi_memory = "4"
pi_processors = "2"
pi_instance_name = "test-vm"
pi_proc_type = "shared"
pi_image_id = "${data.ibm_pi_image.powerimages.id}"
pi_key_pair_name = ibm_pi_key.key.key_id
pi_sys_type = "s922"
pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60"
pi_pin_policy = "none"
pi_health_status = "WARNING"
pi_network {
network_id = data.ibm_pi_public_network.dsnetwork.id
}
# NEW VARIABLE
pi_placement_group_id = "49fba6c9-23f8-40bc-9899-aca322ee7d5b"
}
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
* https://cloud.ibm.com/apidocs/power-cloud#release-notes
* https://cloud.ibm.com/apidocs/power-cloud#pcloud-placementgroups-post
|
1.0
|
Add Placement Group support for PowerVS - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
In June 2021 the PowerVS REST API added support for Placement Groups:
* https://cloud.ibm.com/apidocs/power-cloud#release-notes
Documented here:
* https://cloud.ibm.com/apidocs/power-cloud#pcloud-placementgroups-post
These would be a useful addition to the Terraform provider.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* ibm_pi_placement_group
* ibm_pi_instance
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
Proposed new resource:
```hcl
resource "ibm_pi_placement_group" "placement_group" {
pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60"
pi_placement_name = "some_group_name"
pi_placement_policy = "affinity"
}
```
Proposed new data source:
```hcl
data "ibm_pi_placement_group" "ds_placement_group" {
pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60"
pi_placement_group_name = "some_group_name"
}
```
Addition to existing ibm_pi_instance resource:
```hcl
resource "ibm_pi_instance" "test-instance" {
pi_memory = "4"
pi_processors = "2"
pi_instance_name = "test-vm"
pi_proc_type = "shared"
pi_image_id = "${data.ibm_pi_image.powerimages.id}"
pi_key_pair_name = ibm_pi_key.key.key_id
pi_sys_type = "s922"
pi_cloud_instance_id = "51e1879c-bcbe-4ee1-a008-49cdba0eaf60"
pi_pin_policy = "none"
pi_health_status = "WARNING"
pi_network {
network_id = data.ibm_pi_public_network.dsnetwork.id
}
# NEW VARIABLE
pi_placement_group_id = "49fba6c9-23f8-40bc-9899-aca322ee7d5b"
}
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
* https://cloud.ibm.com/apidocs/power-cloud#release-notes
* https://cloud.ibm.com/apidocs/power-cloud#pcloud-placementgroups-post
|
non_process
|
add placement group support for powervs community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description in june the powervs rest api added support for placement groups documented here these would be a useful addition to the terraform provider new or affected resource s ibm pi placement group ibm pi instance potential terraform configuration proposed new resource hcl resource ibm pi placement group placement group pi cloud instance id bcbe pi placement name some group name pi placement policy affinity proposed new data source hcl data ibm pi placement group ds placement group pi cloud instance id bcbe pi placement group name some group name addition to existing ibm pi instance resource hcl resource ibm pi instance test instance pi memory pi processors pi instance name test vm pi proc type shared pi image id data ibm pi image powerimages id pi key pair name ibm pi key key key id pi sys type pi cloud instance id bcbe pi pin policy none pi health status warning pi network network id data ibm pi public network dsnetwork id new variable pi placement group id references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation
| 0
|
64,044
| 14,657,149,446
|
IssuesEvent
|
2020-12-28 14:59:02
|
turkdevops/grafana
|
https://api.github.com/repos/turkdevops/grafana
|
opened
|
CVE-2019-0205 (High) detected in github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9, github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9
|
security vulnerability
|
## CVE-2019-0205 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b>, <b>github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary>
<p>
<details><summary><b>github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary>
<p>Jaeger Bindings for Go OpenTracing API.</p>
<p>
Dependency Hierarchy:
- github.com/grafana/grafana/pkg/infra/tracing-375e8e4fd051a5c821c2b477634c73166d0694ab (Root Library)
- github.com/uber/jaeger-client-go/config-fe3fa553c313b32f58cc684a59a4d48f03e07df9
- :x: **github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9** (Vulnerable Library)
</details>
<details><summary><b>github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary>
<p>Jaeger Bindings for Go OpenTracing API.</p>
<p>
Dependency Hierarchy:
- github.com/grafana/grafana/pkg/infra/tracing-375e8e4fd051a5c821c2b477634c73166d0694ab (Root Library)
- github.com/uber/jaeger-client-go/config-fe3fa553c313b32f58cc684a59a4d48f03e07df9
- github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9
- github.com/uber/jaeger-client-go/thrift-gen/agent-fe3fa553c313b32f58cc684a59a4d48f03e07df9
- :x: **github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/eaebfb49103a255dc2d75f384b343f6a7f708816">eaebfb49103a255dc2d75f384b343f6a7f708816</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings.
<p>Publish Date: 2019-10-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205>CVE-2019-0205</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205</a></p>
<p>Release Date: 2019-10-29</p>
<p>Fix Resolution: org.apache.thrift:libthrift:0.13.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-0205 (High) detected in github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9, github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9 - ## CVE-2019-0205 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b>, <b>github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary>
<p>
<details><summary><b>github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary>
<p>Jaeger Bindings for Go OpenTracing API.</p>
<p>
Dependency Hierarchy:
- github.com/grafana/grafana/pkg/infra/tracing-375e8e4fd051a5c821c2b477634c73166d0694ab (Root Library)
- github.com/uber/jaeger-client-go/config-fe3fa553c313b32f58cc684a59a4d48f03e07df9
- :x: **github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9** (Vulnerable Library)
</details>
<details><summary><b>github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9</b></p></summary>
<p>Jaeger Bindings for Go OpenTracing API.</p>
<p>
Dependency Hierarchy:
- github.com/grafana/grafana/pkg/infra/tracing-375e8e4fd051a5c821c2b477634c73166d0694ab (Root Library)
- github.com/uber/jaeger-client-go/config-fe3fa553c313b32f58cc684a59a4d48f03e07df9
- github.com/uber/jaeger-client-go-fe3fa553c313b32f58cc684a59a4d48f03e07df9
- github.com/uber/jaeger-client-go/thrift-gen/agent-fe3fa553c313b32f58cc684a59a4d48f03e07df9
- :x: **github.com/uber/jaeger-client-go/thrift-fe3fa553c313b32f58cc684a59a4d48f03e07df9** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/eaebfb49103a255dc2d75f384b343f6a7f708816">eaebfb49103a255dc2d75f384b343f6a7f708816</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Apache Thrift all versions up to and including 0.12.0, a server or client may run into an endless loop when feed with specific input data. Because the issue had already been partially fixed in version 0.11.0, depending on the installed version it affects only certain language bindings.
<p>Publish Date: 2019-10-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0205>CVE-2019-0205</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0205</a></p>
<p>Release Date: 2019-10-29</p>
<p>Fix Resolution: org.apache.thrift:libthrift:0.13.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in github com uber jaeger client go github com uber jaeger client go thrift cve high severity vulnerability vulnerable libraries github com uber jaeger client go github com uber jaeger client go thrift github com uber jaeger client go jaeger bindings for go opentracing api dependency hierarchy github com grafana grafana pkg infra tracing root library github com uber jaeger client go config x github com uber jaeger client go vulnerable library github com uber jaeger client go thrift jaeger bindings for go opentracing api dependency hierarchy github com grafana grafana pkg infra tracing root library github com uber jaeger client go config github com uber jaeger client go github com uber jaeger client go thrift gen agent x github com uber jaeger client go thrift vulnerable library found in head commit a href found in base branch master vulnerability details in apache thrift all versions up to and including a server or client may run into an endless loop when feed with specific input data because the issue had already been partially fixed in version depending on the installed version it affects only certain language bindings publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache thrift libthrift step up your open source security game with whitesource
| 0
|
8,608
| 11,764,726,822
|
IssuesEvent
|
2020-03-14 14:07:00
|
googleapis/google-cloud-cpp
|
https://api.github.com/repos/googleapis/google-cloud-cpp
|
closed
|
Create a CI build using macOS+CMake.
|
type: process
|
We need a build on macOS that uses CMake. There are instructions on the README, and I use CMake routinely, but we also need a CI for it.
- [x] Create new scripts and test them locally
- [x] Enable **only** the continuous cmake and bazel builds on Kokoro.
- [x] Troubleshoot
- [x] Enable presubmit builds and disable old Kokoro builds.
- [x] Remove old scripts, copy `-new` files to their final names.
- [x] Configure Kokoro to use the final version of the files.
- [x] Remove the `-new` files.
|
1.0
|
Create a CI build using macOS+CMake. - We need a build on macOS that uses CMake. There are instructions on the README, and I use CMake routinely, but we also need a CI for it.
- [x] Create new scripts and test them locally
- [x] Enable **only** the continuous cmake and bazel builds on Kokoro.
- [x] Troubleshoot
- [x] Enable presubmit builds and disable old Kokoro builds.
- [x] Remove old scripts, copy `-new` files to their final names.
- [x] Configure Kokoro to use the final version of the files.
- [x] Remove the `-new` files.
|
process
|
create a ci build using macos cmake we need a build on macos that uses cmake there are instructions on the readme and i use cmake routinely but we also need a ci for it create new scripts and test them locally enable only the continuous cmake and bazel builds on kokoro troubleshoot enable presubmit builds and disable old kokoro builds remove old scripts copy new files to their final names configure kokoro to use the final version of the files remove the new files
| 1
|
26,771
| 20,677,559,511
|
IssuesEvent
|
2022-03-10 10:43:36
|
airyhq/airy
|
https://api.github.com/repos/airyhq/airy
|
closed
|
Inconsistencies with the backup/restore process
|
infrastructure docs
|
## This is how we could improve the [documentation](https://docs.airy.co):
Because backups are usually performed on a running Kafka instance, sometimes the restored Kafka can have inconsistencies, particularly in the internal topics.
We should add to the `backup and restore` documentation how to reset the apps and fix these issues.
|
1.0
|
Inconsistencies with the backup/restore process - ## This is how we could improve the [documentation](https://docs.airy.co):
Because backups are usually performed on a running Kafka instance, sometimes the restored Kafka can have inconsistencies, particularly in the internal topics.
We should add to the `backup and restore` documentation how to reset the apps and fix these issues.
|
non_process
|
inconsistencies with the backup restore process this is how we could improve the because backups are usually performed on a running kafka instance sometimes the restored kafka can have inconsistencies particularly in the internal topics we should add to the backup and restore documentation how to reset the apps and fix these issues
| 0
|
50
| 2,513,878,262
|
IssuesEvent
|
2015-01-15 04:33:35
|
GsDevKit/zinc
|
https://api.github.com/repos/GsDevKit/zinc
|
closed
|
NeoJSON loads from master and not from gs_master ?
|
inprocess
|
Why is NeoJSON referenced via "github://GsDevKit/NeoJSON:master/repository" and not via the published "github://GsDevKit/NeoJSON:gs_master/repository" ? All the newly Unicode stuff is then missing ...
Marten
|
1.0
|
NeoJSON loads from master and not from gs_master ? - Why is NeoJSON referenced via "github://GsDevKit/NeoJSON:master/repository" and not via the published "github://GsDevKit/NeoJSON:gs_master/repository" ? All the newly Unicode stuff is then missing ...
Marten
|
process
|
neojson loads from master and not from gs master why is neojson referenced via github gsdevkit neojson master repository and not via the published github gsdevkit neojson gs master repository all the newly unicode stuff is then missing marten
| 1
|
7,639
| 10,736,553,887
|
IssuesEvent
|
2019-10-29 11:08:29
|
googleapis/google-cloud-dotnet
|
https://api.github.com/repos/googleapis/google-cloud-dotnet
|
opened
|
Proposal: remove EF Core provider for Spanner
|
type: process
|
We currently have an alpha version of an Entity Framework Core provider for Spanner, along with a GA version of an ADO.NET provider.
The EF Core provider has not been reviewed and maintained as thoroughly as the ADO.NET provider, and probably needs significant effort to bring it from alpha to beta and then GA. At the moment, we have no plans to put in that effort.
Proposal:
- Delist the packages in NuGet
- Remove the source code from GitHub in a single commit (so we can easily revert it later if we decide to bring the project back)
cc @meteatamel, @JustinBeckwith, @chrisdunelm, @SurferJeffAtGoogle, @jsimonweb, @amanda-tarafa
|
1.0
|
Proposal: remove EF Core provider for Spanner - We currently have an alpha version of an Entity Framework Core provider for Spanner, along with a GA version of an ADO.NET provider.
The EF Core provider has not been reviewed and maintained as thoroughly as the ADO.NET provider, and probably needs significant effort to bring it from alpha to beta and then GA. At the moment, we have no plans to put in that effort.
Proposal:
- Delist the packages in NuGet
- Remove the source code from GitHub in a single commit (so we can easily revert it later if we decide to bring the project back)
cc @meteatamel, @JustinBeckwith, @chrisdunelm, @SurferJeffAtGoogle, @jsimonweb, @amanda-tarafa
|
process
|
proposal remove ef core provider for spanner we currently have an alpha version of an entity framework core provider for spanner along with a ga version of an ado net provider the ef core provider has not been reviewed and maintained as thoroughly as the ado net provider and probably needs significant effort to bring it from alpha to beta and then ga at the moment we have no plans to put in that effort proposal delist the packages in nuget remove the source code from github in a single commit so we can easily revert it later if we decide to bring the project back cc meteatamel justinbeckwith chrisdunelm surferjeffatgoogle jsimonweb amanda tarafa
| 1
|
10,095
| 13,044,162,081
|
IssuesEvent
|
2020-07-29 03:47:29
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `AddDurationAndString` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `AddDurationAndString` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `AddDurationAndString` from TiDB -
## Description
Port the scalar function `AddDurationAndString` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function adddurationandstring from tidb description port the scalar function adddurationandstring from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
7,385
| 10,660,745,520
|
IssuesEvent
|
2019-10-18 10:36:59
|
VincenzoVisconti16/Social-Music
|
https://api.github.com/repos/VincenzoVisconti16/Social-Music
|
opened
|
evaluation system
|
functional requirement
|
The system shall implement an evaluation system*, for rating people samples and answers.
|
1.0
|
evaluation system - The system shall implement an evaluation system*, for rating people samples and answers.
|
non_process
|
evaluation system the system shall implement an evaluation system for rating people samples and answers
| 0
|
173,943
| 6,534,736,894
|
IssuesEvent
|
2017-08-31 12:06:34
|
gaocegege/Processing.R
|
https://api.github.com/repos/gaocegege/Processing.R
|
closed
|
Allow failure in macOS to solve the queueing problem
|
difficulty/low for-new-contributors priority/p2 size/small status/WIP type/enhancement
|
Now we could not get the macOS environment as soon as possible so I think we could set the build to https://docs.travis-ci.com/user/customizing-the-build#Rows-that-are-Allowed-to-Fail

|
1.0
|
Allow failure in macOS to solve the queueing problem - Now we could not get the macOS environment as soon as possible so I think we could set the build to https://docs.travis-ci.com/user/customizing-the-build#Rows-that-are-Allowed-to-Fail

|
non_process
|
allow failure in macos to solve the queueing problem now we could not get the macos environment as soon as possible so i think we could set the build to
| 0
|
796,790
| 28,127,879,532
|
IssuesEvent
|
2023-03-31 19:28:19
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
reopened
|
speech.snippets.quickstart_v2_test: test_quickstart_v2 failed
|
priority: p1 type: bug api: speech samples flakybot: issue flakybot: flaky
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 2b37b6a745661f0a80a75a5ae92077db9d69cb6c
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2e840773-8a31-4c25-9f03-affdb423d789), [Sponge](http://sponge2/2e840773-8a31-4c25-9f03-affdb423d789)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/grpc/_channel.py", line 1030, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking
raise _InactiveRpcError(state) # pytype: disable=not-instantiable
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "The service is currently unavailable."
debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.132.95:443 {grpc_message:"The service is currently unavailable.", grpc_status:14, created_time:"2023-03-31T13:53:04.232449854+00:00"}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/speech/snippets/quickstart_v2_test.py", line 37, in test_quickstart_v2
response = quickstart_v2.quickstart_v2(
File "/workspace/speech/snippets/quickstart_v2.py", line 36, in quickstart_v2
operation = client.create_recognizer(request=request)
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/cloud/speech_v2/services/speech/client.py", line 692, in create_recognizer
response = rpc(
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable
raise exceptions.from_grpc_error(exc) from exc
google.api_core.exceptions.ServiceUnavailable: 503 The service is currently unavailable.</pre></details>
|
1.0
|
speech.snippets.quickstart_v2_test: test_quickstart_v2 failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 2b37b6a745661f0a80a75a5ae92077db9d69cb6c
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2e840773-8a31-4c25-9f03-affdb423d789), [Sponge](http://sponge2/2e840773-8a31-4c25-9f03-affdb423d789)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/grpc/_channel.py", line 1030, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking
raise _InactiveRpcError(state) # pytype: disable=not-instantiable
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "The service is currently unavailable."
debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.132.95:443 {grpc_message:"The service is currently unavailable.", grpc_status:14, created_time:"2023-03-31T13:53:04.232449854+00:00"}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/speech/snippets/quickstart_v2_test.py", line 37, in test_quickstart_v2
response = quickstart_v2.quickstart_v2(
File "/workspace/speech/snippets/quickstart_v2.py", line 36, in quickstart_v2
operation = client.create_recognizer(request=request)
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/cloud/speech_v2/services/speech/client.py", line 692, in create_recognizer
response = rpc(
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/speech/snippets/.nox/py-3-10/lib/python3.10/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable
raise exceptions.from_grpc_error(exc) from exc
google.api_core.exceptions.ServiceUnavailable: 503 The service is currently unavailable.</pre></details>
|
non_process
|
speech snippets quickstart test test quickstart failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace speech snippets nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace speech snippets nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace speech snippets nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state pytype disable not instantiable grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode unavailable details the service is currently unavailable debug error string unknown error received from peer grpc message the service is currently unavailable grpc status created time the above exception was the direct cause of the following exception traceback most recent call last file workspace speech snippets quickstart test py line in test quickstart response quickstart quickstart file workspace speech snippets quickstart py line in quickstart operation client create recognizer request request file workspace speech snippets nox py lib site packages google cloud speech services speech client py line in create recognizer response rpc file workspace speech snippets nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace speech snippets nox py lib site packages google api core grpc helpers py line in error remapped callable raise exceptions from grpc error exc from exc google api core exceptions serviceunavailable the service is currently unavailable
| 0
|
22,460
| 31,237,150,275
|
IssuesEvent
|
2023-08-20 12:21:53
|
Warzone2100/map-submission
|
https://api.github.com/repos/Warzone2100/map-submission
|
opened
|
[MAP]: 10c-UNPNTW5vs5
|
map unprocessed
|
### Upload Map
[10c-UNPNTW5vs5V2.zip](https://github.com/Warzone2100/map-submission/files/12388133/10c-UNPNTW5vs5V2.zip)
### Authorship
Mine: I am the author of this map
### Map Description (optional)
```text
New NTW with large battlefield, scavs and circular mountains.
```
### Notes for Reviewers (optional)
_No response_
|
1.0
|
[MAP]: 10c-UNPNTW5vs5 - ### Upload Map
[10c-UNPNTW5vs5V2.zip](https://github.com/Warzone2100/map-submission/files/12388133/10c-UNPNTW5vs5V2.zip)
### Authorship
Mine: I am the author of this map
### Map Description (optional)
```text
New NTW with large battlefield, scavs and circular mountains.
```
### Notes for Reviewers (optional)
_No response_
|
process
|
upload map authorship mine i am the author of this map map description optional text new ntw with large battlefield scavs and circular mountains notes for reviewers optional no response
| 1
|
775,858
| 27,239,234,432
|
IssuesEvent
|
2023-02-21 18:50:28
|
GoogleChrome/lighthouse
|
https://api.github.com/repos/GoogleChrome/lighthouse
|
closed
|
SyntaxError when attempting to use user-flow from lighthouse v10
|
bug pending-close needs-priority
|
### FAQ
- [X] Yes, my issue is not about [variability](https://github.com/GoogleChrome/lighthouse/blob/main/docs/variability.md) or [throttling](https://github.com/GoogleChrome/lighthouse/blob/main/docs/throttling.md).
- [X] Yes, my issue is not about a specific accessibility audit (file with [axe-core](https://github.com/dequelabs/axe-core) instead).
### URL
https://www.example.com
### What happened?
I am not able to use use user-flow from lighthouse version 10.
I followed the example from the repo [lighthouse/docs/user-flows.md](https://github.com/GoogleChrome/lighthouse/blob/main/docs/user-flows.md#navigations-in-the-nodejs-api) but it throws and error.
Here is a [minimal reproducible example](https://github.com/ChristopherPHolder/mr-lh-uf-v10).
### What did you expect?
Usual behaviour. I expected lighthouse user-flow to run a navigation audit.
### What have you tried?
I am using it inside of a project where I wrote a mini jest test to insure its working, that trow an error so I decided to make a minimal reproducible example repo and that throw the same error.
The error from jest was:
```ts
/Users/christopherholder/Applications/ufo/node_modules/lighthouse/core/index.js:7
import log from 'lighthouse-logger';
^^^^^^
SyntaxError: Cannot use import statement outside a module
> 1 | import { startFlow } from 'lighthouse';
| ^
2 | import { launch } from 'puppeteer';
3 | import { ResultReports } from 'shared';
4 |
at Runtime.createScriptFromCode (../../node_modules/jest-runtime/build/index.js:1796:14)
at Object.<anonymous> (src/lib/user-flow-runner.ts:1:1)
```
The error is the same from the example repo `SyntaxError: Cannot use import statement outside a module` but jest seems to provided more information about its origin.
Error in minimal reproduction:
```
SyntaxError: Cannot use import statement outside a module
at Object.compileFunction (node:vm:360:18)
at wrapSafe (node:internal/modules/cjs/loader:1088:15)
at Module._compile (node:internal/modules/cjs/loader:1123:27)
at Module._extensions..js (node:internal/modules/cjs/loader:1213:10)
at Module.load (node:internal/modules/cjs/loader:1037:32)
at Module._load (node:internal/modules/cjs/loader:878:12)
at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:81:12)
at node:internal/main/run_main_module:23:47
```
### How were you running Lighthouse?
node
### Lighthouse Version
v10.0.1
### Chrome Version
_No response_
### Node Version
v18.12.1 & v16.16.0
### OS
Mac
### Relevant log output
```shell
// In Minimal reproduction example repo
SyntaxError: Cannot use import statement outside a module
at Object.compileFunction (node:vm:360:18)
at wrapSafe (node:internal/modules/cjs/loader:1088:15)
at Module._compile (node:internal/modules/cjs/loader:1123:27)
at Module._extensions..js (node:internal/modules/cjs/loader:1213:10)
at Module.load (node:internal/modules/cjs/loader:1037:32)
at Module._load (node:internal/modules/cjs/loader:878:12)
at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:81:12)
at node:internal/main/run_main_module:23:47
```
|
1.0
|
SyntaxError when attempting to use user-flow from lighthouse v10 - ### FAQ
- [X] Yes, my issue is not about [variability](https://github.com/GoogleChrome/lighthouse/blob/main/docs/variability.md) or [throttling](https://github.com/GoogleChrome/lighthouse/blob/main/docs/throttling.md).
- [X] Yes, my issue is not about a specific accessibility audit (file with [axe-core](https://github.com/dequelabs/axe-core) instead).
### URL
https://www.example.com
### What happened?
I am not able to use use user-flow from lighthouse version 10.
I followed the example from the repo [lighthouse/docs/user-flows.md](https://github.com/GoogleChrome/lighthouse/blob/main/docs/user-flows.md#navigations-in-the-nodejs-api) but it throws and error.
Here is a [minimal reproducible example](https://github.com/ChristopherPHolder/mr-lh-uf-v10).
### What did you expect?
Usual behaviour. I expected lighthouse user-flow to run a navigation audit.
### What have you tried?
I am using it inside of a project where I wrote a mini jest test to insure its working, that trow an error so I decided to make a minimal reproducible example repo and that throw the same error.
The error from jest was:
```ts
/Users/christopherholder/Applications/ufo/node_modules/lighthouse/core/index.js:7
import log from 'lighthouse-logger';
^^^^^^
SyntaxError: Cannot use import statement outside a module
> 1 | import { startFlow } from 'lighthouse';
| ^
2 | import { launch } from 'puppeteer';
3 | import { ResultReports } from 'shared';
4 |
at Runtime.createScriptFromCode (../../node_modules/jest-runtime/build/index.js:1796:14)
at Object.<anonymous> (src/lib/user-flow-runner.ts:1:1)
```
The error is the same from the example repo `SyntaxError: Cannot use import statement outside a module` but jest seems to provided more information about its origin.
Error in minimal reproduction:
```
SyntaxError: Cannot use import statement outside a module
at Object.compileFunction (node:vm:360:18)
at wrapSafe (node:internal/modules/cjs/loader:1088:15)
at Module._compile (node:internal/modules/cjs/loader:1123:27)
at Module._extensions..js (node:internal/modules/cjs/loader:1213:10)
at Module.load (node:internal/modules/cjs/loader:1037:32)
at Module._load (node:internal/modules/cjs/loader:878:12)
at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:81:12)
at node:internal/main/run_main_module:23:47
```
### How were you running Lighthouse?
node
### Lighthouse Version
v10.0.1
### Chrome Version
_No response_
### Node Version
v18.12.1 & v16.16.0
### OS
Mac
### Relevant log output
```shell
// In Minimal reproduction example repo
SyntaxError: Cannot use import statement outside a module
at Object.compileFunction (node:vm:360:18)
at wrapSafe (node:internal/modules/cjs/loader:1088:15)
at Module._compile (node:internal/modules/cjs/loader:1123:27)
at Module._extensions..js (node:internal/modules/cjs/loader:1213:10)
at Module.load (node:internal/modules/cjs/loader:1037:32)
at Module._load (node:internal/modules/cjs/loader:878:12)
at Function.executeUserEntryPoint [as runMain] (node:internal/modules/run_main:81:12)
at node:internal/main/run_main_module:23:47
```
|
non_process
|
syntaxerror when attempting to use user flow from lighthouse faq yes my issue is not about or yes my issue is not about a specific accessibility audit file with instead url what happened i am not able to use use user flow from lighthouse version i followed the example from the repo but it throws and error here is a what did you expect usual behaviour i expected lighthouse user flow to run a navigation audit what have you tried i am using it inside of a project where i wrote a mini jest test to insure its working that trow an error so i decided to make a minimal reproducible example repo and that throw the same error the error from jest was ts users christopherholder applications ufo node modules lighthouse core index js import log from lighthouse logger syntaxerror cannot use import statement outside a module import startflow from lighthouse import launch from puppeteer import resultreports from shared at runtime createscriptfromcode node modules jest runtime build index js at object src lib user flow runner ts the error is the same from the example repo syntaxerror cannot use import statement outside a module but jest seems to provided more information about its origin error in minimal reproduction syntaxerror cannot use import statement outside a module at object compilefunction node vm at wrapsafe node internal modules cjs loader at module compile node internal modules cjs loader at module extensions js node internal modules cjs loader at module load node internal modules cjs loader at module load node internal modules cjs loader at function executeuserentrypoint node internal modules run main at node internal main run main module how were you running lighthouse node lighthouse version chrome version no response node version os mac relevant log output shell in minimal reproduction example repo syntaxerror cannot use import statement outside a module at object compilefunction node vm at wrapsafe node internal modules cjs loader at module compile node internal modules cjs loader at module extensions js node internal modules cjs loader at module load node internal modules cjs loader at module load node internal modules cjs loader at function executeuserentrypoint node internal modules run main at node internal main run main module
| 0
|
4,454
| 7,320,588,318
|
IssuesEvent
|
2018-03-02 08:05:27
|
EBWiki/EBWiki
|
https://api.github.com/repos/EBWiki/EBWiki
|
closed
|
Spec coverage needs to cover failure cases
|
bug process
|
While we have decent spec coverage, we are not covering failure cases enough, where the user enters invalid or incomplete information.
|
1.0
|
Spec coverage needs to cover failure cases - While we have decent spec coverage, we are not covering failure cases enough, where the user enters invalid or incomplete information.
|
process
|
spec coverage needs to cover failure cases while we have decent spec coverage we are not covering failure cases enough where the user enters invalid or incomplete information
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.