Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
21,125
| 10,575,385,587
|
IssuesEvent
|
2019-10-07 15:38:43
|
Harkishen-Singh/Productify
|
https://api.github.com/repos/Harkishen-Singh/Productify
|
closed
|
CVE-2018-3721 Medium Severity Vulnerability detected by WhiteSource
|
security vulnerability
|
## CVE-2018-3721 - Medium Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>path: /tmp/git/Productify/react-mobile-app/node_modules/xmlbuilder/node_modules/lodash/package.json</p>
<p>
<p>Library home page: <a href=http://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz>http://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
Dependency Hierarchy:
- react-native-0.55.4.tgz (Root Library)
- plist-1.2.0.tgz
- xmlbuilder-4.0.0.tgz
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3721>CVE-2018-3721</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3721">https://nvd.nist.gov/vuln/detail/CVE-2018-3721</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 4.17.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-3721 Medium Severity Vulnerability detected by WhiteSource - ## CVE-2018-3721 - Medium Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>path: /tmp/git/Productify/react-mobile-app/node_modules/xmlbuilder/node_modules/lodash/package.json</p>
<p>
<p>Library home page: <a href=http://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz>http://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
Dependency Hierarchy:
- react-native-0.55.4.tgz (Root Library)
- plist-1.2.0.tgz
- xmlbuilder-4.0.0.tgz
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
lodash node module before 4.17.5 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via defaultsDeep, merge, and mergeWith functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3721>CVE-2018-3721</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3721">https://nvd.nist.gov/vuln/detail/CVE-2018-3721</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 4.17.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium severity vulnerability detected by whitesource cve medium severity vulnerability vulnerable library lodash tgz the modern build of lodash modular utilities path tmp git productify react mobile app node modules xmlbuilder node modules lodash package json library home page a href dependency hierarchy react native tgz root library plist tgz xmlbuilder tgz x lodash tgz vulnerable library vulnerability details lodash node module before suffers from a modification of assumed immutable data maid vulnerability via defaultsdeep merge and mergewith functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
31,800
| 6,627,833,883
|
IssuesEvent
|
2017-09-23 09:30:52
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
closed
|
Invalid code emit for left assignment of delegate
|
defect
|
### Steps To Reproduce
https://deck.net/da6492ba392e45d5df14fcd058409f5e
```csharp
public class SomeClass
{
private Action action;
public Action this[string key]
{
get { return action; }
set { action = value; }
}
}
public class Program
{
public static void Main()
{
string msg = null;
var ht = new SomeClass();
ht[""] += () => { msg = "Hello"; };
Console.WriteLine(msg);
}
}
```
### Actual Result
Incorrect JavaScript generated:
```js
ht.setItem("", Bridge.fn.combine(ht.getItem(""), function () {
msg = "Hello";
});
```
Causing:
```
SyntaxError: missing ) after argument list
```
### Expected Result
```js
ht.setItem("", Bridge.fn.combine(ht.getItem(""), function () {
msg = "test";
}));
```
|
1.0
|
Invalid code emit for left assignment of delegate - ### Steps To Reproduce
https://deck.net/da6492ba392e45d5df14fcd058409f5e
```csharp
public class SomeClass
{
private Action action;
public Action this[string key]
{
get { return action; }
set { action = value; }
}
}
public class Program
{
public static void Main()
{
string msg = null;
var ht = new SomeClass();
ht[""] += () => { msg = "Hello"; };
Console.WriteLine(msg);
}
}
```
### Actual Result
Incorrect JavaScript generated:
```js
ht.setItem("", Bridge.fn.combine(ht.getItem(""), function () {
msg = "Hello";
});
```
Causing:
```
SyntaxError: missing ) after argument list
```
### Expected Result
```js
ht.setItem("", Bridge.fn.combine(ht.getItem(""), function () {
msg = "test";
}));
```
|
defect
|
invalid code emit for left assignment of delegate steps to reproduce csharp public class someclass private action action public action this get return action set action value public class program public static void main string msg null var ht new someclass ht msg hello console writeline msg actual result incorrect javascript generated js ht setitem bridge fn combine ht getitem function msg hello causing syntaxerror missing after argument list expected result js ht setitem bridge fn combine ht getitem function msg test
| 1
|
8,467
| 3,184,171,883
|
IssuesEvent
|
2015-09-27 04:02:24
|
BumblebeeBat/FlyingFox
|
https://api.github.com/repos/BumblebeeBat/FlyingFox
|
closed
|
Oracles
|
discussion documentation needs more info
|
We need a new transaction type that allows for the creation of oracles.
The first oracle should be either a single address, or a multisig of several addresses. The participants in the oracle are the people who know the private keys for the addresses that make up the oracle.
If a decision is given to the oracle, then the oracle can profitably answer the decision.
Eventually a mechanism will be added to the oracle to encourage the participants to be honest.
|
1.0
|
Oracles - We need a new transaction type that allows for the creation of oracles.
The first oracle should be either a single address, or a multisig of several addresses. The participants in the oracle are the people who know the private keys for the addresses that make up the oracle.
If a decision is given to the oracle, then the oracle can profitably answer the decision.
Eventually a mechanism will be added to the oracle to encourage the participants to be honest.
|
non_defect
|
oracles we need a new transaction type that allows for the creation of oracles the first oracle should be either a single address or a multisig of several addresses the participants in the oracle are the people who know the private keys for the addresses that make up the oracle if a decision is given to the oracle then the oracle can profitably answer the decision eventually a mechanism will be added to the oracle to encourage the participants to be honest
| 0
|
43,344
| 12,977,309,747
|
IssuesEvent
|
2020-07-21 20:24:37
|
kenferrara/spark
|
https://api.github.com/repos/kenferrara/spark
|
opened
|
CVE-2020-11619 (High) detected in jackson-databind-2.9.8.jar
|
security vulnerability
|
## CVE-2020-11619 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/spark/spark-docker-image-generator/build.gradle</p>
<p>Path to vulnerable library: /tmp/ws-ua_20200721201640_CTTRJV/downloadResource_NEKDHT/20200721202004/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- docker-client-8.16.0.jar (Root Library)
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/spark/commit/3d9cf3cdbd4d6a25eb74453231ea79062c4b6223">3d9cf3cdbd4d6a25eb74453231ea79062c4b6223</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.springframework.aop.config.MethodLocatingFactoryBean (aka spring-aop).
<p>Publish Date: 2020-04-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11619>CVE-2020-11619</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11619">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11619</a></p>
<p>Release Date: 2020-04-07</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","isTransitiveDependency":true,"dependencyTree":"com.spotify:docker-client:8.16.0;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"vulnerabilityIdentifier":"CVE-2020-11619","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.springframework.aop.config.MethodLocatingFactoryBean (aka spring-aop).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11619","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-11619 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2020-11619 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/spark/spark-docker-image-generator/build.gradle</p>
<p>Path to vulnerable library: /tmp/ws-ua_20200721201640_CTTRJV/downloadResource_NEKDHT/20200721202004/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- docker-client-8.16.0.jar (Root Library)
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/spark/commit/3d9cf3cdbd4d6a25eb74453231ea79062c4b6223">3d9cf3cdbd4d6a25eb74453231ea79062c4b6223</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.springframework.aop.config.MethodLocatingFactoryBean (aka spring-aop).
<p>Publish Date: 2020-04-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11619>CVE-2020-11619</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11619">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11619</a></p>
<p>Release Date: 2020-04-07</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","isTransitiveDependency":true,"dependencyTree":"com.spotify:docker-client:8.16.0;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"vulnerabilityIdentifier":"CVE-2020-11619","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.springframework.aop.config.MethodLocatingFactoryBean (aka spring-aop).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11619","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm spark spark docker image generator build gradle path to vulnerable library tmp ws ua cttrjv downloadresource nekdht jackson databind jar dependency hierarchy docker client jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org springframework aop config methodlocatingfactorybean aka spring aop publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org springframework aop config methodlocatingfactorybean aka spring aop vulnerabilityurl
| 0
|
4,207
| 20,679,494,949
|
IssuesEvent
|
2022-03-10 12:34:36
|
BioArchLinux/Packages
|
https://api.github.com/repos/BioArchLinux/Packages
|
closed
|
[BUG] The ugene package does not install everything needed
|
maintain
|
OK this is annoying. I made the package based on my ugene-git package (which works) but apparently is the ugene package in BioArchLinux not installing everything (PREFIX has to be set on both "qmake" and "make install", but even that did not help). When upstream changed how things are supposed to be installed in ugene, it really messed up the package.
Sorry.
I will have to try to figure out what goes wrong in the release version of the package. I will do that sometime in the coming days.
|
True
|
[BUG] The ugene package does not install everything needed - OK this is annoying. I made the package based on my ugene-git package (which works) but apparently is the ugene package in BioArchLinux not installing everything (PREFIX has to be set on both "qmake" and "make install", but even that did not help). When upstream changed how things are supposed to be installed in ugene, it really messed up the package.
Sorry.
I will have to try to figure out what goes wrong in the release version of the package. I will do that sometime in the coming days.
|
non_defect
|
the ugene package does not install everything needed ok this is annoying i made the package based on my ugene git package which works but apparently is the ugene package in bioarchlinux not installing everything prefix has to be set on both qmake and make install but even that did not help when upstream changed how things are supposed to be installed in ugene it really messed up the package sorry i will have to try to figure out what goes wrong in the release version of the package i will do that sometime in the coming days
| 0
|
1,626
| 2,603,968,522
|
IssuesEvent
|
2015-02-24 18:59:39
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳阴茎疱疹会传染吗
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳阴茎疱疹会传染吗〓沈陽軍區政治部醫院性病〓TEL:024-3
1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。�
��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌�
��歷史悠久、設備精良、技術權威、專家云集,是預防、保健
、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲��
�部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、�
��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空
軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體��
�等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:17
|
1.0
|
沈阳阴茎疱疹会传染吗 - ```
沈阳阴茎疱疹会传染吗〓沈陽軍區政治部醫院性病〓TEL:024-3
1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。�
��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌�
��歷史悠久、設備精良、技術權威、專家云集,是預防、保健
、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲��
�部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、�
��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空
軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體��
�等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:17
|
defect
|
沈阳阴茎疱疹会传染吗 沈阳阴茎疱疹会传染吗〓沈陽軍區政治部醫院性病〓tel: 〓 , 。� �� 。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 original issue reported on code google com by gmail com on jun at
| 1
|
318
| 2,525,215,437
|
IssuesEvent
|
2015-01-20 22:58:23
|
AtlasOfLivingAustralia/biocache-hubs
|
https://api.github.com/repos/AtlasOfLivingAustralia/biocache-hubs
|
closed
|
Incorrect taxon matching for SAM records (ignoring higher taxa hint from collectory)
|
bug priority-medium status-new type-defect
|
*migrated from:* https://code.google.com/p/ala/issues/detail?id=722
*date:* Sun Jun 29 17:55:32 2014
*author:* nickdos
---
Alexis identified some records being matched to non-animal higher taxa despite the collectory specifying animalia as the higher taxon hint for this collection.
[http://biocache.ala.org.au/occurrences/5bcd1ec2-211c-4bf6-9993-414d9e815898](http://biocache.ala.org.au/occurrences/5bcd1ec2-211c-4bf6-9993-414d9e815898)
and
[http://biocache.ala.org.au/occurrences/d6fb923e-6c2c-4937-8e2a-08b9187e252b](http://biocache.ala.org.au/occurrences/d6fb923e-6c2c-4937-8e2a-08b9187e252b)
[http://biocache.ala.org.au/occurrences/100c8df7-de72-432c-90ce-9dbe4bf5113d](http://biocache.ala.org.au/occurrences/100c8df7-de72-432c-90ce-9dbe4bf5113d)
[http://biocache.ala.org.au/occurrences/ed6cb62d-3f8d-4cd9-900e-a8d5d7e13abd](http://biocache.ala.org.au/occurrences/ed6cb62d-3f8d-4cd9-900e-a8d5d7e13abd)
which appear to be matching to genus via a fuzzy match despite having a family specified which appears to have been ignored.
|
1.0
|
Incorrect taxon matching for SAM records (ignoring higher taxa hint from collectory) -
*migrated from:* https://code.google.com/p/ala/issues/detail?id=722
*date:* Sun Jun 29 17:55:32 2014
*author:* nickdos
---
Alexis identified some records being matched to non-animal higher taxa despite the collectory specifying animalia as the higher taxon hint for this collection.
[http://biocache.ala.org.au/occurrences/5bcd1ec2-211c-4bf6-9993-414d9e815898](http://biocache.ala.org.au/occurrences/5bcd1ec2-211c-4bf6-9993-414d9e815898)
and
[http://biocache.ala.org.au/occurrences/d6fb923e-6c2c-4937-8e2a-08b9187e252b](http://biocache.ala.org.au/occurrences/d6fb923e-6c2c-4937-8e2a-08b9187e252b)
[http://biocache.ala.org.au/occurrences/100c8df7-de72-432c-90ce-9dbe4bf5113d](http://biocache.ala.org.au/occurrences/100c8df7-de72-432c-90ce-9dbe4bf5113d)
[http://biocache.ala.org.au/occurrences/ed6cb62d-3f8d-4cd9-900e-a8d5d7e13abd](http://biocache.ala.org.au/occurrences/ed6cb62d-3f8d-4cd9-900e-a8d5d7e13abd)
which appear to be matching to genus via a fuzzy match despite having a family specified which appears to have been ignored.
|
defect
|
incorrect taxon matching for sam records ignoring higher taxa hint from collectory migrated from date sun jun author nickdos alexis identified some records being matched to non animal higher taxa despite the collectory specifying animalia as the higher taxon hint for this collection and which appear to be matching to genus via a fuzzy match despite having a family specified which appears to have been ignored
| 1
|
184,806
| 21,784,985,827
|
IssuesEvent
|
2022-05-14 02:03:26
|
kapseliboi/OnlineCV
|
https://api.github.com/repos/kapseliboi/OnlineCV
|
closed
|
CVE-2021-35065 (High) detected in glob-parent-3.1.0.tgz, glob-parent-2.0.0.tgz - autoclosed
|
security vulnerability
|
## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-2.0.0.tgz</b></p></summary>
<p>
<details><summary><b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json,/client/node_modules/chokidar/node_modules/glob-parent/package.json,/admin/node_modules/chokidar/node_modules/glob-parent/package.json,/client/node_modules/fast-glob/node_modules/glob-parent/package.json,/admin/node_modules/fast-glob/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-1.18.8.tgz (Root Library)
- chokidar-2.0.4.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-2.0.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/glob-parent/package.json,/admin/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.3.tgz (Root Library)
- jest-23.6.0.tgz
- jest-cli-23.6.0.tgz
- micromatch-2.3.11.tgz
- parse-glob-3.0.4.tgz
- glob-base-0.3.0.tgz
- :x: **glob-parent-2.0.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/gulpjs/glob-parent/pull/49">https://github.com/gulpjs/glob-parent/pull/49</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution: glob-parent - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-35065 (High) detected in glob-parent-3.1.0.tgz, glob-parent-2.0.0.tgz - autoclosed - ## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-2.0.0.tgz</b></p></summary>
<p>
<details><summary><b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json,/client/node_modules/chokidar/node_modules/glob-parent/package.json,/admin/node_modules/chokidar/node_modules/glob-parent/package.json,/client/node_modules/fast-glob/node_modules/glob-parent/package.json,/admin/node_modules/fast-glob/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-1.18.8.tgz (Root Library)
- chokidar-2.0.4.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-2.0.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/glob-parent/package.json,/admin/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.3.tgz (Root Library)
- jest-23.6.0.tgz
- jest-cli-23.6.0.tgz
- micromatch-2.3.11.tgz
- parse-glob-3.0.4.tgz
- glob-base-0.3.0.tgz
- :x: **glob-parent-2.0.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/gulpjs/glob-parent/pull/49">https://github.com/gulpjs/glob-parent/pull/49</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution: glob-parent - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in glob parent tgz glob parent tgz autoclosed cve high severity vulnerability vulnerable libraries glob parent tgz glob parent tgz glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href path to dependency file package json path to vulnerable library node modules glob parent package json client node modules chokidar node modules glob parent package json admin node modules chokidar node modules glob parent package json client node modules fast glob node modules glob parent package json admin node modules fast glob node modules glob parent package json dependency hierarchy nodemon tgz root library chokidar tgz x glob parent tgz vulnerable library glob parent tgz strips glob magic from a string to provide the parent path library home page a href path to dependency file client package json path to vulnerable library client node modules glob parent package json admin node modules glob parent package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz micromatch tgz parse glob tgz glob base tgz x glob parent tgz vulnerable library found in base branch master vulnerability details the package glob parent before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with whitesource
| 0
|
171,691
| 14,343,146,625
|
IssuesEvent
|
2020-11-28 07:41:12
|
UBC-MDS/covid-19-cases-vs-tests-analysis
|
https://api.github.com/repos/UBC-MDS/covid-19-cases-vs-tests-analysis
|
closed
|
Proposal document
|
documentation
|
- Create a separate Proposal.md file that captures all the content of the Readme.md from Release 0.0.1.
- Put the Proposal.md file under the doc folder.
|
1.0
|
Proposal document - - Create a separate Proposal.md file that captures all the content of the Readme.md from Release 0.0.1.
- Put the Proposal.md file under the doc folder.
|
non_defect
|
proposal document create a separate proposal md file that captures all the content of the readme md from release put the proposal md file under the doc folder
| 0
|
5,197
| 2,610,182,954
|
IssuesEvent
|
2015-02-26 18:58:17
|
chrsmith/quchuseban
|
https://api.github.com/repos/chrsmith/quchuseban
|
opened
|
解密怎么快速去除色斑
|
auto-migrated Priority-Medium Type-Defect
|
```
《摘要》
而这样的夜,我又幻想着怎样把悲伤和迷茫聚集在自己铸造��
�围墙一角,飘飘渺渺的心灵在逃避无边无际的滋长,思绪犹�
��茧丝缠绕,丝丝缠绵。雨,一点一滴的在你心里击打,抹不
掉的是我心里的那份执著;雷,不要吵醒熟睡的人好吗?我��
�等她醒来能够认真的听我说我们在一起的零零落落;电,请�
��我敲开她冷漠如陨石的心扉。我想用我的真情融化你不朽的
冰川,时光把这份情意写在了你我彼此的记忆,你可知道?��
�时的充动让我守候着永久的疼痛,看着你离开的背影,我想�
��手抓住你的双手,不让你走,我怕这一次的转身成了我们一
生的决别,当我微笑着看着你离去,泪水却留在了心里。没��
�了色斑,我有回到了从前美丽的自己,但是没有你的日子我�
��怎么能快乐!怎么快速去除色斑,
《客户案例》
快速去黄褐斑的方法,
大概是天妒我容颜,在我怀孕期间色斑开始在我脸上驰骋,��
�点状到片状,最后毫不留情的占据了我的脸,密密麻麻的,�
��自己看了都感觉身上发凉。刚开始怎么都接受不了自己变成
这个样子,我以前可是很爱美的哦!痛苦万分。去看医生说我�
��分泌失调,雄性激素分泌过多,给我熬了好几包天然精华,
让我喝,足足喝了3个月,斑还是顽强的在那里,对我不离不�
��。
后来一直不停的寻找能祛斑的产品,什么国内的国外的都用��
�,刚开始还是不错的,斑有所减少,但是后来又大面积复发�
��,之后用什么好像都没有作用了,总是反反复复。那些油的
、辣的食物,都不敢使用了.每天就是豆腐白菜,和出家人一�
��.尽管难受,还是忍了。还买来了排毒养颜,每天中午饭,��
�饭后使用。每天睡觉前还喝一杯蜂蜜水。每天光要记得这些�
��,头就觉得很大了.可是斑就是跟我死磕,完全没有办法。</
br>
后来听同事介绍黛芙薇尔不错,研发和生产得力于法国��
�美国、台湾三地的专家通力协作,通过生物高科技手段,全�
��修复老化、病变细胞,彻底逆转受损基因,从而达到了激活
肌肤细胞分裂再生、增强细胞活性、强效美白的功效,特别��
�暗沉、干燥、暗黄、粗糙、肤色不匀的肌肤。建议我试试,�
��就去网上查了查,看到很多患者说是使用黛芙薇尔效果不错
,并且说是涂抹的,用着倒挺方便,我想就试试吧,就订购��
�一个周期。</br>
用了接近20几天的时候就有明显效果,斑点开始浅化,到
后来两个周期快用完的时候,片状的斑,很快就感觉像肤色��
�样了,觉得像奇迹一样,我从来不知道祛斑真的就这么简单�
��他们精油也不错,我以前睡眠不太好,不知道是用了精油还
是觉得斑去了心安,反正睡觉都觉得舒服,呵呵,真的挺不��
�的,我绝对相信天然精华,相信产品带给我的奇迹与喜悦。
阅读了怎么快速去除色斑,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
怎么快速去除色斑,同时为您分享祛斑小方法
1.杏仁(去皮、捣烂)适量,用蛋清将杏仁未调匀,每晚睡前搽�
��,次晨用白酒洗去。
2.柿叶研细末,入熔化的凡士林搅拌为膏,外涂即可。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 2:58
|
1.0
|
解密怎么快速去除色斑 - ```
《摘要》
而这样的夜,我又幻想着怎样把悲伤和迷茫聚集在自己铸造��
�围墙一角,飘飘渺渺的心灵在逃避无边无际的滋长,思绪犹�
��茧丝缠绕,丝丝缠绵。雨,一点一滴的在你心里击打,抹不
掉的是我心里的那份执著;雷,不要吵醒熟睡的人好吗?我��
�等她醒来能够认真的听我说我们在一起的零零落落;电,请�
��我敲开她冷漠如陨石的心扉。我想用我的真情融化你不朽的
冰川,时光把这份情意写在了你我彼此的记忆,你可知道?��
�时的充动让我守候着永久的疼痛,看着你离开的背影,我想�
��手抓住你的双手,不让你走,我怕这一次的转身成了我们一
生的决别,当我微笑着看着你离去,泪水却留在了心里。没��
�了色斑,我有回到了从前美丽的自己,但是没有你的日子我�
��怎么能快乐!怎么快速去除色斑,
《客户案例》
快速去黄褐斑的方法,
大概是天妒我容颜,在我怀孕期间色斑开始在我脸上驰骋,��
�点状到片状,最后毫不留情的占据了我的脸,密密麻麻的,�
��自己看了都感觉身上发凉。刚开始怎么都接受不了自己变成
这个样子,我以前可是很爱美的哦!痛苦万分。去看医生说我�
��分泌失调,雄性激素分泌过多,给我熬了好几包天然精华,
让我喝,足足喝了3个月,斑还是顽强的在那里,对我不离不�
��。
后来一直不停的寻找能祛斑的产品,什么国内的国外的都用��
�,刚开始还是不错的,斑有所减少,但是后来又大面积复发�
��,之后用什么好像都没有作用了,总是反反复复。那些油的
、辣的食物,都不敢使用了.每天就是豆腐白菜,和出家人一�
��.尽管难受,还是忍了。还买来了排毒养颜,每天中午饭,��
�饭后使用。每天睡觉前还喝一杯蜂蜜水。每天光要记得这些�
��,头就觉得很大了.可是斑就是跟我死磕,完全没有办法。</
br>
后来听同事介绍黛芙薇尔不错,研发和生产得力于法国��
�美国、台湾三地的专家通力协作,通过生物高科技手段,全�
��修复老化、病变细胞,彻底逆转受损基因,从而达到了激活
肌肤细胞分裂再生、增强细胞活性、强效美白的功效,特别��
�暗沉、干燥、暗黄、粗糙、肤色不匀的肌肤。建议我试试,�
��就去网上查了查,看到很多患者说是使用黛芙薇尔效果不错
,并且说是涂抹的,用着倒挺方便,我想就试试吧,就订购��
�一个周期。</br>
用了接近20几天的时候就有明显效果,斑点开始浅化,到
后来两个周期快用完的时候,片状的斑,很快就感觉像肤色��
�样了,觉得像奇迹一样,我从来不知道祛斑真的就这么简单�
��他们精油也不错,我以前睡眠不太好,不知道是用了精油还
是觉得斑去了心安,反正睡觉都觉得舒服,呵呵,真的挺不��
�的,我绝对相信天然精华,相信产品带给我的奇迹与喜悦。
阅读了怎么快速去除色斑,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
怎么快速去除色斑,同时为您分享祛斑小方法
1.杏仁(去皮、捣烂)适量,用蛋清将杏仁未调匀,每晚睡前搽�
��,次晨用白酒洗去。
2.柿叶研细末,入熔化的凡士林搅拌为膏,外涂即可。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 2:58
|
defect
|
解密怎么快速去除色斑 《摘要》 而这样的夜,我又幻想着怎样把悲伤和迷茫聚集在自己铸造�� �围墙一角,飘飘渺渺的心灵在逃避无边无际的滋长,思绪犹� ��茧丝缠绕,丝丝缠绵。雨,一点一滴的在你心里击打,抹不 掉的是我心里的那份执著;雷,不要吵醒熟睡的人好吗?我�� �等她醒来能够认真的听我说我们在一起的零零落落;电,请� ��我敲开她冷漠如陨石的心扉。我想用我的真情融化你不朽的 冰川,时光把这份情意写在了你我彼此的记忆,你可知道?�� �时的充动让我守候着永久的疼痛,看着你离开的背影,我想� ��手抓住你的双手,不让你走,我怕这一次的转身成了我们一 生的决别,当我微笑着看着你离去,泪水却留在了心里。没�� �了色斑,我有回到了从前美丽的自己,但是没有你的日子我� ��怎么能快乐!怎么快速去除色斑, 《客户案例》 快速去黄褐斑的方法 大概是天妒我容颜,在我怀孕期间色斑开始在我脸上驰骋,�� �点状到片状,最后毫不留情的占据了我的脸,密密麻麻的,� ��自己看了都感觉身上发凉。刚开始怎么都接受不了自己变成 这个样子,我以前可是很爱美的哦 痛苦万分。去看医生说我� ��分泌失调,雄性激素分泌过多,给我熬了好几包天然精华, 让我喝, ,斑还是顽强的在那里,对我不离不� ��。 后来一直不停的寻找能祛斑的产品,什么国内的国外的都用�� �,刚开始还是不错的,斑有所减少,但是后来又大面积复发� ��,之后用什么好像都没有作用了,总是反反复复。那些油的 、辣的食物,都不敢使用了 每天就是豆腐白菜,和出家人一� �� 尽管难受,还是忍了。还买来了排毒养颜,每天中午饭,�� �饭后使用。每天睡觉前还喝一杯蜂蜜水。每天光要记得这些� ��,头就觉得很大了 可是斑就是跟我死磕,完全没有办法。 br 后来听同事介绍黛芙薇尔不错,研发和生产得力于法国�� �美国、台湾三地的专家通力协作,通过生物高科技手段,全� ��修复老化、病变细胞,彻底逆转受损基因,从而达到了激活 肌肤细胞分裂再生、增强细胞活性、强效美白的功效,特别�� �暗沉、干燥、暗黄、粗糙、肤色不匀的肌肤。建议我试试,� ��就去网上查了查,看到很多患者说是使用黛芙薇尔效果不错 ,并且说是涂抹的,用着倒挺方便,我想就试试吧,就订购�� �一个周期。 ,斑点开始浅化,到 后来两个周期快用完的时候,片状的斑,很快就感觉像肤色�� �样了,觉得像奇迹一样,我从来不知道祛斑真的就这么简单� ��他们精油也不错,我以前睡眠不太好,不知道是用了精油还 是觉得斑去了心安,反正睡觉都觉得舒服,呵呵,真的挺不�� �的,我绝对相信天然精华,相信产品带给我的奇迹与喜悦。 阅读了怎么快速去除色斑,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 怎么快速去除色斑,同时为您分享祛斑小方法 杏仁 去皮、捣烂 适量,用蛋清将杏仁未调匀,每晚睡前搽� ��,次晨用白酒洗去。 柿叶研细末,入熔化的凡士林搅拌为膏,外涂即可。 original issue reported on code google com by additive gmail com on jul at
| 1
|
82,754
| 16,025,057,188
|
IssuesEvent
|
2021-04-21 08:03:16
|
InventorsDev/sdg-004
|
https://api.github.com/repos/InventorsDev/sdg-004
|
closed
|
Work on user registration backend
|
Code
|
Use ajax php for the backend. No page reloading on form submission
|
1.0
|
Work on user registration backend - Use ajax php for the backend. No page reloading on form submission
|
non_defect
|
work on user registration backend use ajax php for the backend no page reloading on form submission
| 0
|
73,961
| 24,887,892,275
|
IssuesEvent
|
2022-10-28 09:21:14
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Changing custom notification sound fails and causes room notification settings window to be inaccessible
|
T-Defect
|
### Steps to reproduce
1. Open the room options menu from the ellipses icon
2. Click "Settings"
3. Under "General", "Security & Privacy", and "Roles & Permissions", click "Notifications"
4. Select a custom notification sound and upload
5. It will fail to set the default notification sound
6. Attempting to access the 'Notifications' tab again will instantly close the entire room settings menu
### Outcome
#### What did you expect?
I expected to be able to set my notification sound, and change it if it fails to set
#### What happened instead?
The menu instantly closes if it failed to upload. This occurs even after a reinstall.
### Operating system
Windows 10
### Application version
Element version 1.11.12, Olm version 3.2.12
### How did you install the app?
From https://element.io/get-started
### Homeserver
matrix.org
### Will you send logs?
No
|
1.0
|
Changing custom notification sound fails and causes room notification settings window to be inaccessible - ### Steps to reproduce
1. Open the room options menu from the ellipses icon
2. Click "Settings"
3. Under "General", "Security & Privacy", and "Roles & Permissions", click "Notifications"
4. Select a custom notification sound and upload
5. It will fail to set the default notification sound
6. Attempting to access the 'Notifications' tab again will instantly close the entire room settings menu
### Outcome
#### What did you expect?
I expected to be able to set my notification sound, and change it if it fails to set
#### What happened instead?
The menu instantly closes if it failed to upload. This occurs even after a reinstall.
### Operating system
Windows 10
### Application version
Element version 1.11.12, Olm version 3.2.12
### How did you install the app?
From https://element.io/get-started
### Homeserver
matrix.org
### Will you send logs?
No
|
defect
|
changing custom notification sound fails and causes room notification settings window to be inaccessible steps to reproduce open the room options menu from the ellipses icon click settings under general security privacy and roles permissions click notifications select a custom notification sound and upload it will fail to set the default notification sound attempting to access the notifications tab again will instantly close the entire room settings menu outcome what did you expect i expected to be able to set my notification sound and change it if it fails to set what happened instead the menu instantly closes if it failed to upload this occurs even after a reinstall operating system windows application version element version olm version how did you install the app from homeserver matrix org will you send logs no
| 1
|
120,735
| 10,132,452,024
|
IssuesEvent
|
2019-08-01 22:30:46
|
tjcsl/ion
|
https://api.github.com/repos/tjcsl/ion
|
opened
|
Test notifications for room changes
|
area/notifications area/testing
|
We should make sure that we are only notifying when we need to.
The tests should give us proper test coverage.
Related to https://github.com/tjcsl/ion/pull/726
|
1.0
|
Test notifications for room changes - We should make sure that we are only notifying when we need to.
The tests should give us proper test coverage.
Related to https://github.com/tjcsl/ion/pull/726
|
non_defect
|
test notifications for room changes we should make sure that we are only notifying when we need to the tests should give us proper test coverage related to
| 0
|
36,640
| 8,042,563,824
|
IssuesEvent
|
2018-07-31 08:34:17
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Dead link to MySQL manual on group by page
|
C: Documentation P: Medium R: Fixed T: Defect
|
This page https://www.jooq.org/doc/latest/manual/sql-building/sql-statements/select-statement/group-by-clause/ contains a link entitled "MySQL manual" (http://dev.mysql.com/doc/refman/5.6/en/group-by-hidden-columns.html) which is dead.
|
1.0
|
Dead link to MySQL manual on group by page - This page https://www.jooq.org/doc/latest/manual/sql-building/sql-statements/select-statement/group-by-clause/ contains a link entitled "MySQL manual" (http://dev.mysql.com/doc/refman/5.6/en/group-by-hidden-columns.html) which is dead.
|
defect
|
dead link to mysql manual on group by page this page contains a link entitled mysql manual which is dead
| 1
|
25,401
| 4,309,975,518
|
IssuesEvent
|
2016-07-21 17:42:57
|
KytechN24/xbox360wirelesschatpad
|
https://api.github.com/repos/KytechN24/xbox360wirelesschatpad
|
closed
|
Remapping buttons for gameplay?
|
auto-migrated Priority-Medium Type-Defect
|
```
ive got this chatpad to work (amazingly by the way) but now id like to game
with it (skyrim specifically)
is there a way to configure the controller with vjoy for each command id like
to do? or should i use another controller program to do so? if so which is best
recommended? thanks!
c
```
Original issue reported on code.google.com by `Communis...@gmail.com` on 14 Mar 2014 at 3:59
|
1.0
|
Remapping buttons for gameplay? - ```
ive got this chatpad to work (amazingly by the way) but now id like to game
with it (skyrim specifically)
is there a way to configure the controller with vjoy for each command id like
to do? or should i use another controller program to do so? if so which is best
recommended? thanks!
c
```
Original issue reported on code.google.com by `Communis...@gmail.com` on 14 Mar 2014 at 3:59
|
defect
|
remapping buttons for gameplay ive got this chatpad to work amazingly by the way but now id like to game with it skyrim specifically is there a way to configure the controller with vjoy for each command id like to do or should i use another controller program to do so if so which is best recommended thanks c original issue reported on code google com by communis gmail com on mar at
| 1
|
603,538
| 18,668,779,552
|
IssuesEvent
|
2021-10-30 09:45:35
|
AY2122S1-CS2103T-W13-2/tp
|
https://api.github.com/repos/AY2122S1-CS2103T-W13-2/tp
|
closed
|
Add jekyll test to our CI pipeline
|
type.Chore priority.Medium
|
> Can we add this test to our CI pipeline for PRs?
Yes we should be able to. I will leave this to @Zhou-Jiahao-1998. There is a guide [here](https://jekyllrb.com/docs/continuous-integration/github-actions/)
_Edit: My bad, the guide is includes publishing, we just need to test build on pr._
_Originally posted by @zhenghanlee in https://github.com/AY2122S1-CS2103T-W13-2/tp/issues/47#issuecomment-937375839_
|
1.0
|
Add jekyll test to our CI pipeline - > Can we add this test to our CI pipeline for PRs?
Yes we should be able to. I will leave this to @Zhou-Jiahao-1998. There is a guide [here](https://jekyllrb.com/docs/continuous-integration/github-actions/)
_Edit: My bad, the guide is includes publishing, we just need to test build on pr._
_Originally posted by @zhenghanlee in https://github.com/AY2122S1-CS2103T-W13-2/tp/issues/47#issuecomment-937375839_
|
non_defect
|
add jekyll test to our ci pipeline can we add this test to our ci pipeline for prs yes we should be able to i will leave this to zhou jiahao there is a guide edit my bad the guide is includes publishing we just need to test build on pr originally posted by zhenghanlee in
| 0
|
18,200
| 5,607,138,918
|
IssuesEvent
|
2017-04-01 02:21:05
|
mcsheaj/SPEasyForms
|
https://api.github.com/repos/mcsheaj/SPEasyForms
|
closed
|
Partial Containers (i.e. a single tab) are not Shown/Hidden Dynamically
|
bug codeplex
|
**This issue was imported from [CodePlex](http://SPEasyForms.codeplex.com/workitem/18)**
**[mcsheaj](http://www.codeplex.com/site/users/view/mcsheaj)** wrote 2015-02-06 at 17:05
If row, tab, or content area of a container is hidden on form load because all of the fields are hidden due to conditional visibility rules, it doesn't get shown when the field is supposed to be shown based on a change in the value of another field in
the form. If it starts off shown, it toggles visibility dynamically like it should. So if it is saved in a state where something should be shown but isn't, reopen it and it will be shown.
**[mcsheaj](http://www.codeplex.com/site/users/view/mcsheaj)** wrote 2015-02-06 at 17:12
Changed to postTransform method of each container to handle this case and modified the even handers for visibility rules to call postTransform on the container collection. Will be delivered in AddOns.00.07.
|
1.0
|
Partial Containers (i.e. a single tab) are not Shown/Hidden Dynamically - **This issue was imported from [CodePlex](http://SPEasyForms.codeplex.com/workitem/18)**
**[mcsheaj](http://www.codeplex.com/site/users/view/mcsheaj)** wrote 2015-02-06 at 17:05
If row, tab, or content area of a container is hidden on form load because all of the fields are hidden due to conditional visibility rules, it doesn't get shown when the field is supposed to be shown based on a change in the value of another field in
the form. If it starts off shown, it toggles visibility dynamically like it should. So if it is saved in a state where something should be shown but isn't, reopen it and it will be shown.
**[mcsheaj](http://www.codeplex.com/site/users/view/mcsheaj)** wrote 2015-02-06 at 17:12
Changed to postTransform method of each container to handle this case and modified the even handers for visibility rules to call postTransform on the container collection. Will be delivered in AddOns.00.07.
|
non_defect
|
partial containers i e a single tab are not shown hidden dynamically this issue was imported from wrote at if row tab or content area of a container is hidden on form load because all of the fields are hidden due to conditional visibility rules it doesn t get shown when the field is supposed to be shown based on a change in the value of another field in the form if it starts off shown it toggles visibility dynamically like it should so if it is saved in a state where something should be shown but isn t reopen it and it will be shown wrote at changed to posttransform method of each container to handle this case and modified the even handers for visibility rules to call posttransform on the container collection will be delivered in addons
| 0
|
8,969
| 27,294,305,200
|
IssuesEvent
|
2023-02-23 18:58:48
|
bcgov/api-services-portal
|
https://api.github.com/repos/bcgov/api-services-portal
|
closed
|
FAILED: Automated Tests(355)
|
automation
|
Stats: {
"suites": 60,
"tests": 478,
"passes": 123,
"pending": 0,
"failures": 355,
"start": "2023-02-22T19:13:24.092Z",
"end": "2023-02-22T20:50:26.231Z",
"duration": 1643245,
"testsRegistered": 478,
"passPercent": 25.732217573221757,
"pendingPercent": 0,
"other": 0,
"hasOther": false,
"skipped": 0,
"hasSkipped": false
}
Failed Tests:
"creates and activates new namespace"
"Verify for invalid namespace name"
"creates a new service account"
"publishes a new API for Dev environment to Kong Gateway"
"creates as new product in the directory"
"Associate Namespace to the organization Unit"
"update the Dataset in BC Data Catelogue to appear the API in the Directory"
"publish product to directory"
"Create a Test environment"
"applies authorization plugin to service published to Kong Gateway"
"activate the service for Test environment"
"activate the service for Dev environment"
"authenticates Janis (api owner)"
"Grant namespace access to Mark (access manager)"
"Grant CredentialIssuer.Admin permission to Janis (API Owner)"
"Grant CredentialIssuer.Admin permission to Janis (API Owner)"
"Collect the credentials"
"Close the popup without collecting credentials"
"authenticates Mark (Access-Manager)"
"Verify that the request status is Pending Approval"
"Collect the credentials"
"Verify that API is not accessible with the generated API Key when the request is not approved"
"authenticates Mark (Access-Manager)"
"verify the request details"
"Add group labels in request details window"
"approves an access request"
"Verify that API is accessible with the generated API Key"
"authenticates Mark (Access-Manager)"
"Navigate to Consumer page and filter the product"
"Click on the first consumer"
"Click on Grant Access button"
"Grant Access to Test environment"
"Verify that API is accessible with the generated API Key for Test environment"
"Apply Key-auth only authorization plugin to Kong Gateway"
"Update the Kong key-auth plugin the new consumer"
"Verify the service is accessibale with API key for free access"
"Verify the service is accessible with API key for elevated access"
"Verify the service is accessibale with API key for free access"
"Verify the service is accessible with API key for elevated access"
"Verify the rate limiting is applied for free access"
"authenticates Mark (Access Manager)"
"Navigate to Consumer page and filter the product"
"Select the consumer from the list"
"set IP address that is not accessible in the network as allowed IP and set Route as scope"
"verify IP Restriction error when the API calls other than the allowed IP"
"set IP address that is not accessible in the network as allowed IP and set service as scope"
"verify IP Restriction error when the API calls other than the allowed IP"
"set IP address that is accessible in the network as allowed IP and set route as scope"
"verify the success stats when the API calls within the allowed IP range"
"set IP address that is accessible in the network as allowed IP and set service as scope"
"verify the success stats when the API calls within the allowed IP range"
"Navigate to Consumer page and filter the product"
"set api ip-restriction to global service level"
"Verify that IP Restriction is set at global service level"
"set IP address that is not accessible in the network as allowed IP and set service as scope"
"verify IP Restriction error when the API calls other than the allowed IP"
"Navigate to Consumer page and filter the product"
"set api ip-restriction to global service level"
"Verify that IP Restriction is set at global service level"
"set IP address that is not accessible in the network as allowed IP and set service as scope"
"verify IP Restriction error when the API calls other than the allowed IP"
"authenticates Mark (Access Manager)"
"Navigate to Consumer page and filter the product"
"Select the consumer from the list "
"set api rate limit as per the test config, Local Policy and Scope as Service"
"verify rate limit error when the API calls beyond the limit"
"set api rate limit as per the test config, Local Policy and Scope as Route"
"verify rate limit error when the API calls beyond the limit"
"set api rate limit as per the test config, Redis Policy and Scope as Service"
"verify rate limit error when the API calls beyond the limit"
"set api rate limit as per the test config, Redis Policy and Scope as Route"
"verify rate limit error when the API calls beyond the limit"
"set api rate limit to global service level"
"Verify that Rate limiting is set at global service level"
"set api rate limit as per the test config, Redis Policy and Scope as Service"
"verify rate limit error when the API calls beyond the limit"
"set api rate limit to global service level"
"Verify that Rate limiting is set at global service level"
"set api rate limit as per the test config, Redis Policy and Scope as Service"
"verify rate limit error when the API calls beyond the limit"
"creates an access request"
"authenticates Mark (Access-Manager)"
"verify the request details"
"Add group labels in request details window"
"approves an access request"
"Verify that API is accessible with the generated API Key"
"authenticates Mark (Access-Manager)"
"verify that consumers are filters as per given parameter"
"authenticates Mark (Access-Manager)"
"Navigate to Consumer page and filter the product"
"Click on the first consumer"
"Verify that labels can be deleted"
"Verify that labels can be updated"
"Verify that labels can be added"
"Creates and activates new namespace"
"Grant namespace access to access manager(Mark)"
"Grant CredentialIssuer.Admin permission to credential issuer(Wendy)"
"Select the namespace created for client credential "
"Creates authorization profile for Client ID/Secret"
"Creates authorization profile for Client ID/Secret - Shared IDP"
"Creates authorization profile for JWT - Generated Key Pair"
"Creates authorization profile for JWKS URL"
"Creates invalid authorization profile"
"Activates namespace for client credential flow tests"
"Creates a new service account"
"Publishes a new API to Kong Gateway"
"Creates a new product in the directory"
"Associate Namespace to the organization Unit"
"Update the Dataset in BC Data Catalogue to appear the API in the Directory"
"Adds environment with Client ID/Secret authenticator to product"
"applies authorization plugin to service published to Kong Gateway"
"Adds environment with JWT - Generated Key Pair authenticator to product"
"Adds environment with JWT - JWKS URL authenticator to product"
"Applies authorization plugin to service published to Kong Gateway"
"activate the service for Test environment"
"Adds environment for invalid authorization profile to other"
"Creates an access request"
"Access Manager logs in"
"Select scopes in Authorization Tab"
"approves an access request"
"Get access token using client ID and secret; make API request"
"Creates an access request"
"Access Manager logs in"
"approves an access request"
"Get access token using JWT key pair; make API request"
"Creates an access request"
"Access Manager logs in"
"approves an access request"
"Get access token using JWT key pair; make API request"
"authenticates Janis (api owner)"
"Expand Gateway service details pane"
"Verify total requests counts"
"Verify the routes details "
"Verify the host details "
"Verify the Tags details "
"authenticates Janis (api owner)"
"verify that Gateway service filters as per given parameter"
"Change Product environment from active to inactive"
"verify that Gateway service filters as per given parameter"
"Change Product environment from inactive to active"
"authenticates Janis (api owner)"
"Grant namespace access to Old User"
"authenticates with old user"
"Get the permission of the user"
"Verify that user account is disabled"
"Get current API Key"
"Regenrate credential"
"Verify that new API key is set to the consumer"
"Verify that only one API key(new key) is set to the consumer in Kong gateway"
"Verify that API is not accessible with the old API Key"
"Regenrate credential client ID and Secret"
"Make sure that the old client ID and Secret is disabled"
"Verify that service is accessible with new client ID and Secret"
"Activates the namespace"
"Deactivate the service for Test environment"
"Update the authorization scope from Kong ACL-API to Client Credential"
"applies authorization plugin to service published to Kong Gateway"
"activate the service for Test environment"
"Verify that service is not accessible with existing Client ID - Secret credentials"
"Raise request access"
"Collect the credentials"
"Access Manager logs in"
"approves an access request"
"Verify that API is accessible with the generated API Key"
"Activates the namespace"
"Create an authorization profile"
"Deactivate the service for Test environment"
"Update the authorization scope from Kong ACL-API to Client Credential"
"applies authorization plugin to service published to Kong Gateway"
"activate the service for Test environment"
"Verify that API is not accessible with the current API Key for Test environment"
"Creates an access request"
"Access Manager logs in"
"approves an access request"
"Get access token using client ID and secret; make API request"
"Activates the namespace"
"Update kong plugin for both the services"
"applies authorization plugin to service published to Kong Gateway"
"Associate another the service for Test environment"
"Get access token using client ID and secret; make API request for test"
"Get access token using client ID and secret; make API request for Dev"
"Creates an access request"
"Access Manager logs in"
"Get access token using client ID and secret; make API request for test"
"Get access token using client ID and secret; make API request for Dev"
"Activates the namespace"
"Change Product environment from active to inactive"
"Change Product environment from inactive to active"
"Edit the namespace from the tree view"
"Navigate to attribute tab"
"Set the Attributes"
"authenticates Janis (api owner) to get the user session token"
"Publish the Shared IDP profile"
"Create an authorization profile and associate it with shared IPD"
"authenticates Janis (api owner) to get the user session token"
"Put the resource and verify the success code in the response"
"Edit the created profile and verify the updated Issuer URL"
"authenticates Janis (api owner) to get the user session token"
"Update the Shared IDP Profile to an active clientID Secret auth environment"
"Creates an access request"
"Get access token using client ID and secret; make API request for test environment"
"Get access token using client ID and secret; make API request for Dev"
"creates and activates new namespace"
"creates a new service account"
"publishes a new API to Kong Gateway"
"creates as new product in the directory"
"Associate Namespace to the organization Unit"
"update the Dataset in BC Data Catelogue to appear the API in the Directory"
"publish product to directory"
"applies authorization plugin to service published to Kong Gateway"
"activate the service for Dev environment"
"authenticates Janis (api owner)"
"Delete Product Environment"
"Delete the Product"
"Delete Service Accounts"
"Delete Namespace"
"authenticates Janis (api owner) to get the user session token"
"Get the resource and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Verify Activity filter for \"Janis Smith\" user"
"Verify Activity filter for \"Harley Jones\" user"
"Verify Activity filter for \"Mark F Mark L\" user"
"Verify Activities filter for consumer"
"Activates namespace for client credential flow tests"
"Get the resource and verify the success code in the response"
"Verify Activity filter foe all the listed activities"
"creates and activates new namespace"
"creates a new service account"
"publishes a new API to Kong Gateway"
"creates as new product in the directory"
"Associate Namespace to the organization Unit"
"update the Dataset in BC Data Catelogue to appear the API in the Directory"
"publish product to directory"
"applies authorization plugin to service published to Kong Gateway"
"authenticates Janis (api owner)"
"grant namespace access to Mark (access manager)"
"Grant permission to Janis (API Owner)"
"Grant permission to Wendy"
"creates an access request"
"authenticates Janis (api owner)"
"Grant \"Access.Manager\" access to Mark (access manager)"
"Authenticates Mark (Access-Manager)"
"Verify that the option to approve request is displayed"
"Authenticates Janis (api owner)"
"Grant only \"Namespace.Manage\" permission to Wendy"
"Authenticates Wendy (Credential-Issuer)"
"Verify that all the namespace options and activities are displayed"
"Authenticates Janis (api owner)"
"Grant only \"CredentialIssuer.Admin\" access to Wendy (access manager)"
"Authenticates Wendy (Credential-Issuer)"
"Verify that only Authorization Profile option is displayed in Namespace page"
"Verify that authorization profile for Client ID/Secret is generated"
"Authenticates Janis (api owner)"
"Grant only \"Namespace.View\" permission to Mark"
"authenticates Mark"
"Verify that service accounts are not created"
"authenticates Janis (api owner)"
"Grant \"GatewayConfig.Publish\" and \"Namespace.View\" access to Wendy (access manager)"
"Authenticates Wendy (Credential-Issuer)"
"Verify that GWA API allows user to publish the API to Kong gateway"
"authenticates Janis (api owner)"
"Get the authorization token for the service account created with out \"Content.Publish\" permission"
"Verify that the document is not published without \"Content.Publish\" permission"
"Verify that application is deleted"
"creates an access request"
"Verify that application is deleted"
"creates an access request"
"authenticates Mark (Access-Manager)"
"approves an access request"
"Verify that application is deleted"
"Verify that API is not accessible with the generated API Key when the application is deleted"
"creates and activates new namespace"
"creates a new service account"
"publishes a new API to Kong Gateway"
"creates as new product in the directory"
"update the Dataset in BC Data Catelogue to appear the API in the Directory"
"applies authorization plugin to service published to Kong Gateway"
"authenticates Janis (api owner)"
"Navigate to Your Product tab in API Directory page"
"Verify that the banner for Preview mode is displayed"
"creates and activates new namespace"
"Get the resource and verify the Organization details in the response"
"Get the resource and verify the org Names in the response"
"authenticates Janis (api owner) to get the user session token"
"Get the resource and verify the success code in the response"
"Compare the scope values in response against the expected values"
"Get the resource and verify the success code in the response"
"Compare the Namespace values in response against the expected values"
"Delete the namespace associated with the organization, organization unit and verify the success code in the response"
"Verify that the deleted Namespace is not displayed in Get Call"
"Add the access of the organization to the specific user and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Compare the Namespace values in response against the expected values"
"authenticates Janis (api owner) to get the user session token"
"Put the resource and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Verify the status code and response message for invalid slugvalue"
"Delete the documentation"
"Delete the documentation"
"Put the resource and verify the success code in the response"
"Verify that document contant is displayed for GET /documentation"
"Verify that document contant is fetch by slug ID"
"authenticates Janis (api owner) to get the user session token"
"Put the resource and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Delete the authorization profile"
"Verify that the authorization profile is deleted"
"Put the resource and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Delete the authorization profile"
"Verify that the authorization profile is deleted"
"Put the resource and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Delete the authorization profile"
"Verify that the authorization profile is deleted"
"authenticates Janis (api owner) to get the user session token"
"Put the resource and verify the success code in the response"
"Get the resource and verify the success code and product name in the response"
"Compare the values in response against the values passed in the request"
"authenticates Janis (api owner) to get the user session token"
"Verify that the product is visible in Manage Product Page"
"Delete the product environment and verify the success code in the response"
"Get the resource and verify that product environment is deleted"
"Delete the product and verify the success code in the response"
"Get the resource and verify that product is deleted"
"authenticates Janis (api owner) to get the user session token"
"Put the resource (/organization/{org}/datasets) and verify the success code in the response"
"Get the resource (/organization/{org}/datasets/{name}) and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Put the resource (/namespaces/{ns}/datasets/{name}) and verify the success code in the response"
"Get the resource (/namespaces/{ns}/datasets/{name}) and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Get the resource (/organizations/{org}/datasets/{name}) and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Get the resource (/organizations/{org}/datasets) and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Get the directory details (/directory) and verify the success code in the response"
"Get the directory details by its ID (/directory/{id}) and verify the success code in the response"
"Get the namespace directory details (/namespaces/{ns}/directory) and verify the success code and empty response for the namespace with no directory"
"Get the namespace directory details (/namespaces/{ns}/directory) and verify the success code in the response"
"Get the namespace directory details by its ID (/namespaces/{ns}/directory/{id}) and verify the success code in the response"
"Get the namespace directory details (/namespaces/{ns}/directory/{id}) for non exist directory ID and verify the response code"
"Delete the dataset (/organizations/{org}/datasets/{name}) and verify the success code in the response"
"Verify that deleted dataset does not display in Get dataset list"
"authenticates Janis (api owner) to get the user session token"
"Get the resource and verify the success code in the response"
"Verify that the selected Namespace is displayed in the Response list in the response"
"Get the resource and verify the success code in the response"
"Get the resource for namespace summary and verify the success code in the response"
"Delete the namespace and verify the Validation to prevent deleting the namespace"
"Force delete the namespace and verify the success code in the response"
Run Link: https://github.com/bcgov/api-services-portal/actions/runs/4245981733
|
1.0
|
FAILED: Automated Tests(355) - Stats: {
"suites": 60,
"tests": 478,
"passes": 123,
"pending": 0,
"failures": 355,
"start": "2023-02-22T19:13:24.092Z",
"end": "2023-02-22T20:50:26.231Z",
"duration": 1643245,
"testsRegistered": 478,
"passPercent": 25.732217573221757,
"pendingPercent": 0,
"other": 0,
"hasOther": false,
"skipped": 0,
"hasSkipped": false
}
Failed Tests:
"creates and activates new namespace"
"Verify for invalid namespace name"
"creates a new service account"
"publishes a new API for Dev environment to Kong Gateway"
"creates as new product in the directory"
"Associate Namespace to the organization Unit"
"update the Dataset in BC Data Catelogue to appear the API in the Directory"
"publish product to directory"
"Create a Test environment"
"applies authorization plugin to service published to Kong Gateway"
"activate the service for Test environment"
"activate the service for Dev environment"
"authenticates Janis (api owner)"
"Grant namespace access to Mark (access manager)"
"Grant CredentialIssuer.Admin permission to Janis (API Owner)"
"Grant CredentialIssuer.Admin permission to Janis (API Owner)"
"Collect the credentials"
"Close the popup without collecting credentials"
"authenticates Mark (Access-Manager)"
"Verify that the request status is Pending Approval"
"Collect the credentials"
"Verify that API is not accessible with the generated API Key when the request is not approved"
"authenticates Mark (Access-Manager)"
"verify the request details"
"Add group labels in request details window"
"approves an access request"
"Verify that API is accessible with the generated API Key"
"authenticates Mark (Access-Manager)"
"Navigate to Consumer page and filter the product"
"Click on the first consumer"
"Click on Grant Access button"
"Grant Access to Test environment"
"Verify that API is accessible with the generated API Key for Test environment"
"Apply Key-auth only authorization plugin to Kong Gateway"
"Update the Kong key-auth plugin the new consumer"
"Verify the service is accessibale with API key for free access"
"Verify the service is accessible with API key for elevated access"
"Verify the service is accessibale with API key for free access"
"Verify the service is accessible with API key for elevated access"
"Verify the rate limiting is applied for free access"
"authenticates Mark (Access Manager)"
"Navigate to Consumer page and filter the product"
"Select the consumer from the list"
"set IP address that is not accessible in the network as allowed IP and set Route as scope"
"verify IP Restriction error when the API calls other than the allowed IP"
"set IP address that is not accessible in the network as allowed IP and set service as scope"
"verify IP Restriction error when the API calls other than the allowed IP"
"set IP address that is accessible in the network as allowed IP and set route as scope"
"verify the success stats when the API calls within the allowed IP range"
"set IP address that is accessible in the network as allowed IP and set service as scope"
"verify the success stats when the API calls within the allowed IP range"
"Navigate to Consumer page and filter the product"
"set api ip-restriction to global service level"
"Verify that IP Restriction is set at global service level"
"set IP address that is not accessible in the network as allowed IP and set service as scope"
"verify IP Restriction error when the API calls other than the allowed IP"
"Navigate to Consumer page and filter the product"
"set api ip-restriction to global service level"
"Verify that IP Restriction is set at global service level"
"set IP address that is not accessible in the network as allowed IP and set service as scope"
"verify IP Restriction error when the API calls other than the allowed IP"
"authenticates Mark (Access Manager)"
"Navigate to Consumer page and filter the product"
"Select the consumer from the list "
"set api rate limit as per the test config, Local Policy and Scope as Service"
"verify rate limit error when the API calls beyond the limit"
"set api rate limit as per the test config, Local Policy and Scope as Route"
"verify rate limit error when the API calls beyond the limit"
"set api rate limit as per the test config, Redis Policy and Scope as Service"
"verify rate limit error when the API calls beyond the limit"
"set api rate limit as per the test config, Redis Policy and Scope as Route"
"verify rate limit error when the API calls beyond the limit"
"set api rate limit to global service level"
"Verify that Rate limiting is set at global service level"
"set api rate limit as per the test config, Redis Policy and Scope as Service"
"verify rate limit error when the API calls beyond the limit"
"set api rate limit to global service level"
"Verify that Rate limiting is set at global service level"
"set api rate limit as per the test config, Redis Policy and Scope as Service"
"verify rate limit error when the API calls beyond the limit"
"creates an access request"
"authenticates Mark (Access-Manager)"
"verify the request details"
"Add group labels in request details window"
"approves an access request"
"Verify that API is accessible with the generated API Key"
"authenticates Mark (Access-Manager)"
"verify that consumers are filters as per given parameter"
"authenticates Mark (Access-Manager)"
"Navigate to Consumer page and filter the product"
"Click on the first consumer"
"Verify that labels can be deleted"
"Verify that labels can be updated"
"Verify that labels can be added"
"Creates and activates new namespace"
"Grant namespace access to access manager(Mark)"
"Grant CredentialIssuer.Admin permission to credential issuer(Wendy)"
"Select the namespace created for client credential "
"Creates authorization profile for Client ID/Secret"
"Creates authorization profile for Client ID/Secret - Shared IDP"
"Creates authorization profile for JWT - Generated Key Pair"
"Creates authorization profile for JWKS URL"
"Creates invalid authorization profile"
"Activates namespace for client credential flow tests"
"Creates a new service account"
"Publishes a new API to Kong Gateway"
"Creates a new product in the directory"
"Associate Namespace to the organization Unit"
"Update the Dataset in BC Data Catalogue to appear the API in the Directory"
"Adds environment with Client ID/Secret authenticator to product"
"applies authorization plugin to service published to Kong Gateway"
"Adds environment with JWT - Generated Key Pair authenticator to product"
"Adds environment with JWT - JWKS URL authenticator to product"
"Applies authorization plugin to service published to Kong Gateway"
"activate the service for Test environment"
"Adds environment for invalid authorization profile to other"
"Creates an access request"
"Access Manager logs in"
"Select scopes in Authorization Tab"
"approves an access request"
"Get access token using client ID and secret; make API request"
"Creates an access request"
"Access Manager logs in"
"approves an access request"
"Get access token using JWT key pair; make API request"
"Creates an access request"
"Access Manager logs in"
"approves an access request"
"Get access token using JWT key pair; make API request"
"authenticates Janis (api owner)"
"Expand Gateway service details pane"
"Verify total requests counts"
"Verify the routes details "
"Verify the host details "
"Verify the Tags details "
"authenticates Janis (api owner)"
"verify that Gateway service filters as per given parameter"
"Change Product environment from active to inactive"
"verify that Gateway service filters as per given parameter"
"Change Product environment from inactive to active"
"authenticates Janis (api owner)"
"Grant namespace access to Old User"
"authenticates with old user"
"Get the permission of the user"
"Verify that user account is disabled"
"Get current API Key"
"Regenrate credential"
"Verify that new API key is set to the consumer"
"Verify that only one API key(new key) is set to the consumer in Kong gateway"
"Verify that API is not accessible with the old API Key"
"Regenrate credential client ID and Secret"
"Make sure that the old client ID and Secret is disabled"
"Verify that service is accessible with new client ID and Secret"
"Activates the namespace"
"Deactivate the service for Test environment"
"Update the authorization scope from Kong ACL-API to Client Credential"
"applies authorization plugin to service published to Kong Gateway"
"activate the service for Test environment"
"Verify that service is not accessible with existing Client ID - Secret credentials"
"Raise request access"
"Collect the credentials"
"Access Manager logs in"
"approves an access request"
"Verify that API is accessible with the generated API Key"
"Activates the namespace"
"Create an authorization profile"
"Deactivate the service for Test environment"
"Update the authorization scope from Kong ACL-API to Client Credential"
"applies authorization plugin to service published to Kong Gateway"
"activate the service for Test environment"
"Verify that API is not accessible with the current API Key for Test environment"
"Creates an access request"
"Access Manager logs in"
"approves an access request"
"Get access token using client ID and secret; make API request"
"Activates the namespace"
"Update kong plugin for both the services"
"applies authorization plugin to service published to Kong Gateway"
"Associate another the service for Test environment"
"Get access token using client ID and secret; make API request for test"
"Get access token using client ID and secret; make API request for Dev"
"Creates an access request"
"Access Manager logs in"
"Get access token using client ID and secret; make API request for test"
"Get access token using client ID and secret; make API request for Dev"
"Activates the namespace"
"Change Product environment from active to inactive"
"Change Product environment from inactive to active"
"Edit the namespace from the tree view"
"Navigate to attribute tab"
"Set the Attributes"
"authenticates Janis (api owner) to get the user session token"
"Publish the Shared IDP profile"
"Create an authorization profile and associate it with shared IPD"
"authenticates Janis (api owner) to get the user session token"
"Put the resource and verify the success code in the response"
"Edit the created profile and verify the updated Issuer URL"
"authenticates Janis (api owner) to get the user session token"
"Update the Shared IDP Profile to an active clientID Secret auth environment"
"Creates an access request"
"Get access token using client ID and secret; make API request for test environment"
"Get access token using client ID and secret; make API request for Dev"
"creates and activates new namespace"
"creates a new service account"
"publishes a new API to Kong Gateway"
"creates as new product in the directory"
"Associate Namespace to the organization Unit"
"update the Dataset in BC Data Catelogue to appear the API in the Directory"
"publish product to directory"
"applies authorization plugin to service published to Kong Gateway"
"activate the service for Dev environment"
"authenticates Janis (api owner)"
"Delete Product Environment"
"Delete the Product"
"Delete Service Accounts"
"Delete Namespace"
"authenticates Janis (api owner) to get the user session token"
"Get the resource and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Verify Activity filter for \"Janis Smith\" user"
"Verify Activity filter for \"Harley Jones\" user"
"Verify Activity filter for \"Mark F Mark L\" user"
"Verify Activities filter for consumer"
"Activates namespace for client credential flow tests"
"Get the resource and verify the success code in the response"
"Verify Activity filter foe all the listed activities"
"creates and activates new namespace"
"creates a new service account"
"publishes a new API to Kong Gateway"
"creates as new product in the directory"
"Associate Namespace to the organization Unit"
"update the Dataset in BC Data Catelogue to appear the API in the Directory"
"publish product to directory"
"applies authorization plugin to service published to Kong Gateway"
"authenticates Janis (api owner)"
"grant namespace access to Mark (access manager)"
"Grant permission to Janis (API Owner)"
"Grant permission to Wendy"
"creates an access request"
"authenticates Janis (api owner)"
"Grant \"Access.Manager\" access to Mark (access manager)"
"Authenticates Mark (Access-Manager)"
"Verify that the option to approve request is displayed"
"Authenticates Janis (api owner)"
"Grant only \"Namespace.Manage\" permission to Wendy"
"Authenticates Wendy (Credential-Issuer)"
"Verify that all the namespace options and activities are displayed"
"Authenticates Janis (api owner)"
"Grant only \"CredentialIssuer.Admin\" access to Wendy (access manager)"
"Authenticates Wendy (Credential-Issuer)"
"Verify that only Authorization Profile option is displayed in Namespace page"
"Verify that authorization profile for Client ID/Secret is generated"
"Authenticates Janis (api owner)"
"Grant only \"Namespace.View\" permission to Mark"
"authenticates Mark"
"Verify that service accounts are not created"
"authenticates Janis (api owner)"
"Grant \"GatewayConfig.Publish\" and \"Namespace.View\" access to Wendy (access manager)"
"Authenticates Wendy (Credential-Issuer)"
"Verify that GWA API allows user to publish the API to Kong gateway"
"authenticates Janis (api owner)"
"Get the authorization token for the service account created with out \"Content.Publish\" permission"
"Verify that the document is not published without \"Content.Publish\" permission"
"Verify that application is deleted"
"creates an access request"
"Verify that application is deleted"
"creates an access request"
"authenticates Mark (Access-Manager)"
"approves an access request"
"Verify that application is deleted"
"Verify that API is not accessible with the generated API Key when the application is deleted"
"creates and activates new namespace"
"creates a new service account"
"publishes a new API to Kong Gateway"
"creates as new product in the directory"
"update the Dataset in BC Data Catelogue to appear the API in the Directory"
"applies authorization plugin to service published to Kong Gateway"
"authenticates Janis (api owner)"
"Navigate to Your Product tab in API Directory page"
"Verify that the banner for Preview mode is displayed"
"creates and activates new namespace"
"Get the resource and verify the Organization details in the response"
"Get the resource and verify the org Names in the response"
"authenticates Janis (api owner) to get the user session token"
"Get the resource and verify the success code in the response"
"Compare the scope values in response against the expected values"
"Get the resource and verify the success code in the response"
"Compare the Namespace values in response against the expected values"
"Delete the namespace associated with the organization, organization unit and verify the success code in the response"
"Verify that the deleted Namespace is not displayed in Get Call"
"Add the access of the organization to the specific user and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Compare the Namespace values in response against the expected values"
"authenticates Janis (api owner) to get the user session token"
"Put the resource and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Verify the status code and response message for invalid slugvalue"
"Delete the documentation"
"Delete the documentation"
"Put the resource and verify the success code in the response"
"Verify that document contant is displayed for GET /documentation"
"Verify that document contant is fetch by slug ID"
"authenticates Janis (api owner) to get the user session token"
"Put the resource and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Delete the authorization profile"
"Verify that the authorization profile is deleted"
"Put the resource and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Delete the authorization profile"
"Verify that the authorization profile is deleted"
"Put the resource and verify the success code in the response"
"Get the resource and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Delete the authorization profile"
"Verify that the authorization profile is deleted"
"authenticates Janis (api owner) to get the user session token"
"Put the resource and verify the success code in the response"
"Get the resource and verify the success code and product name in the response"
"Compare the values in response against the values passed in the request"
"authenticates Janis (api owner) to get the user session token"
"Verify that the product is visible in Manage Product Page"
"Delete the product environment and verify the success code in the response"
"Get the resource and verify that product environment is deleted"
"Delete the product and verify the success code in the response"
"Get the resource and verify that product is deleted"
"authenticates Janis (api owner) to get the user session token"
"Put the resource (/organization/{org}/datasets) and verify the success code in the response"
"Get the resource (/organization/{org}/datasets/{name}) and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Put the resource (/namespaces/{ns}/datasets/{name}) and verify the success code in the response"
"Get the resource (/namespaces/{ns}/datasets/{name}) and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Get the resource (/organizations/{org}/datasets/{name}) and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Get the resource (/organizations/{org}/datasets) and verify the success code in the response"
"Compare the values in response against the values passed in the request"
"Get the directory details (/directory) and verify the success code in the response"
"Get the directory details by its ID (/directory/{id}) and verify the success code in the response"
"Get the namespace directory details (/namespaces/{ns}/directory) and verify the success code and empty response for the namespace with no directory"
"Get the namespace directory details (/namespaces/{ns}/directory) and verify the success code in the response"
"Get the namespace directory details by its ID (/namespaces/{ns}/directory/{id}) and verify the success code in the response"
"Get the namespace directory details (/namespaces/{ns}/directory/{id}) for non exist directory ID and verify the response code"
"Delete the dataset (/organizations/{org}/datasets/{name}) and verify the success code in the response"
"Verify that deleted dataset does not display in Get dataset list"
"authenticates Janis (api owner) to get the user session token"
"Get the resource and verify the success code in the response"
"Verify that the selected Namespace is displayed in the Response list in the response"
"Get the resource and verify the success code in the response"
"Get the resource for namespace summary and verify the success code in the response"
"Delete the namespace and verify the Validation to prevent deleting the namespace"
"Force delete the namespace and verify the success code in the response"
Run Link: https://github.com/bcgov/api-services-portal/actions/runs/4245981733
|
non_defect
|
failed automated tests stats suites tests passes pending failures start end duration testsregistered passpercent pendingpercent other hasother false skipped hasskipped false failed tests creates and activates new namespace verify for invalid namespace name creates a new service account publishes a new api for dev environment to kong gateway creates as new product in the directory associate namespace to the organization unit update the dataset in bc data catelogue to appear the api in the directory publish product to directory create a test environment applies authorization plugin to service published to kong gateway activate the service for test environment activate the service for dev environment authenticates janis api owner grant namespace access to mark access manager grant credentialissuer admin permission to janis api owner grant credentialissuer admin permission to janis api owner collect the credentials close the popup without collecting credentials authenticates mark access manager verify that the request status is pending approval collect the credentials verify that api is not accessible with the generated api key when the request is not approved authenticates mark access manager verify the request details add group labels in request details window approves an access request verify that api is accessible with the generated api key authenticates mark access manager navigate to consumer page and filter the product click on the first consumer click on grant access button grant access to test environment verify that api is accessible with the generated api key for test environment apply key auth only authorization plugin to kong gateway update the kong key auth plugin the new consumer verify the service is accessibale with api key for free access verify the service is accessible with api key for elevated access verify the service is accessibale with api key for free access verify the service is accessible with api key for elevated access verify the rate limiting is applied for free access authenticates mark access manager navigate to consumer page and filter the product select the consumer from the list set ip address that is not accessible in the network as allowed ip and set route as scope verify ip restriction error when the api calls other than the allowed ip set ip address that is not accessible in the network as allowed ip and set service as scope verify ip restriction error when the api calls other than the allowed ip set ip address that is accessible in the network as allowed ip and set route as scope verify the success stats when the api calls within the allowed ip range set ip address that is accessible in the network as allowed ip and set service as scope verify the success stats when the api calls within the allowed ip range navigate to consumer page and filter the product set api ip restriction to global service level verify that ip restriction is set at global service level set ip address that is not accessible in the network as allowed ip and set service as scope verify ip restriction error when the api calls other than the allowed ip navigate to consumer page and filter the product set api ip restriction to global service level verify that ip restriction is set at global service level set ip address that is not accessible in the network as allowed ip and set service as scope verify ip restriction error when the api calls other than the allowed ip authenticates mark access manager navigate to consumer page and filter the product select the consumer from the list set api rate limit as per the test config local policy and scope as service verify rate limit error when the api calls beyond the limit set api rate limit as per the test config local policy and scope as route verify rate limit error when the api calls beyond the limit set api rate limit as per the test config redis policy and scope as service verify rate limit error when the api calls beyond the limit set api rate limit as per the test config redis policy and scope as route verify rate limit error when the api calls beyond the limit set api rate limit to global service level verify that rate limiting is set at global service level set api rate limit as per the test config redis policy and scope as service verify rate limit error when the api calls beyond the limit set api rate limit to global service level verify that rate limiting is set at global service level set api rate limit as per the test config redis policy and scope as service verify rate limit error when the api calls beyond the limit creates an access request authenticates mark access manager verify the request details add group labels in request details window approves an access request verify that api is accessible with the generated api key authenticates mark access manager verify that consumers are filters as per given parameter authenticates mark access manager navigate to consumer page and filter the product click on the first consumer verify that labels can be deleted verify that labels can be updated verify that labels can be added creates and activates new namespace grant namespace access to access manager mark grant credentialissuer admin permission to credential issuer wendy select the namespace created for client credential creates authorization profile for client id secret creates authorization profile for client id secret shared idp creates authorization profile for jwt generated key pair creates authorization profile for jwks url creates invalid authorization profile activates namespace for client credential flow tests creates a new service account publishes a new api to kong gateway creates a new product in the directory associate namespace to the organization unit update the dataset in bc data catalogue to appear the api in the directory adds environment with client id secret authenticator to product applies authorization plugin to service published to kong gateway adds environment with jwt generated key pair authenticator to product adds environment with jwt jwks url authenticator to product applies authorization plugin to service published to kong gateway activate the service for test environment adds environment for invalid authorization profile to other creates an access request access manager logs in select scopes in authorization tab approves an access request get access token using client id and secret make api request creates an access request access manager logs in approves an access request get access token using jwt key pair make api request creates an access request access manager logs in approves an access request get access token using jwt key pair make api request authenticates janis api owner expand gateway service details pane verify total requests counts verify the routes details verify the host details verify the tags details authenticates janis api owner verify that gateway service filters as per given parameter change product environment from active to inactive verify that gateway service filters as per given parameter change product environment from inactive to active authenticates janis api owner grant namespace access to old user authenticates with old user get the permission of the user verify that user account is disabled get current api key regenrate credential verify that new api key is set to the consumer verify that only one api key new key is set to the consumer in kong gateway verify that api is not accessible with the old api key regenrate credential client id and secret make sure that the old client id and secret is disabled verify that service is accessible with new client id and secret activates the namespace deactivate the service for test environment update the authorization scope from kong acl api to client credential applies authorization plugin to service published to kong gateway activate the service for test environment verify that service is not accessible with existing client id secret credentials raise request access collect the credentials access manager logs in approves an access request verify that api is accessible with the generated api key activates the namespace create an authorization profile deactivate the service for test environment update the authorization scope from kong acl api to client credential applies authorization plugin to service published to kong gateway activate the service for test environment verify that api is not accessible with the current api key for test environment creates an access request access manager logs in approves an access request get access token using client id and secret make api request activates the namespace update kong plugin for both the services applies authorization plugin to service published to kong gateway associate another the service for test environment get access token using client id and secret make api request for test get access token using client id and secret make api request for dev creates an access request access manager logs in get access token using client id and secret make api request for test get access token using client id and secret make api request for dev activates the namespace change product environment from active to inactive change product environment from inactive to active edit the namespace from the tree view navigate to attribute tab set the attributes authenticates janis api owner to get the user session token publish the shared idp profile create an authorization profile and associate it with shared ipd authenticates janis api owner to get the user session token put the resource and verify the success code in the response edit the created profile and verify the updated issuer url authenticates janis api owner to get the user session token update the shared idp profile to an active clientid secret auth environment creates an access request get access token using client id and secret make api request for test environment get access token using client id and secret make api request for dev creates and activates new namespace creates a new service account publishes a new api to kong gateway creates as new product in the directory associate namespace to the organization unit update the dataset in bc data catelogue to appear the api in the directory publish product to directory applies authorization plugin to service published to kong gateway activate the service for dev environment authenticates janis api owner delete product environment delete the product delete service accounts delete namespace authenticates janis api owner to get the user session token get the resource and verify the success code in the response get the resource and verify the success code in the response verify activity filter for janis smith user verify activity filter for harley jones user verify activity filter for mark f mark l user verify activities filter for consumer activates namespace for client credential flow tests get the resource and verify the success code in the response verify activity filter foe all the listed activities creates and activates new namespace creates a new service account publishes a new api to kong gateway creates as new product in the directory associate namespace to the organization unit update the dataset in bc data catelogue to appear the api in the directory publish product to directory applies authorization plugin to service published to kong gateway authenticates janis api owner grant namespace access to mark access manager grant permission to janis api owner grant permission to wendy creates an access request authenticates janis api owner grant access manager access to mark access manager authenticates mark access manager verify that the option to approve request is displayed authenticates janis api owner grant only namespace manage permission to wendy authenticates wendy credential issuer verify that all the namespace options and activities are displayed authenticates janis api owner grant only credentialissuer admin access to wendy access manager authenticates wendy credential issuer verify that only authorization profile option is displayed in namespace page verify that authorization profile for client id secret is generated authenticates janis api owner grant only namespace view permission to mark authenticates mark verify that service accounts are not created authenticates janis api owner grant gatewayconfig publish and namespace view access to wendy access manager authenticates wendy credential issuer verify that gwa api allows user to publish the api to kong gateway authenticates janis api owner get the authorization token for the service account created with out content publish permission verify that the document is not published without content publish permission verify that application is deleted creates an access request verify that application is deleted creates an access request authenticates mark access manager approves an access request verify that application is deleted verify that api is not accessible with the generated api key when the application is deleted creates and activates new namespace creates a new service account publishes a new api to kong gateway creates as new product in the directory update the dataset in bc data catelogue to appear the api in the directory applies authorization plugin to service published to kong gateway authenticates janis api owner navigate to your product tab in api directory page verify that the banner for preview mode is displayed creates and activates new namespace get the resource and verify the organization details in the response get the resource and verify the org names in the response authenticates janis api owner to get the user session token get the resource and verify the success code in the response compare the scope values in response against the expected values get the resource and verify the success code in the response compare the namespace values in response against the expected values delete the namespace associated with the organization organization unit and verify the success code in the response verify that the deleted namespace is not displayed in get call add the access of the organization to the specific user and verify the success code in the response get the resource and verify the success code in the response compare the namespace values in response against the expected values authenticates janis api owner to get the user session token put the resource and verify the success code in the response get the resource and verify the success code in the response compare the values in response against the values passed in the request verify the status code and response message for invalid slugvalue delete the documentation delete the documentation put the resource and verify the success code in the response verify that document contant is displayed for get documentation verify that document contant is fetch by slug id authenticates janis api owner to get the user session token put the resource and verify the success code in the response get the resource and verify the success code in the response compare the values in response against the values passed in the request delete the authorization profile verify that the authorization profile is deleted put the resource and verify the success code in the response get the resource and verify the success code in the response compare the values in response against the values passed in the request delete the authorization profile verify that the authorization profile is deleted put the resource and verify the success code in the response get the resource and verify the success code in the response compare the values in response against the values passed in the request delete the authorization profile verify that the authorization profile is deleted authenticates janis api owner to get the user session token put the resource and verify the success code in the response get the resource and verify the success code and product name in the response compare the values in response against the values passed in the request authenticates janis api owner to get the user session token verify that the product is visible in manage product page delete the product environment and verify the success code in the response get the resource and verify that product environment is deleted delete the product and verify the success code in the response get the resource and verify that product is deleted authenticates janis api owner to get the user session token put the resource organization org datasets and verify the success code in the response get the resource organization org datasets name and verify the success code in the response compare the values in response against the values passed in the request put the resource namespaces ns datasets name and verify the success code in the response get the resource namespaces ns datasets name and verify the success code in the response compare the values in response against the values passed in the request get the resource organizations org datasets name and verify the success code in the response compare the values in response against the values passed in the request get the resource organizations org datasets and verify the success code in the response compare the values in response against the values passed in the request get the directory details directory and verify the success code in the response get the directory details by its id directory id and verify the success code in the response get the namespace directory details namespaces ns directory and verify the success code and empty response for the namespace with no directory get the namespace directory details namespaces ns directory and verify the success code in the response get the namespace directory details by its id namespaces ns directory id and verify the success code in the response get the namespace directory details namespaces ns directory id for non exist directory id and verify the response code delete the dataset organizations org datasets name and verify the success code in the response verify that deleted dataset does not display in get dataset list authenticates janis api owner to get the user session token get the resource and verify the success code in the response verify that the selected namespace is displayed in the response list in the response get the resource and verify the success code in the response get the resource for namespace summary and verify the success code in the response delete the namespace and verify the validation to prevent deleting the namespace force delete the namespace and verify the success code in the response run link
| 0
|
296,919
| 9,134,754,340
|
IssuesEvent
|
2019-02-26 01:11:36
|
jwc2790/cuffney.com
|
https://api.github.com/repos/jwc2790/cuffney.com
|
reopened
|
Get CI to Run In Parallel
|
low priority quick
|
See if you can speed up the CI Builds to < 1 minute.
I know this might require a paid plan which probably doesn't make sense for this project...but I want to have a template to refer to for other projects.
I should be able to test this with a free trial at code climate
|
1.0
|
Get CI to Run In Parallel - See if you can speed up the CI Builds to < 1 minute.
I know this might require a paid plan which probably doesn't make sense for this project...but I want to have a template to refer to for other projects.
I should be able to test this with a free trial at code climate
|
non_defect
|
get ci to run in parallel see if you can speed up the ci builds to minute i know this might require a paid plan which probably doesn t make sense for this project but i want to have a template to refer to for other projects i should be able to test this with a free trial at code climate
| 0
|
64,123
| 15,801,082,527
|
IssuesEvent
|
2021-04-03 02:53:05
|
gridcoin-community/Gridcoin-Research
|
https://api.github.com/repos/gridcoin-community/Gridcoin-Research
|
closed
|
SVG in depends folders does not compile
|
build System compatibility gui
|
SVG in depends folder does not compile
If you like I can port my solution for this from https://github.com/navcoin/navcoin-core/pull/557/commits/104023b284b856d2ad3d3bf41b0d73ebaf15a2f8
:+1: do you guys have a bounty program for code?
Or you guys can just take a look at my implementation for the fix and copy it.
|
1.0
|
SVG in depends folders does not compile - SVG in depends folder does not compile
If you like I can port my solution for this from https://github.com/navcoin/navcoin-core/pull/557/commits/104023b284b856d2ad3d3bf41b0d73ebaf15a2f8
:+1: do you guys have a bounty program for code?
Or you guys can just take a look at my implementation for the fix and copy it.
|
non_defect
|
svg in depends folders does not compile svg in depends folder does not compile if you like i can port my solution for this from do you guys have a bounty program for code or you guys can just take a look at my implementation for the fix and copy it
| 0
|
2,346
| 2,595,470,697
|
IssuesEvent
|
2015-02-20 14:32:41
|
RcppCore/Rcpp
|
https://api.github.com/repos/RcppCore/Rcpp
|
opened
|
add abi-compliance-checker to travis script
|
testing
|
see: http://manpages.ubuntu.com/manpages/natty/man1/abi-compliance-checker.1.html
Would give us more accurate notion of whether potential changes actually break the ABI (e.g. https://github.com/RcppCore/Rcpp/pull/260)
|
1.0
|
add abi-compliance-checker to travis script - see: http://manpages.ubuntu.com/manpages/natty/man1/abi-compliance-checker.1.html
Would give us more accurate notion of whether potential changes actually break the ABI (e.g. https://github.com/RcppCore/Rcpp/pull/260)
|
non_defect
|
add abi compliance checker to travis script see would give us more accurate notion of whether potential changes actually break the abi e g
| 0
|
390,014
| 11,520,700,288
|
IssuesEvent
|
2020-02-14 15:17:16
|
earthlab/matplotcheck
|
https://api.github.com/repos/earthlab/matplotcheck
|
closed
|
raster tester can't test for polygon overlays
|
high priority
|
Currently raster plots and vector plots are handled separately. however, often we combine the two via overlays in our assignments.
if the RasterTester object inherited from the VectorTester, then all methods would be available for testing!
|
1.0
|
raster tester can't test for polygon overlays - Currently raster plots and vector plots are handled separately. however, often we combine the two via overlays in our assignments.
if the RasterTester object inherited from the VectorTester, then all methods would be available for testing!
|
non_defect
|
raster tester can t test for polygon overlays currently raster plots and vector plots are handled separately however often we combine the two via overlays in our assignments if the rastertester object inherited from the vectortester then all methods would be available for testing
| 0
|
56,184
| 3,078,457,885
|
IssuesEvent
|
2015-08-21 10:20:11
|
pavel-pimenov/flylinkdc-r5xx
|
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
|
opened
|
Редизайн настроек "Цвета и шрифты"
|
bug Component-UI imported Priority-Medium Usability
|
_From [bobrikov](https://code.google.com/u/bobrikov/) on May 31, 2011 19:06:20_
1. Убрать дубликат кнопки "стиль текста"
2а. Вместо них двух добавить одну "Настройка текста", куда включить все возможные атрибуты (цвета, шрифты, размеры, подчеркивания итд)
2б. убрать кнопку "чернобелый" (сделать такую цветовую схему и хватит имхо)
3. Перенести сюда менеджер тем иконок
4. Уменьшить поле "другие цвета" и перенести кнопки выбора правее. (место позволяет)
**Attachment:** [цвета-и-шрифты.png](http://code.google.com/p/flylinkdc/issues/detail?id=472)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=472_
|
1.0
|
Редизайн настроек "Цвета и шрифты" - _From [bobrikov](https://code.google.com/u/bobrikov/) on May 31, 2011 19:06:20_
1. Убрать дубликат кнопки "стиль текста"
2а. Вместо них двух добавить одну "Настройка текста", куда включить все возможные атрибуты (цвета, шрифты, размеры, подчеркивания итд)
2б. убрать кнопку "чернобелый" (сделать такую цветовую схему и хватит имхо)
3. Перенести сюда менеджер тем иконок
4. Уменьшить поле "другие цвета" и перенести кнопки выбора правее. (место позволяет)
**Attachment:** [цвета-и-шрифты.png](http://code.google.com/p/flylinkdc/issues/detail?id=472)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=472_
|
non_defect
|
редизайн настроек цвета и шрифты from on may убрать дубликат кнопки стиль текста вместо них двух добавить одну настройка текста куда включить все возможные атрибуты цвета шрифты размеры подчеркивания итд убрать кнопку чернобелый сделать такую цветовую схему и хватит имхо перенести сюда менеджер тем иконок уменьшить поле другие цвета и перенести кнопки выбора правее место позволяет attachment original issue
| 0
|
154,136
| 13,539,128,298
|
IssuesEvent
|
2020-09-16 13:04:51
|
MaSyMoS/masymos-seek-middleware
|
https://api.github.com/repos/MaSyMoS/masymos-seek-middleware
|
closed
|
choose unique identifyer for models
|
documentation question
|
# Overview
- json checked in issue #2
## SEEK json
- [example](https://fairdomhub.org/models/20.json)
- `data.id`
- points to the model, ignoring the version (**model id**)
- i.e. `24`
- `data.meta.uuid`
- points to the model, ignoring the version (**model uuid**)
- i.e. `67bb3ce0-caa7-0138-f7f7-0242ac120004`
- `data.links.self`
- points to the model + version without domain dependency (**model relative link**)
- i.e. `/models/24?version=3`
- `data.attributes.version`
- provides the current version number (**model version**)
- i.e. `3`
- `data.attributes.versions.#.url` (_# is version number_)
- provides the model + version with domain dependency (**model absolute link**)
- i.e. `https://sandbox2.fairdomhub.org/models/24.json?version=3`
## possible uses
- full absolute URL to model, i.e. `https://sandbox2.fairdomhub.org/models/24`
- **implies**: masymos-database needs a reset on domain change
- **benefits**:
- no configuration of seek domain needed
- no stored metadata needed on masymos side
- human-readable unique identifier
- `data.id` OR `data.meta.uuid`
- **requires**: MaSeMiWa must have the SEEK-domain configured
- **implies**: MaSeMiWa could identify, if model version is newer by saving `data.attributes.version` in graph meta data
- `data.links.self`
- **requires**: MaSeMiWa must have the SEEK-domain configured
- **possible benefits**:
- MaSeMiWa could only use the ID on masymos-side
- is domain independent
# ToDo
- [x] decide what UID to use → **full URL**
- [x] documentation
|
1.0
|
choose unique identifyer for models - # Overview
- json checked in issue #2
## SEEK json
- [example](https://fairdomhub.org/models/20.json)
- `data.id`
- points to the model, ignoring the version (**model id**)
- i.e. `24`
- `data.meta.uuid`
- points to the model, ignoring the version (**model uuid**)
- i.e. `67bb3ce0-caa7-0138-f7f7-0242ac120004`
- `data.links.self`
- points to the model + version without domain dependency (**model relative link**)
- i.e. `/models/24?version=3`
- `data.attributes.version`
- provides the current version number (**model version**)
- i.e. `3`
- `data.attributes.versions.#.url` (_# is version number_)
- provides the model + version with domain dependency (**model absolute link**)
- i.e. `https://sandbox2.fairdomhub.org/models/24.json?version=3`
## possible uses
- full absolute URL to model, i.e. `https://sandbox2.fairdomhub.org/models/24`
- **implies**: masymos-database needs a reset on domain change
- **benefits**:
- no configuration of seek domain needed
- no stored metadata needed on masymos side
- human-readable unique identifier
- `data.id` OR `data.meta.uuid`
- **requires**: MaSeMiWa must have the SEEK-domain configured
- **implies**: MaSeMiWa could identify, if model version is newer by saving `data.attributes.version` in graph meta data
- `data.links.self`
- **requires**: MaSeMiWa must have the SEEK-domain configured
- **possible benefits**:
- MaSeMiWa could only use the ID on masymos-side
- is domain independent
# ToDo
- [x] decide what UID to use → **full URL**
- [x] documentation
|
non_defect
|
choose unique identifyer for models overview json checked in issue seek json data id points to the model ignoring the version model id i e data meta uuid points to the model ignoring the version model uuid i e data links self points to the model version without domain dependency model relative link i e models version data attributes version provides the current version number model version i e data attributes versions url is version number provides the model version with domain dependency model absolute link i e possible uses full absolute url to model i e implies masymos database needs a reset on domain change benefits no configuration of seek domain needed no stored metadata needed on masymos side human readable unique identifier data id or data meta uuid requires masemiwa must have the seek domain configured implies masemiwa could identify if model version is newer by saving data attributes version in graph meta data data links self requires masemiwa must have the seek domain configured possible benefits masemiwa could only use the id on masymos side is domain independent todo decide what uid to use → full url documentation
| 0
|
320,862
| 9,790,367,861
|
IssuesEvent
|
2019-06-10 12:33:28
|
openshift/odo
|
https://api.github.com/repos/openshift/odo
|
closed
|
odo component create interactive mode: default value and prompt are confusing for source location
|
kind/bug priority/High
|
[kind/bug]
This tripped up quite a few people at the Summit lab, as they were blindly accepting the defaults in interactive mode, despite the instructions saying to enter '/' for the source location in interactive mode to try to avoid this issue.
## What versions of software are you using?
- Operating System: macOS
- Output of `odo version`: v1.0.0-beta2 (also present in beta1)
## How did you run odo exactly?
```
git clone https://github.com/openshift-labs/beercan-shooter-game.git bc2
cd bc2
odo component create
```
choose nodejs for component type, 10 for version, and then look at the next step where it asks for location.
## Actual behavior
Location of context component, relative to '/opt/app-root/src/bc2' (/opt/app-root/src/bc2)
If you accept this default, it appends the current path to the current path, resulting in:
```
/opt/app-root-src-bc2/opt/app-root/src/bc2
```
However, that is not clear from odo config view:
```
odo config view
COMPONENT SETTINGS
------------------------------------------------
PARAMETER CURRENT_VALUE
Type nodejs:10
Application app
Project developing-with-odo-odobeta2-6bh0t
SourceType local
Ref
SourceLocation /opt/app-root/src/bc2
Ports
Name nodejs-yyzt
MinMemory
MaxMemory
Ignore
MinCPU
MaxCPU
```
But when you run odo push, you'll see that it's looking in /opt/app-root/src/bc2/opt/app-root/src/bc2:
```
master ? $ odo push
✓ Checking component
✓ Checking component version
✓ Creating nodejs:10 component with name nodejs-yyzt
✓ Initializing 'nodejs-yyzt' component
✓ Creating component nodejs-yyzt
✓ Successfully created component nodejs-yyzt
✓ Applying component settings to component: nodejs-yyzt
✓ Successfully updated component with name: nodejs-yyzt
✓ Pushing changes to component: nodejs-yyzt of type local
✗ open /opt/app-root/src/bc2/opt/app-root/src/bc2: no such file or directory
```
To get it to actually work, I have to enter '/' as the source location. Note that entering '.' does not work, which was my first inclination.
## Expected behavior
If the value entered for source location is in fact supposed to be relative to the current directory, then change the default value, so that if the user accepts the default, the source location is the current directory.
For example:
Location of context component, relative to '/opt/app-root/src/bc2' (/)
or
Location of context component, relative to '/opt/app-root/src/bc2' (.)
## Any logs, error output, etc?
|
1.0
|
odo component create interactive mode: default value and prompt are confusing for source location - [kind/bug]
This tripped up quite a few people at the Summit lab, as they were blindly accepting the defaults in interactive mode, despite the instructions saying to enter '/' for the source location in interactive mode to try to avoid this issue.
## What versions of software are you using?
- Operating System: macOS
- Output of `odo version`: v1.0.0-beta2 (also present in beta1)
## How did you run odo exactly?
```
git clone https://github.com/openshift-labs/beercan-shooter-game.git bc2
cd bc2
odo component create
```
choose nodejs for component type, 10 for version, and then look at the next step where it asks for location.
## Actual behavior
Location of context component, relative to '/opt/app-root/src/bc2' (/opt/app-root/src/bc2)
If you accept this default, it appends the current path to the current path, resulting in:
```
/opt/app-root-src-bc2/opt/app-root/src/bc2
```
However, that is not clear from odo config view:
```
odo config view
COMPONENT SETTINGS
------------------------------------------------
PARAMETER CURRENT_VALUE
Type nodejs:10
Application app
Project developing-with-odo-odobeta2-6bh0t
SourceType local
Ref
SourceLocation /opt/app-root/src/bc2
Ports
Name nodejs-yyzt
MinMemory
MaxMemory
Ignore
MinCPU
MaxCPU
```
But when you run odo push, you'll see that it's looking in /opt/app-root/src/bc2/opt/app-root/src/bc2:
```
master ? $ odo push
✓ Checking component
✓ Checking component version
✓ Creating nodejs:10 component with name nodejs-yyzt
✓ Initializing 'nodejs-yyzt' component
✓ Creating component nodejs-yyzt
✓ Successfully created component nodejs-yyzt
✓ Applying component settings to component: nodejs-yyzt
✓ Successfully updated component with name: nodejs-yyzt
✓ Pushing changes to component: nodejs-yyzt of type local
✗ open /opt/app-root/src/bc2/opt/app-root/src/bc2: no such file or directory
```
To get it to actually work, I have to enter '/' as the source location. Note that entering '.' does not work, which was my first inclination.
## Expected behavior
If the value entered for source location is in fact supposed to be relative to the current directory, then change the default value, so that if the user accepts the default, the source location is the current directory.
For example:
Location of context component, relative to '/opt/app-root/src/bc2' (/)
or
Location of context component, relative to '/opt/app-root/src/bc2' (.)
## Any logs, error output, etc?
|
non_defect
|
odo component create interactive mode default value and prompt are confusing for source location this tripped up quite a few people at the summit lab as they were blindly accepting the defaults in interactive mode despite the instructions saying to enter for the source location in interactive mode to try to avoid this issue what versions of software are you using operating system macos output of odo version also present in how did you run odo exactly git clone cd odo component create choose nodejs for component type for version and then look at the next step where it asks for location actual behavior location of context component relative to opt app root src opt app root src if you accept this default it appends the current path to the current path resulting in opt app root src opt app root src however that is not clear from odo config view odo config view component settings parameter current value type nodejs application app project developing with odo sourcetype local ref sourcelocation opt app root src ports name nodejs yyzt minmemory maxmemory ignore mincpu maxcpu but when you run odo push you ll see that it s looking in opt app root src opt app root src master odo push ✓ checking component ✓ checking component version ✓ creating nodejs component with name nodejs yyzt ✓ initializing nodejs yyzt component ✓ creating component nodejs yyzt ✓ successfully created component nodejs yyzt ✓ applying component settings to component nodejs yyzt ✓ successfully updated component with name nodejs yyzt ✓ pushing changes to component nodejs yyzt of type local ✗ open opt app root src opt app root src no such file or directory to get it to actually work i have to enter as the source location note that entering does not work which was my first inclination expected behavior if the value entered for source location is in fact supposed to be relative to the current directory then change the default value so that if the user accepts the default the source location is the current directory for example location of context component relative to opt app root src or location of context component relative to opt app root src any logs error output etc
| 0
|
511,019
| 14,851,896,802
|
IssuesEvent
|
2021-01-18 07:45:28
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
m.999.md - see bug description
|
browser-firefox-mobile engine-gecko ml-needsdiagnosis-false priority-normal
|
<!-- @browser: Firefox Mobile 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/65789 -->
**URL**: https://m.999.md/ru/
**Browser / Version**: Firefox Mobile 84.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: web content associated with previous account-related search from Google Chrome browser!
**Steps to Reproduce**:
There are some kind of cross-browser cookies or Smith(I'm not a developer) which violates my anonimity.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
m.999.md - see bug description - <!-- @browser: Firefox Mobile 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/65789 -->
**URL**: https://m.999.md/ru/
**Browser / Version**: Firefox Mobile 84.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: web content associated with previous account-related search from Google Chrome browser!
**Steps to Reproduce**:
There are some kind of cross-browser cookies or Smith(I'm not a developer) which violates my anonimity.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
m md see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description web content associated with previous account related search from google chrome browser steps to reproduce there are some kind of cross browser cookies or smith i m not a developer which violates my anonimity browser configuration none from with ❤️
| 0
|
116,118
| 9,820,524,098
|
IssuesEvent
|
2019-06-14 03:02:08
|
sylabs/singularity
|
https://api.github.com/repos/sylabs/singularity
|
closed
|
Improve test coverage of pkg/image
|
Release 3.3.0 Testing
|
## Version of Singularity:
Improve test coverage of pkg/image
### Expected behavior
Get close to 100% coverage.
### Actual behavior
`
$ go test -coverprofile=coverage.out && go tool cover -func=coverage.out
PASS
coverage: 38.2% of statements
ok github.com/sylabs/singularity/pkg/image 2.706s
github.com/sylabs/singularity/pkg/image/ext3.go:43: CheckExt3Header 0.0%
github.com/sylabs/singularity/pkg/image/ext3.go:76: initializer 0.0%
github.com/sylabs/singularity/pkg/image/ext3.go:100: openMode 0.0%
github.com/sylabs/singularity/pkg/image/image.go:78: AuthorizedPath 0.0%
github.com/sylabs/singularity/pkg/image/image.go:96: AuthorizedOwner 0.0%
github.com/sylabs/singularity/pkg/image/image.go:117: AuthorizedGroup 0.0%
github.com/sylabs/singularity/pkg/image/image.go:138: HasRootFs 75.0%
github.com/sylabs/singularity/pkg/image/image.go:148: ResolvePath 71.4%
github.com/sylabs/singularity/pkg/image/image.go:161: Init 59.4%
github.com/sylabs/singularity/pkg/image/reader.go:16: Error 0.0%
github.com/sylabs/singularity/pkg/image/reader.go:25: checkImage 100.0%
github.com/sylabs/singularity/pkg/image/reader.go:35: getSectionReader 100.0%
github.com/sylabs/singularity/pkg/image/reader.go:39: commonSectionReader 100.0%
github.com/sylabs/singularity/pkg/image/reader.go:79: NewPartitionReader 100.0%
github.com/sylabs/singularity/pkg/image/reader.go:87: NewSectionReader 100.0%
github.com/sylabs/singularity/pkg/image/sandbox.go:15: initializer 0.0%
github.com/sylabs/singularity/pkg/image/sandbox.go:30: openMode 0.0%
github.com/sylabs/singularity/pkg/image/sif.go:25: initializer 60.0%
github.com/sylabs/singularity/pkg/image/sif.go:153: openMode 66.7%
github.com/sylabs/singularity/pkg/image/squashfs.go:40: CheckSquashfsHeader 0.0%
github.com/sylabs/singularity/pkg/image/squashfs.go:79: initializer 0.0%
github.com/sylabs/singularity/pkg/image/squashfs.go:109: openMode 0.0%
total: (statements) 38.2%
`
### Steps to reproduce behavior
Write here.
|
1.0
|
Improve test coverage of pkg/image - ## Version of Singularity:
Improve test coverage of pkg/image
### Expected behavior
Get close to 100% coverage.
### Actual behavior
`
$ go test -coverprofile=coverage.out && go tool cover -func=coverage.out
PASS
coverage: 38.2% of statements
ok github.com/sylabs/singularity/pkg/image 2.706s
github.com/sylabs/singularity/pkg/image/ext3.go:43: CheckExt3Header 0.0%
github.com/sylabs/singularity/pkg/image/ext3.go:76: initializer 0.0%
github.com/sylabs/singularity/pkg/image/ext3.go:100: openMode 0.0%
github.com/sylabs/singularity/pkg/image/image.go:78: AuthorizedPath 0.0%
github.com/sylabs/singularity/pkg/image/image.go:96: AuthorizedOwner 0.0%
github.com/sylabs/singularity/pkg/image/image.go:117: AuthorizedGroup 0.0%
github.com/sylabs/singularity/pkg/image/image.go:138: HasRootFs 75.0%
github.com/sylabs/singularity/pkg/image/image.go:148: ResolvePath 71.4%
github.com/sylabs/singularity/pkg/image/image.go:161: Init 59.4%
github.com/sylabs/singularity/pkg/image/reader.go:16: Error 0.0%
github.com/sylabs/singularity/pkg/image/reader.go:25: checkImage 100.0%
github.com/sylabs/singularity/pkg/image/reader.go:35: getSectionReader 100.0%
github.com/sylabs/singularity/pkg/image/reader.go:39: commonSectionReader 100.0%
github.com/sylabs/singularity/pkg/image/reader.go:79: NewPartitionReader 100.0%
github.com/sylabs/singularity/pkg/image/reader.go:87: NewSectionReader 100.0%
github.com/sylabs/singularity/pkg/image/sandbox.go:15: initializer 0.0%
github.com/sylabs/singularity/pkg/image/sandbox.go:30: openMode 0.0%
github.com/sylabs/singularity/pkg/image/sif.go:25: initializer 60.0%
github.com/sylabs/singularity/pkg/image/sif.go:153: openMode 66.7%
github.com/sylabs/singularity/pkg/image/squashfs.go:40: CheckSquashfsHeader 0.0%
github.com/sylabs/singularity/pkg/image/squashfs.go:79: initializer 0.0%
github.com/sylabs/singularity/pkg/image/squashfs.go:109: openMode 0.0%
total: (statements) 38.2%
`
### Steps to reproduce behavior
Write here.
|
non_defect
|
improve test coverage of pkg image version of singularity improve test coverage of pkg image expected behavior get close to coverage actual behavior go test coverprofile coverage out go tool cover func coverage out pass coverage of statements ok github com sylabs singularity pkg image github com sylabs singularity pkg image go github com sylabs singularity pkg image go initializer github com sylabs singularity pkg image go openmode github com sylabs singularity pkg image image go authorizedpath github com sylabs singularity pkg image image go authorizedowner github com sylabs singularity pkg image image go authorizedgroup github com sylabs singularity pkg image image go hasrootfs github com sylabs singularity pkg image image go resolvepath github com sylabs singularity pkg image image go init github com sylabs singularity pkg image reader go error github com sylabs singularity pkg image reader go checkimage github com sylabs singularity pkg image reader go getsectionreader github com sylabs singularity pkg image reader go commonsectionreader github com sylabs singularity pkg image reader go newpartitionreader github com sylabs singularity pkg image reader go newsectionreader github com sylabs singularity pkg image sandbox go initializer github com sylabs singularity pkg image sandbox go openmode github com sylabs singularity pkg image sif go initializer github com sylabs singularity pkg image sif go openmode github com sylabs singularity pkg image squashfs go checksquashfsheader github com sylabs singularity pkg image squashfs go initializer github com sylabs singularity pkg image squashfs go openmode total statements steps to reproduce behavior write here
| 0
|
719,918
| 24,773,836,194
|
IssuesEvent
|
2022-10-23 13:36:52
|
AY2223S1-CS2103T-F12-1/tp
|
https://api.github.com/repos/AY2223S1-CS2103T-F12-1/tp
|
closed
|
Refactor PriceRange class
|
enhancement priority.high
|
Refactor PriceRange class to contain two Price objects instead of doubles, also add custom comparator for Price class
|
1.0
|
Refactor PriceRange class - Refactor PriceRange class to contain two Price objects instead of doubles, also add custom comparator for Price class
|
non_defect
|
refactor pricerange class refactor pricerange class to contain two price objects instead of doubles also add custom comparator for price class
| 0
|
48,687
| 5,966,724,091
|
IssuesEvent
|
2017-05-30 14:35:19
|
telerik/kendo-ui-core
|
https://api.github.com/repos/telerik/kendo-ui-core
|
closed
|
Missing Spreadsheet toolbar buttons localization
|
Bug C: Spreadsheet In Testing SEV: Medium
|
### Bug report
Alignment buttons, freeze buttons, merge buttons cannot be localized through messages. The same tools can be localized properly through the dialog messages options, when they are displayed in dialogs.
### Reproduction of the problem
1. Use the following [Dojo](http://dojo.telerik.com/ULEVi/3)
### Current behavior
The buttons are not localized according to the provided messages.
### Expected/desired behavior
The buttons are localized according to the provided messages.
### Environment
* **Kendo UI version:** 2017.2.504
* **Browser:** [all]
|
1.0
|
Missing Spreadsheet toolbar buttons localization - ### Bug report
Alignment buttons, freeze buttons, merge buttons cannot be localized through messages. The same tools can be localized properly through the dialog messages options, when they are displayed in dialogs.
### Reproduction of the problem
1. Use the following [Dojo](http://dojo.telerik.com/ULEVi/3)
### Current behavior
The buttons are not localized according to the provided messages.
### Expected/desired behavior
The buttons are localized according to the provided messages.
### Environment
* **Kendo UI version:** 2017.2.504
* **Browser:** [all]
|
non_defect
|
missing spreadsheet toolbar buttons localization bug report alignment buttons freeze buttons merge buttons cannot be localized through messages the same tools can be localized properly through the dialog messages options when they are displayed in dialogs reproduction of the problem use the following current behavior the buttons are not localized according to the provided messages expected desired behavior the buttons are localized according to the provided messages environment kendo ui version browser
| 0
|
189,845
| 15,208,453,828
|
IssuesEvent
|
2021-02-17 02:41:26
|
jpmorganchase/regular-table
|
https://api.github.com/repos/jpmorganchase/regular-table
|
closed
|
How do I add a regular-table to a flex layout without breaking it?
|
documentation question
|
## Support Question
I'm creating a panel element that has a few widgets at the top and then a `<regular-table>` on the bottom. I want to use a flex-column layout for the panel, like so:

Following suggestions online, I tried setting the CSS to the following:
```css
.myapp-panel {
display: flex;
flex-flow: column;
}
.myapp-breadcrumbs {
height: 16px;
width: 100%;
}
.myapp-filter {
height: 16px;
width: 100%;
}
regular-table.myapp-regular-table {
flex: 1;
}
```
and then I ran into trouble. `regular-table` ships a stylesheet in its shadow root that applies `position: absolute` to its `:host` (ie the `regular-table` element itself). `position: absolute` does not seem to be compatible with having `display: flex` in the parent panel; these styles together fling the `regular-element` out of the visible area of the document.
Next I tried changing the `position` styling of the `regular-table` by adding more CSS:
```css
regular-table.myapp-regular-table {
flex: 1;
position: relative;
}
```
However, changing the `position: absolute` styling seems to completely break regular-table; on initial render it only displays one or two rows, and it seems to break in stranger ways as well:

(where the heck does that nub come from?)
Any suggestions on how I can achieve my desired layout (or reasonable facsimile) without breaking regular-table would be greatly appreciated.
For more context, here's a qualitative description of what I'm trying to achieve in terms of layout:
- The `.myapp-panel` element serves as the outer container for the other elements. Its height and width should each be able to grow to fill available space. Ideally, either of height/width should alternatively be fixed without breaking any child element behavior and/or CSS
- The `.myapp-breadcrumbs` element should have a fixed height, and it's width should grow to fit its parent's width
- The `.myapp-filter` elements should be the same (fixed height, grow width)
- The `.myapp-regular-table` element should grow in height to fill the remainder of the panel, and its width should be set by the usual regular-table autosizing logic, up to the width of the containing panel
|
1.0
|
How do I add a regular-table to a flex layout without breaking it? - ## Support Question
I'm creating a panel element that has a few widgets at the top and then a `<regular-table>` on the bottom. I want to use a flex-column layout for the panel, like so:

Following suggestions online, I tried setting the CSS to the following:
```css
.myapp-panel {
display: flex;
flex-flow: column;
}
.myapp-breadcrumbs {
height: 16px;
width: 100%;
}
.myapp-filter {
height: 16px;
width: 100%;
}
regular-table.myapp-regular-table {
flex: 1;
}
```
and then I ran into trouble. `regular-table` ships a stylesheet in its shadow root that applies `position: absolute` to its `:host` (ie the `regular-table` element itself). `position: absolute` does not seem to be compatible with having `display: flex` in the parent panel; these styles together fling the `regular-element` out of the visible area of the document.
Next I tried changing the `position` styling of the `regular-table` by adding more CSS:
```css
regular-table.myapp-regular-table {
flex: 1;
position: relative;
}
```
However, changing the `position: absolute` styling seems to completely break regular-table; on initial render it only displays one or two rows, and it seems to break in stranger ways as well:

(where the heck does that nub come from?)
Any suggestions on how I can achieve my desired layout (or reasonable facsimile) without breaking regular-table would be greatly appreciated.
For more context, here's a qualitative description of what I'm trying to achieve in terms of layout:
- The `.myapp-panel` element serves as the outer container for the other elements. Its height and width should each be able to grow to fill available space. Ideally, either of height/width should alternatively be fixed without breaking any child element behavior and/or CSS
- The `.myapp-breadcrumbs` element should have a fixed height, and it's width should grow to fit its parent's width
- The `.myapp-filter` elements should be the same (fixed height, grow width)
- The `.myapp-regular-table` element should grow in height to fill the remainder of the panel, and its width should be set by the usual regular-table autosizing logic, up to the width of the containing panel
|
non_defect
|
how do i add a regular table to a flex layout without breaking it support question i m creating a panel element that has a few widgets at the top and then a on the bottom i want to use a flex column layout for the panel like so following suggestions online i tried setting the css to the following css myapp panel display flex flex flow column myapp breadcrumbs height width myapp filter height width regular table myapp regular table flex and then i ran into trouble regular table ships a stylesheet in its shadow root that applies position absolute to its host ie the regular table element itself position absolute does not seem to be compatible with having display flex in the parent panel these styles together fling the regular element out of the visible area of the document next i tried changing the position styling of the regular table by adding more css css regular table myapp regular table flex position relative however changing the position absolute styling seems to completely break regular table on initial render it only displays one or two rows and it seems to break in stranger ways as well where the heck does that nub come from any suggestions on how i can achieve my desired layout or reasonable facsimile without breaking regular table would be greatly appreciated for more context here s a qualitative description of what i m trying to achieve in terms of layout the myapp panel element serves as the outer container for the other elements its height and width should each be able to grow to fill available space ideally either of height width should alternatively be fixed without breaking any child element behavior and or css the myapp breadcrumbs element should have a fixed height and it s width should grow to fit its parent s width the myapp filter elements should be the same fixed height grow width the myapp regular table element should grow in height to fill the remainder of the panel and its width should be set by the usual regular table autosizing logic up to the width of the containing panel
| 0
|
18,420
| 10,227,417,052
|
IssuesEvent
|
2019-08-16 20:43:12
|
pcrane70/ksa
|
https://api.github.com/repos/pcrane70/ksa
|
opened
|
CVE-2015-8858 (High) detected in uglify-js-1.2.6.tgz
|
security vulnerability
|
## CVE-2015-8858 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>uglify-js-1.2.6.tgz</b></p></summary>
<p>JavaScript parser and compressor/beautifier toolkit</p>
<p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-1.2.6.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-1.2.6.tgz</a></p>
<p>Path to dependency file: /ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json</p>
<p>Path to vulnerable library: /ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/node_modules/uglify-js/package.json</p>
<p>
Dependency Hierarchy:
- :x: **uglify-js-1.2.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/pcrane70/ksa/commit/b419088c3be90c5d97780b23bdd08ca48569d01b">b419088c3be90c5d97780b23bdd08ca48569d01b</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The uglify-js package before 2.6.0 for Node.js allows attackers to cause a denial of service (CPU consumption) via crafted input in a parse call, aka a "regular expression denial of service (ReDoS)."
<p>Publish Date: 2017-01-23
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-8858>CVE-2015-8858</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-8858">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-8858</a></p>
<p>Release Date: 2018-12-15</p>
<p>Fix Resolution: v2.6.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"uglify-js","packageVersion":"1.2.6","isTransitiveDependency":false,"dependencyTree":"uglify-js:1.2.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v2.6.0"}],"vulnerabilityIdentifier":"CVE-2015-8858","vulnerabilityDetails":"The uglify-js package before 2.6.0 for Node.js allows attackers to cause a denial of service (CPU consumption) via crafted input in a parse call, aka a \"regular expression denial of service (ReDoS).\"","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2015-8858","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2015-8858 (High) detected in uglify-js-1.2.6.tgz - ## CVE-2015-8858 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>uglify-js-1.2.6.tgz</b></p></summary>
<p>JavaScript parser and compressor/beautifier toolkit</p>
<p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-1.2.6.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-1.2.6.tgz</a></p>
<p>Path to dependency file: /ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json</p>
<p>Path to vulnerable library: /ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/node_modules/uglify-js/package.json</p>
<p>
Dependency Hierarchy:
- :x: **uglify-js-1.2.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/pcrane70/ksa/commit/b419088c3be90c5d97780b23bdd08ca48569d01b">b419088c3be90c5d97780b23bdd08ca48569d01b</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The uglify-js package before 2.6.0 for Node.js allows attackers to cause a denial of service (CPU consumption) via crafted input in a parse call, aka a "regular expression denial of service (ReDoS)."
<p>Publish Date: 2017-01-23
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-8858>CVE-2015-8858</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-8858">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-8858</a></p>
<p>Release Date: 2018-12-15</p>
<p>Fix Resolution: v2.6.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"uglify-js","packageVersion":"1.2.6","isTransitiveDependency":false,"dependencyTree":"uglify-js:1.2.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v2.6.0"}],"vulnerabilityIdentifier":"CVE-2015-8858","vulnerabilityDetails":"The uglify-js package before 2.6.0 for Node.js allows attackers to cause a denial of service (CPU consumption) via crafted input in a parse call, aka a \"regular expression denial of service (ReDoS).\"","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2015-8858","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve high detected in uglify js tgz cve high severity vulnerability vulnerable library uglify js tgz javascript parser and compressor beautifier toolkit library home page a href path to dependency file ksa ksa web root ksa web src main webapp rs bootstrap package json path to vulnerable library ksa ksa web root ksa web src main webapp rs bootstrap node modules uglify js package json dependency hierarchy x uglify js tgz vulnerable library found in head commit a href vulnerability details the uglify js package before for node js allows attackers to cause a denial of service cpu consumption via crafted input in a parse call aka a regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails the uglify js package before for node js allows attackers to cause a denial of service cpu consumption via crafted input in a parse call aka a regular expression denial of service redos vulnerabilityurl
| 0
|
70,818
| 3,343,870,992
|
IssuesEvent
|
2015-11-15 20:59:20
|
ianopolous/Peergos
|
https://api.github.com/repos/ianopolous/Peergos
|
opened
|
change colour of your files and folders that are shared with someone else (to orange?)
|
priority:Normal
|
Consider labelling them with "Shared with X friends"
|
1.0
|
change colour of your files and folders that are shared with someone else (to orange?) - Consider labelling them with "Shared with X friends"
|
non_defect
|
change colour of your files and folders that are shared with someone else to orange consider labelling them with shared with x friends
| 0
|
39,949
| 5,165,099,628
|
IssuesEvent
|
2017-01-17 12:41:30
|
nextcloud/nextcloud.com
|
https://api.github.com/repos/nextcloud/nextcloud.com
|
opened
|
Animation of logo a bit often
|
design
|
So the animation of the logo is nice, but doing it on every single page change _and_ on hovering it is a bit much. Can we do it so it only happens once? cc @Espina2 @nextcloud/designers otherwise the effect becomes more annoying instead of a nice detail :\
|
1.0
|
Animation of logo a bit often - So the animation of the logo is nice, but doing it on every single page change _and_ on hovering it is a bit much. Can we do it so it only happens once? cc @Espina2 @nextcloud/designers otherwise the effect becomes more annoying instead of a nice detail :\
|
non_defect
|
animation of logo a bit often so the animation of the logo is nice but doing it on every single page change and on hovering it is a bit much can we do it so it only happens once cc nextcloud designers otherwise the effect becomes more annoying instead of a nice detail
| 0
|
70,742
| 23,299,099,940
|
IssuesEvent
|
2022-08-07 03:12:15
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Dark Theme: colored text is hard to read in the browser
|
T-Defect
|
Closed because this is an unrelated issue.
|
1.0
|
Dark Theme: colored text is hard to read in the browser - Closed because this is an unrelated issue.
|
defect
|
dark theme colored text is hard to read in the browser closed because this is an unrelated issue
| 1
|
2,877
| 2,607,964,199
|
IssuesEvent
|
2015-02-26 00:41:34
|
chrsmithdemos/leveldb
|
https://api.github.com/repos/chrsmithdemos/leveldb
|
opened
|
Feature Request: more control over, better information about disk space utilization
|
auto-migrated Priority-Medium Type-Defect
|
```
In an embedded system or a handheld computer like an iPhone, space is at a
premium. It's often important to maintain strict control over database size
growth.
Please enhance leveldb to provide better runtime control over the amount of
disk space used and/or provide more complete, up-to-date information about how
much space is bing used, so that it's possible to remove entries to make more
space. The current size estimate functionality is often dramatically wrong,
even after closing and re-opening the database (I see values off by 75% or more
versus file system measurements).
Also, please provide a reliable compaction interface for mere mortals who don't
know how the Internals work. The current DB::Compact() method has the opposite
to the intended effect: the database grows dramatically each time it's invoked.
The only way to get actual compaction behaviour seems to be to close and
re-open the database.
```
-----
Original issue reported on code.google.com by `bgulanow...@gmail.com` on 3 Nov 2011 at 8:46
|
1.0
|
Feature Request: more control over, better information about disk space utilization - ```
In an embedded system or a handheld computer like an iPhone, space is at a
premium. It's often important to maintain strict control over database size
growth.
Please enhance leveldb to provide better runtime control over the amount of
disk space used and/or provide more complete, up-to-date information about how
much space is bing used, so that it's possible to remove entries to make more
space. The current size estimate functionality is often dramatically wrong,
even after closing and re-opening the database (I see values off by 75% or more
versus file system measurements).
Also, please provide a reliable compaction interface for mere mortals who don't
know how the Internals work. The current DB::Compact() method has the opposite
to the intended effect: the database grows dramatically each time it's invoked.
The only way to get actual compaction behaviour seems to be to close and
re-open the database.
```
-----
Original issue reported on code.google.com by `bgulanow...@gmail.com` on 3 Nov 2011 at 8:46
|
defect
|
feature request more control over better information about disk space utilization in an embedded system or a handheld computer like an iphone space is at a premium it s often important to maintain strict control over database size growth please enhance leveldb to provide better runtime control over the amount of disk space used and or provide more complete up to date information about how much space is bing used so that it s possible to remove entries to make more space the current size estimate functionality is often dramatically wrong even after closing and re opening the database i see values off by or more versus file system measurements also please provide a reliable compaction interface for mere mortals who don t know how the internals work the current db compact method has the opposite to the intended effect the database grows dramatically each time it s invoked the only way to get actual compaction behaviour seems to be to close and re open the database original issue reported on code google com by bgulanow gmail com on nov at
| 1
|
52,480
| 13,224,774,974
|
IssuesEvent
|
2020-08-17 19:49:20
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
[IceHive] ...is not working anymore (Trac #2290)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2290">https://code.icecube.wisc.edu/projects/icecube/ticket/2290</a>, reported by grenziand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2020-06-24T12:31:42",
"_ts": "1593001902142004",
"description": "The code below is working with less recent combo versions, not anymore with the newest ones\n\n\n{{{\n from icecube import IceHive\n\n singleRings = IceHive.RingLimits()\n singleRings.AddLimitPair(IceHive.LimitPair(-255., 255.))\n singleRings.AddLimitPair(IceHive.LimitPair(-272.7, 272.7))\n singleRings.AddLimitPair(IceHive.LimitPair(-165.8, 165.8))\n doubleRings = IceHive.RingLimits()\n doubleRings.AddLimitPair(IceHive.LimitPair(-70., 70.))\n doubleRings.AddLimitPair(IceHive.LimitPair(-131.5, 131.5))\n doubleRings.AddLimitPair(IceHive.LimitPair(-40.8, 40.8))\n tripleRings = IceHive.RingLimits()\n tripleRings.AddLimitPair(IceHive.LimitPair(-70., 70.))\n tripleRings.AddLimitPair(IceHive.LimitPair(-144.1, 144.1))\n tripleRings.AddLimitPair(IceHive.LimitPair(-124.7, 124.7))\n tripleRings.AddLimitPair(IceHive.LimitPair(-82.8, 82.8))\n singleVicinity = IceHive.RingLimits()\n singleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n singleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n doubleVicinity = IceHive.RingLimits()\n doubleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n doubleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n doubleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n tripleVicinity = IceHive.RingLimits()\n tripleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n tripleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n tripleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n tripleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n\n tray.AddModule(\"I3IceHive<I3RecoPulse>\",\"IceHiveSplit\",\n InputName=\"InIcePulses\",#\"OfflinePulses\",\n OutputName=\"MaskedInIcePulses\",#\"MaskedOfflinePulses\",\n SubEventStreamName=\"IceHiveSplit\",\n Multiplicity=4,\n TimeWindow=2000.*I3Units.ns,\n TimeCVMinus=300.*I3Units.ns,\n TimeCVPlus=300.*I3Units.ns,\n TimeCNMinus=200.*I3Units.ns,\n TimeCNPlus=200.*I3Units.ns,\n TimeStatic=200.*I3Units.ns,\n SingleDenseRingLimits=singleRings,\n DoubleDenseRingLimits=doubleRings,\n TripleDenseRingLimits=tripleRings,\n SingleDenseRingVicinity=singleVicinity,\n DoubleDenseRingVicinity=doubleVicinity,\n TripleDenseRingVicinity=tripleVicinity,\n ReadoutWindowMinus = 4000.*I3Units.ns,\n ReadoutWindowPlus = 6000.*I3Units.ns,\n SaveSplitCount=True)\n\n cleanVicinity = IceHive.RingLimits()\n cleanVicinity.AddLimitPair(IceHive.LimitPair(-70., 70.))\n cleanVicinity.AddLimitPair(IceHive.LimitPair(-70., 70.))\n \n #clean the output pulses in the P-frame\n tray.AddModule(\"I3HiveCleaning<I3RecoPulse>\",\"HiveClean\",\n InputName = \"MaskedInIcePulses\",#\"MaskedOfflinePulses\",\n OutputName = \"HC\"+\"MaskedInIcePulses\",#+\"MaskedOfflinePulses\",\n TimeStaticMinus= 600.*I3Units.ns,\n TimeStaticPlus= 600.*I3Units.ns,\n SingleDenseRingVicinity=cleanVicinity,\n DoubleDenseRingVicinity=cleanVicinity,\n TripleDenseRingVicinity=cleanVicinity,\n Stream = icetray.I3Frame.Physics,\n If = lambda f: True)\n}}}\n\nThe error I get in running this is \n\n\n{{{\nFATAL (IceHive): File '/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC2/IceHive/resources/data/IC86Topology.dat' seems not to have the correct format; please review the IceHive documentation (Hive-lib.cxx:451 in builder::Topology builder::ReadTopologyFromFile(const string&))\nERROR (I3Tray): Exception thrown while configuring module \"IceHiveSplit\". (I3Tray.cxx:384 in void I3Tray::Configure())\nIceHiveSplit (I3IceHive<I3RecoPulse>)\n...\nTraceback (most recent call last):\n File \"/home/grenzi/L3b/level3_Master.py\", line 259, in <module>\n tray.Execute()\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC2/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\nRuntimeError: File '/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC2/IceHive/resources/data/IC86Topology.dat' seems not to have the correct format; please review the IceHive documentation (in builder::Topology builder::ReadTopologyFromFile(const string&))\n}}}\n\nThe file IC86Topology.dat seems to have been the same for 5 years",
"reporter": "grenzi",
"cc": "",
"resolution": "fixed",
"time": "2019-05-21T13:55:50",
"component": "combo reconstruction",
"summary": "[IceHive] ...is not working anymore",
"priority": "blocker",
"keywords": "",
"milestone": "Autumnal Equinox 2020",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[IceHive] ...is not working anymore (Trac #2290) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2290">https://code.icecube.wisc.edu/projects/icecube/ticket/2290</a>, reported by grenziand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2020-06-24T12:31:42",
"_ts": "1593001902142004",
"description": "The code below is working with less recent combo versions, not anymore with the newest ones\n\n\n{{{\n from icecube import IceHive\n\n singleRings = IceHive.RingLimits()\n singleRings.AddLimitPair(IceHive.LimitPair(-255., 255.))\n singleRings.AddLimitPair(IceHive.LimitPair(-272.7, 272.7))\n singleRings.AddLimitPair(IceHive.LimitPair(-165.8, 165.8))\n doubleRings = IceHive.RingLimits()\n doubleRings.AddLimitPair(IceHive.LimitPair(-70., 70.))\n doubleRings.AddLimitPair(IceHive.LimitPair(-131.5, 131.5))\n doubleRings.AddLimitPair(IceHive.LimitPair(-40.8, 40.8))\n tripleRings = IceHive.RingLimits()\n tripleRings.AddLimitPair(IceHive.LimitPair(-70., 70.))\n tripleRings.AddLimitPair(IceHive.LimitPair(-144.1, 144.1))\n tripleRings.AddLimitPair(IceHive.LimitPair(-124.7, 124.7))\n tripleRings.AddLimitPair(IceHive.LimitPair(-82.8, 82.8))\n singleVicinity = IceHive.RingLimits()\n singleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n singleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n doubleVicinity = IceHive.RingLimits()\n doubleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n doubleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n doubleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n tripleVicinity = IceHive.RingLimits()\n tripleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n tripleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n tripleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n tripleVicinity.AddLimitPair(IceHive.LimitPair(-100.,100.))\n\n tray.AddModule(\"I3IceHive<I3RecoPulse>\",\"IceHiveSplit\",\n InputName=\"InIcePulses\",#\"OfflinePulses\",\n OutputName=\"MaskedInIcePulses\",#\"MaskedOfflinePulses\",\n SubEventStreamName=\"IceHiveSplit\",\n Multiplicity=4,\n TimeWindow=2000.*I3Units.ns,\n TimeCVMinus=300.*I3Units.ns,\n TimeCVPlus=300.*I3Units.ns,\n TimeCNMinus=200.*I3Units.ns,\n TimeCNPlus=200.*I3Units.ns,\n TimeStatic=200.*I3Units.ns,\n SingleDenseRingLimits=singleRings,\n DoubleDenseRingLimits=doubleRings,\n TripleDenseRingLimits=tripleRings,\n SingleDenseRingVicinity=singleVicinity,\n DoubleDenseRingVicinity=doubleVicinity,\n TripleDenseRingVicinity=tripleVicinity,\n ReadoutWindowMinus = 4000.*I3Units.ns,\n ReadoutWindowPlus = 6000.*I3Units.ns,\n SaveSplitCount=True)\n\n cleanVicinity = IceHive.RingLimits()\n cleanVicinity.AddLimitPair(IceHive.LimitPair(-70., 70.))\n cleanVicinity.AddLimitPair(IceHive.LimitPair(-70., 70.))\n \n #clean the output pulses in the P-frame\n tray.AddModule(\"I3HiveCleaning<I3RecoPulse>\",\"HiveClean\",\n InputName = \"MaskedInIcePulses\",#\"MaskedOfflinePulses\",\n OutputName = \"HC\"+\"MaskedInIcePulses\",#+\"MaskedOfflinePulses\",\n TimeStaticMinus= 600.*I3Units.ns,\n TimeStaticPlus= 600.*I3Units.ns,\n SingleDenseRingVicinity=cleanVicinity,\n DoubleDenseRingVicinity=cleanVicinity,\n TripleDenseRingVicinity=cleanVicinity,\n Stream = icetray.I3Frame.Physics,\n If = lambda f: True)\n}}}\n\nThe error I get in running this is \n\n\n{{{\nFATAL (IceHive): File '/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC2/IceHive/resources/data/IC86Topology.dat' seems not to have the correct format; please review the IceHive documentation (Hive-lib.cxx:451 in builder::Topology builder::ReadTopologyFromFile(const string&))\nERROR (I3Tray): Exception thrown while configuring module \"IceHiveSplit\". (I3Tray.cxx:384 in void I3Tray::Configure())\nIceHiveSplit (I3IceHive<I3RecoPulse>)\n...\nTraceback (most recent call last):\n File \"/home/grenzi/L3b/level3_Master.py\", line 259, in <module>\n tray.Execute()\n File \"/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC2/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\nRuntimeError: File '/cvmfs/icecube.opensciencegrid.org/py2-v3.1.1/RHEL_7_x86_64/metaprojects/combo/V00-00-00-RC2/IceHive/resources/data/IC86Topology.dat' seems not to have the correct format; please review the IceHive documentation (in builder::Topology builder::ReadTopologyFromFile(const string&))\n}}}\n\nThe file IC86Topology.dat seems to have been the same for 5 years",
"reporter": "grenzi",
"cc": "",
"resolution": "fixed",
"time": "2019-05-21T13:55:50",
"component": "combo reconstruction",
"summary": "[IceHive] ...is not working anymore",
"priority": "blocker",
"keywords": "",
"milestone": "Autumnal Equinox 2020",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
defect
|
is not working anymore trac migrated from json status closed changetime ts description the code below is working with less recent combo versions not anymore with the newest ones n n n n from icecube import icehive n n singlerings icehive ringlimits n singlerings addlimitpair icehive limitpair n singlerings addlimitpair icehive limitpair n singlerings addlimitpair icehive limitpair n doublerings icehive ringlimits n doublerings addlimitpair icehive limitpair n doublerings addlimitpair icehive limitpair n doublerings addlimitpair icehive limitpair n triplerings icehive ringlimits n triplerings addlimitpair icehive limitpair n triplerings addlimitpair icehive limitpair n triplerings addlimitpair icehive limitpair n triplerings addlimitpair icehive limitpair n singlevicinity icehive ringlimits n singlevicinity addlimitpair icehive limitpair n singlevicinity addlimitpair icehive limitpair n doublevicinity icehive ringlimits n doublevicinity addlimitpair icehive limitpair n doublevicinity addlimitpair icehive limitpair n doublevicinity addlimitpair icehive limitpair n triplevicinity icehive ringlimits n triplevicinity addlimitpair icehive limitpair n triplevicinity addlimitpair icehive limitpair n triplevicinity addlimitpair icehive limitpair n triplevicinity addlimitpair icehive limitpair n n tray addmodule icehivesplit n inputname inicepulses offlinepulses n outputname maskedinicepulses maskedofflinepulses n subeventstreamname icehivesplit n multiplicity n timewindow ns n timecvminus ns n timecvplus ns n timecnminus ns n timecnplus ns n timestatic ns n singledenseringlimits singlerings n doubledenseringlimits doublerings n tripledenseringlimits triplerings n singledenseringvicinity singlevicinity n doubledenseringvicinity doublevicinity n tripledenseringvicinity triplevicinity n readoutwindowminus ns n readoutwindowplus ns n savesplitcount true n n cleanvicinity icehive ringlimits n cleanvicinity addlimitpair icehive limitpair n cleanvicinity addlimitpair icehive limitpair n n clean the output pulses in the p frame n tray addmodule hiveclean n inputname maskedinicepulses maskedofflinepulses n outputname hc maskedinicepulses maskedofflinepulses n timestaticminus ns n timestaticplus ns n singledenseringvicinity cleanvicinity n doubledenseringvicinity cleanvicinity n tripledenseringvicinity cleanvicinity n stream icetray physics n if lambda f true n n nthe error i get in running this is n n n nfatal icehive file cvmfs icecube opensciencegrid org rhel metaprojects combo icehive resources data dat seems not to have the correct format please review the icehive documentation hive lib cxx in builder topology builder readtopologyfromfile const string nerror exception thrown while configuring module icehivesplit cxx in void configure nicehivesplit n ntraceback most recent call last n file home grenzi master py line in n tray execute n file cvmfs icecube opensciencegrid org rhel metaprojects combo lib py line in execute n super self execute nruntimeerror file cvmfs icecube opensciencegrid org rhel metaprojects combo icehive resources data dat seems not to have the correct format please review the icehive documentation in builder topology builder readtopologyfromfile const string n n nthe file dat seems to have been the same for years reporter grenzi cc resolution fixed time component combo reconstruction summary is not working anymore priority blocker keywords milestone autumnal equinox owner jvansanten type defect
| 1
|
63,738
| 17,876,840,863
|
IssuesEvent
|
2021-09-07 05:48:35
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
opened
|
ExtraNodesetGenerator can’t find correct node using cord
|
T: defect P: normal
|
## Bug Description
<!--A clear and concise description of the problem (Note: A missing feature is not a bug).-->
When the mesh dimension is lower than spatial dimension, say, mesh dimension is 2, spatial dimension is 3. The code will only copy 2 coordinate components of “c” into “p”, which is not correct, since c[2] may not be 0.
https://github.com/idaholab/moose/blob/e3c48025ba7d3169cd5de39de0fcfd25ca443aca/framework/src/meshgenerators/ExtraNodesetGenerator.C#L114
## Steps to Reproduce
<!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)-->
## Impact
<!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
Yes
|
1.0
|
ExtraNodesetGenerator can’t find correct node using cord - ## Bug Description
<!--A clear and concise description of the problem (Note: A missing feature is not a bug).-->
When the mesh dimension is lower than spatial dimension, say, mesh dimension is 2, spatial dimension is 3. The code will only copy 2 coordinate components of “c” into “p”, which is not correct, since c[2] may not be 0.
https://github.com/idaholab/moose/blob/e3c48025ba7d3169cd5de39de0fcfd25ca443aca/framework/src/meshgenerators/ExtraNodesetGenerator.C#L114
## Steps to Reproduce
<!--Steps to reproduce the behavior (input file, or modifications to an existing input file, etc.)-->
## Impact
<!--Does this prevent you from getting your work done, or is it more of an annoyance?-->
Yes
|
defect
|
extranodesetgenerator can’t find correct node using cord bug description when the mesh dimension is lower than spatial dimension say mesh dimension is spatial dimension is the code will only copy coordinate components of “c” into “p” which is not correct since c may not be steps to reproduce impact yes
| 1
|
154,035
| 24,232,105,531
|
IssuesEvent
|
2022-09-26 19:15:56
|
CDCgov/prime-simplereport
|
https://api.github.com/repos/CDCgov/prime-simplereport
|
closed
|
"All facilities" breadcrumb naming not accessible
|
Content design Accessibility Critical Facility settings
|
## Area or Component
Facility form
## Quick Summary
The breadcrumb back to look at the list of facilities is labelled "all facilities", but this name is not accessible enough to screenreader users without the back arrow context (the arrow is aria-hidden.)
### How to Replicate
1. Go to settings page
2. Choose "manage facilities"
3. Click on a facility
4. Inspect the "all facilities" link
## Accessibility Violation
Name: The element's name is missing or incorrect: https://cdc-devtools.dequecloud.com/rule-help/en/aria-name-missing-incorrect
## Action Requested
Change the name of the link to something more descriptive (e.g. with an action), like "View all facilities" or "back to all facilities".
## Screenshots (optional)
<img width="982" alt="Screen Shot 2022-07-27 at 4 26 26 PM" src="https://user-images.githubusercontent.com/80282552/181389412-e36c4694-d01c-4de6-802c-0c0e6d9e5919.png">
## Original Deque Issue
https://docs.google.com/spreadsheets/d/1b6xjubAN5Uw11DV4_B5zvSu7HeDI6Z-DpV4DMcToqAY/edit#gid=1346058064&range=8:8
|
1.0
|
"All facilities" breadcrumb naming not accessible - ## Area or Component
Facility form
## Quick Summary
The breadcrumb back to look at the list of facilities is labelled "all facilities", but this name is not accessible enough to screenreader users without the back arrow context (the arrow is aria-hidden.)
### How to Replicate
1. Go to settings page
2. Choose "manage facilities"
3. Click on a facility
4. Inspect the "all facilities" link
## Accessibility Violation
Name: The element's name is missing or incorrect: https://cdc-devtools.dequecloud.com/rule-help/en/aria-name-missing-incorrect
## Action Requested
Change the name of the link to something more descriptive (e.g. with an action), like "View all facilities" or "back to all facilities".
## Screenshots (optional)
<img width="982" alt="Screen Shot 2022-07-27 at 4 26 26 PM" src="https://user-images.githubusercontent.com/80282552/181389412-e36c4694-d01c-4de6-802c-0c0e6d9e5919.png">
## Original Deque Issue
https://docs.google.com/spreadsheets/d/1b6xjubAN5Uw11DV4_B5zvSu7HeDI6Z-DpV4DMcToqAY/edit#gid=1346058064&range=8:8
|
non_defect
|
all facilities breadcrumb naming not accessible area or component facility form quick summary the breadcrumb back to look at the list of facilities is labelled all facilities but this name is not accessible enough to screenreader users without the back arrow context the arrow is aria hidden how to replicate go to settings page choose manage facilities click on a facility inspect the all facilities link accessibility violation name the element s name is missing or incorrect action requested change the name of the link to something more descriptive e g with an action like view all facilities or back to all facilities screenshots optional img width alt screen shot at pm src original deque issue
| 0
|
8,206
| 2,611,471,008
|
IssuesEvent
|
2015-02-27 05:15:48
|
chrsmith/hedgewars
|
https://api.github.com/repos/chrsmith/hedgewars
|
closed
|
nemo always wins :(
|
auto-migrated Priority-Critical Type-Defect Usability
|
```
What steps will reproduce the problem?
1. challenge nemo to a game
2. play against him
3. lose
What is the expected output? What do you see instead?
I lose, but I want to win :(
Please use labels and text to provide additional information.
Makes me a sad panda.
XD
```
Original issue reported on code.google.com by `sheepyluva` on 12 Apr 2011 at 8:35
|
1.0
|
nemo always wins :( - ```
What steps will reproduce the problem?
1. challenge nemo to a game
2. play against him
3. lose
What is the expected output? What do you see instead?
I lose, but I want to win :(
Please use labels and text to provide additional information.
Makes me a sad panda.
XD
```
Original issue reported on code.google.com by `sheepyluva` on 12 Apr 2011 at 8:35
|
defect
|
nemo always wins what steps will reproduce the problem challenge nemo to a game play against him lose what is the expected output what do you see instead i lose but i want to win please use labels and text to provide additional information makes me a sad panda xd original issue reported on code google com by sheepyluva on apr at
| 1
|
22,262
| 3,619,690,517
|
IssuesEvent
|
2016-02-08 16:56:52
|
miracle091/transmission-remote-dotnet
|
https://api.github.com/repos/miracle091/transmission-remote-dotnet
|
closed
|
Disconned in a web request
|
Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
I have added a public torrent and transmission give me this problem:
Error Say: exception in a web request
What version of the products are you using?
OS: Vista
Transmission: 3.24 (build 3)
Remote: 2.42 (rev XXX)
I lost the connection, i try reconnect and or last only few seconds or isen´t
possible connect. This is not for transmission-daemon because i can connect
with others Guis at the same time. I have this problem today(15/11). I attach
the log file and other file from other day (29/10) for other problem (this i
think is not as important because only was that day), but i attach too...
I attach too 2 images of the program messagge, one is previous to lost
connection an say "Failed request 3:exception in a web request", the other is
later and say "Disconnected. Exceed maximun number of failed request".
All this is because (i think) there is a tracker that say response code
401(Unauthorized).
```
Original issue reported on code.google.com by `H3rm...@gmail.com` on 15 Nov 2011 at 6:28
Attachments:
* [trdcrash_20111114_171428.log](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-421/comment-0/trdcrash_20111114_171428.log)
* [trdcrash_20111029_160701.log](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-421/comment-0/trdcrash_20111029_160701.log)
* [disconnected-previous.jpg](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-421/comment-0/disconnected-previous.jpg)
* [disconnected.jpg](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-421/comment-0/disconnected.jpg)
|
1.0
|
Disconned in a web request - ```
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
I have added a public torrent and transmission give me this problem:
Error Say: exception in a web request
What version of the products are you using?
OS: Vista
Transmission: 3.24 (build 3)
Remote: 2.42 (rev XXX)
I lost the connection, i try reconnect and or last only few seconds or isen´t
possible connect. This is not for transmission-daemon because i can connect
with others Guis at the same time. I have this problem today(15/11). I attach
the log file and other file from other day (29/10) for other problem (this i
think is not as important because only was that day), but i attach too...
I attach too 2 images of the program messagge, one is previous to lost
connection an say "Failed request 3:exception in a web request", the other is
later and say "Disconnected. Exceed maximun number of failed request".
All this is because (i think) there is a tracker that say response code
401(Unauthorized).
```
Original issue reported on code.google.com by `H3rm...@gmail.com` on 15 Nov 2011 at 6:28
Attachments:
* [trdcrash_20111114_171428.log](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-421/comment-0/trdcrash_20111114_171428.log)
* [trdcrash_20111029_160701.log](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-421/comment-0/trdcrash_20111029_160701.log)
* [disconnected-previous.jpg](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-421/comment-0/disconnected-previous.jpg)
* [disconnected.jpg](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-421/comment-0/disconnected.jpg)
|
defect
|
disconned in a web request what steps will reproduce the problem what is the expected output what do you see instead i have added a public torrent and transmission give me this problem error say exception in a web request what version of the products are you using os vista transmission build remote rev xxx i lost the connection i try reconnect and or last only few seconds or isen´t possible connect this is not for transmission daemon because i can connect with others guis at the same time i have this problem today i attach the log file and other file from other day for other problem this i think is not as important because only was that day but i attach too i attach too images of the program messagge one is previous to lost connection an say failed request exception in a web request the other is later and say disconnected exceed maximun number of failed request all this is because i think there is a tracker that say response code unauthorized original issue reported on code google com by gmail com on nov at attachments
| 1
|
319,353
| 9,742,626,395
|
IssuesEvent
|
2019-06-02 18:41:41
|
teambit/bit
|
https://api.github.com/repos/teambit/bit
|
closed
|
Add machine name flag to bit login command
|
area/login priority/low type/feature
|
## Expected Behavior
Ability to specify the machine name in the `bit login` command to be able to create a token for a specific machine.
That will prevent a case when someone generates a token in his machine to use in CI system then after some time he re-login and the old token revoked.
|
1.0
|
Add machine name flag to bit login command - ## Expected Behavior
Ability to specify the machine name in the `bit login` command to be able to create a token for a specific machine.
That will prevent a case when someone generates a token in his machine to use in CI system then after some time he re-login and the old token revoked.
|
non_defect
|
add machine name flag to bit login command expected behavior ability to specify the machine name in the bit login command to be able to create a token for a specific machine that will prevent a case when someone generates a token in his machine to use in ci system then after some time he re login and the old token revoked
| 0
|
4,898
| 2,610,160,214
|
IssuesEvent
|
2015-02-26 18:50:51
|
chrsmith/republic-at-war
|
https://api.github.com/repos/chrsmith/republic-at-war
|
closed
|
Map Issue
|
auto-migrated Priority-Medium Type-Defect
|
```
Geonosis
Rescale Planet rings
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 31 Jan 2011 at 2:21
|
1.0
|
Map Issue - ```
Geonosis
Rescale Planet rings
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 31 Jan 2011 at 2:21
|
defect
|
map issue geonosis rescale planet rings original issue reported on code google com by gmail com on jan at
| 1
|
5,965
| 2,610,218,666
|
IssuesEvent
|
2015-02-26 19:09:32
|
chrsmith/somefinders
|
https://api.github.com/repos/chrsmith/somefinders
|
opened
|
Mss32 dll для call of duty modern warfare 3 .pdf
|
auto-migrated Priority-Medium Type-Defect
|
```
'''Вилий Брагин'''
День добрый никак не могу найти .Mss32 dll для
call of duty modern warfare 3 .pdf. где то видел уже
'''Ананий Туров'''
Вот хороший сайт где можно скачать
http://bit.ly/Hx2RuI
'''Велимир Константинов'''
Спасибо вроде то но просит телефон вводить
'''Анри Филиппов'''
Не это не влияет на баланс
'''Артём Зайцев'''
Неа все ок у меня ничего не списало
Информация о файле: Mss32 dll для call of duty modern
warfare 3 .pdf
Загружен: В этом месяце
Скачан раз: 327
Рейтинг: 1270
Средняя скорость скачивания: 114
Похожих файлов: 33
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 3:28
|
1.0
|
Mss32 dll для call of duty modern warfare 3 .pdf - ```
'''Вилий Брагин'''
День добрый никак не могу найти .Mss32 dll для
call of duty modern warfare 3 .pdf. где то видел уже
'''Ананий Туров'''
Вот хороший сайт где можно скачать
http://bit.ly/Hx2RuI
'''Велимир Константинов'''
Спасибо вроде то но просит телефон вводить
'''Анри Филиппов'''
Не это не влияет на баланс
'''Артём Зайцев'''
Неа все ок у меня ничего не списало
Информация о файле: Mss32 dll для call of duty modern
warfare 3 .pdf
Загружен: В этом месяце
Скачан раз: 327
Рейтинг: 1270
Средняя скорость скачивания: 114
Похожих файлов: 33
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 3:28
|
defect
|
dll для call of duty modern warfare pdf вилий брагин день добрый никак не могу найти dll для call of duty modern warfare pdf где то видел уже ананий туров вот хороший сайт где можно скачать велимир константинов спасибо вроде то но просит телефон вводить анри филиппов не это не влияет на баланс артём зайцев неа все ок у меня ничего не списало информация о файле dll для call of duty modern warfare pdf загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
| 1
|
461,519
| 13,231,592,424
|
IssuesEvent
|
2020-08-18 11:57:38
|
mozilla/addons-server
|
https://api.github.com/repos/mozilla/addons-server
|
closed
|
Multiple Homepage shelves can occupy the same shelf Position at the same time
|
component: admin tools priority: p4
|
Follow up for #14874
### Describe the problem and steps to reproduce it:
1. Open the [Homepage shelves](https://addons-internal.dev.mozaws.net/en-US/admin/models/discovery/homepageshelves/add/) admin tool with an admin account
2. Select one of the available shelves from the list and set a Position
3. Save this homepage shelf and add another one, selecting the same Position as for the previous shelf
4. Save the second homepage shelf
### What happened?
The second shelf is successfully saved
### What did you expect to happen?
The form should not allow for the enabling of multiple homepage shelves with the same Position
example of multiple [homepage shelves](https://addons-internal.dev.mozaws.net/en-US/admin/models/discovery/homepageshelves/), all with Position 1

|
1.0
|
Multiple Homepage shelves can occupy the same shelf Position at the same time - Follow up for #14874
### Describe the problem and steps to reproduce it:
1. Open the [Homepage shelves](https://addons-internal.dev.mozaws.net/en-US/admin/models/discovery/homepageshelves/add/) admin tool with an admin account
2. Select one of the available shelves from the list and set a Position
3. Save this homepage shelf and add another one, selecting the same Position as for the previous shelf
4. Save the second homepage shelf
### What happened?
The second shelf is successfully saved
### What did you expect to happen?
The form should not allow for the enabling of multiple homepage shelves with the same Position
example of multiple [homepage shelves](https://addons-internal.dev.mozaws.net/en-US/admin/models/discovery/homepageshelves/), all with Position 1

|
non_defect
|
multiple homepage shelves can occupy the same shelf position at the same time follow up for describe the problem and steps to reproduce it open the admin tool with an admin account select one of the available shelves from the list and set a position save this homepage shelf and add another one selecting the same position as for the previous shelf save the second homepage shelf what happened the second shelf is successfully saved what did you expect to happen the form should not allow for the enabling of multiple homepage shelves with the same position example of multiple all with position
| 0
|
4,530
| 3,391,916,276
|
IssuesEvent
|
2015-11-30 17:20:35
|
htacg/tidy-html5
|
https://api.github.com/repos/htacg/tidy-html5
|
closed
|
exe don't work
|
Build/Install/Distribute Technical Support
|
I'm trying to compile tidy library on Windows with MinGW .
When I try to compile with Eclipse and Cmake I have an error with this line:
if ( !GetFileSizeEx( fp, (LARGE_INTEGER*)&fin->size )
In mappedio.c line 207. GetFileSizeEx is unknown.
If I comment this code, I can compile without errors the library, but when I try to execute the exe it gets locked or freeze. There isn't any output.
Anyone can help me with this?
Thank you very much.
|
1.0
|
exe don't work - I'm trying to compile tidy library on Windows with MinGW .
When I try to compile with Eclipse and Cmake I have an error with this line:
if ( !GetFileSizeEx( fp, (LARGE_INTEGER*)&fin->size )
In mappedio.c line 207. GetFileSizeEx is unknown.
If I comment this code, I can compile without errors the library, but when I try to execute the exe it gets locked or freeze. There isn't any output.
Anyone can help me with this?
Thank you very much.
|
non_defect
|
exe don t work i m trying to compile tidy library on windows with mingw when i try to compile with eclipse and cmake i have an error with this line if getfilesizeex fp large integer fin size in mappedio c line getfilesizeex is unknown if i comment this code i can compile without errors the library but when i try to execute the exe it gets locked or freeze there isn t any output anyone can help me with this thank you very much
| 0
|
12,018
| 18,770,912,658
|
IssuesEvent
|
2021-11-06 20:28:22
|
renovatebot/renovate
|
https://api.github.com/repos/renovatebot/renovate
|
opened
|
Optimize Ubuntu versioning
|
type:feature status:requirements priority-5-triage versioning:ubuntu
|
### What would you like Renovate to be able to do?
Quote from @viceice:
> I think we need a new issue to track Ubuntu versioning optimization.
We recently merged a "hacky" PR #12528 which makes it so we consider Ubuntu `22.04` unstable (right now this is released a "early testing" build, not the proper stable release).
### If you have any ideas on how this should be implemented, please tell us here.
How about we create a `.json` file and put the Ubuntu release schedule in it?
I got the idea from the Node.js schedule file that we used to mirror in the Renovate repository. [^node-schedule]
Node.js schedule JSON example:
```json
"v16": {
"start": "2021-04-20",
"lts": "2021-10-26",
"maintenance": "2022-10-18",
"end": "2024-04-30",
"codename": "Gallium"
},
```
Possible JSON format for the Ubuntu versioning:
```json
"v20.04": {
"initial-release": "2020-04-23",
"initial-lts-release": "2020-04-23",
"end-of-normal-support": "2025-04-23",
"end-of-extended-security-maintenance": "2030-04-23",
"codename": "Focal Fossa"
},
```
I haven't found a schedule like this from Ubuntu themselves, at least not in a form that a computer can understand/parse. I did find a list of releases on the Ubuntu wiki though. [^ubuntu-wiki]
[^node-schedule]: https://github.com/nodejs/Release/blob/main/schedule.json
[^ubuntu-wiki]: https://wiki.ubuntu.com/Releases
### Is this a feature you are interested in implementing yourself?
No
|
1.0
|
Optimize Ubuntu versioning - ### What would you like Renovate to be able to do?
Quote from @viceice:
> I think we need a new issue to track Ubuntu versioning optimization.
We recently merged a "hacky" PR #12528 which makes it so we consider Ubuntu `22.04` unstable (right now this is released a "early testing" build, not the proper stable release).
### If you have any ideas on how this should be implemented, please tell us here.
How about we create a `.json` file and put the Ubuntu release schedule in it?
I got the idea from the Node.js schedule file that we used to mirror in the Renovate repository. [^node-schedule]
Node.js schedule JSON example:
```json
"v16": {
"start": "2021-04-20",
"lts": "2021-10-26",
"maintenance": "2022-10-18",
"end": "2024-04-30",
"codename": "Gallium"
},
```
Possible JSON format for the Ubuntu versioning:
```json
"v20.04": {
"initial-release": "2020-04-23",
"initial-lts-release": "2020-04-23",
"end-of-normal-support": "2025-04-23",
"end-of-extended-security-maintenance": "2030-04-23",
"codename": "Focal Fossa"
},
```
I haven't found a schedule like this from Ubuntu themselves, at least not in a form that a computer can understand/parse. I did find a list of releases on the Ubuntu wiki though. [^ubuntu-wiki]
[^node-schedule]: https://github.com/nodejs/Release/blob/main/schedule.json
[^ubuntu-wiki]: https://wiki.ubuntu.com/Releases
### Is this a feature you are interested in implementing yourself?
No
|
non_defect
|
optimize ubuntu versioning what would you like renovate to be able to do quote from viceice i think we need a new issue to track ubuntu versioning optimization we recently merged a hacky pr which makes it so we consider ubuntu unstable right now this is released a early testing build not the proper stable release if you have any ideas on how this should be implemented please tell us here how about we create a json file and put the ubuntu release schedule in it i got the idea from the node js schedule file that we used to mirror in the renovate repository node js schedule json example json start lts maintenance end codename gallium possible json format for the ubuntu versioning json initial release initial lts release end of normal support end of extended security maintenance codename focal fossa i haven t found a schedule like this from ubuntu themselves at least not in a form that a computer can understand parse i did find a list of releases on the ubuntu wiki though is this a feature you are interested in implementing yourself no
| 0
|
220,754
| 16,984,841,504
|
IssuesEvent
|
2021-06-30 13:21:03
|
usnistgov/OSCAL
|
https://api.github.com/repos/usnistgov/OSCAL
|
opened
|
Tutorial: SSP System Characteristics - Part 2 of 3
|
Scope: Documentation User Story enhancement
|
# User Story:
As an OSCAL SSP author, I need to understand how to document system characteristics so that I can create well-formed OSCAL SSPs with the appropriate system attributes specified.
## Goals:
This tutorial (part 2 of 3) should provide in-depth documentation showing how the following system characteristics properties are set and any constraints:
- security-sensitivity-level
- system-information
- security-impact-level
- status
- any associated props and links
## Dependencies:
Issue #975.
## Acceptance Criteria
- [ ] All [OSCAL website](https://pages.nist.gov/OSCAL) and readme documentation affected by the changes in this issue have been updated. Changes to the OSCAL website can be made in the docs/content directory of your branch.
- [ ] A Pull Request (PR) is submitted that fully addresses the goals of this User Story. This issue is referenced in the PR.
- [ ] The CI-CD build process runs without any reported errors on the PR. This can be confirmed by reviewing that all checks have passed in the PR.
{The items above are general acceptance criteria for all User Stories. Please describe anything else that must be completed for this issue to be considered resolved.}
|
1.0
|
Tutorial: SSP System Characteristics - Part 2 of 3 - # User Story:
As an OSCAL SSP author, I need to understand how to document system characteristics so that I can create well-formed OSCAL SSPs with the appropriate system attributes specified.
## Goals:
This tutorial (part 2 of 3) should provide in-depth documentation showing how the following system characteristics properties are set and any constraints:
- security-sensitivity-level
- system-information
- security-impact-level
- status
- any associated props and links
## Dependencies:
Issue #975.
## Acceptance Criteria
- [ ] All [OSCAL website](https://pages.nist.gov/OSCAL) and readme documentation affected by the changes in this issue have been updated. Changes to the OSCAL website can be made in the docs/content directory of your branch.
- [ ] A Pull Request (PR) is submitted that fully addresses the goals of this User Story. This issue is referenced in the PR.
- [ ] The CI-CD build process runs without any reported errors on the PR. This can be confirmed by reviewing that all checks have passed in the PR.
{The items above are general acceptance criteria for all User Stories. Please describe anything else that must be completed for this issue to be considered resolved.}
|
non_defect
|
tutorial ssp system characteristics part of user story as an oscal ssp author i need to understand how to document system characteristics so that i can create well formed oscal ssps with the appropriate system attributes specified goals this tutorial part of should provide in depth documentation showing how the following system characteristics properties are set and any constraints security sensitivity level system information security impact level status any associated props and links dependencies issue acceptance criteria all and readme documentation affected by the changes in this issue have been updated changes to the oscal website can be made in the docs content directory of your branch a pull request pr is submitted that fully addresses the goals of this user story this issue is referenced in the pr the ci cd build process runs without any reported errors on the pr this can be confirmed by reviewing that all checks have passed in the pr the items above are general acceptance criteria for all user stories please describe anything else that must be completed for this issue to be considered resolved
| 0
|
34,925
| 12,306,531,019
|
IssuesEvent
|
2020-05-12 01:44:38
|
allofphysicsgraph/proofofconcept
|
https://api.github.com/repos/allofphysicsgraph/proofofconcept
|
opened
|
prevent malicious users from using up services
|
enhancement security web frontend
|
use fail2ban to limit resource use associated with both SSH and Nginx
https://www.digitalocean.com/community/tutorials/how-to-protect-ssh-with-fail2ban-on-ubuntu-14-04
https://www.digitalocean.com/community/tutorials/how-to-protect-an-nginx-server-with-fail2ban-on-ubuntu-14-04
https://www.google.com/search?&q=fail2ban+ubuntu+18.04
|
True
|
prevent malicious users from using up services - use fail2ban to limit resource use associated with both SSH and Nginx
https://www.digitalocean.com/community/tutorials/how-to-protect-ssh-with-fail2ban-on-ubuntu-14-04
https://www.digitalocean.com/community/tutorials/how-to-protect-an-nginx-server-with-fail2ban-on-ubuntu-14-04
https://www.google.com/search?&q=fail2ban+ubuntu+18.04
|
non_defect
|
prevent malicious users from using up services use to limit resource use associated with both ssh and nginx
| 0
|
433,284
| 30,321,458,481
|
IssuesEvent
|
2023-07-10 19:35:10
|
apache/beam
|
https://api.github.com/repos/apache/beam
|
opened
|
[Task][SDK Authoring Guide]: Document source and sink urns.
|
protobuf task documentation
|
### What needs to happen?
The beam protos don't describe the urns "beam:runner:source:v1" "beam:runner:sink:v1" which are necessary for the implementation of a Portable SDK. While there are now several examples, it would be good to have them be concretely described in the protocol.
### Issue Priority
Priority: 3 (nice-to-have improvement)
### Issue Components
- [ ] Component: Python SDK
- [ ] Component: Java SDK
- [ ] Component: Go SDK
- [ ] Component: Typescript SDK
- [ ] Component: IO connector
- [ ] Component: Beam examples
- [ ] Component: Beam playground
- [ ] Component: Beam katas
- [ ] Component: Website
- [ ] Component: Spark Runner
- [ ] Component: Flink Runner
- [ ] Component: Samza Runner
- [ ] Component: Twister2 Runner
- [ ] Component: Hazelcast Jet Runner
- [ ] Component: Google Cloud Dataflow Runner
|
1.0
|
[Task][SDK Authoring Guide]: Document source and sink urns. - ### What needs to happen?
The beam protos don't describe the urns "beam:runner:source:v1" "beam:runner:sink:v1" which are necessary for the implementation of a Portable SDK. While there are now several examples, it would be good to have them be concretely described in the protocol.
### Issue Priority
Priority: 3 (nice-to-have improvement)
### Issue Components
- [ ] Component: Python SDK
- [ ] Component: Java SDK
- [ ] Component: Go SDK
- [ ] Component: Typescript SDK
- [ ] Component: IO connector
- [ ] Component: Beam examples
- [ ] Component: Beam playground
- [ ] Component: Beam katas
- [ ] Component: Website
- [ ] Component: Spark Runner
- [ ] Component: Flink Runner
- [ ] Component: Samza Runner
- [ ] Component: Twister2 Runner
- [ ] Component: Hazelcast Jet Runner
- [ ] Component: Google Cloud Dataflow Runner
|
non_defect
|
document source and sink urns what needs to happen the beam protos don t describe the urns beam runner source beam runner sink which are necessary for the implementation of a portable sdk while there are now several examples it would be good to have them be concretely described in the protocol issue priority priority nice to have improvement issue components component python sdk component java sdk component go sdk component typescript sdk component io connector component beam examples component beam playground component beam katas component website component spark runner component flink runner component samza runner component runner component hazelcast jet runner component google cloud dataflow runner
| 0
|
7,346
| 3,082,727,909
|
IssuesEvent
|
2015-08-24 00:47:47
|
california-civic-data-coalition/django-calaccess-raw-data
|
https://api.github.com/repos/california-civic-data-coalition/django-calaccess-raw-data
|
opened
|
Add documentation for the ``recip_zip4`` field on the ``LccmCd`` database model
|
documentation enhancement small
|
## Your mission
Add documentation for the ``recip_zip4`` field on the ``LccmCd`` database model.
## Here's how
**Step 1**: Claim this ticket by leaving a comment below. Tell everyone you're ON IT!
**Step 2**: Open up the file that contains this model. It should be in <a href="https://github.com/california-civic-data-coalition/django-calaccess-raw-data/blob/master/calaccess_raw/models/lobbying.py">calaccess_raw.models.lobbying.py</a>.
**Step 3**: Hit the little pencil button in the upper-right corner of the code box to begin editing the file.

**Step 4**: Find this model and field in the file. (Clicking into the box and searching with CTRL-F can help you here.) Once you find it, we expect the field to lack the ``help_text`` field typically used in Django to explain what a field contains.
```python
effect_dt = fields.DateField(
null=True,
db_column="EFFECT_DT"
)
```
**Step 5**: In a separate tab, open up the <a href="Quilmes">official state documentation</a> and find the page that defines all the fields in this model.

**Step 6**: Find the row in that table's definition table that spells out what this field contains. If it lacks documentation. Note that in the ticket and close it now.

**Step 7**: Return to the GitHub tab.
**Step 8**: Add the state's label explaining what's in the field, to our field definition by inserting it a ``help_text`` argument. That should look something like this:
```python
effect_dt = fields.DateField(
null=True,
db_column="EFFECT_DT",
# Add a help_text argument like the one here, but put your string in instead.
help_text="The other values in record were effective as of this date"
)
```
**Step 9**: Scroll down below the code box and describe the change you've made in the commit message. Press the button below.

**Step 10**: Review your changes and create a pull request submitting them to the core team for inclusion.

That's it! Mission accomplished!
|
1.0
|
Add documentation for the ``recip_zip4`` field on the ``LccmCd`` database model -
## Your mission
Add documentation for the ``recip_zip4`` field on the ``LccmCd`` database model.
## Here's how
**Step 1**: Claim this ticket by leaving a comment below. Tell everyone you're ON IT!
**Step 2**: Open up the file that contains this model. It should be in <a href="https://github.com/california-civic-data-coalition/django-calaccess-raw-data/blob/master/calaccess_raw/models/lobbying.py">calaccess_raw.models.lobbying.py</a>.
**Step 3**: Hit the little pencil button in the upper-right corner of the code box to begin editing the file.

**Step 4**: Find this model and field in the file. (Clicking into the box and searching with CTRL-F can help you here.) Once you find it, we expect the field to lack the ``help_text`` field typically used in Django to explain what a field contains.
```python
effect_dt = fields.DateField(
null=True,
db_column="EFFECT_DT"
)
```
**Step 5**: In a separate tab, open up the <a href="Quilmes">official state documentation</a> and find the page that defines all the fields in this model.

**Step 6**: Find the row in that table's definition table that spells out what this field contains. If it lacks documentation. Note that in the ticket and close it now.

**Step 7**: Return to the GitHub tab.
**Step 8**: Add the state's label explaining what's in the field, to our field definition by inserting it a ``help_text`` argument. That should look something like this:
```python
effect_dt = fields.DateField(
null=True,
db_column="EFFECT_DT",
# Add a help_text argument like the one here, but put your string in instead.
help_text="The other values in record were effective as of this date"
)
```
**Step 9**: Scroll down below the code box and describe the change you've made in the commit message. Press the button below.

**Step 10**: Review your changes and create a pull request submitting them to the core team for inclusion.

That's it! Mission accomplished!
|
non_defect
|
add documentation for the recip field on the lccmcd database model your mission add documentation for the recip field on the lccmcd database model here s how step claim this ticket by leaving a comment below tell everyone you re on it step open up the file that contains this model it should be in a href step hit the little pencil button in the upper right corner of the code box to begin editing the file step find this model and field in the file clicking into the box and searching with ctrl f can help you here once you find it we expect the field to lack the help text field typically used in django to explain what a field contains python effect dt fields datefield null true db column effect dt step in a separate tab open up the official state documentation and find the page that defines all the fields in this model step find the row in that table s definition table that spells out what this field contains if it lacks documentation note that in the ticket and close it now step return to the github tab step add the state s label explaining what s in the field to our field definition by inserting it a help text argument that should look something like this python effect dt fields datefield null true db column effect dt add a help text argument like the one here but put your string in instead help text the other values in record were effective as of this date step scroll down below the code box and describe the change you ve made in the commit message press the button below step review your changes and create a pull request submitting them to the core team for inclusion that s it mission accomplished
| 0
|
2,813
| 2,607,946,536
|
IssuesEvent
|
2015-02-26 00:33:30
|
chrsmithdemos/switchlist
|
https://api.github.com/repos/chrsmithdemos/switchlist
|
closed
|
Ship Waybill switchlist
|
auto-migrated Priority-Medium Type-Defect
|
```
The Waybill switchlist is neat. It needs the following to be reasonable for
external use:
1) Use intermediate stops to name exits/entrances on layout. (All stops on
layouts would be another possibility, but intermediate probably works ok.
2) Get rid of San Jose specific stuff, or make easy to customize by naming
default routing for certain entrances.
3) Figure out how to color-code. Add outline when printing on printer.
```
-----
Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 27 Jul 2014 at 4:54
|
1.0
|
Ship Waybill switchlist - ```
The Waybill switchlist is neat. It needs the following to be reasonable for
external use:
1) Use intermediate stops to name exits/entrances on layout. (All stops on
layouts would be another possibility, but intermediate probably works ok.
2) Get rid of San Jose specific stuff, or make easy to customize by naming
default routing for certain entrances.
3) Figure out how to color-code. Add outline when printing on printer.
```
-----
Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 27 Jul 2014 at 4:54
|
defect
|
ship waybill switchlist the waybill switchlist is neat it needs the following to be reasonable for external use use intermediate stops to name exits entrances on layout all stops on layouts would be another possibility but intermediate probably works ok get rid of san jose specific stuff or make easy to customize by naming default routing for certain entrances figure out how to color code add outline when printing on printer original issue reported on code google com by rwbowdi gmail com on jul at
| 1
|
97,687
| 8,665,850,430
|
IssuesEvent
|
2018-11-29 01:12:02
|
datalad/datalad
|
https://api.github.com/repos/datalad/datalad
|
closed
|
test__version__ is now failing on some py3 runs
|
failing-test fails-on-travis
|
Observed [here](https://travis-ci.org/datalad/datalad/jobs/460830536#L3487) and [here](https://travis-ci.org/datalad/datalad/jobs/460830537#L2820).
```
======================================================================
ERROR: datalad.tests.test_version.test__version__
----------------------------------------------------------------------
Traceback (most recent call last):
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/nose/case.py", line 198, in runTest
self.test(*self.arg)
File "/home/travis/build/datalad/datalad/datalad/tests/test_version.py", line 43, in test__version__
if not line.startswith('## '):
TypeError: startswith first arg must be bytes or a tuple of bytes, not str
```
|
1.0
|
test__version__ is now failing on some py3 runs - Observed [here](https://travis-ci.org/datalad/datalad/jobs/460830536#L3487) and [here](https://travis-ci.org/datalad/datalad/jobs/460830537#L2820).
```
======================================================================
ERROR: datalad.tests.test_version.test__version__
----------------------------------------------------------------------
Traceback (most recent call last):
File "/home/travis/virtualenv/python3.6.3/lib/python3.6/site-packages/nose/case.py", line 198, in runTest
self.test(*self.arg)
File "/home/travis/build/datalad/datalad/datalad/tests/test_version.py", line 43, in test__version__
if not line.startswith('## '):
TypeError: startswith first arg must be bytes or a tuple of bytes, not str
```
|
non_defect
|
test version is now failing on some runs observed and error datalad tests test version test version traceback most recent call last file home travis virtualenv lib site packages nose case py line in runtest self test self arg file home travis build datalad datalad datalad tests test version py line in test version if not line startswith typeerror startswith first arg must be bytes or a tuple of bytes not str
| 0
|
185,016
| 14,292,764,336
|
IssuesEvent
|
2020-11-24 01:55:29
|
github-vet/rangeclosure-findings
|
https://api.github.com/repos/github-vet/rangeclosure-findings
|
closed
|
oren/pokemon: vendor/github.com/gogo/protobuf/proto/text_test.go; 8 LoC
|
fresh test tiny
|
Found a possible issue in [oren/pokemon](https://www.github.com/oren/pokemon) at [vendor/github.com/gogo/protobuf/proto/text_test.go](https://github.com/oren/pokemon/blob/ba215f93a8cc2a1c86eb6771c5b2378f07be4c73/vendor/github.com/gogo/protobuf/proto/text_test.go#L404-L411)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/oren/pokemon/blob/ba215f93a8cc2a1c86eb6771c5b2378f07be4c73/vendor/github.com/gogo/protobuf/proto/text_test.go#L404-L411)
<details>
<summary>Click here to show the 8 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range tests {
msg := &pb.FloatingPoint{F: &test.f}
got := strings.TrimSpace(msg.String())
want := `f:` + test.want
if got != want {
t.Errorf("f=%f: got %q, want %q", test.f, got, want)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: ba215f93a8cc2a1c86eb6771c5b2378f07be4c73
|
1.0
|
oren/pokemon: vendor/github.com/gogo/protobuf/proto/text_test.go; 8 LoC -
Found a possible issue in [oren/pokemon](https://www.github.com/oren/pokemon) at [vendor/github.com/gogo/protobuf/proto/text_test.go](https://github.com/oren/pokemon/blob/ba215f93a8cc2a1c86eb6771c5b2378f07be4c73/vendor/github.com/gogo/protobuf/proto/text_test.go#L404-L411)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/oren/pokemon/blob/ba215f93a8cc2a1c86eb6771c5b2378f07be4c73/vendor/github.com/gogo/protobuf/proto/text_test.go#L404-L411)
<details>
<summary>Click here to show the 8 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range tests {
msg := &pb.FloatingPoint{F: &test.f}
got := strings.TrimSpace(msg.String())
want := `f:` + test.want
if got != want {
t.Errorf("f=%f: got %q, want %q", test.f, got, want)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: ba215f93a8cc2a1c86eb6771c5b2378f07be4c73
|
non_defect
|
oren pokemon vendor github com gogo protobuf proto text test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for test range tests msg pb floatingpoint f test f got strings trimspace msg string want f test want if got want t errorf f f got q want q test f got want leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 0
|
13,835
| 2,787,231,626
|
IssuesEvent
|
2015-05-08 03:05:50
|
cfpb/hmda-pilot
|
https://api.github.com/repos/cfpb/hmda-pilot
|
closed
|
About page issues
|
defect ui
|
- [x] The link to ffiec.gov/hmda is broken. Also, the Confirm Navigation dialog appears twice when clicking the link.
- [x] Two different styles of quotation marks are being used around the word "clean" (3rd and last paragraph).
|
1.0
|
About page issues - - [x] The link to ffiec.gov/hmda is broken. Also, the Confirm Navigation dialog appears twice when clicking the link.
- [x] Two different styles of quotation marks are being used around the word "clean" (3rd and last paragraph).
|
defect
|
about page issues the link to ffiec gov hmda is broken also the confirm navigation dialog appears twice when clicking the link two different styles of quotation marks are being used around the word clean and last paragraph
| 1
|
637,659
| 20,674,682,781
|
IssuesEvent
|
2022-03-10 08:01:31
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.facebook.com - desktop site instead of mobile site
|
browser-chrome priority-critical
|
<!-- @browser: Chrome 99.0.4844 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/99.0.4844.51 Safari/537.36 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/100764 -->
**URL**: https://www.facebook.com/
**Browser / Version**: Chrome 99.0.4844
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
page not found, request denied,
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/3/3e87ac4d-5488-4b51-a002-951802587652.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.facebook.com - desktop site instead of mobile site - <!-- @browser: Chrome 99.0.4844 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/99.0.4844.51 Safari/537.36 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/100764 -->
**URL**: https://www.facebook.com/
**Browser / Version**: Chrome 99.0.4844
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
page not found, request denied,
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/3/3e87ac4d-5488-4b51-a002-951802587652.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
desktop site instead of mobile site url browser version chrome operating system windows tested another browser yes chrome problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce page not found request denied view the screenshot img alt screenshot src browser configuration none from with ❤️
| 0
|
24,775
| 4,103,739,158
|
IssuesEvent
|
2016-06-04 21:59:34
|
google/google-api-dotnet-client
|
https://api.github.com/repos/google/google-api-dotnet-client
|
closed
|
Provide support for Mono/Xamarin Mobile Platforms (iOS & Android)
|
auto-migrated Component-Platforms Priority-Medium Type-Defect
|
```
External references, such as a standards document, or specification?
http://xamarin.com/monoforandroid
http://xamarin.com/ios
Target platform (e.g. Windows, Mono, Silverlight, WP7, All)?
Xamarin.Android / Xamarin.iOS
Much like the WinRT request
(https://code.google.com/p/google-api-dotnet-client/issues/detail?id=149&q=mono&
colspec=ID%20Type%20Component%20Status%20Priority%20Milestone%20Owner%20Summary)
support for the mobile Mono/.Net based platforms of Xamarin would be great.
They provide great support for .Net on both platforms and having the Google API
client available on these platforms would be quite awesome.
```
Original issue reported on code.google.com by `j...@joergbattermann.com` on 30 Jun 2013 at 11:06
|
1.0
|
Provide support for Mono/Xamarin Mobile Platforms (iOS & Android) - ```
External references, such as a standards document, or specification?
http://xamarin.com/monoforandroid
http://xamarin.com/ios
Target platform (e.g. Windows, Mono, Silverlight, WP7, All)?
Xamarin.Android / Xamarin.iOS
Much like the WinRT request
(https://code.google.com/p/google-api-dotnet-client/issues/detail?id=149&q=mono&
colspec=ID%20Type%20Component%20Status%20Priority%20Milestone%20Owner%20Summary)
support for the mobile Mono/.Net based platforms of Xamarin would be great.
They provide great support for .Net on both platforms and having the Google API
client available on these platforms would be quite awesome.
```
Original issue reported on code.google.com by `j...@joergbattermann.com` on 30 Jun 2013 at 11:06
|
defect
|
provide support for mono xamarin mobile platforms ios android external references such as a standards document or specification target platform e g windows mono silverlight all xamarin android xamarin ios much like the winrt request colspec id support for the mobile mono net based platforms of xamarin would be great they provide great support for net on both platforms and having the google api client available on these platforms would be quite awesome original issue reported on code google com by j joergbattermann com on jun at
| 1
|
34,736
| 7,459,775,565
|
IssuesEvent
|
2018-03-30 16:47:15
|
kerdokullamae/test_koik_issued
|
https://api.github.com/repos/kerdokullamae/test_koik_issued
|
closed
|
Keskkondade deploymisel viga
|
C: AIS P: highest R: fixed T: defect
|
**Reported by sven syld on 28 Dec 2015 16:51 UTC**
PropelMigration_1445864995 topicu järjekorda seadmine annab vea, sest vahepeal on mudel muutunud.
|
1.0
|
Keskkondade deploymisel viga - **Reported by sven syld on 28 Dec 2015 16:51 UTC**
PropelMigration_1445864995 topicu järjekorda seadmine annab vea, sest vahepeal on mudel muutunud.
|
defect
|
keskkondade deploymisel viga reported by sven syld on dec utc propelmigration topicu järjekorda seadmine annab vea sest vahepeal on mudel muutunud
| 1
|
86,517
| 8,038,683,037
|
IssuesEvent
|
2018-07-30 16:02:11
|
seqan/seqan3
|
https://api.github.com/repos/seqan/seqan3
|
closed
|
make core/concept/core_detail.hpp a proper test case
|
good first issue test case
|
`/include/seqan3/core/concept/core_detail.hpp` tests core functionality within the header via `static_assert`s, this should be transformed into a test case.
|
1.0
|
make core/concept/core_detail.hpp a proper test case - `/include/seqan3/core/concept/core_detail.hpp` tests core functionality within the header via `static_assert`s, this should be transformed into a test case.
|
non_defect
|
make core concept core detail hpp a proper test case include core concept core detail hpp tests core functionality within the header via static assert s this should be transformed into a test case
| 0
|
65,296
| 19,325,382,145
|
IssuesEvent
|
2021-12-14 10:43:09
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
In-Call dialpad is off the right of the screen
|
T-Defect
|
### Steps to reproduce
1. Place a call to an endpoint that advertises dialpad support (eg. a sip / pstn bridge)
2. Wait for other side to answer
3. Click dialpad button
4. See dial pad hanging off the right of the screen
### Outcome
#### What did you expect?
Dial pad on left of call window
#### What happened instead?
<img width="757" alt="Screenshot 2021-12-14 at 10 34 43" src="https://user-images.githubusercontent.com/986903/145983222-15223cbf-cac9-49b4-82d3-b1751cb99203.png">
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
app.element.io
### Application version
v1.9.7
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
In-Call dialpad is off the right of the screen - ### Steps to reproduce
1. Place a call to an endpoint that advertises dialpad support (eg. a sip / pstn bridge)
2. Wait for other side to answer
3. Click dialpad button
4. See dial pad hanging off the right of the screen
### Outcome
#### What did you expect?
Dial pad on left of call window
#### What happened instead?
<img width="757" alt="Screenshot 2021-12-14 at 10 34 43" src="https://user-images.githubusercontent.com/986903/145983222-15223cbf-cac9-49b4-82d3-b1751cb99203.png">
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
app.element.io
### Application version
v1.9.7
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
in call dialpad is off the right of the screen steps to reproduce place a call to an endpoint that advertises dialpad support eg a sip pstn bridge wait for other side to answer click dialpad button see dial pad hanging off the right of the screen outcome what did you expect dial pad on left of call window what happened instead img width alt screenshot at src operating system no response browser information no response url for webapp app element io application version homeserver no response will you send logs no
| 1
|
511,205
| 14,856,161,495
|
IssuesEvent
|
2021-01-18 13:47:23
|
gnosis/conditional-tokens-explorer
|
https://api.github.com/repos/gnosis/conditional-tokens-explorer
|
opened
|
Impossible to merge positions when 'Use CPK Address' option is switched off
|
High priority bug
|
Related to #735, #782
1. Switch off 'Use CPK Address' option
2. split positions
3. Open merge positions page and try me merge just created positions
**AR**: Merge with section always shows position with 0 balance --> impossible to start the merge

**ER**: user is able to merge positions
|
1.0
|
Impossible to merge positions when 'Use CPK Address' option is switched off - Related to #735, #782
1. Switch off 'Use CPK Address' option
2. split positions
3. Open merge positions page and try me merge just created positions
**AR**: Merge with section always shows position with 0 balance --> impossible to start the merge

**ER**: user is able to merge positions
|
non_defect
|
impossible to merge positions when use cpk address option is switched off related to switch off use cpk address option split positions open merge positions page and try me merge just created positions ar merge with section always shows position with balance impossible to start the merge er user is able to merge positions
| 0
|
6,787
| 2,610,279,374
|
IssuesEvent
|
2015-02-26 19:29:19
|
chrsmith/scribefire-chrome
|
https://api.github.com/repos/chrsmith/scribefire-chrome
|
closed
|
Change the <!--more--> label when updating the blogspot post
|
auto-migrated Priority-Medium Type-Defect
|
```
What's the problem?
The blogspot support <!--more--> tag in the html source, and let your post
expandable/hide the long post with a "read more" link.
It is successful when I use the scribefire to post, but when I updating the old
post, I found scribefire download the post, and display the source code as "<a
tag=>more</a>". So the "read more" couldn't be implemented, I've to change it
manually again.
What browser are you using?
Chrome up to date
What version of ScribeFire are you running?
1.7
```
-----
Original issue reported on code.google.com by `leelei...@gmail.com` on 16 Jun 2011 at 3:37
* Merged into: #335
|
1.0
|
Change the <!--more--> label when updating the blogspot post - ```
What's the problem?
The blogspot support <!--more--> tag in the html source, and let your post
expandable/hide the long post with a "read more" link.
It is successful when I use the scribefire to post, but when I updating the old
post, I found scribefire download the post, and display the source code as "<a
tag=>more</a>". So the "read more" couldn't be implemented, I've to change it
manually again.
What browser are you using?
Chrome up to date
What version of ScribeFire are you running?
1.7
```
-----
Original issue reported on code.google.com by `leelei...@gmail.com` on 16 Jun 2011 at 3:37
* Merged into: #335
|
defect
|
change the label when updating the blogspot post what s the problem the blogspot support tag in the html source and let your post expandable hide the long post with a read more link it is successful when i use the scribefire to post but when i updating the old post i found scribefire download the post and display the source code as a tag more so the read more couldn t be implemented i ve to change it manually again what browser are you using chrome up to date what version of scribefire are you running original issue reported on code google com by leelei gmail com on jun at merged into
| 1
|
15,480
| 2,856,531,111
|
IssuesEvent
|
2015-06-02 15:21:39
|
contao/core
|
https://api.github.com/repos/contao/core
|
closed
|
Nachrichtenarchiv-Bearbeitungslink (Stift-Icon) obwohl keine Rechte
|
defect
|
Kleiner Schönheitsfehler:
Wenn ein Benutzer keinerlei Rechte zur Bearbeitung eines Nachrichtenarchivs hat, so erscheint in der BE-Nachrichtenübersicht oben im Archiv-Header trotzdem der "Stift"; und führt dann etwas unschön zu einer leeren Bearbeitungsmaske... In der Archivübersicht zu Anfang ist das entspr. Bearbeitungssymbol hingegen ordnungsgemäß inaktiv.
Auch der "Mehrere bearbeiten"-Link sollte überall ggf. besser deaktiviert / ausgeblendet werden, insofern es keine Einträge gibt oder der User keinerlei Rechte zur Bearbeitung hat.




|
1.0
|
Nachrichtenarchiv-Bearbeitungslink (Stift-Icon) obwohl keine Rechte - Kleiner Schönheitsfehler:
Wenn ein Benutzer keinerlei Rechte zur Bearbeitung eines Nachrichtenarchivs hat, so erscheint in der BE-Nachrichtenübersicht oben im Archiv-Header trotzdem der "Stift"; und führt dann etwas unschön zu einer leeren Bearbeitungsmaske... In der Archivübersicht zu Anfang ist das entspr. Bearbeitungssymbol hingegen ordnungsgemäß inaktiv.
Auch der "Mehrere bearbeiten"-Link sollte überall ggf. besser deaktiviert / ausgeblendet werden, insofern es keine Einträge gibt oder der User keinerlei Rechte zur Bearbeitung hat.




|
defect
|
nachrichtenarchiv bearbeitungslink stift icon obwohl keine rechte kleiner schönheitsfehler wenn ein benutzer keinerlei rechte zur bearbeitung eines nachrichtenarchivs hat so erscheint in der be nachrichtenübersicht oben im archiv header trotzdem der stift und führt dann etwas unschön zu einer leeren bearbeitungsmaske in der archivübersicht zu anfang ist das entspr bearbeitungssymbol hingegen ordnungsgemäß inaktiv auch der mehrere bearbeiten link sollte überall ggf besser deaktiviert ausgeblendet werden insofern es keine einträge gibt oder der user keinerlei rechte zur bearbeitung hat
| 1
|
46,284
| 13,055,885,358
|
IssuesEvent
|
2020-07-30 03:01:23
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
PROPOSAL build fail due to missing tables in I3_TESTDATA (Trac #923)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
Migrated from https://code.icecube.wisc.edu/ticket/923
```json
{
"status": "closed",
"changetime": "2015-04-13T16:20:55",
"description": "when trying to build PROPOSAL in DEBUG mode compiler will fail due to missing directory in I3_TESTDATA that is tried to be accessed before tables were built:\n\n{{{\n[ 50%] Generating resources/tables/.tables.auto_generated\nFATAL (I3PropagatorService): /usr/local/test-data/PROPOSAL/resources/tables does not exist (I3PropagatorServicePROPOSAL.cxx:117 in static std::string I3PropagatorServicePROPOSAL::GetDefaultTableDir())\nTraceback (most recent call last):\n File \"resources/tables/generate_tables.py\", line 6, in <module>\n from icecube.PROPOSAL import I3PropagatorServicePROPOSAL\nRuntimeError: /usr/local/test-data/PROPOSAL/resources/tables does not exist (in static std::string I3PropagatorServicePROPOSAL::GetDefaultTableDir())\nmake[2]: *** [PROPOSAL/resources/tables/.tables.auto_generated] Error 1\nmake[1]: *** [PROPOSAL/CMakeFiles/PROPOSAL-tables.dir/all] Error 2\nmake: *** [all] Error 2\n}}}",
"reporter": "musner",
"cc": "",
"resolution": "invalid",
"_ts": "1428942055132091",
"component": "combo simulation",
"summary": "PROPOSAL build fail due to missing tables in I3_TESTDATA",
"priority": "normal",
"keywords": "",
"time": "2015-04-10T11:11:48",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
|
1.0
|
PROPOSAL build fail due to missing tables in I3_TESTDATA (Trac #923) - Migrated from https://code.icecube.wisc.edu/ticket/923
```json
{
"status": "closed",
"changetime": "2015-04-13T16:20:55",
"description": "when trying to build PROPOSAL in DEBUG mode compiler will fail due to missing directory in I3_TESTDATA that is tried to be accessed before tables were built:\n\n{{{\n[ 50%] Generating resources/tables/.tables.auto_generated\nFATAL (I3PropagatorService): /usr/local/test-data/PROPOSAL/resources/tables does not exist (I3PropagatorServicePROPOSAL.cxx:117 in static std::string I3PropagatorServicePROPOSAL::GetDefaultTableDir())\nTraceback (most recent call last):\n File \"resources/tables/generate_tables.py\", line 6, in <module>\n from icecube.PROPOSAL import I3PropagatorServicePROPOSAL\nRuntimeError: /usr/local/test-data/PROPOSAL/resources/tables does not exist (in static std::string I3PropagatorServicePROPOSAL::GetDefaultTableDir())\nmake[2]: *** [PROPOSAL/resources/tables/.tables.auto_generated] Error 1\nmake[1]: *** [PROPOSAL/CMakeFiles/PROPOSAL-tables.dir/all] Error 2\nmake: *** [all] Error 2\n}}}",
"reporter": "musner",
"cc": "",
"resolution": "invalid",
"_ts": "1428942055132091",
"component": "combo simulation",
"summary": "PROPOSAL build fail due to missing tables in I3_TESTDATA",
"priority": "normal",
"keywords": "",
"time": "2015-04-10T11:11:48",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
|
defect
|
proposal build fail due to missing tables in testdata trac migrated from json status closed changetime description when trying to build proposal in debug mode compiler will fail due to missing directory in testdata that is tried to be accessed before tables were built n n n generating resources tables tables auto generated nfatal usr local test data proposal resources tables does not exist cxx in static std string getdefaulttabledir ntraceback most recent call last n file resources tables generate tables py line in n from icecube proposal import nruntimeerror usr local test data proposal resources tables does not exist in static std string getdefaulttabledir nmake error nmake error nmake error n reporter musner cc resolution invalid ts component combo simulation summary proposal build fail due to missing tables in testdata priority normal keywords time milestone owner nega type defect
| 1
|
396,990
| 11,716,780,692
|
IssuesEvent
|
2020-03-09 16:10:15
|
Noggog/Mutagen
|
https://api.github.com/repos/Noggog/Mutagen
|
opened
|
Remove LoquiBinaryTranslation Calls for Simple Objects
|
Low Priority enhancement
|
Lots of calls to simple Loqui objects bounce off LoquiBinaryTranslation. This isn't too expensive, but can probably be trimmed out.
|
1.0
|
Remove LoquiBinaryTranslation Calls for Simple Objects - Lots of calls to simple Loqui objects bounce off LoquiBinaryTranslation. This isn't too expensive, but can probably be trimmed out.
|
non_defect
|
remove loquibinarytranslation calls for simple objects lots of calls to simple loqui objects bounce off loquibinarytranslation this isn t too expensive but can probably be trimmed out
| 0
|
564,933
| 16,745,188,926
|
IssuesEvent
|
2021-06-11 14:42:09
|
ampproject/amphtml
|
https://api.github.com/repos/ampproject/amphtml
|
opened
|
`amp-story-desktop-one-panel` Confirm calculation of `--story-page-` CSS variables
|
P1: High Priority Type: Feature Request WG: stories
|
### Description
Confirm calculation of `--story-page-` CSS variables when the `amp-story-desktop-one-panel` experiment is active.
### Alternatives Considered
N / A
### Additional Context
_No response_
|
1.0
|
`amp-story-desktop-one-panel` Confirm calculation of `--story-page-` CSS variables - ### Description
Confirm calculation of `--story-page-` CSS variables when the `amp-story-desktop-one-panel` experiment is active.
### Alternatives Considered
N / A
### Additional Context
_No response_
|
non_defect
|
amp story desktop one panel confirm calculation of story page css variables description confirm calculation of story page css variables when the amp story desktop one panel experiment is active alternatives considered n a additional context no response
| 0
|
86,378
| 15,755,619,208
|
IssuesEvent
|
2021-03-31 02:05:57
|
crossbrowsertesting/cbt-tunnel-nodejs
|
https://api.github.com/repos/crossbrowsertesting/cbt-tunnel-nodejs
|
opened
|
CVE-2021-28918 (Medium) detected in netmask-1.0.6.tgz
|
security vulnerability
|
## CVE-2021-28918 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netmask-1.0.6.tgz</b></p></summary>
<p>Parse and lookup IP network blocks</p>
<p>Library home page: <a href="https://registry.npmjs.org/netmask/-/netmask-1.0.6.tgz">https://registry.npmjs.org/netmask/-/netmask-1.0.6.tgz</a></p>
<p>Path to dependency file: cbt-tunnel-nodejs/package.json</p>
<p>Path to vulnerable library: cbt-tunnel-nodejs/node_modules/netmask/package.json</p>
<p>
Dependency Hierarchy:
- pac-resolver-3.0.1.tgz (Root Library)
- :x: **netmask-1.0.6.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
netmask npm package is vulnerable to octal input data. This may lead to server-side request forgery, remote file inclusion, local file inclusion, and other vulnerabilities.
<p>Publish Date: 2021-03-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28918>CVE-2021-28918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-pch5-whg9-qr2r">https://github.com/advisories/GHSA-pch5-whg9-qr2r</a></p>
<p>Release Date: 2021-03-19</p>
<p>Fix Resolution: netmask - 2.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"netmask","packageVersion":"1.0.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@crossbrowsertesting/pac-resolver:3.0.1;netmask:1.0.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"netmask - 2.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-28918","vulnerabilityDetails":"netmask npm package is vulnerable to octal input data. This may lead to server-side request forgery, remote file inclusion, local file inclusion, and other vulnerabilities.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28918","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-28918 (Medium) detected in netmask-1.0.6.tgz - ## CVE-2021-28918 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netmask-1.0.6.tgz</b></p></summary>
<p>Parse and lookup IP network blocks</p>
<p>Library home page: <a href="https://registry.npmjs.org/netmask/-/netmask-1.0.6.tgz">https://registry.npmjs.org/netmask/-/netmask-1.0.6.tgz</a></p>
<p>Path to dependency file: cbt-tunnel-nodejs/package.json</p>
<p>Path to vulnerable library: cbt-tunnel-nodejs/node_modules/netmask/package.json</p>
<p>
Dependency Hierarchy:
- pac-resolver-3.0.1.tgz (Root Library)
- :x: **netmask-1.0.6.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
netmask npm package is vulnerable to octal input data. This may lead to server-side request forgery, remote file inclusion, local file inclusion, and other vulnerabilities.
<p>Publish Date: 2021-03-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28918>CVE-2021-28918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-pch5-whg9-qr2r">https://github.com/advisories/GHSA-pch5-whg9-qr2r</a></p>
<p>Release Date: 2021-03-19</p>
<p>Fix Resolution: netmask - 2.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"netmask","packageVersion":"1.0.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@crossbrowsertesting/pac-resolver:3.0.1;netmask:1.0.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"netmask - 2.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-28918","vulnerabilityDetails":"netmask npm package is vulnerable to octal input data. This may lead to server-side request forgery, remote file inclusion, local file inclusion, and other vulnerabilities.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28918","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve medium detected in netmask tgz cve medium severity vulnerability vulnerable library netmask tgz parse and lookup ip network blocks library home page a href path to dependency file cbt tunnel nodejs package json path to vulnerable library cbt tunnel nodejs node modules netmask package json dependency hierarchy pac resolver tgz root library x netmask tgz vulnerable library found in base branch master vulnerability details netmask npm package is vulnerable to octal input data this may lead to server side request forgery remote file inclusion local file inclusion and other vulnerabilities publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution netmask isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree crossbrowsertesting pac resolver netmask isminimumfixversionavailable true minimumfixversion netmask basebranches vulnerabilityidentifier cve vulnerabilitydetails netmask npm package is vulnerable to octal input data this may lead to server side request forgery remote file inclusion local file inclusion and other vulnerabilities vulnerabilityurl
| 0
|
24,395
| 3,972,024,304
|
IssuesEvent
|
2016-05-04 14:07:00
|
buildo/react-components
|
https://api.github.com/repos/buildo/react-components
|
opened
|
[DropdownMenu] Menu item is not really vertical centered
|
defect
|
## description
The actual `menu-item` render in `Menu` doesn't align its content due to use of `span`.
<img width="245" alt="screen shot 2016-05-04 at 15 39 20" src="https://cloud.githubusercontent.com/assets/1838567/15015843/6e2607d4-120e-11e6-8d42-3d98defd34ea.png">
## how to reproduce
- {optional: describe steps to reproduce defect}
## specs
- replace `span`s with `FlexView`
## misc
{optional: other useful info}
|
1.0
|
[DropdownMenu] Menu item is not really vertical centered - ## description
The actual `menu-item` render in `Menu` doesn't align its content due to use of `span`.
<img width="245" alt="screen shot 2016-05-04 at 15 39 20" src="https://cloud.githubusercontent.com/assets/1838567/15015843/6e2607d4-120e-11e6-8d42-3d98defd34ea.png">
## how to reproduce
- {optional: describe steps to reproduce defect}
## specs
- replace `span`s with `FlexView`
## misc
{optional: other useful info}
|
defect
|
menu item is not really vertical centered description the actual menu item render in menu doesn t align its content due to use of span img width alt screen shot at src how to reproduce optional describe steps to reproduce defect specs replace span s with flexview misc optional other useful info
| 1
|
823,572
| 31,024,786,929
|
IssuesEvent
|
2023-08-10 08:24:18
|
yugabyte/yugabyte-db
|
https://api.github.com/repos/yugabyte/yugabyte-db
|
closed
|
Adding waiting mechanisms for GetChangesFromCDC() call, in the integration tests CDCSDKYsqlTests
|
kind/new-feature priority/low area/cdcsdk jira-originated
|
Jira Link: [DB-7499](https://yugabyte.atlassian.net/browse/DB-7499)
[DB-7499]: https://yugabyte.atlassian.net/browse/DB-7499?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
|
1.0
|
Adding waiting mechanisms for GetChangesFromCDC() call, in the integration tests CDCSDKYsqlTests - Jira Link: [DB-7499](https://yugabyte.atlassian.net/browse/DB-7499)
[DB-7499]: https://yugabyte.atlassian.net/browse/DB-7499?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
|
non_defect
|
adding waiting mechanisms for getchangesfromcdc call in the integration tests cdcsdkysqltests jira link
| 0
|
133,538
| 12,543,765,727
|
IssuesEvent
|
2020-06-05 16:05:34
|
cilium/cilium
|
https://api.github.com/repos/cilium/cilium
|
closed
|
Clarify potential for connection disruption due to --native-routing-cidr flag
|
area/documentation priority/release-blocker
|
In Cilium 1.8, we intend to change the masquerade logic to more accurately determine when to apply masquerade (ie SNAT) to traffic in direct-routing mode, to ensure that traffic is only masqueraded when the destination is not directly reachable. To do this, we solicit the directly routable CIDR range from the user via the `--native-routing-cidr` option.
The upgrade docs already describe this briefly:
https://github.com/cilium/cilium/blob/master/Documentation/install/upgrade.rst#important-changes-required-before-upgrading-to-180
By my understanding, if this option is not configured, this will result in behaviour change for direct routing users which may lead to connection disruption. We should briefly extend it to clearly define the consequences of failing to configure the option to ensure that users treat this option with the severity it requires.
Questions to guide this documentation:
* [ ] Are all existing direct-routing users already explicitly configuring `--ipv4-cluster-cidr-mask-size` or is it automatically configured? If it is automatically configured, users may not be aware that they were running with this option.
* By my initial grep for this option in the Cilium-1.7 codebase, it looks like no instructions inform the user that they must specify this CIDR.
* If it's autodetected today, then we should make this clear that users who have no knowledge of this setting, but are running `--tunnel=disabled` will *still* need to read the paragraph and take action. This can be achieved with a short sentence that states what the default behaviour is, eg "Cilium 1.7 and below inferred the setting for this value if it was not explicitly set."
* [ ] What are the consequences of failing to configure this option?
* For example, cilium-managed pods may be unable to establish connections to the outside world(?), (or to non-Cilium-managed IPs within the cluster?).
* [ ] Why does the user need to consider this now?
* For example, "Previous Cilium versions inferred the cluster range to choose when to perform masquerading, however in some cases it would mistakenly masquerade traffic that is destined outside the cluster".
|
1.0
|
Clarify potential for connection disruption due to --native-routing-cidr flag - In Cilium 1.8, we intend to change the masquerade logic to more accurately determine when to apply masquerade (ie SNAT) to traffic in direct-routing mode, to ensure that traffic is only masqueraded when the destination is not directly reachable. To do this, we solicit the directly routable CIDR range from the user via the `--native-routing-cidr` option.
The upgrade docs already describe this briefly:
https://github.com/cilium/cilium/blob/master/Documentation/install/upgrade.rst#important-changes-required-before-upgrading-to-180
By my understanding, if this option is not configured, this will result in behaviour change for direct routing users which may lead to connection disruption. We should briefly extend it to clearly define the consequences of failing to configure the option to ensure that users treat this option with the severity it requires.
Questions to guide this documentation:
* [ ] Are all existing direct-routing users already explicitly configuring `--ipv4-cluster-cidr-mask-size` or is it automatically configured? If it is automatically configured, users may not be aware that they were running with this option.
* By my initial grep for this option in the Cilium-1.7 codebase, it looks like no instructions inform the user that they must specify this CIDR.
* If it's autodetected today, then we should make this clear that users who have no knowledge of this setting, but are running `--tunnel=disabled` will *still* need to read the paragraph and take action. This can be achieved with a short sentence that states what the default behaviour is, eg "Cilium 1.7 and below inferred the setting for this value if it was not explicitly set."
* [ ] What are the consequences of failing to configure this option?
* For example, cilium-managed pods may be unable to establish connections to the outside world(?), (or to non-Cilium-managed IPs within the cluster?).
* [ ] Why does the user need to consider this now?
* For example, "Previous Cilium versions inferred the cluster range to choose when to perform masquerading, however in some cases it would mistakenly masquerade traffic that is destined outside the cluster".
|
non_defect
|
clarify potential for connection disruption due to native routing cidr flag in cilium we intend to change the masquerade logic to more accurately determine when to apply masquerade ie snat to traffic in direct routing mode to ensure that traffic is only masqueraded when the destination is not directly reachable to do this we solicit the directly routable cidr range from the user via the native routing cidr option the upgrade docs already describe this briefly by my understanding if this option is not configured this will result in behaviour change for direct routing users which may lead to connection disruption we should briefly extend it to clearly define the consequences of failing to configure the option to ensure that users treat this option with the severity it requires questions to guide this documentation are all existing direct routing users already explicitly configuring cluster cidr mask size or is it automatically configured if it is automatically configured users may not be aware that they were running with this option by my initial grep for this option in the cilium codebase it looks like no instructions inform the user that they must specify this cidr if it s autodetected today then we should make this clear that users who have no knowledge of this setting but are running tunnel disabled will still need to read the paragraph and take action this can be achieved with a short sentence that states what the default behaviour is eg cilium and below inferred the setting for this value if it was not explicitly set what are the consequences of failing to configure this option for example cilium managed pods may be unable to establish connections to the outside world or to non cilium managed ips within the cluster why does the user need to consider this now for example previous cilium versions inferred the cluster range to choose when to perform masquerading however in some cases it would mistakenly masquerade traffic that is destined outside the cluster
| 0
|
624,581
| 19,701,986,588
|
IssuesEvent
|
2022-01-12 17:30:44
|
BTAA-Geospatial-Data-Project/geoportal
|
https://api.github.com/repos/BTAA-Geospatial-Data-Project/geoportal
|
opened
|
Capitalize [Creator not found] on Cite widget
|
interface:item page priority:medium
|
When there is no creator, the tool inserts the default string [creator not found]. Change to [Creator not found]
|
1.0
|
Capitalize [Creator not found] on Cite widget - When there is no creator, the tool inserts the default string [creator not found]. Change to [Creator not found]
|
non_defect
|
capitalize on cite widget when there is no creator the tool inserts the default string change to
| 0
|
601,976
| 18,444,279,240
|
IssuesEvent
|
2021-10-14 22:29:11
|
bcgov/entity
|
https://api.github.com/repos/bcgov/entity
|
closed
|
Inconsistent login and log out scenarios in case of Staff login using IDIR
|
bug Priority2 Relationships Director Search
|
**Describe the bug in current situation**
When a Staff user login in to the application, it uses a different login screen whereas when they log out the are navigated to welcome to cooperatives screen which is inconsistent
**Impact of this bug**
Staff user will have to copy paste the url again to login using IDIR as when they logout they are navigated to a different screen which is pretty inconsistent
**Chance of Occurring (high/medium/low/very low)**
High
**Pre Conditions: which Env, any pre-requesites or assumptions to execute steps?**
User is acting as Staff and loggin in with IDIR
**Steps to Reproduce**
1. For staff user login -https://test.bcregistry.ca/cooperatives/auth/signin/idir
2. Now enter the IDIR cred and login to access the staff dashbaord
3. Now logout from the application and notice where the user is navigated to
**Actual/ observed behavior/ results**
Staff is navigated to welcome to cooperatives online screen which is inconsistent
**Expected behavior**
Staff should be navigated to the same screen from where it logs in to avoid confusion
**Screenshots/ Visual Reference/ Source**
**Login Screen**

**Logout Screen**

FYI @saravankumarpa
|
1.0
|
Inconsistent login and log out scenarios in case of Staff login using IDIR - **Describe the bug in current situation**
When a Staff user login in to the application, it uses a different login screen whereas when they log out the are navigated to welcome to cooperatives screen which is inconsistent
**Impact of this bug**
Staff user will have to copy paste the url again to login using IDIR as when they logout they are navigated to a different screen which is pretty inconsistent
**Chance of Occurring (high/medium/low/very low)**
High
**Pre Conditions: which Env, any pre-requesites or assumptions to execute steps?**
User is acting as Staff and loggin in with IDIR
**Steps to Reproduce**
1. For staff user login -https://test.bcregistry.ca/cooperatives/auth/signin/idir
2. Now enter the IDIR cred and login to access the staff dashbaord
3. Now logout from the application and notice where the user is navigated to
**Actual/ observed behavior/ results**
Staff is navigated to welcome to cooperatives online screen which is inconsistent
**Expected behavior**
Staff should be navigated to the same screen from where it logs in to avoid confusion
**Screenshots/ Visual Reference/ Source**
**Login Screen**

**Logout Screen**

FYI @saravankumarpa
|
non_defect
|
inconsistent login and log out scenarios in case of staff login using idir describe the bug in current situation when a staff user login in to the application it uses a different login screen whereas when they log out the are navigated to welcome to cooperatives screen which is inconsistent impact of this bug staff user will have to copy paste the url again to login using idir as when they logout they are navigated to a different screen which is pretty inconsistent chance of occurring high medium low very low high pre conditions which env any pre requesites or assumptions to execute steps user is acting as staff and loggin in with idir steps to reproduce for staff user login now enter the idir cred and login to access the staff dashbaord now logout from the application and notice where the user is navigated to actual observed behavior results staff is navigated to welcome to cooperatives online screen which is inconsistent expected behavior staff should be navigated to the same screen from where it logs in to avoid confusion screenshots visual reference source login screen logout screen fyi saravankumarpa
| 0
|
15,863
| 9,121,625,160
|
IssuesEvent
|
2019-02-23 00:22:44
|
Microsoft/azuredatastudio
|
https://api.github.com/repos/Microsoft/azuredatastudio
|
closed
|
Intellisense delayed
|
bug intellisense performance
|

<!-- Use Help > Report Issue to prefill these. -->
- Azure Data Studio Version: 1.3.9
Steps to Reproduce:
1. Download latest version of Azure Data Studio (1.3.9)
2. Connect to localhost SQL Server 2017
3. Open New Query Window
4. Type S
5. Intellisense has a clear delay
I have tried a fresh install and restarting computer, and refreshing intellisense cache. Computer is 16gb and 4 cores.
|
True
|
Intellisense delayed - 
<!-- Use Help > Report Issue to prefill these. -->
- Azure Data Studio Version: 1.3.9
Steps to Reproduce:
1. Download latest version of Azure Data Studio (1.3.9)
2. Connect to localhost SQL Server 2017
3. Open New Query Window
4. Type S
5. Intellisense has a clear delay
I have tried a fresh install and restarting computer, and refreshing intellisense cache. Computer is 16gb and 4 cores.
|
non_defect
|
intellisense delayed report issue to prefill these azure data studio version steps to reproduce download latest version of azure data studio connect to localhost sql server open new query window type s intellisense has a clear delay i have tried a fresh install and restarting computer and refreshing intellisense cache computer is and cores
| 0
|
26,923
| 4,832,631,732
|
IssuesEvent
|
2016-11-08 08:09:05
|
TNGSB/eWallet
|
https://api.github.com/repos/TNGSB/eWallet
|
opened
|
eWallet_MobileApp(Registration)_04112016 #021
|
Defect - High (Sev-2)
|
[Defect_Mobile App #021.xlsx](https://github.com/TNGSB/eWallet/files/577617/Defect_Mobile.App.021.xlsx)
Test Description : To verify the input of expired OTP - exceed 3 minutes after receiving the OTP
Expected Result : "System to display error message as it is pre-defined when user entered a expired OTP
Note: User need to set up PIN and click on ""Submit"" button in order to allow the system to do input validation"
Actual Result : System allow user to insert expired OTP and proceed with successful registration
Attached document for your perusal
|
1.0
|
eWallet_MobileApp(Registration)_04112016 #021 - [Defect_Mobile App #021.xlsx](https://github.com/TNGSB/eWallet/files/577617/Defect_Mobile.App.021.xlsx)
Test Description : To verify the input of expired OTP - exceed 3 minutes after receiving the OTP
Expected Result : "System to display error message as it is pre-defined when user entered a expired OTP
Note: User need to set up PIN and click on ""Submit"" button in order to allow the system to do input validation"
Actual Result : System allow user to insert expired OTP and proceed with successful registration
Attached document for your perusal
|
defect
|
ewallet mobileapp registration test description to verify the input of expired otp exceed minutes after receiving the otp expected result system to display error message as it is pre defined when user entered a expired otp note user need to set up pin and click on submit button in order to allow the system to do input validation actual result system allow user to insert expired otp and proceed with successful registration attached document for your perusal
| 1
|
79,387
| 28,142,252,875
|
IssuesEvent
|
2023-04-02 03:50:13
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
closed
|
Rocky/RHEL: Cannot boot from legacy root-filesystem - cmdline: root=ZFS=pool/ROOT/rocky-8.4
|
Type: Defect Status: Stale
|
<!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Rocky Linux
Distribution Version | 8.4
Kernel Version | 4.18
Architecture | x84_64
OpenZFS Version | 2.0.6
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
Mounting legacy root filesystems under Rocky-8.4 does not work. and I believe this is the case for all RHEL based systems
On Debian and Ubuntu I use **root=ZFS=pool/ROOT/ubuntu-20.04** on the cmdline to boot into Linux with root-on-ZFS, the dataset has the mountpoint set to legacy. I have been using this setup for 10 year. My ZFS-filesystem can hence be booted from different OS-es including FreeBSD and Illumos in the past.
Now that RHEL is available for ordinary people again via Rocky 8.4 I thought I would check it out. Unfortunately, I could not boot Rocky Linux from a legacy-mounted dataset.
The culprit is dracut (initramfs) which always want to mount the rootfs-dataset with **-o zfutils**, which again means you cannot use legacy mount.
The Rocky-8.4 version of **/lib/dracut/modules.d/90zfs** seems to be an older version of your **zfs/contrib/dracut/90zfs/**
I have described how I got Rocky-Linux to bringup my ZFS-file-system on my blog http://peter.lorenzen.us/zfs/installing-rocky-8-4-rhel-into-a-live-zfs-filesystem
To get Rocky-84 to boot from a legacy dataset I had to apply the patch below, on the Rocky root-file-system, I believe the patch would also work on your current version, although the proper way to implement this should rely on access to the online pool/dataset (which is not the case when zfs-generator.sh is called now).
```
--- /lib/dracut/modules.d/90zfs/zfs-generator.sh.orig 2021-09-16 20:22:33.000000000 -0400
+++ /lib/dracut/modules.d/90zfs/zfs-generator.sh 2021-10-29 19:34:19.271560868 -0400
@@ -28,11 +28,26 @@
exit 0
rootfstype=zfs
-case ",${rootflags}," in
- *,zfsutil,*) ;;
- ,,) rootflags=zfsutil ;;
- *) rootflags="zfsutil,${rootflags}" ;;
-esac
+mountpoint=""
+if [ x"$(zpool list -Ho name)" = "x" ]; then
+ echo "zfs-generator: zpool is not imported yet so we cannot see if mountpoint is legacy" >> /
dev/kmsg
+ if [ "${root##ZFS=}" != "$root" ]; then
+ echo "zfs-generator: root=$root - assume legacy if root=ZFS=" >> /dev/kmsg
+ mountpoint="legacy"
+ fi
+else
+ mountpoint="$(zfs get -Ho value mountpoint ${root##ZFS=})"
+fi
+if [ "$mountpoint" = legacy ]; then
+ echo "zfs-generator: $root is legacy mount" >> /dev/kmsg
+else
+ echo "zfs-generator: $root is not legacy setting zfutil" >> /dev/kmsg
+ case ",${rootflags}," in
+ *,zfsutil,*) ;;
+ ,,) rootflags=zfsutil ;;
+ *) rootflags="zfsutil,${rootflags}" ;;
+ esac
+fi
```
|
1.0
|
Rocky/RHEL: Cannot boot from legacy root-filesystem - cmdline: root=ZFS=pool/ROOT/rocky-8.4 - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Rocky Linux
Distribution Version | 8.4
Kernel Version | 4.18
Architecture | x84_64
OpenZFS Version | 2.0.6
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
Mounting legacy root filesystems under Rocky-8.4 does not work. and I believe this is the case for all RHEL based systems
On Debian and Ubuntu I use **root=ZFS=pool/ROOT/ubuntu-20.04** on the cmdline to boot into Linux with root-on-ZFS, the dataset has the mountpoint set to legacy. I have been using this setup for 10 year. My ZFS-filesystem can hence be booted from different OS-es including FreeBSD and Illumos in the past.
Now that RHEL is available for ordinary people again via Rocky 8.4 I thought I would check it out. Unfortunately, I could not boot Rocky Linux from a legacy-mounted dataset.
The culprit is dracut (initramfs) which always want to mount the rootfs-dataset with **-o zfutils**, which again means you cannot use legacy mount.
The Rocky-8.4 version of **/lib/dracut/modules.d/90zfs** seems to be an older version of your **zfs/contrib/dracut/90zfs/**
I have described how I got Rocky-Linux to bringup my ZFS-file-system on my blog http://peter.lorenzen.us/zfs/installing-rocky-8-4-rhel-into-a-live-zfs-filesystem
To get Rocky-84 to boot from a legacy dataset I had to apply the patch below, on the Rocky root-file-system, I believe the patch would also work on your current version, although the proper way to implement this should rely on access to the online pool/dataset (which is not the case when zfs-generator.sh is called now).
```
--- /lib/dracut/modules.d/90zfs/zfs-generator.sh.orig 2021-09-16 20:22:33.000000000 -0400
+++ /lib/dracut/modules.d/90zfs/zfs-generator.sh 2021-10-29 19:34:19.271560868 -0400
@@ -28,11 +28,26 @@
exit 0
rootfstype=zfs
-case ",${rootflags}," in
- *,zfsutil,*) ;;
- ,,) rootflags=zfsutil ;;
- *) rootflags="zfsutil,${rootflags}" ;;
-esac
+mountpoint=""
+if [ x"$(zpool list -Ho name)" = "x" ]; then
+ echo "zfs-generator: zpool is not imported yet so we cannot see if mountpoint is legacy" >> /
dev/kmsg
+ if [ "${root##ZFS=}" != "$root" ]; then
+ echo "zfs-generator: root=$root - assume legacy if root=ZFS=" >> /dev/kmsg
+ mountpoint="legacy"
+ fi
+else
+ mountpoint="$(zfs get -Ho value mountpoint ${root##ZFS=})"
+fi
+if [ "$mountpoint" = legacy ]; then
+ echo "zfs-generator: $root is legacy mount" >> /dev/kmsg
+else
+ echo "zfs-generator: $root is not legacy setting zfutil" >> /dev/kmsg
+ case ",${rootflags}," in
+ *,zfsutil,*) ;;
+ ,,) rootflags=zfsutil ;;
+ *) rootflags="zfsutil,${rootflags}" ;;
+ esac
+fi
```
|
defect
|
rocky rhel cannot boot from legacy root filesystem cmdline root zfs pool root rocky thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name rocky linux distribution version kernel version architecture openzfs version command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd mounting legacy root filesystems under rocky does not work and i believe this is the case for all rhel based systems on debian and ubuntu i use root zfs pool root ubuntu on the cmdline to boot into linux with root on zfs the dataset has the mountpoint set to legacy i have been using this setup for year my zfs filesystem can hence be booted from different os es including freebsd and illumos in the past now that rhel is available for ordinary people again via rocky i thought i would check it out unfortunately i could not boot rocky linux from a legacy mounted dataset the culprit is dracut initramfs which always want to mount the rootfs dataset with o zfutils which again means you cannot use legacy mount the rocky version of lib dracut modules d seems to be an older version of your zfs contrib dracut i have described how i got rocky linux to bringup my zfs file system on my blog to get rocky to boot from a legacy dataset i had to apply the patch below on the rocky root file system i believe the patch would also work on your current version although the proper way to implement this should rely on access to the online pool dataset which is not the case when zfs generator sh is called now lib dracut modules d zfs generator sh orig lib dracut modules d zfs generator sh exit rootfstype zfs case rootflags in zfsutil rootflags zfsutil rootflags zfsutil rootflags esac mountpoint if then echo zfs generator zpool is not imported yet so we cannot see if mountpoint is legacy dev kmsg if then echo zfs generator root root assume legacy if root zfs dev kmsg mountpoint legacy fi else mountpoint zfs get ho value mountpoint root zfs fi if then echo zfs generator root is legacy mount dev kmsg else echo zfs generator root is not legacy setting zfutil dev kmsg case rootflags in zfsutil rootflags zfsutil rootflags zfsutil rootflags esac fi
| 1
|
79,987
| 29,810,964,173
|
IssuesEvent
|
2023-06-16 15:01:45
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
opened
|
Component: Title pTooltip not working with version 16.0.0
|
:lady_beetle: defect :bangbang: needs-triage
|
### Describe the bug
There is not any tooltip when directive "pTooltip" is added. You can see the error in console - Uncaught Cannot append null to [object HTMLDivElement]
### Reproducer
This is demo
[https://codesandbox.io/s/rykwlw?file=/package.json:1854-1861](url)
### Expected behavior
There is tooltip when the directive "pTooltip" is added
### PrimeFaces edition
Community
### PrimeFaces version
16.0.0
### Theme
saga-blue
### JSF implementation
Mojarra
### JSF version
19.0.2
### Java version
19.0.2
### Browser(s)
Version 114.0.5735.134 (Official Build) (64-bit)
|
1.0
|
Component: Title pTooltip not working with version 16.0.0 - ### Describe the bug
There is not any tooltip when directive "pTooltip" is added. You can see the error in console - Uncaught Cannot append null to [object HTMLDivElement]
### Reproducer
This is demo
[https://codesandbox.io/s/rykwlw?file=/package.json:1854-1861](url)
### Expected behavior
There is tooltip when the directive "pTooltip" is added
### PrimeFaces edition
Community
### PrimeFaces version
16.0.0
### Theme
saga-blue
### JSF implementation
Mojarra
### JSF version
19.0.2
### Java version
19.0.2
### Browser(s)
Version 114.0.5735.134 (Official Build) (64-bit)
|
defect
|
component title ptooltip not working with version describe the bug there is not any tooltip when directive ptooltip is added you can see the error in console uncaught cannot append null to reproducer this is demo url expected behavior there is tooltip when the directive ptooltip is added primefaces edition community primefaces version theme saga blue jsf implementation mojarra jsf version java version browser s version official build bit
| 1
|
76,517
| 26,473,566,903
|
IssuesEvent
|
2023-01-17 09:26:35
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Delete in kotlin coroutine reactive transaction fails silently
|
T: Defect C: Functionality P: High E: All Editions
|
Deletes issued within a reactive transaction, using the Kotlin extension `DSLContext.transactionCoroutine` fail silently; no exception is thrown, the expected record count is returned, but records are not deleted. It's entirely possible that I'm doing something wrong, but I can't find it. The repository at https://github.com/wfhartford/jooq-already-resumed is a simple reproduction, and explains more in the README.
### Expected behavior
A delete issued in a reactive transaction from Kotlin should delete matching records
### Actual behavior
The delete appears to succeed, returning the expected number of records, but subsequent operations show that the records were not deleted. The reactor package logs many errors and warnings.
### Steps to reproduce the problem
I was unaware of the jOOQ MCVE project, so do not have a reproduction sample in that form, but there is a self-contained reproduction project at https://github.com/wfhartford/jooq-already-resumed. If a reproduction based on the MCVE project would be more helpful, I'd be happy to create that as well.
### Versions
- jOOQ: 3.17.2
- Java: openjdk version "17.0.3" 2022-04-19
- Database (include vendor): postgresql 14.2
- OS: Linux Pop!_OS 22.04 LTS
- JDBC Driver (include name if unofficial driver): org.postgresql:r2dbc-postgresql:0.9.1.RELEASE
- Kotlin: 1.6.20
- Kotlin Coroutines: 1.6.2
|
1.0
|
Delete in kotlin coroutine reactive transaction fails silently - Deletes issued within a reactive transaction, using the Kotlin extension `DSLContext.transactionCoroutine` fail silently; no exception is thrown, the expected record count is returned, but records are not deleted. It's entirely possible that I'm doing something wrong, but I can't find it. The repository at https://github.com/wfhartford/jooq-already-resumed is a simple reproduction, and explains more in the README.
### Expected behavior
A delete issued in a reactive transaction from Kotlin should delete matching records
### Actual behavior
The delete appears to succeed, returning the expected number of records, but subsequent operations show that the records were not deleted. The reactor package logs many errors and warnings.
### Steps to reproduce the problem
I was unaware of the jOOQ MCVE project, so do not have a reproduction sample in that form, but there is a self-contained reproduction project at https://github.com/wfhartford/jooq-already-resumed. If a reproduction based on the MCVE project would be more helpful, I'd be happy to create that as well.
### Versions
- jOOQ: 3.17.2
- Java: openjdk version "17.0.3" 2022-04-19
- Database (include vendor): postgresql 14.2
- OS: Linux Pop!_OS 22.04 LTS
- JDBC Driver (include name if unofficial driver): org.postgresql:r2dbc-postgresql:0.9.1.RELEASE
- Kotlin: 1.6.20
- Kotlin Coroutines: 1.6.2
|
defect
|
delete in kotlin coroutine reactive transaction fails silently deletes issued within a reactive transaction using the kotlin extension dslcontext transactioncoroutine fail silently no exception is thrown the expected record count is returned but records are not deleted it s entirely possible that i m doing something wrong but i can t find it the repository at is a simple reproduction and explains more in the readme expected behavior a delete issued in a reactive transaction from kotlin should delete matching records actual behavior the delete appears to succeed returning the expected number of records but subsequent operations show that the records were not deleted the reactor package logs many errors and warnings steps to reproduce the problem i was unaware of the jooq mcve project so do not have a reproduction sample in that form but there is a self contained reproduction project at if a reproduction based on the mcve project would be more helpful i d be happy to create that as well versions jooq java openjdk version database include vendor postgresql os linux pop os lts jdbc driver include name if unofficial driver org postgresql postgresql release kotlin kotlin coroutines
| 1
|
30,715
| 6,255,681,966
|
IssuesEvent
|
2017-07-14 07:58:05
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
[reliable-topic] [ringbuffer] NullPointerException ReliableTopic / Ringbuffer init
|
Team: Core Type: Critical Type: Defect
|
in this multi-version job
https://hazelcast-l337.ci.cloudbees.com/view/multi-version/job/multi-version-stable-All/10/console
```
/disk1/jenkins/workspace/multi-version-stable-All/3.9-SNAPSHOT/2017_06_15-13_19_49/stable/destroy/create-use-dist-destroy Failed
```
http://54.163.63.218/~jenkins/workspace/multi-version-stable-All/3.9-SNAPSHOT/2017_06_15-13_19_49/stable/destroy/create-use-dist-destroy
cluster versions
```
20:22:14 HzMember2HZ Started on 10.0.0.25 797@ip-10-0-0-25 /home/ec2-user/hz-root/hz-lib/3.8
20:22:14 HzMember1HZ Started on 10.0.0.140 1101@ip-10-0-0-140 /home/ec2-user/hz-root/hz-lib/3.8
20:22:14 HzClient1HZ Started on 10.0.0.140 1222@ip-10-0-0-140 /home/ec2-user/hz-root/hz-lib/3.8
20:22:14 HzClient2HZ Started on 10.0.0.25 918@ip-10-0-0-25 /home/ec2-user/hz-root/hz-lib/3.8
20:22:14 HzMember4HZ Started on 10.0.0.5 793@ip-10-0-0-5 /home/ec2-user/hz-root/hz-lib/3.8.1
20:22:14 HzMember3HZ Started on 10.0.0.223 1117@ip-10-0-0-223 /home/ec2-user/hz-root/hz-lib/3.8.1
20:22:14 HzClient3HZ Started on 10.0.0.223 1244@ip-10-0-0-223 /home/ec2-user/hz-root/hz-lib/3.8.1
20:22:14 HzClient4HZ Started on 10.0.0.5 918@ip-10-0-0-5 /home/ec2-user/hz-root/hz-lib/3.8.1
20:22:14 HzMember6HZ Started on 10.0.0.165 1341@ip-10-0-0-165 /home/ec2-user/hz-root/hz-lib/3.8.2
20:22:14 HzMember5HZ Started on 10.0.0.35 1166@ip-10-0-0-35 /home/ec2-user/hz-root/hz-lib/3.8.2
20:22:14 HzClient5HZ Started on 10.0.0.35 1293@ip-10-0-0-35 /home/ec2-user/hz-root/hz-lib/3.8.2
20:22:14 HzClient6HZ Started on 10.0.0.165 1469@ip-10-0-0-165 /home/ec2-user/hz-root/hz-lib/3.8.2
20:22:14 HzMember7HZ Started on 10.0.0.85 1532@ip-10-0-0-85 /home/ec2-user/hz-root/hz-lib/3.9-SNAPSHOT
20:22:14 HzMember8HZ Started on 10.0.0.33 965@ip-10-0-0-33 /home/ec2-user/hz-root/hz-lib/3.9-SNAPSHOT
20:22:14 HzClient8HZ Started on 10.0.0.33 1090@ip-10-0-0-33 /home/ec2-user/hz-root/hz-lib/3.9-SNAPSHOT
20:22:14 HzClient7HZ Started on 10.0.0.85 1660@ip-10-0-0-85 /home/ec2-user/hz-root/hz-lib/3.9-SNAPSHOT
```
HzMember4HZ _create-use-dist-destroy_createUseDistDestroy_reliableTopic hzcmd.topic.CreateUse threadId=0 java.lang.NullPointerException
http://54.163.63.218/~jenkins/workspace/multi-version-stable-All/3.9-SNAPSHOT/2017_06_15-13_19_49/stable/destroy/create-use-dist-destroy/output/HZ/HzMember4HZ/exception.txt
```
java.lang.NullPointerException
at com.hazelcast.ringbuffer.impl.RingbufferContainer.<init>(RingbufferContainer.java:99)
at com.hazelcast.ringbuffer.impl.RingbufferService.getContainer(RingbufferService.java:159)
at com.hazelcast.ringbuffer.impl.operations.AbstractRingBufferOperation.getRingBufferContainer(AbstractRingBufferOperation.java:78)
at com.hazelcast.ringbuffer.impl.operations.AddOperation.run(AddOperation.java:56)
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:194)
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:409)
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:115)
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.run(OperationThread.java:100)
at ------ submitted from ------.(Unknown Source)
at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolve(InvocationFuture.java:114)
at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolveAndThrowIfException(InvocationFuture.java:75)
at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:155)
at com.hazelcast.topic.impl.reliable.ReliableTopicProxy.addWithBackoff(ReliableTopicProxy.java:190)
at com.hazelcast.topic.impl.reliable.ReliableTopicProxy.publish(ReliableTopicProxy.java:163)
at hzcmd.topic.CreateUse.timeStep(CreateUse.java:21)
at remote.bench.marker.MetricsMarker.flatOut(MetricsMarker.java:53)
at remote.bench.marker.MetricsMarker.bench(MetricsMarker.java:40)
at remote.bench.BenchThread.call(BenchThread.java:38)
at remote.bench.BenchThread.call(BenchThread.java:12)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:473)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:748)
```
|
1.0
|
[reliable-topic] [ringbuffer] NullPointerException ReliableTopic / Ringbuffer init -
in this multi-version job
https://hazelcast-l337.ci.cloudbees.com/view/multi-version/job/multi-version-stable-All/10/console
```
/disk1/jenkins/workspace/multi-version-stable-All/3.9-SNAPSHOT/2017_06_15-13_19_49/stable/destroy/create-use-dist-destroy Failed
```
http://54.163.63.218/~jenkins/workspace/multi-version-stable-All/3.9-SNAPSHOT/2017_06_15-13_19_49/stable/destroy/create-use-dist-destroy
cluster versions
```
20:22:14 HzMember2HZ Started on 10.0.0.25 797@ip-10-0-0-25 /home/ec2-user/hz-root/hz-lib/3.8
20:22:14 HzMember1HZ Started on 10.0.0.140 1101@ip-10-0-0-140 /home/ec2-user/hz-root/hz-lib/3.8
20:22:14 HzClient1HZ Started on 10.0.0.140 1222@ip-10-0-0-140 /home/ec2-user/hz-root/hz-lib/3.8
20:22:14 HzClient2HZ Started on 10.0.0.25 918@ip-10-0-0-25 /home/ec2-user/hz-root/hz-lib/3.8
20:22:14 HzMember4HZ Started on 10.0.0.5 793@ip-10-0-0-5 /home/ec2-user/hz-root/hz-lib/3.8.1
20:22:14 HzMember3HZ Started on 10.0.0.223 1117@ip-10-0-0-223 /home/ec2-user/hz-root/hz-lib/3.8.1
20:22:14 HzClient3HZ Started on 10.0.0.223 1244@ip-10-0-0-223 /home/ec2-user/hz-root/hz-lib/3.8.1
20:22:14 HzClient4HZ Started on 10.0.0.5 918@ip-10-0-0-5 /home/ec2-user/hz-root/hz-lib/3.8.1
20:22:14 HzMember6HZ Started on 10.0.0.165 1341@ip-10-0-0-165 /home/ec2-user/hz-root/hz-lib/3.8.2
20:22:14 HzMember5HZ Started on 10.0.0.35 1166@ip-10-0-0-35 /home/ec2-user/hz-root/hz-lib/3.8.2
20:22:14 HzClient5HZ Started on 10.0.0.35 1293@ip-10-0-0-35 /home/ec2-user/hz-root/hz-lib/3.8.2
20:22:14 HzClient6HZ Started on 10.0.0.165 1469@ip-10-0-0-165 /home/ec2-user/hz-root/hz-lib/3.8.2
20:22:14 HzMember7HZ Started on 10.0.0.85 1532@ip-10-0-0-85 /home/ec2-user/hz-root/hz-lib/3.9-SNAPSHOT
20:22:14 HzMember8HZ Started on 10.0.0.33 965@ip-10-0-0-33 /home/ec2-user/hz-root/hz-lib/3.9-SNAPSHOT
20:22:14 HzClient8HZ Started on 10.0.0.33 1090@ip-10-0-0-33 /home/ec2-user/hz-root/hz-lib/3.9-SNAPSHOT
20:22:14 HzClient7HZ Started on 10.0.0.85 1660@ip-10-0-0-85 /home/ec2-user/hz-root/hz-lib/3.9-SNAPSHOT
```
HzMember4HZ _create-use-dist-destroy_createUseDistDestroy_reliableTopic hzcmd.topic.CreateUse threadId=0 java.lang.NullPointerException
http://54.163.63.218/~jenkins/workspace/multi-version-stable-All/3.9-SNAPSHOT/2017_06_15-13_19_49/stable/destroy/create-use-dist-destroy/output/HZ/HzMember4HZ/exception.txt
```
java.lang.NullPointerException
at com.hazelcast.ringbuffer.impl.RingbufferContainer.<init>(RingbufferContainer.java:99)
at com.hazelcast.ringbuffer.impl.RingbufferService.getContainer(RingbufferService.java:159)
at com.hazelcast.ringbuffer.impl.operations.AbstractRingBufferOperation.getRingBufferContainer(AbstractRingBufferOperation.java:78)
at com.hazelcast.ringbuffer.impl.operations.AddOperation.run(AddOperation.java:56)
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:194)
at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:409)
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:115)
at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.run(OperationThread.java:100)
at ------ submitted from ------.(Unknown Source)
at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolve(InvocationFuture.java:114)
at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolveAndThrowIfException(InvocationFuture.java:75)
at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:155)
at com.hazelcast.topic.impl.reliable.ReliableTopicProxy.addWithBackoff(ReliableTopicProxy.java:190)
at com.hazelcast.topic.impl.reliable.ReliableTopicProxy.publish(ReliableTopicProxy.java:163)
at hzcmd.topic.CreateUse.timeStep(CreateUse.java:21)
at remote.bench.marker.MetricsMarker.flatOut(MetricsMarker.java:53)
at remote.bench.marker.MetricsMarker.bench(MetricsMarker.java:40)
at remote.bench.BenchThread.call(BenchThread.java:38)
at remote.bench.BenchThread.call(BenchThread.java:12)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:473)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:748)
```
|
defect
|
nullpointerexception reliabletopic ringbuffer init in this multi version job jenkins workspace multi version stable all snapshot stable destroy create use dist destroy failed cluster versions started on ip home user hz root hz lib started on ip home user hz root hz lib started on ip home user hz root hz lib started on ip home user hz root hz lib started on ip home user hz root hz lib started on ip home user hz root hz lib started on ip home user hz root hz lib started on ip home user hz root hz lib started on ip home user hz root hz lib started on ip home user hz root hz lib started on ip home user hz root hz lib started on ip home user hz root hz lib started on ip home user hz root hz lib snapshot started on ip home user hz root hz lib snapshot started on ip home user hz root hz lib snapshot started on ip home user hz root hz lib snapshot create use dist destroy createusedistdestroy reliabletopic hzcmd topic createuse threadid java lang nullpointerexception java lang nullpointerexception at com hazelcast ringbuffer impl ringbuffercontainer ringbuffercontainer java at com hazelcast ringbuffer impl ringbufferservice getcontainer ringbufferservice java at com hazelcast ringbuffer impl operations abstractringbufferoperation getringbuffercontainer abstractringbufferoperation java at com hazelcast ringbuffer impl operations addoperation run addoperation java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationexecutor impl operationthread process operationthread java at com hazelcast spi impl operationexecutor impl operationthread run operationthread java at submitted from unknown source at com hazelcast spi impl operationservice impl invocationfuture resolve invocationfuture java at com hazelcast spi impl operationservice impl invocationfuture resolveandthrowifexception invocationfuture java at com hazelcast spi impl abstractinvocationfuture get abstractinvocationfuture java at com hazelcast topic impl reliable reliabletopicproxy addwithbackoff reliabletopicproxy java at com hazelcast topic impl reliable reliabletopicproxy publish reliabletopicproxy java at hzcmd topic createuse timestep createuse java at remote bench marker metricsmarker flatout metricsmarker java at remote bench marker metricsmarker bench metricsmarker java at remote bench benchthread call benchthread java at remote bench benchthread call benchthread java at java util concurrent futuretask run futuretask java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java
| 1
|
317,309
| 9,663,219,589
|
IssuesEvent
|
2019-05-20 23:33:55
|
RobotLocomotion/drake
|
https://api.github.com/repos/RobotLocomotion/drake
|
closed
|
direct feedthrough behavior on VectorSystem and CompassGait caused confusion
|
priority: medium team: dynamics type: bug
|
From a student
> Hi, I'm getting a really weird error with the compass gait. After creating a controller and connecting the compass gait's output port to the controller's input port, I printed out the input to the controller's DoCalcVectorOutput function, but the input was empty. I've tried modeling the code after the code in the psets. Is there any reason why this should behave differently? (code below)
It turns out that a few things were wrong:
1) he was declaring the controller as direct_feedthrough false... which meant that VectorSystem was not writing the input values to inp. This is reasonable behavior, but i can see that it was very confusing (took me a few minutes and some printfs to realize it, too). Is there anything we can do to make that better?
2) the real problem was that my commit https://github.com/RobotLocomotion/drake/pull/11472 -- which added an input to the compass gait -- the automatic direct feedthrough detection seems to have failed. I will add the proper dependency ticket now, but i think that the symbolic deduction should have gotten this one correct.
```
import argparse
import math
import numpy as np
from pydrake.all import (Box,
DiagramBuilder,
FindResourceOrThrow,
FloatingBaseType,
Isometry3,
RigidBodyTree,
SignalLogger,
Simulator,
VisualElement,
PortDataType,
LeafSystem,
VectorSystem)
from pydrake.examples.compass_gait import (CompassGait, CompassGaitParams)
from underactuated import (PlanarRigidBodyVisualizer)
class CompassGaitController(VectorSystem):
def __init__(self, compass_gait):
self.compass_gait = compass_gait
self.num_states = compass_gait.get_output_port(0).size()
self.compass_gait_context = self.compass_gait.CreateDefaultContext()
VectorSystem.__init__(self, self.num_states, 1, direct_feedthrough = False)
def _DoCalcVectorOutput(self, context, inp, state, out):
self.compass_gait_context = context
print_stuff = True
if print_stuff:
print("inp", inp)
print("state", state)
print("out", out)
out[0] = 0
return 0
def Simulate2dCompassGait(duration):
tree = RigidBodyTree(FindResourceOrThrow(
"drake/examples/compass_gait/CompassGait.urdf"),
FloatingBaseType.kRollPitchYaw)
params = CompassGaitParams()
R = np.identity(3)
R[0, 0] = math.cos(params.slope())
R[0, 2] = math.sin(params.slope())
R[2, 0] = -math.sin(params.slope())
R[2, 2] = math.cos(params.slope())
X = Isometry3(rotation=R, translation=[0, 0, -5.])
color = np.array([0.9297, 0.7930, 0.6758, 1])
tree.world().AddVisualElement(VisualElement(Box([100., 1., 10.]), X, color))
tree.compile()
# initialize builder and compass gait plant
builder = DiagramBuilder()
compass_gait = builder.AddSystem(CompassGait())
# Create a logger to log at 30hz
state_dim = compass_gait.get_output_port(1).size()
state_log = builder.AddSystem(SignalLogger(state_dim))
state_log.DeclarePeriodicPublish(0.0333, 0.0) # 30hz logging
builder.Connect(compass_gait.get_output_port(1), state_log.get_input_port(0))
# Create a controller
controller = builder.AddSystem(
CompassGaitController(compass_gait))
builder.Connect(compass_gait.get_output_port(0), controller.get_input_port(0))
builder.Connect(controller.get_output_port(0), compass_gait.get_input_port(0))
visualizer = builder.AddSystem(PlanarRigidBodyVisualizer(tree,
xlim=[-1., 5.],
ylim=[-1., 2.],
figsize_multiplier=2))
builder.Connect(compass_gait.get_output_port(1), visualizer.get_input_port(0))
#print(compass_gait.get_input_port(0))
diagram = builder.Build()
simulator = Simulator(diagram)
simulator.Initialize()
simulator.set_target_realtime_rate(1.0)
context = simulator.get_mutable_context()
diagram.Publish(context) # draw once to get the window open
context.set_accuracy(1e-4)
context.SetContinuousState([0., 0., 0.4, -2.])
#context.SetContinuousState([0, 0, 1, -3.])
simulator.StepTo(duration)
return visualizer, state_log
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument("-T", "--duration",
type=float,
help="Duration to run sim.",
default=10.0)
args = parser.parse_args()
Simulate2dCompassGait(args.duration)
```
|
1.0
|
direct feedthrough behavior on VectorSystem and CompassGait caused confusion - From a student
> Hi, I'm getting a really weird error with the compass gait. After creating a controller and connecting the compass gait's output port to the controller's input port, I printed out the input to the controller's DoCalcVectorOutput function, but the input was empty. I've tried modeling the code after the code in the psets. Is there any reason why this should behave differently? (code below)
It turns out that a few things were wrong:
1) he was declaring the controller as direct_feedthrough false... which meant that VectorSystem was not writing the input values to inp. This is reasonable behavior, but i can see that it was very confusing (took me a few minutes and some printfs to realize it, too). Is there anything we can do to make that better?
2) the real problem was that my commit https://github.com/RobotLocomotion/drake/pull/11472 -- which added an input to the compass gait -- the automatic direct feedthrough detection seems to have failed. I will add the proper dependency ticket now, but i think that the symbolic deduction should have gotten this one correct.
```
import argparse
import math
import numpy as np
from pydrake.all import (Box,
DiagramBuilder,
FindResourceOrThrow,
FloatingBaseType,
Isometry3,
RigidBodyTree,
SignalLogger,
Simulator,
VisualElement,
PortDataType,
LeafSystem,
VectorSystem)
from pydrake.examples.compass_gait import (CompassGait, CompassGaitParams)
from underactuated import (PlanarRigidBodyVisualizer)
class CompassGaitController(VectorSystem):
def __init__(self, compass_gait):
self.compass_gait = compass_gait
self.num_states = compass_gait.get_output_port(0).size()
self.compass_gait_context = self.compass_gait.CreateDefaultContext()
VectorSystem.__init__(self, self.num_states, 1, direct_feedthrough = False)
def _DoCalcVectorOutput(self, context, inp, state, out):
self.compass_gait_context = context
print_stuff = True
if print_stuff:
print("inp", inp)
print("state", state)
print("out", out)
out[0] = 0
return 0
def Simulate2dCompassGait(duration):
tree = RigidBodyTree(FindResourceOrThrow(
"drake/examples/compass_gait/CompassGait.urdf"),
FloatingBaseType.kRollPitchYaw)
params = CompassGaitParams()
R = np.identity(3)
R[0, 0] = math.cos(params.slope())
R[0, 2] = math.sin(params.slope())
R[2, 0] = -math.sin(params.slope())
R[2, 2] = math.cos(params.slope())
X = Isometry3(rotation=R, translation=[0, 0, -5.])
color = np.array([0.9297, 0.7930, 0.6758, 1])
tree.world().AddVisualElement(VisualElement(Box([100., 1., 10.]), X, color))
tree.compile()
# initialize builder and compass gait plant
builder = DiagramBuilder()
compass_gait = builder.AddSystem(CompassGait())
# Create a logger to log at 30hz
state_dim = compass_gait.get_output_port(1).size()
state_log = builder.AddSystem(SignalLogger(state_dim))
state_log.DeclarePeriodicPublish(0.0333, 0.0) # 30hz logging
builder.Connect(compass_gait.get_output_port(1), state_log.get_input_port(0))
# Create a controller
controller = builder.AddSystem(
CompassGaitController(compass_gait))
builder.Connect(compass_gait.get_output_port(0), controller.get_input_port(0))
builder.Connect(controller.get_output_port(0), compass_gait.get_input_port(0))
visualizer = builder.AddSystem(PlanarRigidBodyVisualizer(tree,
xlim=[-1., 5.],
ylim=[-1., 2.],
figsize_multiplier=2))
builder.Connect(compass_gait.get_output_port(1), visualizer.get_input_port(0))
#print(compass_gait.get_input_port(0))
diagram = builder.Build()
simulator = Simulator(diagram)
simulator.Initialize()
simulator.set_target_realtime_rate(1.0)
context = simulator.get_mutable_context()
diagram.Publish(context) # draw once to get the window open
context.set_accuracy(1e-4)
context.SetContinuousState([0., 0., 0.4, -2.])
#context.SetContinuousState([0, 0, 1, -3.])
simulator.StepTo(duration)
return visualizer, state_log
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument("-T", "--duration",
type=float,
help="Duration to run sim.",
default=10.0)
args = parser.parse_args()
Simulate2dCompassGait(args.duration)
```
|
non_defect
|
direct feedthrough behavior on vectorsystem and compassgait caused confusion from a student hi i m getting a really weird error with the compass gait after creating a controller and connecting the compass gait s output port to the controller s input port i printed out the input to the controller s docalcvectoroutput function but the input was empty i ve tried modeling the code after the code in the psets is there any reason why this should behave differently code below it turns out that a few things were wrong he was declaring the controller as direct feedthrough false which meant that vectorsystem was not writing the input values to inp this is reasonable behavior but i can see that it was very confusing took me a few minutes and some printfs to realize it too is there anything we can do to make that better the real problem was that my commit which added an input to the compass gait the automatic direct feedthrough detection seems to have failed i will add the proper dependency ticket now but i think that the symbolic deduction should have gotten this one correct import argparse import math import numpy as np from pydrake all import box diagrambuilder findresourceorthrow floatingbasetype rigidbodytree signallogger simulator visualelement portdatatype leafsystem vectorsystem from pydrake examples compass gait import compassgait compassgaitparams from underactuated import planarrigidbodyvisualizer class compassgaitcontroller vectorsystem def init self compass gait self compass gait compass gait self num states compass gait get output port size self compass gait context self compass gait createdefaultcontext vectorsystem init self self num states direct feedthrough false def docalcvectoroutput self context inp state out self compass gait context context print stuff true if print stuff print inp inp print state state print out out out return def duration tree rigidbodytree findresourceorthrow drake examples compass gait compassgait urdf floatingbasetype krollpitchyaw params compassgaitparams r np identity r math cos params slope r math sin params slope r math sin params slope r math cos params slope x rotation r translation color np array tree world addvisualelement visualelement box x color tree compile initialize builder and compass gait plant builder diagrambuilder compass gait builder addsystem compassgait create a logger to log at state dim compass gait get output port size state log builder addsystem signallogger state dim state log declareperiodicpublish logging builder connect compass gait get output port state log get input port create a controller controller builder addsystem compassgaitcontroller compass gait builder connect compass gait get output port controller get input port builder connect controller get output port compass gait get input port visualizer builder addsystem planarrigidbodyvisualizer tree xlim ylim figsize multiplier builder connect compass gait get output port visualizer get input port print compass gait get input port diagram builder build simulator simulator diagram simulator initialize simulator set target realtime rate context simulator get mutable context diagram publish context draw once to get the window open context set accuracy context setcontinuousstate context setcontinuousstate simulator stepto duration return visualizer state log if name main parser argparse argumentparser parser add argument t duration type float help duration to run sim default args parser parse args args duration
| 0
|
439,464
| 12,683,083,875
|
IssuesEvent
|
2020-06-19 18:53:26
|
SD2E/experimental-intent-parser
|
https://api.github.com/repos/SD2E/experimental-intent-parser
|
closed
|
Figure out how to handle coming across a term that is already linked
|
priority #3
|
In GitLab by @jakebeal on Feb 1, 2019, 20:57
If a term is already linked, what should we do with it?
Some possibilities:
- Ignore the term
- Ignore the link and offer to replace it
- Check if the link is to a definition source (initially SynBioHub, but there might be more later) and if not, ask if they want to replace it.
- Something else?
|
1.0
|
Figure out how to handle coming across a term that is already linked - In GitLab by @jakebeal on Feb 1, 2019, 20:57
If a term is already linked, what should we do with it?
Some possibilities:
- Ignore the term
- Ignore the link and offer to replace it
- Check if the link is to a definition source (initially SynBioHub, but there might be more later) and if not, ask if they want to replace it.
- Something else?
|
non_defect
|
figure out how to handle coming across a term that is already linked in gitlab by jakebeal on feb if a term is already linked what should we do with it some possibilities ignore the term ignore the link and offer to replace it check if the link is to a definition source initially synbiohub but there might be more later and if not ask if they want to replace it something else
| 0
|
35,247
| 7,673,974,609
|
IssuesEvent
|
2018-05-15 01:05:22
|
adamhope/testing-github-and-projects
|
https://api.github.com/repos/adamhope/testing-github-and-projects
|
closed
|
Error occurs when exporting Prod Schedule from Showcase and QA
|
Defect campaign delivery
|
https://trello.com/c/uK95iHLG/32-error-occurs-when-exporting-prod-schedule-from-showcase-and-qa
You are currently not able to export the Production Schedule from both the QA and Showcase OTP environments. An error occurs saying it failed to fetch installers. See attached screenshot.
Note: This works in the OTP Production environment even after a recent release so it is possibly caused by changes to the inventory service or CRM?
|
1.0
|
Error occurs when exporting Prod Schedule from Showcase and QA - https://trello.com/c/uK95iHLG/32-error-occurs-when-exporting-prod-schedule-from-showcase-and-qa
You are currently not able to export the Production Schedule from both the QA and Showcase OTP environments. An error occurs saying it failed to fetch installers. See attached screenshot.
Note: This works in the OTP Production environment even after a recent release so it is possibly caused by changes to the inventory service or CRM?
|
defect
|
error occurs when exporting prod schedule from showcase and qa you are currently not able to export the production schedule from both the qa and showcase otp environments an error occurs saying it failed to fetch installers see attached screenshot note this works in the otp production environment even after a recent release so it is possibly caused by changes to the inventory service or crm
| 1
|
17,720
| 23,625,311,230
|
IssuesEvent
|
2022-08-25 02:51:33
|
MPMG-DCC-UFMG/C01
|
https://api.github.com/repos/MPMG-DCC-UFMG/C01
|
closed
|
Campo para XPath não aparece no passo "Executar em iframe"
|
[1] Bug [2] Alta Prioridade [0] Desenvolvimento [3] Processamento Dinâmico
|
## Comportamento Esperado
Ao utilizar o passo "Executar em iframe" em coletas dinâmicas, ser possível especificar o XPath referente ao iframe em que os processos devem ser executados.
## Comportamento Atual
Ao selecionar o tipo de passo como "Executar em iframe", não aparece um campo para que o XPath do iframe possa ser especificado, inviabilizando o uso dessa ação.
## Passos para reproduzir o erro
1. Criar um novo coletor
2. Marcar a opção de Processamento dinâmico
3. Adicionar passo Executar em iframe
4. Observar que não aparece um campo para especificação do XPath
## Screenshots
Como a ação de "Executar em iframe" aparece na visualização de passos no Processamento dinâmico:

|
1.0
|
Campo para XPath não aparece no passo "Executar em iframe" - ## Comportamento Esperado
Ao utilizar o passo "Executar em iframe" em coletas dinâmicas, ser possível especificar o XPath referente ao iframe em que os processos devem ser executados.
## Comportamento Atual
Ao selecionar o tipo de passo como "Executar em iframe", não aparece um campo para que o XPath do iframe possa ser especificado, inviabilizando o uso dessa ação.
## Passos para reproduzir o erro
1. Criar um novo coletor
2. Marcar a opção de Processamento dinâmico
3. Adicionar passo Executar em iframe
4. Observar que não aparece um campo para especificação do XPath
## Screenshots
Como a ação de "Executar em iframe" aparece na visualização de passos no Processamento dinâmico:

|
non_defect
|
campo para xpath não aparece no passo executar em iframe comportamento esperado ao utilizar o passo executar em iframe em coletas dinâmicas ser possível especificar o xpath referente ao iframe em que os processos devem ser executados comportamento atual ao selecionar o tipo de passo como executar em iframe não aparece um campo para que o xpath do iframe possa ser especificado inviabilizando o uso dessa ação passos para reproduzir o erro criar um novo coletor marcar a opção de processamento dinâmico adicionar passo executar em iframe observar que não aparece um campo para especificação do xpath screenshots como a ação de executar em iframe aparece na visualização de passos no processamento dinâmico
| 0
|
84,094
| 15,720,832,205
|
IssuesEvent
|
2021-03-29 01:21:11
|
ysmanohar/myRetail
|
https://api.github.com/repos/ysmanohar/myRetail
|
opened
|
CVE-2020-10672 (High) detected in jackson-databind-2.7.4.jar
|
security vulnerability
|
## CVE-2020-10672 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /myRetail/myRetail/pom.xml</p>
<p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.7.4/jackson-databind-2.7.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.4.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.aries.transaction.jms.internal.XaPooledConnectionFactory (aka aries.transaction.jms).
<p>Publish Date: 2020-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10672>CVE-2020-10672</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10672">https://nvd.nist.gov/vuln/detail/CVE-2020-10672</a></p>
<p>Release Date: 2020-03-18</p>
<p>Fix Resolution: jackson-databind-2.9.10.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-10672 (High) detected in jackson-databind-2.7.4.jar - ## CVE-2020-10672 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /myRetail/myRetail/pom.xml</p>
<p>Path to vulnerable library: 2/repository/com/fasterxml/jackson/core/jackson-databind/2.7.4/jackson-databind-2.7.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.4.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.aries.transaction.jms.internal.XaPooledConnectionFactory (aka aries.transaction.jms).
<p>Publish Date: 2020-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10672>CVE-2020-10672</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10672">https://nvd.nist.gov/vuln/detail/CVE-2020-10672</a></p>
<p>Release Date: 2020-03-18</p>
<p>Fix Resolution: jackson-databind-2.9.10.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file myretail myretail pom xml path to vulnerable library repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache aries transaction jms internal xapooledconnectionfactory aka aries transaction jms publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jackson databind step up your open source security game with whitesource
| 0
|
61,042
| 17,023,586,497
|
IssuesEvent
|
2021-07-03 02:47:34
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
tirex-backend-manager crashes when trying to see help
|
Component: tirex Priority: minor Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 3.06am, Saturday, 8th May 2010]**
tirex-backend-manager --help results in:
```
Undefined subroutine &main::pod2usage called at /usr/bin/tirex-backend-manager line 59.
```
|
1.0
|
tirex-backend-manager crashes when trying to see help - **[Submitted to the original trac issue database at 3.06am, Saturday, 8th May 2010]**
tirex-backend-manager --help results in:
```
Undefined subroutine &main::pod2usage called at /usr/bin/tirex-backend-manager line 59.
```
|
defect
|
tirex backend manager crashes when trying to see help tirex backend manager help results in undefined subroutine main called at usr bin tirex backend manager line
| 1
|
31,537
| 6,547,518,893
|
IssuesEvent
|
2017-09-04 15:08:13
|
jccastillo0007/eFacturaT
|
https://api.github.com/repos/jccastillo0007/eFacturaT
|
opened
|
Web CFDI v3.3 - no jala la conversión de monedas al momento de facturar.
|
bug defect
|
Ejemplo, en el catálogo de productos, la moneda asociada es MXN para el producto X.
Si para ese producto X, en la pantalla de facturación elijo una moneda distinta a MXN, se genera el siguiente error:
El TipoDeComprobante es I,E o N, el importe registrado en el campo no es igual a la suma de los importes de los conceptos registrados. (3)
Esto es no jala la conversión de monedas.
|
1.0
|
Web CFDI v3.3 - no jala la conversión de monedas al momento de facturar. - Ejemplo, en el catálogo de productos, la moneda asociada es MXN para el producto X.
Si para ese producto X, en la pantalla de facturación elijo una moneda distinta a MXN, se genera el siguiente error:
El TipoDeComprobante es I,E o N, el importe registrado en el campo no es igual a la suma de los importes de los conceptos registrados. (3)
Esto es no jala la conversión de monedas.
|
defect
|
web cfdi no jala la conversión de monedas al momento de facturar ejemplo en el catálogo de productos la moneda asociada es mxn para el producto x si para ese producto x en la pantalla de facturación elijo una moneda distinta a mxn se genera el siguiente error el tipodecomprobante es i e o n el importe registrado en el campo no es igual a la suma de los importes de los conceptos registrados esto es no jala la conversión de monedas
| 1
|
259,115
| 22,391,517,841
|
IssuesEvent
|
2022-06-17 08:11:32
|
SonarSource/sonar-scanner-msbuild
|
https://api.github.com/repos/SonarSource/sonar-scanner-msbuild
|
closed
|
Remove unused package references from UTs
|
Testing
|
`Microsoft.Build.Framework` and `Microsoft.Build.Utilities.Core` references are not used anymore in `SonarScanner.MSBuild.Tasks.UnitTest.csproj`.
|
1.0
|
Remove unused package references from UTs - `Microsoft.Build.Framework` and `Microsoft.Build.Utilities.Core` references are not used anymore in `SonarScanner.MSBuild.Tasks.UnitTest.csproj`.
|
non_defect
|
remove unused package references from uts microsoft build framework and microsoft build utilities core references are not used anymore in sonarscanner msbuild tasks unittest csproj
| 0
|
306,409
| 9,392,750,114
|
IssuesEvent
|
2019-04-07 04:30:22
|
bigbluebutton/bigbluebutton
|
https://api.github.com/repos/bigbluebutton/bigbluebutton
|
closed
|
Add a connection quality indicator to the client
|
Enhancement Flash Client Normal Priority
|
Originally reported on Google Code with ID 1004
```
> * It would be nice to have connection quality indicator on the client
> (let say green dot for good connection, yellow, slow, red terrible,
> gray or black - no connection with the server - after timeout -
> session is closed).
Posted in this thread
http://groups.google.com/group/bigbluebutton-setup/browse_thread/thread/dfcf1fb5d310b120#
```
Reported by `ffdixon` on 2011-09-18 13:28:15
|
1.0
|
Add a connection quality indicator to the client - Originally reported on Google Code with ID 1004
```
> * It would be nice to have connection quality indicator on the client
> (let say green dot for good connection, yellow, slow, red terrible,
> gray or black - no connection with the server - after timeout -
> session is closed).
Posted in this thread
http://groups.google.com/group/bigbluebutton-setup/browse_thread/thread/dfcf1fb5d310b120#
```
Reported by `ffdixon` on 2011-09-18 13:28:15
|
non_defect
|
add a connection quality indicator to the client originally reported on google code with id it would be nice to have connection quality indicator on the client let say green dot for good connection yellow slow red terrible gray or black no connection with the server after timeout session is closed posted in this thread reported by ffdixon on
| 0
|
49,860
| 13,187,282,166
|
IssuesEvent
|
2020-08-13 02:55:14
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
PGF77 bug in Sibyll 2.3c datasets (Trac #2190)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2190">https://code.icecube.wisc.edu/ticket/2190</a>, reported by juancarlos and owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-03-18T15:58:38",
"description": "Linking problem with PGF77 CORSIKA withe gfortran PYTHIA.\n\nDatasets affected: 20243, 20222, 20209, 20208.\n\nDatasets should be deleted and replaced.",
"reporter": "juancarlos",
"cc": "",
"resolution": "fixed",
"_ts": "1552924718605825",
"component": "combo simulation",
"summary": "PGF77 bug in Sibyll 2.3c datasets",
"priority": "normal",
"keywords": "",
"time": "2018-09-23T09:28:32",
"milestone": "Vernal Equinox 2019",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
PGF77 bug in Sibyll 2.3c datasets (Trac #2190) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2190">https://code.icecube.wisc.edu/ticket/2190</a>, reported by juancarlos and owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-03-18T15:58:38",
"description": "Linking problem with PGF77 CORSIKA withe gfortran PYTHIA.\n\nDatasets affected: 20243, 20222, 20209, 20208.\n\nDatasets should be deleted and replaced.",
"reporter": "juancarlos",
"cc": "",
"resolution": "fixed",
"_ts": "1552924718605825",
"component": "combo simulation",
"summary": "PGF77 bug in Sibyll 2.3c datasets",
"priority": "normal",
"keywords": "",
"time": "2018-09-23T09:28:32",
"milestone": "Vernal Equinox 2019",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
|
defect
|
bug in sibyll datasets trac migrated from json status closed changetime description linking problem with corsika withe gfortran pythia n ndatasets affected n ndatasets should be deleted and replaced reporter juancarlos cc resolution fixed ts component combo simulation summary bug in sibyll datasets priority normal keywords time milestone vernal equinox owner juancarlos type defect
| 1
|
51,308
| 13,207,428,856
|
IssuesEvent
|
2020-08-14 23:04:04
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
i3Monitoring ROOT output file names needs an additional identifier (Trac #187)
|
Incomplete Migration Migrated from Trac defect jeb + pnf
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/187">https://code.icecube.wisc.edu/projects/icecube/ticket/187</a>, reported by blaufussand owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:57",
"_ts": "1416713877066511",
"description": "In current system, if JEB system is stopped and restarted, existing ROOT files will be overwritten. To avoid this, and to generate multiple ROOT files\nper client in this case, need to add a unique ID (in this case the PID will\nlikely work well) to each filename, so that:\n\nEvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1_Run00109709_Subrun00000000.root\n\nis \n\nEvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1.PID_Run00109709_Subrun00000000.root\n\nwhere PID is the actual Process ID, which should be easily available to the process.",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"time": "2009-12-07T22:38:21",
"component": "jeb + pnf",
"summary": "i3Monitoring ROOT output file names needs an additional identifier",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
i3Monitoring ROOT output file names needs an additional identifier (Trac #187) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/187">https://code.icecube.wisc.edu/projects/icecube/ticket/187</a>, reported by blaufussand owned by blaufuss</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:57",
"_ts": "1416713877066511",
"description": "In current system, if JEB system is stopped and restarted, existing ROOT files will be overwritten. To avoid this, and to generate multiple ROOT files\nper client in this case, need to add a unique ID (in this case the PID will\nlikely work well) to each filename, so that:\n\nEvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1_Run00109709_Subrun00000000.root\n\nis \n\nEvtMon_PhysicsData_PhysicsFiltering_PFClient.sps-fpslave01.client1.PID_Run00109709_Subrun00000000.root\n\nwhere PID is the actual Process ID, which should be easily available to the process.",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"time": "2009-12-07T22:38:21",
"component": "jeb + pnf",
"summary": "i3Monitoring ROOT output file names needs an additional identifier",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
defect
|
root output file names needs an additional identifier trac migrated from json status closed changetime ts description in current system if jeb system is stopped and restarted existing root files will be overwritten to avoid this and to generate multiple root files nper client in this case need to add a unique id in this case the pid will nlikely work well to each filename so that n nevtmon physicsdata physicsfiltering pfclient sps root n nis n nevtmon physicsdata physicsfiltering pfclient sps pid root n nwhere pid is the actual process id which should be easily available to the process reporter blaufuss cc resolution fixed time component jeb pnf summary root output file names needs an additional identifier priority normal keywords milestone owner blaufuss type defect
| 1
|
314,454
| 9,597,286,899
|
IssuesEvent
|
2019-05-09 20:53:03
|
minio/mc
|
https://api.github.com/repos/minio/mc
|
closed
|
MC should not modify my .zshrc without special permission
|
community priority: medium triage
|
## Expected behavior
Program asks to install or not completion.
And if it is allowed to install, it should use correct binary name.
Because this mc conflicts with Midnight Commander i renamed binary to `minio-mc`, but completion is set for `mc`.
## Actual behavior
Program writes to shell rc file without any approval.
Always uses `mc` name for completion.
## Steps to reproduce the behavior
Run any mc command, e.g. `mc version`.
## mc version
Version: 2019-04-15T17:01:49Z
Release-tag: DEVELOPMENT.2019-04-15T17-01-49Z
Commit-id: 5729a01e691cc05142bcb257163e887062dbecd3
## System information
Distributor ID: Ubuntu
Description: Ubuntu 18.10
Release: 18.10
Codename: cosmic
|
1.0
|
MC should not modify my .zshrc without special permission - ## Expected behavior
Program asks to install or not completion.
And if it is allowed to install, it should use correct binary name.
Because this mc conflicts with Midnight Commander i renamed binary to `minio-mc`, but completion is set for `mc`.
## Actual behavior
Program writes to shell rc file without any approval.
Always uses `mc` name for completion.
## Steps to reproduce the behavior
Run any mc command, e.g. `mc version`.
## mc version
Version: 2019-04-15T17:01:49Z
Release-tag: DEVELOPMENT.2019-04-15T17-01-49Z
Commit-id: 5729a01e691cc05142bcb257163e887062dbecd3
## System information
Distributor ID: Ubuntu
Description: Ubuntu 18.10
Release: 18.10
Codename: cosmic
|
non_defect
|
mc should not modify my zshrc without special permission expected behavior program asks to install or not completion and if it is allowed to install it should use correct binary name because this mc conflicts with midnight commander i renamed binary to minio mc but completion is set for mc actual behavior program writes to shell rc file without any approval always uses mc name for completion steps to reproduce the behavior run any mc command e g mc version mc version version release tag development commit id system information distributor id ubuntu description ubuntu release codename cosmic
| 0
|
25,394
| 4,309,853,959
|
IssuesEvent
|
2016-07-21 17:19:30
|
KytechN24/xbox360wirelesschatpad
|
https://api.github.com/repos/KytechN24/xbox360wirelesschatpad
|
closed
|
for help or hinderence here is my suggestion
|
auto-migrated Priority-Medium Type-Defect
|
```
I haven't been able to look through EVERYTHING related to this project so I may
be a bit late to say this but I believe you should be looking more into
chatpad-super-driver (https://code.google.com/p/chatpad-super-driver/). It is a
fairly solid implementation of a chatpad driver including remapping and I
believe it either works alongside the native driver or reimplements it (xinput
and force feedback worked). It had it's issues (wired only, unsigned driver,
driver not "letting go" of the device) but depending on how deep you want to go
you might learn something from it. This probably isn't anything new to you
(especially since it's the only other known working driver) but there's my 2
cents.
```
Original issue reported on code.google.com by `crimsonspecter91` on 10 Oct 2014 at 12:40
|
1.0
|
for help or hinderence here is my suggestion - ```
I haven't been able to look through EVERYTHING related to this project so I may
be a bit late to say this but I believe you should be looking more into
chatpad-super-driver (https://code.google.com/p/chatpad-super-driver/). It is a
fairly solid implementation of a chatpad driver including remapping and I
believe it either works alongside the native driver or reimplements it (xinput
and force feedback worked). It had it's issues (wired only, unsigned driver,
driver not "letting go" of the device) but depending on how deep you want to go
you might learn something from it. This probably isn't anything new to you
(especially since it's the only other known working driver) but there's my 2
cents.
```
Original issue reported on code.google.com by `crimsonspecter91` on 10 Oct 2014 at 12:40
|
defect
|
for help or hinderence here is my suggestion i haven t been able to look through everything related to this project so i may be a bit late to say this but i believe you should be looking more into chatpad super driver it is a fairly solid implementation of a chatpad driver including remapping and i believe it either works alongside the native driver or reimplements it xinput and force feedback worked it had it s issues wired only unsigned driver driver not letting go of the device but depending on how deep you want to go you might learn something from it this probably isn t anything new to you especially since it s the only other known working driver but there s my cents original issue reported on code google com by on oct at
| 1
|
12,095
| 9,594,694,517
|
IssuesEvent
|
2019-05-09 14:30:41
|
laopinion/siteLaopinion
|
https://api.github.com/repos/laopinion/siteLaopinion
|
opened
|
Clean tpl drupal 7
|
Doing Infrastructure
|
# Clean tpls drupal 7
Files that terminate with extensions tpl.php
## Example tpl
> node-article.tpl.php
```
<link href="<?php print base_path(); ?>sites/default/themes/op/dist/public/css/article.css" rel="stylesheet" type="text/css"/>
<?php
if (arg(0) == 'node' && is_numeric(arg(1))) $nodeid = arg(1);
// print $nodeid; // show me your nid!
include("pautas/clickio_inarticle.php");
$taxonomy_menu_block = block_load('taxonomy_menu_block','1'); $taxtmprender = _block_render_blocks(array($taxonomy_menu_block)); $taxtmprender2 = _block_get_renderable_array($taxtmprender); print drupal_render($taxtmprender2);
?>
<div id="app" data-nodeid="<?php print $nodeid; ?>"></div>
<script async src="https://platform.twitter.com/widgets.js" charset="utf-8"></script>
<script src="<?php print base_path(); ?>sites/default/themes/op/dist/public/js/article.js"></script>
```
It is loaded alone files css, js and nodeid of node drupal 7.
|
1.0
|
Clean tpl drupal 7 - # Clean tpls drupal 7
Files that terminate with extensions tpl.php
## Example tpl
> node-article.tpl.php
```
<link href="<?php print base_path(); ?>sites/default/themes/op/dist/public/css/article.css" rel="stylesheet" type="text/css"/>
<?php
if (arg(0) == 'node' && is_numeric(arg(1))) $nodeid = arg(1);
// print $nodeid; // show me your nid!
include("pautas/clickio_inarticle.php");
$taxonomy_menu_block = block_load('taxonomy_menu_block','1'); $taxtmprender = _block_render_blocks(array($taxonomy_menu_block)); $taxtmprender2 = _block_get_renderable_array($taxtmprender); print drupal_render($taxtmprender2);
?>
<div id="app" data-nodeid="<?php print $nodeid; ?>"></div>
<script async src="https://platform.twitter.com/widgets.js" charset="utf-8"></script>
<script src="<?php print base_path(); ?>sites/default/themes/op/dist/public/js/article.js"></script>
```
It is loaded alone files css, js and nodeid of node drupal 7.
|
non_defect
|
clean tpl drupal clean tpls drupal files that terminate with extensions tpl php example tpl node article tpl php sites default themes op dist public css article css rel stylesheet type text css php if arg node is numeric arg nodeid arg print nodeid show me your nid include pautas clickio inarticle php taxonomy menu block block load taxonomy menu block taxtmprender block render blocks array taxonomy menu block block get renderable array taxtmprender print drupal render sites default themes op dist public js article js it is loaded alone files css js and nodeid of node drupal
| 0
|
43,796
| 11,849,820,299
|
IssuesEvent
|
2020-03-24 15:46:20
|
idaholab/raven
|
https://api.github.com/repos/idaholab/raven
|
closed
|
[DEFECT] Fix failed heavy test
|
defect priority_normal
|
--------
Defect Description
--------
**Describe the defect**
tests/framework/ROM/TimeSeries/ARMA/VARMAheavy failed due to time out, this can be fixed by increasing the max_time.
##### What did you expect to see happen?
##### What did you see instead?
##### Do you have a suggested fix for the development team?
**Describe how to Reproduce**
Steps to reproduce the behavior:
1.
2.
3.
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or task?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [x] 1. If the issue is a defect, is the defect fixed?
- [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
|
1.0
|
[DEFECT] Fix failed heavy test - --------
Defect Description
--------
**Describe the defect**
tests/framework/ROM/TimeSeries/ARMA/VARMAheavy failed due to time out, this can be fixed by increasing the max_time.
##### What did you expect to see happen?
##### What did you see instead?
##### Do you have a suggested fix for the development team?
**Describe how to Reproduce**
Steps to reproduce the behavior:
1.
2.
3.
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or task?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [x] 1. If the issue is a defect, is the defect fixed?
- [x] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [x] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [x] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [x] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
|
defect
|
fix failed heavy test defect description describe the defect tests framework rom timeseries arma varmaheavy failed due to time out this can be fixed by increasing the max time what did you expect to see happen what did you see instead do you have a suggested fix for the development team describe how to reproduce steps to reproduce the behavior screenshots and input files please attach the input file s that generate this error the simpler the input the faster we can find the issue platform please complete the following information os version dependencies installation for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided
| 1
|
44,512
| 12,219,044,276
|
IssuesEvent
|
2020-05-01 20:44:25
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
[ZOOM]: VAOS - CONSIDER Upcoming appointments tab goes to two lines at 400%
|
508-defect-4 508-issue-mobile-design 508/Accessibility frontend vaos
|
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
The Upcoming Appointments tab rolls over to two lines at 400% zoom by 1280px browser width. This is a low priority, consider fix only. Screen shot attached below.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Trevor_
## Acceptance Criteria
<!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. -->
- [ ] Label stays on one line at 400% zoom, assuming 1280px screen width
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->
<img width="1237" alt="Screen Shot 2020-04-17 at 11 00 21 AM" src="https://user-images.githubusercontent.com/934879/79589457-0c210c00-809b-11ea-8ac9-75a5d56ce386.png">
|
1.0
|
[ZOOM]: VAOS - CONSIDER Upcoming appointments tab goes to two lines at 400% - **Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
The Upcoming Appointments tab rolls over to two lines at 400% zoom by 1280px browser width. This is a low priority, consider fix only. Screen shot attached below.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Trevor_
## Acceptance Criteria
<!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. -->
- [ ] Label stays on one line at 400% zoom, assuming 1280px screen width
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->
<img width="1237" alt="Screen Shot 2020-04-17 at 11 00 21 AM" src="https://user-images.githubusercontent.com/934879/79589457-0c210c00-809b-11ea-8ac9-75a5d56ce386.png">
|
defect
|
vaos consider upcoming appointments tab goes to two lines at feedback framework ❗️ must for if the feedback must be applied ⚠️should if the feedback is best practice ✔️ consider for suggestions enhancements description the upcoming appointments tab rolls over to two lines at zoom by browser width this is a low priority consider fix only screen shot attached below point of contact if this issue is being opened by a vfs team member please add a point of contact usually this is the same person who enters the issue ticket vfs point of contact trevor acceptance criteria label stays on one line at zoom assuming screen width screenshots or trace logs img width alt screen shot at am src
| 1
|
36,049
| 7,853,405,899
|
IssuesEvent
|
2018-06-20 17:19:13
|
laurentperrinet/NeuroTools
|
https://api.github.com/repos/laurentperrinet/NeuroTools
|
closed
|
ParameterSet labels are not preserved on save and load
|
defect
|
see the `test_save_and_load()` test in `test_parameters.py`
---
### Imported from Trac ticket:33
**Opened**: 2008-10-31 13:49:59
**Last modified**: 2008-11-03 17:52:55
**Component**: parameters
**Priority**: minor
**Owner**: apdavison
**Reporter**: apdavison
|
1.0
|
ParameterSet labels are not preserved on save and load - see the `test_save_and_load()` test in `test_parameters.py`
---
### Imported from Trac ticket:33
**Opened**: 2008-10-31 13:49:59
**Last modified**: 2008-11-03 17:52:55
**Component**: parameters
**Priority**: minor
**Owner**: apdavison
**Reporter**: apdavison
|
defect
|
parameterset labels are not preserved on save and load see the test save and load test in test parameters py imported from trac ticket opened last modified component parameters priority minor owner apdavison reporter apdavison
| 1
|
18,293
| 3,041,023,907
|
IssuesEvent
|
2015-08-07 18:44:24
|
ariya/phantomjs
|
https://api.github.com/repos/ariya/phantomjs
|
closed
|
PhantomJS has crashed.
|
old.Priority-Medium old.Status-New old.Type-Defect
|
_**[eugene.m...@gmail.com](http://code.google.com/u/105310361067299319588/) commented:**_
> > Which version of PhantomJS are you using?
> 1.6.0
>
> > What steps will reproduce the problem?
> 1. Run "phantomjs test.js" in your terminal. Find test.js file in attach.
>
> > What is the expected output? What do you see instead?
> Expected output is html page.
> "Segmentation fault (core dumped)" I see instead of it.
>
> > Which operating system are you using?
> Ubuntu Desktop 12.04
>
> > Did you use binary PhantomJS or did you compile it from source?
> Binary
>
> > Please provide any additional information below.
>
> See the crash dump file in attach.
**Disclaimer:**
This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #675](http://code.google.com/p/phantomjs/issues/detail?id=675).
:star2: **2** people had starred this issue at the time of migration.
|
1.0
|
PhantomJS has crashed. - _**[eugene.m...@gmail.com](http://code.google.com/u/105310361067299319588/) commented:**_
> > Which version of PhantomJS are you using?
> 1.6.0
>
> > What steps will reproduce the problem?
> 1. Run "phantomjs test.js" in your terminal. Find test.js file in attach.
>
> > What is the expected output? What do you see instead?
> Expected output is html page.
> "Segmentation fault (core dumped)" I see instead of it.
>
> > Which operating system are you using?
> Ubuntu Desktop 12.04
>
> > Did you use binary PhantomJS or did you compile it from source?
> Binary
>
> > Please provide any additional information below.
>
> See the crash dump file in attach.
**Disclaimer:**
This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #675](http://code.google.com/p/phantomjs/issues/detail?id=675).
:star2: **2** people had starred this issue at the time of migration.
|
defect
|
phantomjs has crashed commented gt which version of phantomjs are you using gt what steps will reproduce the problem run quot phantomjs test js quot in your terminal find test js file in attach gt what is the expected output what do you see instead expected output is html page quot segmentation fault core dumped quot i see instead of it gt which operating system are you using ubuntu desktop gt did you use binary phantomjs or did you compile it from source binary gt please provide any additional information below see the crash dump file in attach disclaimer this issue was migrated on from the project s former issue tracker on google code nbsp people had starred this issue at the time of migration
| 1
|
20,333
| 3,343,309,962
|
IssuesEvent
|
2015-11-15 11:36:54
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
closed
|
TypedArray.Prototype should not be in C# and in emitted JavaScript
|
defect up-for-grabs
|
The issue is discovered due to this [comment](https://github.com/bridgedotnet/Bridge/issues/182#issuecomment-152838336).
**C#**:
* Actual: `float32Array.Prototype.Buffer`
* Expected: `float32Array.Buffer`
The `.Prototype` thing should not be required in C# code.
**JavaScript**:
* Actual: `float32Array.prototype.buffer` - it throws an error that `prototype` is not defined.
* Expected: `float32Array.buffer`
The [live example](http://live.bridge.net/#0fce2de66ff629f504f6).
It is actual for all the TypedArray classes.
|
1.0
|
TypedArray.Prototype should not be in C# and in emitted JavaScript - The issue is discovered due to this [comment](https://github.com/bridgedotnet/Bridge/issues/182#issuecomment-152838336).
**C#**:
* Actual: `float32Array.Prototype.Buffer`
* Expected: `float32Array.Buffer`
The `.Prototype` thing should not be required in C# code.
**JavaScript**:
* Actual: `float32Array.prototype.buffer` - it throws an error that `prototype` is not defined.
* Expected: `float32Array.buffer`
The [live example](http://live.bridge.net/#0fce2de66ff629f504f6).
It is actual for all the TypedArray classes.
|
defect
|
typedarray prototype should not be in c and in emitted javascript the issue is discovered due to this c actual prototype buffer expected buffer the prototype thing should not be required in c code javascript actual prototype buffer it throws an error that prototype is not defined expected buffer the it is actual for all the typedarray classes
| 1
|
17,941
| 3,013,800,253
|
IssuesEvent
|
2015-07-29 11:19:29
|
yawlfoundation/yawl
|
https://api.github.com/repos/yawlfoundation/yawl
|
closed
|
'Choose Completer of Task:' Constraint will not remove
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Create a Net with two Tasks, the first with resourcing offered and allocated
to a user.
2. On the second task's Resourcing, check the 'Choose completer(s) of task:'
box so that the previous task is selected. OK
3. Re-open the Resourcing, uncheck the constraint, close and reopen.
What is the expected output? What do you see instead?
The Constraint should be removed, but it is not.
What version of the product are you using? On what operating system?
YAWL Editor 3.0a build 401, OS X 10.9.2, and Windows XP.
Please provide any additional information below.
In the source code for the workflow,
the '<familiarParticipant taskID="Task1" />' remains. If that is manually
removed, the Constraint becomes unchecked successfully.
Further, the taskID property in familiarParticipant is not updated when the ID
of the task is changed. That is, the name of a task was changed, and the
familiarParticipant caused an error in the workflow because the taskID could
not be found, nor the reference removed.
```
Original issue reported on code.google.com by `c.cowlr...@gmail.com` on 21 Mar 2014 at 2:50
|
1.0
|
'Choose Completer of Task:' Constraint will not remove - ```
What steps will reproduce the problem?
1. Create a Net with two Tasks, the first with resourcing offered and allocated
to a user.
2. On the second task's Resourcing, check the 'Choose completer(s) of task:'
box so that the previous task is selected. OK
3. Re-open the Resourcing, uncheck the constraint, close and reopen.
What is the expected output? What do you see instead?
The Constraint should be removed, but it is not.
What version of the product are you using? On what operating system?
YAWL Editor 3.0a build 401, OS X 10.9.2, and Windows XP.
Please provide any additional information below.
In the source code for the workflow,
the '<familiarParticipant taskID="Task1" />' remains. If that is manually
removed, the Constraint becomes unchecked successfully.
Further, the taskID property in familiarParticipant is not updated when the ID
of the task is changed. That is, the name of a task was changed, and the
familiarParticipant caused an error in the workflow because the taskID could
not be found, nor the reference removed.
```
Original issue reported on code.google.com by `c.cowlr...@gmail.com` on 21 Mar 2014 at 2:50
|
defect
|
choose completer of task constraint will not remove what steps will reproduce the problem create a net with two tasks the first with resourcing offered and allocated to a user on the second task s resourcing check the choose completer s of task box so that the previous task is selected ok re open the resourcing uncheck the constraint close and reopen what is the expected output what do you see instead the constraint should be removed but it is not what version of the product are you using on what operating system yawl editor build os x and windows xp please provide any additional information below in the source code for the workflow the remains if that is manually removed the constraint becomes unchecked successfully further the taskid property in familiarparticipant is not updated when the id of the task is changed that is the name of a task was changed and the familiarparticipant caused an error in the workflow because the taskid could not be found nor the reference removed original issue reported on code google com by c cowlr gmail com on mar at
| 1
|
70,564
| 30,693,877,698
|
IssuesEvent
|
2023-07-26 17:02:14
|
hashicorp/terraform-provider-azurerm
|
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
|
closed
|
azurerm_subnet should be recreated when modifying delegation
|
bug service/subnets v/2.x (legacy)
|
<!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform (and AzureRM Provider) Version
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
Terraform v0.14.5
+ provider registry.terraform.io/hashicorp/azurerm v2.44.0
+ provider registry.terraform.io/hashicorp/null v3.0.0
+ provider registry.terraform.io/hashicorp/random v3.0.1
I can't upgrade to azurerm 2.45.1 due to #10371
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* `azurerm_subnet`
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
data "azurerm_resource_group" "main" {
name = var.resource_group_name
}
data "azurerm_virtual_network" "main" {
name = var.virtual_network_name
resource_group_name = data.azurerm_resource_group.main.name
}
resource "azurerm_subnet" "main" {
name = "subnet"
resource_group_name = data.azurerm_resource_group.main.name
virtual_network_name = data.azurerm_virtual_network.main.name
address_prefixes = [var.address_prefix]
enforce_private_link_endpoint_network_policies = true
delegation {
name = "delegation"
service_delegation {
name = "Microsoft.Web/serverFarms"
}
}
}
```
### Expected Behaviour
Plan is applied correctly
### Actual Behaviour
Applying plan result in an error :
```
Error: Error updating Subnet "subnet" (Virtual Network " vnet" / Resource Group "resourcegroup"): network.SubnetsClient#CreateOrUpdate: Failure sending request: StatusCode=400 -- Original Error: Code="SubnetDelegationsCannotChangeWhenSubnetUsedByResource" Message="Delegations of subnet /subscriptions/593cfa7e-d3b2-4aab-9653-b1ddd2f5a7db/resourceGroups/resourcegroup/providers/Microsoft.Network/virtualNetworks/vnet/subnets/subnet cannot be changed from [] to [Microsoft.Web/serverFarms] because it is being used by the resource /subscriptions/593cfa7e-d3b2-4aab-9653-b1ddd2f5a7db/resourceGroups/resourcegroup/providers/Microsoft.Network/networkInterfaces/privateendpoint.nic.aa70ab99-e8a7-4bc7-bd8a-29177edd608b/ipConfigurations/postgresqlServer-postgresqlServer.privateEndpoint." Details=[]
```
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. `terraform apply`
|
1.0
|
azurerm_subnet should be recreated when modifying delegation - <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform (and AzureRM Provider) Version
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
Terraform v0.14.5
+ provider registry.terraform.io/hashicorp/azurerm v2.44.0
+ provider registry.terraform.io/hashicorp/null v3.0.0
+ provider registry.terraform.io/hashicorp/random v3.0.1
I can't upgrade to azurerm 2.45.1 due to #10371
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* `azurerm_subnet`
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
data "azurerm_resource_group" "main" {
name = var.resource_group_name
}
data "azurerm_virtual_network" "main" {
name = var.virtual_network_name
resource_group_name = data.azurerm_resource_group.main.name
}
resource "azurerm_subnet" "main" {
name = "subnet"
resource_group_name = data.azurerm_resource_group.main.name
virtual_network_name = data.azurerm_virtual_network.main.name
address_prefixes = [var.address_prefix]
enforce_private_link_endpoint_network_policies = true
delegation {
name = "delegation"
service_delegation {
name = "Microsoft.Web/serverFarms"
}
}
}
```
### Expected Behaviour
Plan is applied correctly
### Actual Behaviour
Applying plan result in an error :
```
Error: Error updating Subnet "subnet" (Virtual Network " vnet" / Resource Group "resourcegroup"): network.SubnetsClient#CreateOrUpdate: Failure sending request: StatusCode=400 -- Original Error: Code="SubnetDelegationsCannotChangeWhenSubnetUsedByResource" Message="Delegations of subnet /subscriptions/593cfa7e-d3b2-4aab-9653-b1ddd2f5a7db/resourceGroups/resourcegroup/providers/Microsoft.Network/virtualNetworks/vnet/subnets/subnet cannot be changed from [] to [Microsoft.Web/serverFarms] because it is being used by the resource /subscriptions/593cfa7e-d3b2-4aab-9653-b1ddd2f5a7db/resourceGroups/resourcegroup/providers/Microsoft.Network/networkInterfaces/privateendpoint.nic.aa70ab99-e8a7-4bc7-bd8a-29177edd608b/ipConfigurations/postgresqlServer-postgresqlServer.privateEndpoint." Details=[]
```
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. `terraform apply`
|
non_defect
|
azurerm subnet should be recreated when modifying delegation please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform and azurerm provider version terraform provider registry terraform io hashicorp azurerm provider registry terraform io hashicorp null provider registry terraform io hashicorp random i can t upgrade to azurerm due to affected resource s azurerm subnet terraform configuration files hcl data azurerm resource group main name var resource group name data azurerm virtual network main name var virtual network name resource group name data azurerm resource group main name resource azurerm subnet main name subnet resource group name data azurerm resource group main name virtual network name data azurerm virtual network main name address prefixes enforce private link endpoint network policies true delegation name delegation service delegation name microsoft web serverfarms expected behaviour plan is applied correctly actual behaviour applying plan result in an error error error updating subnet subnet virtual network vnet resource group resourcegroup network subnetsclient createorupdate failure sending request statuscode original error code subnetdelegationscannotchangewhensubnetusedbyresource message delegations of subnet subscriptions resourcegroups resourcegroup providers microsoft network virtualnetworks vnet subnets subnet cannot be changed from to because it is being used by the resource subscriptions resourcegroups resourcegroup providers microsoft network networkinterfaces privateendpoint nic ipconfigurations postgresqlserver postgresqlserver privateendpoint details steps to reproduce terraform apply
| 0
|
307,551
| 23,203,997,173
|
IssuesEvent
|
2022-08-02 02:06:25
|
polywrap/documentation
|
https://api.github.com/repos/polywrap/documentation
|
closed
|
Documentation: wrapper development
|
documentation
|
Needed For Completion:
- Improved documentation on:
- [x] manifests
- [x] schema imports
- [x] base schema types
- [ ] AssemblyScript and Rust support
- [x] WASM limitations
- [x] default plugins
|
1.0
|
Documentation: wrapper development - Needed For Completion:
- Improved documentation on:
- [x] manifests
- [x] schema imports
- [x] base schema types
- [ ] AssemblyScript and Rust support
- [x] WASM limitations
- [x] default plugins
|
non_defect
|
documentation wrapper development needed for completion improved documentation on manifests schema imports base schema types assemblyscript and rust support wasm limitations default plugins
| 0
|
3,961
| 2,610,084,811
|
IssuesEvent
|
2015-02-26 18:25:50
|
chrsmith/dsdsdaadf
|
https://api.github.com/repos/chrsmith/dsdsdaadf
|
opened
|
深圳彩光怎么治痘痘
|
auto-migrated Priority-Medium Type-Defect
|
```
深圳彩光怎么治痘痘【深圳韩方科颜全国热线400-869-1818,24小
时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国��
�方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩�
��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”
健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专��
�治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的�
��痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:00
|
1.0
|
深圳彩光怎么治痘痘 - ```
深圳彩光怎么治痘痘【深圳韩方科颜全国热线400-869-1818,24小
时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国��
�方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩�
��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”
健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专��
�治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的�
��痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:00
|
defect
|
深圳彩光怎么治痘痘 深圳彩光怎么治痘痘【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国�� �方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩� ��科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹” 健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专�� �治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的� ��痘。 original issue reported on code google com by szft com on may at
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.