Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
9,588 | 2,615,163,170 | IssuesEvent | 2015-03-01 06:42:41 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | DIR-655 B1 WPS lock down | auto-migrated Priority-Triage Type-Defect | ```
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.): Latest (Backtrack5 R2)
1. What operating system are you using (Linux is the only supported OS)?
Backtrack5 R2
2. Is your wireless card in monitor mode (yes/no)? Yes
3. What is the signal strength of the Access Point you are trying to crack? 35
to 40
4. What is the manufacturer and model # of the device you are trying to
crack? DLink DIR-655 B1
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b **my BSSID** -vv --lock-delay=630
6. Please describe what you think the issue is.
After 0.19% progress, I got the warning "Detected AP rate limiting, waiting 60
seconds before re-checking. " So, I checked with wash command if the router had
actually locked down WPS. It turned out that WPS was indeed locked. So, I
increased the lock delay to 315 seconds, but it did not help. I increased it
further to 630 seconds, but that didn't help either. I guess the issue is that
we need to know the WPS lock duration/timeout for DIR 655 B1 router. Please
help.
7. Paste the output from Reaver below.
Detected AP rate limiting, waiting 60 seconds before re-checking.
```
Original issue reported on code.google.com by `varmasuy...@gmail.com` on 29 Jun 2012 at 10:22 | 1.0 | DIR-655 B1 WPS lock down - ```
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.): Latest (Backtrack5 R2)
1. What operating system are you using (Linux is the only supported OS)?
Backtrack5 R2
2. Is your wireless card in monitor mode (yes/no)? Yes
3. What is the signal strength of the Access Point you are trying to crack? 35
to 40
4. What is the manufacturer and model # of the device you are trying to
crack? DLink DIR-655 B1
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b **my BSSID** -vv --lock-delay=630
6. Please describe what you think the issue is.
After 0.19% progress, I got the warning "Detected AP rate limiting, waiting 60
seconds before re-checking. " So, I checked with wash command if the router had
actually locked down WPS. It turned out that WPS was indeed locked. So, I
increased the lock delay to 315 seconds, but it did not help. I increased it
further to 630 seconds, but that didn't help either. I guess the issue is that
we need to know the WPS lock duration/timeout for DIR 655 B1 router. Please
help.
7. Paste the output from Reaver below.
Detected AP rate limiting, waiting 60 seconds before re-checking.
```
Original issue reported on code.google.com by `varmasuy...@gmail.com` on 29 Jun 2012 at 10:22 | defect | dir wps lock down what version of reaver are you using only defects against the latest version will be considered latest what operating system are you using linux is the only supported os is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack to what is the manufacturer and model of the device you are trying to crack dlink dir what is the entire command line string you are supplying to reaver reaver i b my bssid vv lock delay please describe what you think the issue is after progress i got the warning detected ap rate limiting waiting seconds before re checking so i checked with wash command if the router had actually locked down wps it turned out that wps was indeed locked so i increased the lock delay to seconds but it did not help i increased it further to seconds but that didn t help either i guess the issue is that we need to know the wps lock duration timeout for dir router please help paste the output from reaver below detected ap rate limiting waiting seconds before re checking original issue reported on code google com by varmasuy gmail com on jun at | 1 |
277,499 | 30,659,260,374 | IssuesEvent | 2023-07-25 14:03:57 | rsoreq/zenbot | https://api.github.com/repos/rsoreq/zenbot | closed | CVE-2022-3517 (High) detected in minimatch-3.0.4.tgz - autoclosed | Mend: dependency security vulnerability | ## CVE-2022-3517 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p>
<p>
Dependency Hierarchy:
- shelljs-0.8.4.tgz (Root Library)
- glob-7.1.6.tgz
- :x: **minimatch-3.0.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rsoreq/zenbot/commit/7a24c0d7b98ee76e6bac827974cff490a7694378">7a24c0d7b98ee76e6bac827974cff490a7694378</a></p>
<p>Found in base branch: <b>unstable</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3517>CVE-2022-3517</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: minimatch - 3.0.5</p>
</p>
</details>
<p></p>
| True | CVE-2022-3517 (High) detected in minimatch-3.0.4.tgz - autoclosed - ## CVE-2022-3517 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p>
<p>
Dependency Hierarchy:
- shelljs-0.8.4.tgz (Root Library)
- glob-7.1.6.tgz
- :x: **minimatch-3.0.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rsoreq/zenbot/commit/7a24c0d7b98ee76e6bac827974cff490a7694378">7a24c0d7b98ee76e6bac827974cff490a7694378</a></p>
<p>Found in base branch: <b>unstable</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3517>CVE-2022-3517</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: minimatch - 3.0.5</p>
</p>
</details>
<p></p>
| non_defect | cve high detected in minimatch tgz autoclosed cve high severity vulnerability vulnerable library minimatch tgz a glob matcher in javascript library home page a href dependency hierarchy shelljs tgz root library glob tgz x minimatch tgz vulnerable library found in head commit a href found in base branch unstable vulnerability details a vulnerability was found in the minimatch package this flaw allows a regular expression denial of service redos when calling the braceexpand function with specific arguments resulting in a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimatch | 0 |
22,233 | 6,229,921,655 | IssuesEvent | 2017-07-11 06:18:00 | XceedBoucherS/TestImport5 | https://api.github.com/repos/XceedBoucherS/TestImport5 | closed | Unable to commit changes | CodePlex | <b>Nimgoble[CodePlex]</b> <br />Hello,
I'm unable to commit changes that I've made.
I keep receiving a Commit Failed error. I've attached a screenshot of the error.
I'm using Visual Studio 2010 with VisualSVN.
Thank you
| 1.0 | Unable to commit changes - <b>Nimgoble[CodePlex]</b> <br />Hello,
I'm unable to commit changes that I've made.
I keep receiving a Commit Failed error. I've attached a screenshot of the error.
I'm using Visual Studio 2010 with VisualSVN.
Thank you
| non_defect | unable to commit changes nimgoble hello i m unable to commit changes that i ve made i keep receiving a commit failed error i ve attached a screenshot of the error i m using visual studio with visualsvn thank you | 0 |
38,042 | 8,639,913,268 | IssuesEvent | 2018-11-23 22:39:15 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | Comparer<string>.Default not using System.String.compare | defect in-progress | Comparer<string>.Default.Compare(str1, str2) gives a different result the str1CompareTo(str2).
The first case calls Bridge.compare, and seems to perform char sorting.
The second case calls System.String.compare and seems to correctly sort text.
https://deck.net/42fe359ec3524e6024107106ec9c9386
I would expect that Comparer<string>.Default would call System.String.compare as System.String implements IComparer<System.String>.
| 1.0 | Comparer<string>.Default not using System.String.compare - Comparer<string>.Default.Compare(str1, str2) gives a different result the str1CompareTo(str2).
The first case calls Bridge.compare, and seems to perform char sorting.
The second case calls System.String.compare and seems to correctly sort text.
https://deck.net/42fe359ec3524e6024107106ec9c9386
I would expect that Comparer<string>.Default would call System.String.compare as System.String implements IComparer<System.String>.
| defect | comparer default not using system string compare comparer default compare gives a different result the the first case calls bridge compare and seems to perform char sorting the second case calls system string compare and seems to correctly sort text i would expect that comparer default would call system string compare as system string implements icomparer | 1 |
32,090 | 8,793,887,899 | IssuesEvent | 2018-12-21 22:01:11 | angular/angular-cli | https://api.github.com/repos/angular/angular-cli | closed | Less Sourcemap only references files inside 'styles' config | comp: devkit/build-angular freq2: medium severity2: inconvenient type: bug/fix | <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request
```
- [x] bug report
- [ ] feature request
```
### Versions.
<!--
Output from: `ng --version`.
If nothing, output from: `node --version` and `npm --version`.
Windows (7/8/10). Linux (incl. distribution). macOS (El Capitan? Sierra?)
-->
@angular/cli: 1.0.0-rc.2
macOS El Capitan
### Repro steps.
<!--
Simple steps to reproduce this bug.
Please include: commands run, packages added, related code changes.
A link to a sample repo would help too.
-->
My `angular-cli.json`:
```
"styles": [
"styles.less"
],
```
My `styles.less`
```
@import "less/vars";
@import "less/base";
@import "less/layout";
// ...etc.
```
When I run `ng serve -ec -sm` I can see the sourcemap file generated, but Chrome's devtools lists `styles.less` as the sole source file instead of `base.less` or `layout.less`, etc.
### The log given by the failure.
<!-- Normally this include a stack trace and some more information. -->
Here's what I skimmed out of the generated sourcemap file. You can see there's no mention of the imported files whatsoever.
```
{"version":3,"sources":["webpack:///./src/styles.less"], ... \n\n\n\n// WEBPACK FOOTER //\n// ./src/styles.less"],"sourceRoot":""}
```
### Desired functionality.
<!--
What would like to see implemented?
What is the usecase?
-->
I should see the individual imported files in the Chrome devtools and not just the main file.
### Mention any other details that might be useful.
<!-- Please include a link to the repo if this is related to an OSS project. -->
Wonder if this is solved by today's release of [less-loader v4.0.0](https://github.com/webpack-contrib/less-loader/releases/tag/v4.0.0)?
| 1.0 | Less Sourcemap only references files inside 'styles' config - <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request
```
- [x] bug report
- [ ] feature request
```
### Versions.
<!--
Output from: `ng --version`.
If nothing, output from: `node --version` and `npm --version`.
Windows (7/8/10). Linux (incl. distribution). macOS (El Capitan? Sierra?)
-->
@angular/cli: 1.0.0-rc.2
macOS El Capitan
### Repro steps.
<!--
Simple steps to reproduce this bug.
Please include: commands run, packages added, related code changes.
A link to a sample repo would help too.
-->
My `angular-cli.json`:
```
"styles": [
"styles.less"
],
```
My `styles.less`
```
@import "less/vars";
@import "less/base";
@import "less/layout";
// ...etc.
```
When I run `ng serve -ec -sm` I can see the sourcemap file generated, but Chrome's devtools lists `styles.less` as the sole source file instead of `base.less` or `layout.less`, etc.
### The log given by the failure.
<!-- Normally this include a stack trace and some more information. -->
Here's what I skimmed out of the generated sourcemap file. You can see there's no mention of the imported files whatsoever.
```
{"version":3,"sources":["webpack:///./src/styles.less"], ... \n\n\n\n// WEBPACK FOOTER //\n// ./src/styles.less"],"sourceRoot":""}
```
### Desired functionality.
<!--
What would like to see implemented?
What is the usecase?
-->
I should see the individual imported files in the Chrome devtools and not just the main file.
### Mention any other details that might be useful.
<!-- Please include a link to the repo if this is related to an OSS project. -->
Wonder if this is solved by today's release of [less-loader v4.0.0](https://github.com/webpack-contrib/less-loader/releases/tag/v4.0.0)?
| non_defect | less sourcemap only references files inside styles config if you don t fill out the following information your issue might be closed without investigating bug report or feature request bug report feature request versions output from ng version if nothing output from node version and npm version windows linux incl distribution macos el capitan sierra angular cli rc macos el capitan repro steps simple steps to reproduce this bug please include commands run packages added related code changes a link to a sample repo would help too my angular cli json styles styles less my styles less import less vars import less base import less layout etc when i run ng serve ec sm i can see the sourcemap file generated but chrome s devtools lists styles less as the sole source file instead of base less or layout less etc the log given by the failure here s what i skimmed out of the generated sourcemap file you can see there s no mention of the imported files whatsoever version sources n n n n webpack footer n src styles less sourceroot desired functionality what would like to see implemented what is the usecase i should see the individual imported files in the chrome devtools and not just the main file mention any other details that might be useful wonder if this is solved by today s release of | 0 |
114,668 | 17,258,891,998 | IssuesEvent | 2021-07-22 02:56:10 | nforgeio/neonKUBE | https://api.github.com/repos/nforgeio/neonKUBE | closed | neon-cluster-manager DB credentials are hardcoded and insecure | bug cluster-setup neon-kube security | I was going to copy/paste some code from **neon-cluster-manager** and noticed a couple issues:
- [ ] @marcusbooyah: You define `Service.connString` as a hardcoded string with an insecure password. You should create the database user with a random password when creating the cluster and then loading that password via a secret k8s configuration passed to the service. I think you should be able to do this using the following steps while setting up the cluster:
* Generate a random password
* Persist the secret to the cluster configuration (this will require #1084)
* Create the database and user (you shouldn't be using the `postgres` user because its the **superuser**)
* You grant only the permissions required
* Persist the connection string as a k8s secret
* Modify the **neon-cluster-manager** to read the configuration
- [x] I noticed that the `Service` class (that I renamed earlier as a simplifying convention) is `partial` but that the other parts of the definition were hard to find because their file names weren't prefixed by **"Service.**. I've been using the convention for `partial` classes is that the file with the constructor will be have the base name **Service.cs** in this case and that the remaining files will be named with the base name, followed by a dot and a name that identifies what's implemented inside.
I went ahead and modified the other files to be named **Service.Kibana.cs**, **Service.LogPurger**, and **Service.Setup**. This makes it much easier to grok what's going on. | True | neon-cluster-manager DB credentials are hardcoded and insecure - I was going to copy/paste some code from **neon-cluster-manager** and noticed a couple issues:
- [ ] @marcusbooyah: You define `Service.connString` as a hardcoded string with an insecure password. You should create the database user with a random password when creating the cluster and then loading that password via a secret k8s configuration passed to the service. I think you should be able to do this using the following steps while setting up the cluster:
* Generate a random password
* Persist the secret to the cluster configuration (this will require #1084)
* Create the database and user (you shouldn't be using the `postgres` user because its the **superuser**)
* You grant only the permissions required
* Persist the connection string as a k8s secret
* Modify the **neon-cluster-manager** to read the configuration
- [x] I noticed that the `Service` class (that I renamed earlier as a simplifying convention) is `partial` but that the other parts of the definition were hard to find because their file names weren't prefixed by **"Service.**. I've been using the convention for `partial` classes is that the file with the constructor will be have the base name **Service.cs** in this case and that the remaining files will be named with the base name, followed by a dot and a name that identifies what's implemented inside.
I went ahead and modified the other files to be named **Service.Kibana.cs**, **Service.LogPurger**, and **Service.Setup**. This makes it much easier to grok what's going on. | non_defect | neon cluster manager db credentials are hardcoded and insecure i was going to copy paste some code from neon cluster manager and noticed a couple issues marcusbooyah you define service connstring as a hardcoded string with an insecure password you should create the database user with a random password when creating the cluster and then loading that password via a secret configuration passed to the service i think you should be able to do this using the following steps while setting up the cluster generate a random password persist the secret to the cluster configuration this will require create the database and user you shouldn t be using the postgres user because its the superuser you grant only the permissions required persist the connection string as a secret modify the neon cluster manager to read the configuration nbsp i noticed that the service class that i renamed earlier as a simplifying convention is partial but that the other parts of the definition were hard to find because their file names weren t prefixed by service i ve been using the convention for partial classes is that the file with the constructor will be have the base name service cs in this case and that the remaining files will be named with the base name followed by a dot and a name that identifies what s implemented inside i went ahead and modified the other files to be named service kibana cs service logpurger and service setup this makes it much easier to grok what s going on | 0 |
72,951 | 24,381,715,324 | IssuesEvent | 2022-10-04 08:24:27 | PowerDNS/pdns | https://api.github.com/repos/PowerDNS/pdns | closed | rec: Inconsistency in logging between lua and protobuf | rec defect | <!-- Hi! Thanks for filing an issue. It will be read with care by human beings. Can we ask you to please fill out this template and not simply demand new features or send in complaints? Thanks! -->
<!-- Also please search the existing issues (both open and closed) to see if your report might be duplicate -->
<!-- Please don't file an issue when you have a support question, send support questions to the mailinglist or ask them on IRC (https://www.powerdns.com/opensource.html) -->
<!-- Tell us what is issue is about -->
- Program: Recursor
- Issue type: Bug report
### Short description
Just a quick email to say that we're seeing an inconsistency in the PowerDNS (4.3.0 beta 2) output between lua and protobuf. When we request a domain that is listed in a whitelist RPZ, the response.appliedPolicy is not being set in the protobuf output. The lua scripting environment does appear to see the policy set, as shown by logging dq.appliedPolicy.policyName. Is it possible to confirm if this is a known and / or expected behaviour? Our assumption is that any RPZ policy matching applied and recorded in dq.appliedPolicy.policyName should also be reflected in the protobuf 'response' output.
### Environment
<!-- Tell us about the environment -->
- Operating system: Not sure, Centos I think
- Software version: 4.3.0 beta2
- Software source: PowerDNS repository
### Steps to reproduce
See above.
| 1.0 | rec: Inconsistency in logging between lua and protobuf - <!-- Hi! Thanks for filing an issue. It will be read with care by human beings. Can we ask you to please fill out this template and not simply demand new features or send in complaints? Thanks! -->
<!-- Also please search the existing issues (both open and closed) to see if your report might be duplicate -->
<!-- Please don't file an issue when you have a support question, send support questions to the mailinglist or ask them on IRC (https://www.powerdns.com/opensource.html) -->
<!-- Tell us what is issue is about -->
- Program: Recursor
- Issue type: Bug report
### Short description
Just a quick email to say that we're seeing an inconsistency in the PowerDNS (4.3.0 beta 2) output between lua and protobuf. When we request a domain that is listed in a whitelist RPZ, the response.appliedPolicy is not being set in the protobuf output. The lua scripting environment does appear to see the policy set, as shown by logging dq.appliedPolicy.policyName. Is it possible to confirm if this is a known and / or expected behaviour? Our assumption is that any RPZ policy matching applied and recorded in dq.appliedPolicy.policyName should also be reflected in the protobuf 'response' output.
### Environment
<!-- Tell us about the environment -->
- Operating system: Not sure, Centos I think
- Software version: 4.3.0 beta2
- Software source: PowerDNS repository
### Steps to reproduce
See above.
| defect | rec inconsistency in logging between lua and protobuf program recursor issue type bug report short description just a quick email to say that we re seeing an inconsistency in the powerdns beta output between lua and protobuf when we request a domain that is listed in a whitelist rpz the response appliedpolicy is not being set in the protobuf output the lua scripting environment does appear to see the policy set as shown by logging dq appliedpolicy policyname is it possible to confirm if this is a known and or expected behaviour our assumption is that any rpz policy matching applied and recorded in dq appliedpolicy policyname should also be reflected in the protobuf response output environment operating system not sure centos i think software version software source powerdns repository steps to reproduce see above | 1 |
90,007 | 3,808,117,188 | IssuesEvent | 2016-03-25 13:22:17 | iSoron/uhabits | https://api.github.com/repos/iSoron/uhabits | closed | Import data from HabitBull | low-priority new-feature | [HabitBull](https://play.google.com/store/apps/details?id=com.oristats.habitbull) is one of the most popular habit apps on Android and iOS. It does allow to export habits into a .csv file. Would be great, if Loop could import them. | 1.0 | Import data from HabitBull - [HabitBull](https://play.google.com/store/apps/details?id=com.oristats.habitbull) is one of the most popular habit apps on Android and iOS. It does allow to export habits into a .csv file. Would be great, if Loop could import them. | non_defect | import data from habitbull is one of the most popular habit apps on android and ios it does allow to export habits into a csv file would be great if loop could import them | 0 |
33,751 | 27,781,591,028 | IssuesEvent | 2023-03-16 21:35:55 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | opened | Unnecessary ApiCompat CP0001 errors when type forwards used | bug area-Infrastructure-libraries |
See https://github.com/dotnet/runtime/pull/82453#discussion_r1139284637 for more detail.
Local errors for when the https://github.com/dotnet/runtime/tree/main/src/libraries/System.DirectoryServices/src/CompatibilitySuppressions.xml file is removed (please delete that file when this issue is fixed):
```
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error : API compatibility errors between 'lib/net6.0/System.DirectoryServices.dll' (C:\Users\sharter\.nuget\packages\
system.directoryservices\7.0.0\system.directoryservices.7.0.0.nupkg) and 'lib/net6.0/System.DirectoryServices.dll' (C:\git\runtime1\artifacts\packages\Debug\Shipping\System.DirectoryServices.8.0.0-dev.nupkg): [C:\git\runtime1\src\libraries\System.DirectoryServices\src\
System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermission' exists on [Baseline] lib/net6.0/System.Dire
ctoryServices.dll but not on lib/net6.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAccess' exists on [Baseline] lib/net6.0/Syste
m.DirectoryServices.dll but not on lib/net6.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAttribute' exists on [Baseline] lib/net6.0/Sy
stem.DirectoryServices.dll but not on lib/net6.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntry' exists on [Baseline] lib/net6.0/System
.DirectoryServices.dll but not on lib/net6.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntryCollection' exists on [Baseline] lib/net
6.0/System.DirectoryServices.dll but not on lib/net6.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error : API compatibility errors between 'lib/net7.0/System.DirectoryServices.dll' (C:\Users\sharter\.nuget\packages\
system.directoryservices\7.0.0\system.directoryservices.7.0.0.nupkg) and 'lib/net7.0/System.DirectoryServices.dll' (C:\git\runtime1\artifacts\packages\Debug\Shipping\System.DirectoryServices.8.0.0-dev.nupkg): [C:\git\runtime1\src\libraries\System.DirectoryServices\src\
System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermission' exists on [Baseline] lib/net7.0/System.Dire
ctoryServices.dll but not on lib/net7.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAccess' exists on [Baseline] lib/net7.0/Syste
m.DirectoryServices.dll but not on lib/net7.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAttribute' exists on [Baseline] lib/net7.0/Sy
stem.DirectoryServices.dll but not on lib/net7.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntry' exists on [Baseline] lib/net7.0/System
.DirectoryServices.dll but not on lib/net7.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntryCollection' exists on [Baseline] lib/net
7.0/System.DirectoryServices.dll but not on lib/net7.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error : API compatibility errors between 'lib/netstandard2.0/System.DirectoryServices.dll' (C:\Users\sharter\.nuget\p
ackages\system.directoryservices\7.0.0\system.directoryservices.7.0.0.nupkg) and 'lib/netstandard2.0/System.DirectoryServices.dll' (C:\git\runtime1\artifacts\packages\Debug\Shipping\System.DirectoryServices.8.0.0-dev.nupkg): [C:\git\runtime1\src\libraries\System.Direct
oryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermission' exists on [Baseline] lib/netstandard2.0/Sys
tem.DirectoryServices.dll but not on lib/netstandard2.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAccess' exists on [Baseline] lib/netstandard2
.0/System.DirectoryServices.dll but not on lib/netstandard2.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAttribute' exists on [Baseline] lib/netstanda
rd2.0/System.DirectoryServices.dll but not on lib/netstandard2.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntry' exists on [Baseline] lib/netstandard2.
0/System.DirectoryServices.dll but not on lib/netstandard2.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntryCollection' exists on [Baseline] lib/net
standard2.0/System.DirectoryServices.dll but not on lib/netstandard2.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
``` | 1.0 | Unnecessary ApiCompat CP0001 errors when type forwards used -
See https://github.com/dotnet/runtime/pull/82453#discussion_r1139284637 for more detail.
Local errors for when the https://github.com/dotnet/runtime/tree/main/src/libraries/System.DirectoryServices/src/CompatibilitySuppressions.xml file is removed (please delete that file when this issue is fixed):
```
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error : API compatibility errors between 'lib/net6.0/System.DirectoryServices.dll' (C:\Users\sharter\.nuget\packages\
system.directoryservices\7.0.0\system.directoryservices.7.0.0.nupkg) and 'lib/net6.0/System.DirectoryServices.dll' (C:\git\runtime1\artifacts\packages\Debug\Shipping\System.DirectoryServices.8.0.0-dev.nupkg): [C:\git\runtime1\src\libraries\System.DirectoryServices\src\
System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermission' exists on [Baseline] lib/net6.0/System.Dire
ctoryServices.dll but not on lib/net6.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAccess' exists on [Baseline] lib/net6.0/Syste
m.DirectoryServices.dll but not on lib/net6.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAttribute' exists on [Baseline] lib/net6.0/Sy
stem.DirectoryServices.dll but not on lib/net6.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntry' exists on [Baseline] lib/net6.0/System
.DirectoryServices.dll but not on lib/net6.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntryCollection' exists on [Baseline] lib/net
6.0/System.DirectoryServices.dll but not on lib/net6.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error : API compatibility errors between 'lib/net7.0/System.DirectoryServices.dll' (C:\Users\sharter\.nuget\packages\
system.directoryservices\7.0.0\system.directoryservices.7.0.0.nupkg) and 'lib/net7.0/System.DirectoryServices.dll' (C:\git\runtime1\artifacts\packages\Debug\Shipping\System.DirectoryServices.8.0.0-dev.nupkg): [C:\git\runtime1\src\libraries\System.DirectoryServices\src\
System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermission' exists on [Baseline] lib/net7.0/System.Dire
ctoryServices.dll but not on lib/net7.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAccess' exists on [Baseline] lib/net7.0/Syste
m.DirectoryServices.dll but not on lib/net7.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAttribute' exists on [Baseline] lib/net7.0/Sy
stem.DirectoryServices.dll but not on lib/net7.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntry' exists on [Baseline] lib/net7.0/System
.DirectoryServices.dll but not on lib/net7.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntryCollection' exists on [Baseline] lib/net
7.0/System.DirectoryServices.dll but not on lib/net7.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error : API compatibility errors between 'lib/netstandard2.0/System.DirectoryServices.dll' (C:\Users\sharter\.nuget\p
ackages\system.directoryservices\7.0.0\system.directoryservices.7.0.0.nupkg) and 'lib/netstandard2.0/System.DirectoryServices.dll' (C:\git\runtime1\artifacts\packages\Debug\Shipping\System.DirectoryServices.8.0.0-dev.nupkg): [C:\git\runtime1\src\libraries\System.Direct
oryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermission' exists on [Baseline] lib/netstandard2.0/Sys
tem.DirectoryServices.dll but not on lib/netstandard2.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAccess' exists on [Baseline] lib/netstandard2
.0/System.DirectoryServices.dll but not on lib/netstandard2.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionAttribute' exists on [Baseline] lib/netstanda
rd2.0/System.DirectoryServices.dll but not on lib/netstandard2.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntry' exists on [Baseline] lib/netstandard2.
0/System.DirectoryServices.dll but not on lib/netstandard2.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
C:\Users\sharter\.nuget\packages\microsoft.dotnet.apicompat.task\8.0.100-preview.2.23107.1\build\Microsoft.NET.ApiCompat.ValidatePackage.targets(39,5): error CP0001: Type 'System.DirectoryServices.DirectoryServicesPermissionEntryCollection' exists on [Baseline] lib/net
standard2.0/System.DirectoryServices.dll but not on lib/netstandard2.0/System.DirectoryServices.dll [C:\git\runtime1\src\libraries\System.DirectoryServices\src\System.DirectoryServices.csproj]
``` | non_defect | unnecessary apicompat errors when type forwards used see for more detail local errors for when the file is removed please delete that file when this issue is fixed c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error api compatibility errors between lib system directoryservices dll c users sharter nuget packages system directoryservices system directoryservices nupkg and lib system directoryservices dll c git artifacts packages debug shipping system directoryservices dev nupkg c git src libraries system directoryservices src system directoryservices csproj c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermission exists on lib system dire ctoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionaccess exists on lib syste m directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionattribute exists on lib sy stem directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionentry exists on lib system directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionentrycollection exists on lib net system directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error api compatibility errors between lib system directoryservices dll c users sharter nuget packages system directoryservices system directoryservices nupkg and lib system directoryservices dll c git artifacts packages debug shipping system directoryservices dev nupkg c git src libraries system directoryservices src system directoryservices csproj c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermission exists on lib system dire ctoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionaccess exists on lib syste m directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionattribute exists on lib sy stem directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionentry exists on lib system directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionentrycollection exists on lib net system directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error api compatibility errors between lib system directoryservices dll c users sharter nuget p ackages system directoryservices system directoryservices nupkg and lib system directoryservices dll c git artifacts packages debug shipping system directoryservices dev nupkg c git src libraries system direct oryservices src system directoryservices csproj c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermission exists on lib sys tem directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionaccess exists on lib system directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionattribute exists on lib netstanda system directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionentry exists on lib system directoryservices dll but not on lib system directoryservices dll c users sharter nuget packages microsoft dotnet apicompat task preview build microsoft net apicompat validatepackage targets error type system directoryservices directoryservicespermissionentrycollection exists on lib net system directoryservices dll but not on lib system directoryservices dll | 0 |
151,913 | 13,438,549,892 | IssuesEvent | 2020-09-07 18:24:50 | featherity/featherity | https://api.github.com/repos/featherity/featherity | opened | Update CONTRIBUTING file with rule for one PR for each icon or icon series. | documentation | Regarding to the thread in #57 @llaenowyd said this:
> I think after re-reading the submission guidelines , maybe there should be one PR for each icon.
I certain way I think this is a good idea to write this down in CONTRIBUTING file.
For me it's a little-bit of an unwritten rule that you need to submit icons or "icon series" in separate pull-requests for several reasons:
- It's easier to review
- Keeps the discussions and reviews about the icon(s) scoped and in one thread.
- All points above contribute to less time to merge PR's
I'm curious what you think. | 1.0 | Update CONTRIBUTING file with rule for one PR for each icon or icon series. - Regarding to the thread in #57 @llaenowyd said this:
> I think after re-reading the submission guidelines , maybe there should be one PR for each icon.
I certain way I think this is a good idea to write this down in CONTRIBUTING file.
For me it's a little-bit of an unwritten rule that you need to submit icons or "icon series" in separate pull-requests for several reasons:
- It's easier to review
- Keeps the discussions and reviews about the icon(s) scoped and in one thread.
- All points above contribute to less time to merge PR's
I'm curious what you think. | non_defect | update contributing file with rule for one pr for each icon or icon series regarding to the thread in llaenowyd said this i think after re reading the submission guidelines maybe there should be one pr for each icon i certain way i think this is a good idea to write this down in contributing file for me it s a little bit of an unwritten rule that you need to submit icons or icon series in separate pull requests for several reasons it s easier to review keeps the discussions and reviews about the icon s scoped and in one thread all points above contribute to less time to merge pr s i m curious what you think | 0 |
36,601 | 8,030,931,627 | IssuesEvent | 2018-07-27 21:41:47 | IBM/CAST | https://api.github.com/repos/IBM/CAST | opened | Possible Bug: csm_allocation_query_active_all | Comp: CSM Comp: CSM.api PhaseFound: Development Sev: 2 Status: Open Type: Defect | **Describe the bug**
There are some inconsistencies and mismatches between `csm_allocation_query_active_all` VS `csm_allocation_query` and `csm_cluster_query_state` and the CSM database. The number of nodes in an allocation seems to be off sometimes. and sometimes the `compute_nodes` field may not show all the nodes inside that allocation.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Expected behavior**
`csm_allocation_query_active_all` should match the output of other APIs and CSM database information.
**Issue Source:**
@fpizzano noticed this issue on the summit cluster.
Issue ToDo:
- [ ] Nick please investigate.
- [ ] figure out why this behavior is occurring.
- [ ] ???
- [ ] Profit
| 1.0 | Possible Bug: csm_allocation_query_active_all - **Describe the bug**
There are some inconsistencies and mismatches between `csm_allocation_query_active_all` VS `csm_allocation_query` and `csm_cluster_query_state` and the CSM database. The number of nodes in an allocation seems to be off sometimes. and sometimes the `compute_nodes` field may not show all the nodes inside that allocation.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Expected behavior**
`csm_allocation_query_active_all` should match the output of other APIs and CSM database information.
**Issue Source:**
@fpizzano noticed this issue on the summit cluster.
Issue ToDo:
- [ ] Nick please investigate.
- [ ] figure out why this behavior is occurring.
- [ ] ???
- [ ] Profit
| defect | possible bug csm allocation query active all describe the bug there are some inconsistencies and mismatches between csm allocation query active all vs csm allocation query and csm cluster query state and the csm database the number of nodes in an allocation seems to be off sometimes and sometimes the compute nodes field may not show all the nodes inside that allocation to reproduce steps to reproduce the behavior go to click on scroll down to see error expected behavior csm allocation query active all should match the output of other apis and csm database information issue source fpizzano noticed this issue on the summit cluster issue todo nick please investigate figure out why this behavior is occurring profit | 1 |
134,763 | 10,928,811,435 | IssuesEvent | 2019-11-22 19:56:15 | status-im/nim-beacon-chain | https://api.github.com/repos/status-im/nim-beacon-chain | opened | Reenable BLS nd shuffling tests | test suite | See https://github.com/status-im/nim-beacon-chain/pull/585 it seems like we removed the json tests too early.
So we need to update the shuffling and BLS tests to the yaml format. | 1.0 | Reenable BLS nd shuffling tests - See https://github.com/status-im/nim-beacon-chain/pull/585 it seems like we removed the json tests too early.
So we need to update the shuffling and BLS tests to the yaml format. | non_defect | reenable bls nd shuffling tests see it seems like we removed the json tests too early so we need to update the shuffling and bls tests to the yaml format | 0 |
32,595 | 12,131,321,359 | IssuesEvent | 2020-04-23 04:21:17 | kenferrara/spark | https://api.github.com/repos/kenferrara/spark | opened | WS-2018-0021 (Medium) detected in bootstrap-2.1.0.min.js, bootstrap-2.1.0.js | security vulnerability | ## WS-2018-0021 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bootstrap-2.1.0.min.js</b>, <b>bootstrap-2.1.0.js</b></p></summary>
<p>
<details><summary><b>bootstrap-2.1.0.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.min.js</a></p>
<p>Path to vulnerable library: /spark/docs/js/vendor/bootstrap.min.js,/spark/docs/js/vendor/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-2.1.0.min.js** (Vulnerable Library)
</details>
<details><summary><b>bootstrap-2.1.0.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js</a></p>
<p>Path to vulnerable library: /spark/docs/js/vendor/bootstrap.js,/spark/docs/js/vendor/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-2.1.0.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/spark/commit/486b2d6a475cb2e6ab8e936905d6ceec71fc60ce">486b2d6a475cb2e6ab8e936905d6ceec71fc60ce</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XSS in data-target in bootstrap (3.3.7 and before)
<p>Publish Date: 2017-06-27
<p>URL: <a href=https://github.com/twbs/bootstrap/issues/20184>WS-2018-0021</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p>
<p>Release Date: 2019-06-12</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"2.1.0","isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:2.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"2.1.0","isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:2.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"vulnerabilityIdentifier":"WS-2018-0021","vulnerabilityDetails":"XSS in data-target in bootstrap (3.3.7 and before)","vulnerabilityUrl":"https://github.com/twbs/bootstrap/issues/20184","cvss2Severity":"medium","cvss2Score":"6.5","extraData":{}}</REMEDIATE> --> | True | WS-2018-0021 (Medium) detected in bootstrap-2.1.0.min.js, bootstrap-2.1.0.js - ## WS-2018-0021 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bootstrap-2.1.0.min.js</b>, <b>bootstrap-2.1.0.js</b></p></summary>
<p>
<details><summary><b>bootstrap-2.1.0.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.min.js</a></p>
<p>Path to vulnerable library: /spark/docs/js/vendor/bootstrap.min.js,/spark/docs/js/vendor/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-2.1.0.min.js** (Vulnerable Library)
</details>
<details><summary><b>bootstrap-2.1.0.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/2.1.0/bootstrap.js</a></p>
<p>Path to vulnerable library: /spark/docs/js/vendor/bootstrap.js,/spark/docs/js/vendor/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-2.1.0.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/spark/commit/486b2d6a475cb2e6ab8e936905d6ceec71fc60ce">486b2d6a475cb2e6ab8e936905d6ceec71fc60ce</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XSS in data-target in bootstrap (3.3.7 and before)
<p>Publish Date: 2017-06-27
<p>URL: <a href=https://github.com/twbs/bootstrap/issues/20184>WS-2018-0021</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p>
<p>Release Date: 2019-06-12</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"2.1.0","isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:2.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"twitter-bootstrap","packageVersion":"2.1.0","isTransitiveDependency":false,"dependencyTree":"twitter-bootstrap:2.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"vulnerabilityIdentifier":"WS-2018-0021","vulnerabilityDetails":"XSS in data-target in bootstrap (3.3.7 and before)","vulnerabilityUrl":"https://github.com/twbs/bootstrap/issues/20184","cvss2Severity":"medium","cvss2Score":"6.5","extraData":{}}</REMEDIATE> --> | non_defect | ws medium detected in bootstrap min js bootstrap js ws medium severity vulnerability vulnerable libraries bootstrap min js bootstrap js bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library spark docs js vendor bootstrap min js spark docs js vendor bootstrap min js dependency hierarchy x bootstrap min js vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library spark docs js vendor bootstrap js spark docs js vendor bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href vulnerability details xss in data target in bootstrap and before publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails xss in data target in bootstrap and before vulnerabilityurl | 0 |
44,632 | 12,301,008,942 | IssuesEvent | 2020-05-11 14:48:21 | contao/contao | https://api.github.com/repos/contao/contao | closed | Overridden default template not shown in template selection | defect | **Affected version(s)**
Contao 4.8.5
**Description**
Currently the default template is not shown in the custom template selection, e.g. for a content element. However, this also includes default templates, that have been overriden in the `templates/` folder.


This is confusing, since now you cannot see, that your overridden default template is actually used - _and from which theme_ (wherever applicable).
**How to reproduce**
1. Log into the back end.
2. Go to _Templates_.
3. Create a new `ce_text` template and rename it to `ce_text_foo`.
4. Create a new default `ce_text` template.
5. Go to _Articles_ and create a new _Text_ content element.
The custom template selection will not show the custom default template.
| 1.0 | Overridden default template not shown in template selection - **Affected version(s)**
Contao 4.8.5
**Description**
Currently the default template is not shown in the custom template selection, e.g. for a content element. However, this also includes default templates, that have been overriden in the `templates/` folder.


This is confusing, since now you cannot see, that your overridden default template is actually used - _and from which theme_ (wherever applicable).
**How to reproduce**
1. Log into the back end.
2. Go to _Templates_.
3. Create a new `ce_text` template and rename it to `ce_text_foo`.
4. Create a new default `ce_text` template.
5. Go to _Articles_ and create a new _Text_ content element.
The custom template selection will not show the custom default template.
| defect | overridden default template not shown in template selection affected version s contao description currently the default template is not shown in the custom template selection e g for a content element however this also includes default templates that have been overriden in the templates folder this is confusing since now you cannot see that your overridden default template is actually used and from which theme wherever applicable how to reproduce log into the back end go to templates create a new ce text template and rename it to ce text foo create a new default ce text template go to articles and create a new text content element the custom template selection will not show the custom default template | 1 |
39,061 | 9,188,365,656 | IssuesEvent | 2019-03-06 07:09:24 | beefproject/beef | https://api.github.com/repos/beefproject/beef | closed | beefproject on Ubuntu can't run beef -error | Defect | Verify first that your issue/request has not been posted previously:
* https://github.com/beefproject/beef/issues
* https://github.com/beefproject/beef/wiki/FAQ
Ensure you're using the [latest version of BeEF](https://github.com/beefproject/beef/releases/tag/beef-0.4.7.1).
#### Environment
What version/revision of BeEF are you using? latest version
On what version of Ruby?
On what browser? Firefox
On what operating system? Ubuntu
#### Configuration
Are you using a non-default configuration? No
Have you enabled or disabled any BeEF extensions? No
#### Summary
Please provide a summary of the issue.
I cannot run the beEF project on Ubuntu. the usual ./beef command is not responding and an error message is popping instead:
Traceback (most recent call last):
1: from ./beef:44:in '<main>'
./beef:47:in 'rescue in <main>: undefined method ' print_error' for main:Object (NoMethodError)
#### Expected Behaviour
What was the expected result?
once ./beef command is entered it should run the interface and show a link to the hookpage
#### Actual Behaviour
What was the actual result?
error message
#### Steps to Reproduce
Please provide steps to reproduce this issue.
The first thing: install Ruby on my Ubuntu 18.04.2 image:
sudo apt install ruby ruby-dev
The next installed git:
sudo apt install git
clone the git project:
git clone https://github.com/beefproject/beef
.Next, move into the beef directory:
cd beef
On Ubuntu run:
sudo ./install
Once the installation was successfully completed, I ran:
./beef
I was quickly greeted with the following error. To fix this, I tried to tell BeEF that it needs this gem, so I modified the Gemfile file following these steps:
rm Gemfile.lock. Click Y to remove it.
sudo nano Gemfile
In the file, add the following line:
gem ‘xmlrpc’
Save the file and re-run the installation:
sudo ./install
At this point, the installation should be successful. Try running the following command:
./beef
But another error message popped and I am now stocked:
Traceback (most recent call last):
1: from ./beef:44:in '<main>'
./beef:47:in 'rescue in <main>: undefined method ' print_error' for main:Object (NoMethodError)


#### Additional Information
Please provide any additional information which may be useful in resolving this issue, such as debugging output and relevant screen shots. Debug output can be enabled by specifying `debug: true` in the `config.yaml` configuration file.
| 1.0 | beefproject on Ubuntu can't run beef -error - Verify first that your issue/request has not been posted previously:
* https://github.com/beefproject/beef/issues
* https://github.com/beefproject/beef/wiki/FAQ
Ensure you're using the [latest version of BeEF](https://github.com/beefproject/beef/releases/tag/beef-0.4.7.1).
#### Environment
What version/revision of BeEF are you using? latest version
On what version of Ruby?
On what browser? Firefox
On what operating system? Ubuntu
#### Configuration
Are you using a non-default configuration? No
Have you enabled or disabled any BeEF extensions? No
#### Summary
Please provide a summary of the issue.
I cannot run the beEF project on Ubuntu. the usual ./beef command is not responding and an error message is popping instead:
Traceback (most recent call last):
1: from ./beef:44:in '<main>'
./beef:47:in 'rescue in <main>: undefined method ' print_error' for main:Object (NoMethodError)
#### Expected Behaviour
What was the expected result?
once ./beef command is entered it should run the interface and show a link to the hookpage
#### Actual Behaviour
What was the actual result?
error message
#### Steps to Reproduce
Please provide steps to reproduce this issue.
The first thing: install Ruby on my Ubuntu 18.04.2 image:
sudo apt install ruby ruby-dev
The next installed git:
sudo apt install git
clone the git project:
git clone https://github.com/beefproject/beef
.Next, move into the beef directory:
cd beef
On Ubuntu run:
sudo ./install
Once the installation was successfully completed, I ran:
./beef
I was quickly greeted with the following error. To fix this, I tried to tell BeEF that it needs this gem, so I modified the Gemfile file following these steps:
rm Gemfile.lock. Click Y to remove it.
sudo nano Gemfile
In the file, add the following line:
gem ‘xmlrpc’
Save the file and re-run the installation:
sudo ./install
At this point, the installation should be successful. Try running the following command:
./beef
But another error message popped and I am now stocked:
Traceback (most recent call last):
1: from ./beef:44:in '<main>'
./beef:47:in 'rescue in <main>: undefined method ' print_error' for main:Object (NoMethodError)


#### Additional Information
Please provide any additional information which may be useful in resolving this issue, such as debugging output and relevant screen shots. Debug output can be enabled by specifying `debug: true` in the `config.yaml` configuration file.
| defect | beefproject on ubuntu can t run beef error verify first that your issue request has not been posted previously ensure you re using the environment what version revision of beef are you using latest version on what version of ruby on what browser firefox on what operating system ubuntu configuration are you using a non default configuration no have you enabled or disabled any beef extensions no summary please provide a summary of the issue i cannot run the beef project on ubuntu the usual beef command is not responding and an error message is popping instead traceback most recent call last from beef in beef in rescue in undefined method print error for main object nomethoderror expected behaviour what was the expected result once beef command is entered it should run the interface and show a link to the hookpage actual behaviour what was the actual result error message steps to reproduce please provide steps to reproduce this issue the first thing install ruby on my ubuntu image sudo apt install ruby ruby dev the next installed git sudo apt install git clone the git project git clone next move into the beef directory cd beef on ubuntu run sudo install once the installation was successfully completed i ran beef i was quickly greeted with the following error to fix this i tried to tell beef that it needs this gem so i modified the gemfile file following these steps rm gemfile lock click y to remove it sudo nano gemfile in the file add the following line gem ‘xmlrpc’ save the file and re run the installation sudo install at this point the installation should be successful try running the following command beef but another error message popped and i am now stocked traceback most recent call last from beef in beef in rescue in undefined method print error for main object nomethoderror additional information please provide any additional information which may be useful in resolving this issue such as debugging output and relevant screen shots debug output can be enabled by specifying debug true in the config yaml configuration file | 1 |
58,957 | 16,964,809,966 | IssuesEvent | 2021-06-29 09:34:54 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | UISIs after restart | T-Defect | This is almost certainly captured elsewhere, but:
- I went into the office for a day and didn't use my eleweb nightly on my desktop
- I came home and used it the next day
- I clicked update to update to the latest version
- It's 20 minutes later and a good chunk of my encrypted rooms are still full of UISIs
I suspect it'll sort itself out, but in the absence of any feedback or progress it feels pretty broken. | 1.0 | UISIs after restart - This is almost certainly captured elsewhere, but:
- I went into the office for a day and didn't use my eleweb nightly on my desktop
- I came home and used it the next day
- I clicked update to update to the latest version
- It's 20 minutes later and a good chunk of my encrypted rooms are still full of UISIs
I suspect it'll sort itself out, but in the absence of any feedback or progress it feels pretty broken. | defect | uisis after restart this is almost certainly captured elsewhere but i went into the office for a day and didn t use my eleweb nightly on my desktop i came home and used it the next day i clicked update to update to the latest version it s minutes later and a good chunk of my encrypted rooms are still full of uisis i suspect it ll sort itself out but in the absence of any feedback or progress it feels pretty broken | 1 |
265,408 | 20,092,703,312 | IssuesEvent | 2022-02-06 02:08:55 | atsign-foundation/at_server | https://api.github.com/repos/atsign-foundation/at_server | closed | Review for Missing Documentation - at_persistence_spec (API Reference) | documentation | - [ ] Review for any additional documentation needed on [API Reference](https://pub.dev/documentation/at_persistence_spec/latest/). Right nav bar (samples and examples)
- [ ] Broken [documentation link](https://atsign.dev/at_docs-dev_env/at_persistence_spec/index.html) needs to be fixed
Please review library links as well (current, samples and examples)
LIBRARIES
- [ ] [at_persistence_spec](https://pub.dev/documentation/at_persistence_spec/latest/at_persistence_spec/at_persistence_spec-library.html)
| 1.0 | Review for Missing Documentation - at_persistence_spec (API Reference) - - [ ] Review for any additional documentation needed on [API Reference](https://pub.dev/documentation/at_persistence_spec/latest/). Right nav bar (samples and examples)
- [ ] Broken [documentation link](https://atsign.dev/at_docs-dev_env/at_persistence_spec/index.html) needs to be fixed
Please review library links as well (current, samples and examples)
LIBRARIES
- [ ] [at_persistence_spec](https://pub.dev/documentation/at_persistence_spec/latest/at_persistence_spec/at_persistence_spec-library.html)
| non_defect | review for missing documentation at persistence spec api reference review for any additional documentation needed on right nav bar samples and examples broken needs to be fixed please review library links as well current samples and examples libraries | 0 |
24,145 | 3,917,074,302 | IssuesEvent | 2016-04-21 06:25:10 | irnawansuprapti/openbiz-cubi | https://api.github.com/repos/irnawansuprapti/openbiz-cubi | closed | >>>>>>>>>>>>>>>>health 5756 | auto-migrated Priority-Medium spam Type-Defect | ```
In order to wake up early, you need to go to bed a little early as well. Most
people sit before the television set till they get sleepy. They assume that
they need some entertainment after a hard day's work. But in reality, most
people watch shows that get them emotionally charged or drained, which adds
more stress to their mind.
Have a healthy eating habit. Learn to keep your portion sizes reasonably small
and limited. Never skip your meals in an effort to follow Women health tips to
lose weight. Skipping meals to control your weight could make your hunger go
out of control. You would end up eating and putting on more weight instead of
losing it. You can eat light snacks between meals to reduce your intake during
meals. http://www.strongmenmuscle.com/revtest/
```
Original issue reported on code.google.com by `GaylBuc...@gmail.com` on 16 Apr 2015 at 6:07 | 1.0 | >>>>>>>>>>>>>>>>health 5756 - ```
In order to wake up early, you need to go to bed a little early as well. Most
people sit before the television set till they get sleepy. They assume that
they need some entertainment after a hard day's work. But in reality, most
people watch shows that get them emotionally charged or drained, which adds
more stress to their mind.
Have a healthy eating habit. Learn to keep your portion sizes reasonably small
and limited. Never skip your meals in an effort to follow Women health tips to
lose weight. Skipping meals to control your weight could make your hunger go
out of control. You would end up eating and putting on more weight instead of
losing it. You can eat light snacks between meals to reduce your intake during
meals. http://www.strongmenmuscle.com/revtest/
```
Original issue reported on code.google.com by `GaylBuc...@gmail.com` on 16 Apr 2015 at 6:07 | defect | health in order to wake up early you need to go to bed a little early as well most people sit before the television set till they get sleepy they assume that they need some entertainment after a hard day s work but in reality most people watch shows that get them emotionally charged or drained which adds more stress to their mind have a healthy eating habit learn to keep your portion sizes reasonably small and limited never skip your meals in an effort to follow women health tips to lose weight skipping meals to control your weight could make your hunger go out of control you would end up eating and putting on more weight instead of losing it you can eat light snacks between meals to reduce your intake during meals original issue reported on code google com by gaylbuc gmail com on apr at | 1 |
32,268 | 2,751,749,752 | IssuesEvent | 2015-04-24 11:16:36 | NatLibFi/Finto-data | https://api.github.com/repos/NatLibFi/Finto-data | closed | YSAn seuraaja-suhde ei näy SKOSissa | bug imported InCharge-Technical Onto-YSA Priority-Low sv | _From [osma.suominen@helsinki.fi](https://code.google.com/u/osma.suominen@helsinki.fi/) on December 19, 2013 14:58:41_
YSAssa on seuraaja-suhde joidenkin käsitteiden välillä, esim. atk-ohjelmat -> tietokoneohjelmat: http://vesa.lib.helsinki.fi/cgi-bin/ysa.pl?h=atk-ohjelmat Tämä ei kuitenkaan mitenkään näy SKOS-muodossa jolloin tätä suhdetta ei myöskään voida näyttää Fintossa. Sopiva suhde olisikin varmaan dct:isReplacedBy.
Pitäisi korjata YSAn MARC->SKOS muunnosohjelmaa huomioimaan tämä tieto.
_Original issue: http://code.google.com/p/onki-data/issues/detail?id=38_ | 1.0 | YSAn seuraaja-suhde ei näy SKOSissa - _From [osma.suominen@helsinki.fi](https://code.google.com/u/osma.suominen@helsinki.fi/) on December 19, 2013 14:58:41_
YSAssa on seuraaja-suhde joidenkin käsitteiden välillä, esim. atk-ohjelmat -> tietokoneohjelmat: http://vesa.lib.helsinki.fi/cgi-bin/ysa.pl?h=atk-ohjelmat Tämä ei kuitenkaan mitenkään näy SKOS-muodossa jolloin tätä suhdetta ei myöskään voida näyttää Fintossa. Sopiva suhde olisikin varmaan dct:isReplacedBy.
Pitäisi korjata YSAn MARC->SKOS muunnosohjelmaa huomioimaan tämä tieto.
_Original issue: http://code.google.com/p/onki-data/issues/detail?id=38_ | non_defect | ysan seuraaja suhde ei näy skosissa from on december ysassa on seuraaja suhde joidenkin käsitteiden välillä esim atk ohjelmat tietokoneohjelmat tämä ei kuitenkaan mitenkään näy skos muodossa jolloin tätä suhdetta ei myöskään voida näyttää fintossa sopiva suhde olisikin varmaan dct isreplacedby pitäisi korjata ysan marc skos muunnosohjelmaa huomioimaan tämä tieto original issue | 0 |
12,092 | 9,593,931,976 | IssuesEvent | 2019-05-09 12:52:19 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | failed to unmarshal jolokia JSON response | :infrastructure Metricbeat bug | Jolokia module fails to parse some responses fetched using the GET method with the error:
```
failed to unmarshal jolokia JSON response '{"request":{"mbean":"java.lang:type=OperatingSystem","attribute":"ProcessCpuLoad","type":"read"},"value":2.519526329050139E-4,"timestamp":1557157281,"status":200}': json: cannot unmarshal number into Go struct field Entry.Value of type map[string]interface {}
```
For confirmed bugs, please report:
Version: >= 6.6.2 (only seems to affect GET method)
Steps to Reproduce:
Use this config with the jolokia module:
```
http_method: 'GET'
jmx.mappings:
- mbean: 'java.lang:type=OperatingSystem'
attributes:
- attr: ProcessCpuLoad
field: jvm.process.cpu.load
``` | 1.0 | failed to unmarshal jolokia JSON response - Jolokia module fails to parse some responses fetched using the GET method with the error:
```
failed to unmarshal jolokia JSON response '{"request":{"mbean":"java.lang:type=OperatingSystem","attribute":"ProcessCpuLoad","type":"read"},"value":2.519526329050139E-4,"timestamp":1557157281,"status":200}': json: cannot unmarshal number into Go struct field Entry.Value of type map[string]interface {}
```
For confirmed bugs, please report:
Version: >= 6.6.2 (only seems to affect GET method)
Steps to Reproduce:
Use this config with the jolokia module:
```
http_method: 'GET'
jmx.mappings:
- mbean: 'java.lang:type=OperatingSystem'
attributes:
- attr: ProcessCpuLoad
field: jvm.process.cpu.load
``` | non_defect | failed to unmarshal jolokia json response jolokia module fails to parse some responses fetched using the get method with the error failed to unmarshal jolokia json response request mbean java lang type operatingsystem attribute processcpuload type read value timestamp status json cannot unmarshal number into go struct field entry value of type map interface for confirmed bugs please report version only seems to affect get method steps to reproduce use this config with the jolokia module http method get jmx mappings mbean java lang type operatingsystem attributes attr processcpuload field jvm process cpu load | 0 |
39,656 | 9,603,117,171 | IssuesEvent | 2019-05-10 16:09:59 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | Negative zero literal not supported | defect | The compiler will turn a -0 literal to 0. This is not a big issue, but not strictly correct C# semantics.
https://deck.net/e8fd971e74153e169bc375b9e4db6f11
```csharp
public class Test
{
public static void Main()
{
double negativezero = -0.0; // fails
System.Console.WriteLine(1 / negativezero);
}
}
```
Output:
```
Infinity
```
Expected output:
```
-Infinity
```
## See Also
- #3899 | 1.0 | Negative zero literal not supported - The compiler will turn a -0 literal to 0. This is not a big issue, but not strictly correct C# semantics.
https://deck.net/e8fd971e74153e169bc375b9e4db6f11
```csharp
public class Test
{
public static void Main()
{
double negativezero = -0.0; // fails
System.Console.WriteLine(1 / negativezero);
}
}
```
Output:
```
Infinity
```
Expected output:
```
-Infinity
```
## See Also
- #3899 | defect | negative zero literal not supported the compiler will turn a literal to this is not a big issue but not strictly correct c semantics csharp public class test public static void main double negativezero fails system console writeline negativezero output infinity expected output infinity see also | 1 |
11,467 | 3,005,058,477 | IssuesEvent | 2015-07-26 15:41:31 | cgstudiomap/cgstudiomap | https://api.github.com/repos/cgstudiomap/cgstudiomap | closed | Navbar other pages | design Development | Put a class on homepage in order to prevent the padding in the other pages. | 1.0 | Navbar other pages - Put a class on homepage in order to prevent the padding in the other pages. | non_defect | navbar other pages put a class on homepage in order to prevent the padding in the other pages | 0 |
297,837 | 9,182,295,516 | IssuesEvent | 2019-03-05 12:29:13 | servicemesher/istio-official-translation | https://api.github.com/repos/servicemesher/istio-official-translation | closed | content/about/contribute/github/index.md | lang/zh pending priority/P0 sync/update version/1.1 | 文件路径:content/about/contribute/github/index.md
[源码](https://github.com/istio/istio.github.io/tree/master/content/about/contribute/github/index.md)
[网址](https://istio.io//about/contribute/github/index.htm)
```diff
diff --git a/content/about/contribute/github/index.md b/content/about/contribute/github/index.md
index 649e5ade..fcd930df 100644
--- a/content/about/contribute/github/index.md
+++ b/content/about/contribute/github/index.md
@@ -9,7 +9,7 @@ aliases:
- /about/contribute/creating-a-pull-request
- /about/contribute/editing
- /about/contribute/staging-your-changes
-keywords: [contribute, community, GitHub, PR]
+keywords: [contribute]
---
We're excited that you're interested in contributing to improve and expand
@@ -28,79 +28,30 @@ The documentation is published under the [Apache
## How to contribute
-There are three ways you can contribute to the Istio documentation:
+There are two ways you can contribute to the Istio documentation:
* If you want to edit an existing page, you can open up the page in your
browser and select the **Edit This Page on GitHub** option from the gear menu
at the top right of each page. This takes you to GitHub to edit and
submit the changes.
-* If you want to work on the site in general, follow the steps in our
- [How to add content section](#add).
+* If you want to work on the site in general, you must create a fork of the
+ repository. Click the button below to visit the GitHub repository. Then, you
+ must click the **Fork** button in the upper-right corner of the screen to
+ create a copy of our repository in your GitHub account. Create a clone of
+ your fork and make any changes you want. When you are ready to send those
+ changes to us, push the changes to your fork, go to the index page for your
+ fork, and click **New Pull Request** to let us know about it.
-* If you want to review an existing pull request (PR), follow the steps in our
- [How to review content section](#review)
+<a class="btn btn-istio"
+href="https://github.com/istio/istio.io/">Browse this site's source
+code</a>
Once your changes are merged, they show up immediately on
`preliminary.istio.io`. However, the changes only
show up on `istio.io` the next time we produce a new
release, which happens around once a quarter.
-### How to add content {#add}
-
-To add content you must create a fork of the repository and a PR from
-your fork to the docs main repository. The following steps describe the
-process:
-
-<a class="btn btn-istio"
-href="https://github.com/istio/istio.io/">Browse this site's source
-code</a>
-
-1. Click the button above to visit the GitHub repository.
-
-1. Click the **Fork** button in the upper-right corner of the screen to
- create a copy of our repository in your GitHub account.
-
-1. Create a clone of your fork and make any changes you want.
-1. When you are ready to send those changes to us, push the changes to your
- fork.
-1. Go to the index page for your fork, and click **New Pull Request** to let
- us know about it.
-
-### How to review content {#review}
-
-If your review is small, simply comment on the PR directly. If you review the
-content in detail, follow these steps:
-
-1. Leave a comment on the PR with the text `/hold`. This command prevents the
- PR from being merged before you are able to complete your review.
-
-1. Perform your detailed review. When possible leave specific comments
- directly on the files and lines affected.
-
-1. Provide suggestions to the PR owner in your comments when appropriate. For
- example:
-
- {{< text markdown >}}
- Use present tense to avoid verb congruence issues and
- to make the text easier to understand:
-
- ```suggestion
-
- Pilot maintains an abstract model of the mesh.
-
- ```
- {{< /text >}}
-
-1. Publish your review to share your comments and suggestions with us and the
- PR owner. Request changes as the review warrants.
-
- {{< warning_icon >}} If you do not publish your review, the PR owner and
- the community cannot see your comments.
-
-1. Once you publish your review, leave a comment with the text:
- `/hold cancel`. That command unblocks the PR from being merged.
-
## Previewing your work
When you submit a pull request, your PR page on GitHub shows a link to a
@@ -112,15 +63,3 @@ If you created a fork of the repository, you can preview your changes locally.
See this
[README](https://github.com/istio/istio.io/blob/master/README.md) for
instructions.
-
-## Istio community roles
-
-Depending on your contributions and responsibilities, there are several roles
-you can assume.
-
-Visit our [role summary page](https://github.com/istio/community/blob/master/ROLES.md#role-summary)
-to learn about the roles, the related requirements and responsibilities, and
-the privileges associated with the roles.
-
-Visit our [community page](https://github.com/istio/community) to learn more
-about the Istio community in general.
```
| 1.0 | content/about/contribute/github/index.md - 文件路径:content/about/contribute/github/index.md
[源码](https://github.com/istio/istio.github.io/tree/master/content/about/contribute/github/index.md)
[网址](https://istio.io//about/contribute/github/index.htm)
```diff
diff --git a/content/about/contribute/github/index.md b/content/about/contribute/github/index.md
index 649e5ade..fcd930df 100644
--- a/content/about/contribute/github/index.md
+++ b/content/about/contribute/github/index.md
@@ -9,7 +9,7 @@ aliases:
- /about/contribute/creating-a-pull-request
- /about/contribute/editing
- /about/contribute/staging-your-changes
-keywords: [contribute, community, GitHub, PR]
+keywords: [contribute]
---
We're excited that you're interested in contributing to improve and expand
@@ -28,79 +28,30 @@ The documentation is published under the [Apache
## How to contribute
-There are three ways you can contribute to the Istio documentation:
+There are two ways you can contribute to the Istio documentation:
* If you want to edit an existing page, you can open up the page in your
browser and select the **Edit This Page on GitHub** option from the gear menu
at the top right of each page. This takes you to GitHub to edit and
submit the changes.
-* If you want to work on the site in general, follow the steps in our
- [How to add content section](#add).
+* If you want to work on the site in general, you must create a fork of the
+ repository. Click the button below to visit the GitHub repository. Then, you
+ must click the **Fork** button in the upper-right corner of the screen to
+ create a copy of our repository in your GitHub account. Create a clone of
+ your fork and make any changes you want. When you are ready to send those
+ changes to us, push the changes to your fork, go to the index page for your
+ fork, and click **New Pull Request** to let us know about it.
-* If you want to review an existing pull request (PR), follow the steps in our
- [How to review content section](#review)
+<a class="btn btn-istio"
+href="https://github.com/istio/istio.io/">Browse this site's source
+code</a>
Once your changes are merged, they show up immediately on
`preliminary.istio.io`. However, the changes only
show up on `istio.io` the next time we produce a new
release, which happens around once a quarter.
-### How to add content {#add}
-
-To add content you must create a fork of the repository and a PR from
-your fork to the docs main repository. The following steps describe the
-process:
-
-<a class="btn btn-istio"
-href="https://github.com/istio/istio.io/">Browse this site's source
-code</a>
-
-1. Click the button above to visit the GitHub repository.
-
-1. Click the **Fork** button in the upper-right corner of the screen to
- create a copy of our repository in your GitHub account.
-
-1. Create a clone of your fork and make any changes you want.
-1. When you are ready to send those changes to us, push the changes to your
- fork.
-1. Go to the index page for your fork, and click **New Pull Request** to let
- us know about it.
-
-### How to review content {#review}
-
-If your review is small, simply comment on the PR directly. If you review the
-content in detail, follow these steps:
-
-1. Leave a comment on the PR with the text `/hold`. This command prevents the
- PR from being merged before you are able to complete your review.
-
-1. Perform your detailed review. When possible leave specific comments
- directly on the files and lines affected.
-
-1. Provide suggestions to the PR owner in your comments when appropriate. For
- example:
-
- {{< text markdown >}}
- Use present tense to avoid verb congruence issues and
- to make the text easier to understand:
-
- ```suggestion
-
- Pilot maintains an abstract model of the mesh.
-
- ```
- {{< /text >}}
-
-1. Publish your review to share your comments and suggestions with us and the
- PR owner. Request changes as the review warrants.
-
- {{< warning_icon >}} If you do not publish your review, the PR owner and
- the community cannot see your comments.
-
-1. Once you publish your review, leave a comment with the text:
- `/hold cancel`. That command unblocks the PR from being merged.
-
## Previewing your work
When you submit a pull request, your PR page on GitHub shows a link to a
@@ -112,15 +63,3 @@ If you created a fork of the repository, you can preview your changes locally.
See this
[README](https://github.com/istio/istio.io/blob/master/README.md) for
instructions.
-
-## Istio community roles
-
-Depending on your contributions and responsibilities, there are several roles
-you can assume.
-
-Visit our [role summary page](https://github.com/istio/community/blob/master/ROLES.md#role-summary)
-to learn about the roles, the related requirements and responsibilities, and
-the privileges associated with the roles.
-
-Visit our [community page](https://github.com/istio/community) to learn more
-about the Istio community in general.
```
| non_defect | content about contribute github index md 文件路径:content about contribute github index md diff diff git a content about contribute github index md b content about contribute github index md index a content about contribute github index md b content about contribute github index md aliases about contribute creating a pull request about contribute editing about contribute staging your changes keywords keywords we re excited that you re interested in contributing to improve and expand the documentation is published under the apache how to contribute there are three ways you can contribute to the istio documentation there are two ways you can contribute to the istio documentation if you want to edit an existing page you can open up the page in your browser and select the edit this page on github option from the gear menu at the top right of each page this takes you to github to edit and submit the changes if you want to work on the site in general follow the steps in our add if you want to work on the site in general you must create a fork of the repository click the button below to visit the github repository then you must click the fork button in the upper right corner of the screen to create a copy of our repository in your github account create a clone of your fork and make any changes you want when you are ready to send those changes to us push the changes to your fork go to the index page for your fork and click new pull request to let us know about it if you want to review an existing pull request pr follow the steps in our review a class btn btn istio href this site s source code once your changes are merged they show up immediately on preliminary istio io however the changes only show up on istio io the next time we produce a new release which happens around once a quarter how to add content add to add content you must create a fork of the repository and a pr from your fork to the docs main repository the following steps describe the process a class btn btn istio href this site s source code click the button above to visit the github repository click the fork button in the upper right corner of the screen to create a copy of our repository in your github account create a clone of your fork and make any changes you want when you are ready to send those changes to us push the changes to your fork go to the index page for your fork and click new pull request to let us know about it how to review content review if your review is small simply comment on the pr directly if you review the content in detail follow these steps leave a comment on the pr with the text hold this command prevents the pr from being merged before you are able to complete your review perform your detailed review when possible leave specific comments directly on the files and lines affected provide suggestions to the pr owner in your comments when appropriate for example use present tense to avoid verb congruence issues and to make the text easier to understand suggestion pilot maintains an abstract model of the mesh publish your review to share your comments and suggestions with us and the pr owner request changes as the review warrants if you do not publish your review the pr owner and the community cannot see your comments once you publish your review leave a comment with the text hold cancel that command unblocks the pr from being merged previewing your work when you submit a pull request your pr page on github shows a link to a if you created a fork of the repository you can preview your changes locally see this for instructions istio community roles depending on your contributions and responsibilities there are several roles you can assume visit our to learn about the roles the related requirements and responsibilities and the privileges associated with the roles visit our to learn more about the istio community in general | 0 |
23,136 | 3,770,013,514 | IssuesEvent | 2016-03-16 13:09:50 | gstreamer-java/gstreamer-java | https://api.github.com/repos/gstreamer-java/gstreamer-java | closed | Making gstreamer-java work with binaries from gstreamer.com | auto-migrated Priority-Medium Type-Defect | ```
I recently found http://www.gstreamer.com/ which offer what appears to be
Windows binaries for gstreamer that are not over a year old. It gives me some
new hope of integrating gstreamer in my program (as the ossbuild binaries crash
too often).
I'm trying to get this to work with gstreamer-java.
The first thing I noticed is that their main dll is called
"libgstreamer-0.10-0.dll". GStreamer-java doesn't pick this up, renaming it to
"libgstreamer-0.10.dll" works, but then I get this output:
(javaw.exe:6692): GStreamer-WARNING **: Failed to load plugin
'C:\Local\GStreamer\2012.5\x86\lib\gstreamer-0.10\libgsta52dec.dll':
`C:\Local\GStreamer\2012.5\x86\lib\gstreamer-0.10\libgsta52dec.dll': The
specified module could not be found.
(javaw.exe:6692): GStreamer-WARNING **: Failed to load plugin
'C:\Local\GStreamer\2012.5\x86\lib\gstreamer-0.10\libgstadder.dll':
`C:\Local\GStreamer\2012.5\x86\lib\gstreamer-0.10\libgstadder.dll': The
specified module could not be found.
(and 50 orso more for all the other plugins).
The files are in the location it says they are, but it doesn't want to load
them apparently. Any idea what this could be?
```
Original issue reported on code.google.com by `john.hen...@gmail.com` on 22 Jun 2012 at 11:42 | 1.0 | Making gstreamer-java work with binaries from gstreamer.com - ```
I recently found http://www.gstreamer.com/ which offer what appears to be
Windows binaries for gstreamer that are not over a year old. It gives me some
new hope of integrating gstreamer in my program (as the ossbuild binaries crash
too often).
I'm trying to get this to work with gstreamer-java.
The first thing I noticed is that their main dll is called
"libgstreamer-0.10-0.dll". GStreamer-java doesn't pick this up, renaming it to
"libgstreamer-0.10.dll" works, but then I get this output:
(javaw.exe:6692): GStreamer-WARNING **: Failed to load plugin
'C:\Local\GStreamer\2012.5\x86\lib\gstreamer-0.10\libgsta52dec.dll':
`C:\Local\GStreamer\2012.5\x86\lib\gstreamer-0.10\libgsta52dec.dll': The
specified module could not be found.
(javaw.exe:6692): GStreamer-WARNING **: Failed to load plugin
'C:\Local\GStreamer\2012.5\x86\lib\gstreamer-0.10\libgstadder.dll':
`C:\Local\GStreamer\2012.5\x86\lib\gstreamer-0.10\libgstadder.dll': The
specified module could not be found.
(and 50 orso more for all the other plugins).
The files are in the location it says they are, but it doesn't want to load
them apparently. Any idea what this could be?
```
Original issue reported on code.google.com by `john.hen...@gmail.com` on 22 Jun 2012 at 11:42 | defect | making gstreamer java work with binaries from gstreamer com i recently found which offer what appears to be windows binaries for gstreamer that are not over a year old it gives me some new hope of integrating gstreamer in my program as the ossbuild binaries crash too often i m trying to get this to work with gstreamer java the first thing i noticed is that their main dll is called libgstreamer dll gstreamer java doesn t pick this up renaming it to libgstreamer dll works but then i get this output javaw exe gstreamer warning failed to load plugin c local gstreamer lib gstreamer dll c local gstreamer lib gstreamer dll the specified module could not be found javaw exe gstreamer warning failed to load plugin c local gstreamer lib gstreamer libgstadder dll c local gstreamer lib gstreamer libgstadder dll the specified module could not be found and orso more for all the other plugins the files are in the location it says they are but it doesn t want to load them apparently any idea what this could be original issue reported on code google com by john hen gmail com on jun at | 1 |
112,243 | 17,080,843,846 | IssuesEvent | 2021-07-08 04:48:40 | MohamedElashri/Zotero-Docker | https://api.github.com/repos/MohamedElashri/Zotero-Docker | opened | CVE-2018-11694 (High) detected in node-sass-4.14.1.tgz, opennmsopennms-source-26.0.0-1 | security vulnerability | ## CVE-2018-11694 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b>, <b>opennmsopennms-source-26.0.0-1</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: Zotero-Docker/web-library/package.json</p>
<p>Path to vulnerable library: Zotero-Docker/web-library/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/MohamedElashri/Zotero-Docker/commit/d9b3b08f65f152a5478ffe8bd588c83a3a7d0758">d9b3b08f65f152a5478ffe8bd588c83a3a7d0758</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.4. A NULL pointer dereference was found in the function Sass::Functions::selector_append which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11694>CVE-2018-11694</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11694">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11694</a></p>
<p>Release Date: 2018-06-04</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-11694 (High) detected in node-sass-4.14.1.tgz, opennmsopennms-source-26.0.0-1 - ## CVE-2018-11694 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b>, <b>opennmsopennms-source-26.0.0-1</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: Zotero-Docker/web-library/package.json</p>
<p>Path to vulnerable library: Zotero-Docker/web-library/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/MohamedElashri/Zotero-Docker/commit/d9b3b08f65f152a5478ffe8bd588c83a3a7d0758">d9b3b08f65f152a5478ffe8bd588c83a3a7d0758</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.4. A NULL pointer dereference was found in the function Sass::Functions::selector_append which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11694>CVE-2018-11694</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11694">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11694</a></p>
<p>Release Date: 2018-06-04</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in node sass tgz opennmsopennms source cve high severity vulnerability vulnerable libraries node sass tgz opennmsopennms source node sass tgz wrapper around libsass library home page a href path to dependency file zotero docker web library package json path to vulnerable library zotero docker web library node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href found in base branch main vulnerability details an issue was discovered in libsass through a null pointer dereference was found in the function sass functions selector append which could be leveraged by an attacker to cause a denial of service application crash or possibly have unspecified other impact publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource | 0 |
49,644 | 13,187,244,774 | IssuesEvent | 2020-08-13 02:48:20 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | [pybindings] Stack address stored into global variable (Trac #1800) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1800">https://code.icecube.wisc.edu/ticket/1800</a>, reported by kjmeagher and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-07-27T09:34:51",
"description": "static analysis has issue with these pybindings, as the maintainer of ipdf i tried to figure this one out but couldn't, someone who is better at boost::python should take a look at this.\n\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-270315.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-48765c.html#EndPath",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1469612091058248",
"component": "cmake",
"summary": "[pybindings] Stack address stored into global variable",
"priority": "normal",
"keywords": "",
"time": "2016-07-27T08:05:15",
"milestone": "Long-Term Future",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [pybindings] Stack address stored into global variable (Trac #1800) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1800">https://code.icecube.wisc.edu/ticket/1800</a>, reported by kjmeagher and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-07-27T09:34:51",
"description": "static analysis has issue with these pybindings, as the maintainer of ipdf i tried to figure this one out but couldn't, someone who is better at boost::python should take a look at this.\n\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-270315.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-48765c.html#EndPath",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1469612091058248",
"component": "cmake",
"summary": "[pybindings] Stack address stored into global variable",
"priority": "normal",
"keywords": "",
"time": "2016-07-27T08:05:15",
"milestone": "Long-Term Future",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | stack address stored into global variable trac migrated from json status closed changetime description static analysis has issue with these pybindings as the maintainer of ipdf i tried to figure this one out but couldn t someone who is better at boost python should take a look at this n n reporter kjmeagher cc resolution fixed ts component cmake summary stack address stored into global variable priority normal keywords time milestone long term future owner nega type defect | 1 |
430,989 | 12,468,361,116 | IssuesEvent | 2020-05-28 18:43:15 | betheluniversity/cascade | https://api.github.com/repos/betheluniversity/cascade | closed | Add a link to add an event to your calendar on Event Pages | low priority | We need to
- Add a link to add the event to your calendar (your personal calendar on your iPhone, for example)
- only after RSVP if there is an RSVP form.
| 1.0 | Add a link to add an event to your calendar on Event Pages - We need to
- Add a link to add the event to your calendar (your personal calendar on your iPhone, for example)
- only after RSVP if there is an RSVP form.
| non_defect | add a link to add an event to your calendar on event pages we need to add a link to add the event to your calendar your personal calendar on your iphone for example only after rsvp if there is an rsvp form | 0 |
401,048 | 27,314,099,668 | IssuesEvent | 2023-02-24 14:27:12 | rrousselGit/riverpod | https://api.github.com/repos/rrousselGit/riverpod | closed | Pub example app fail in web but OK in mobile | bug documentation | https://github.com/rrousselGit/riverpod/tree/master/examples/pub Demo cannot be launched in a web browser.
It throws
```The following JSRangeError was thrown building:
Invalid argument: Maximum call stack size exceeded
...
======== Exception caught by SVG ===================================================================
The following ProgressEvent$ object was thrown resolving a single-frame picture stream:
[object ProgressEvent]
When the exception was thrown, this was the stack:
Picture provider: NetworkPicture("https://pub.dev/static/hash-6pt3begn/img/pub-dev-logo.svg", headers: null, colorFilter: null)
Picture key: PictureKey(Instance of 'NetworkPictureKeyData', colorFilter: null, theme: SvgTheme(currentColor: Color(0xff000000), fontSize: 20, xHeight: 10))
====================================================================================================
```
| 1.0 | Pub example app fail in web but OK in mobile - https://github.com/rrousselGit/riverpod/tree/master/examples/pub Demo cannot be launched in a web browser.
It throws
```The following JSRangeError was thrown building:
Invalid argument: Maximum call stack size exceeded
...
======== Exception caught by SVG ===================================================================
The following ProgressEvent$ object was thrown resolving a single-frame picture stream:
[object ProgressEvent]
When the exception was thrown, this was the stack:
Picture provider: NetworkPicture("https://pub.dev/static/hash-6pt3begn/img/pub-dev-logo.svg", headers: null, colorFilter: null)
Picture key: PictureKey(Instance of 'NetworkPictureKeyData', colorFilter: null, theme: SvgTheme(currentColor: Color(0xff000000), fontSize: 20, xHeight: 10))
====================================================================================================
```
| non_defect | pub example app fail in web but ok in mobile demo cannot be launched in a web browser it throws the following jsrangeerror was thrown building invalid argument maximum call stack size exceeded exception caught by svg the following progressevent object was thrown resolving a single frame picture stream when the exception was thrown this was the stack picture provider networkpicture headers null colorfilter null picture key picturekey instance of networkpicturekeydata colorfilter null theme svgtheme currentcolor color fontsize xheight | 0 |
79,904 | 10,144,832,209 | IssuesEvent | 2019-08-05 00:45:54 | cornellius-gp/gpytorch | https://api.github.com/repos/cornellius-gp/gpytorch | closed | error in example - new data | documentation | Hi @dougalsutherland @jmhessel
Hi I am trying the below example with new data
https://github.com/cornellius-gp/gpytorch/blob/master/examples/03_Multitask_GP_Regression/Multitask_GP_Regression.ipynb
model.train()
likelihood.train()
Use the adam optimizer
optimizer = torch.optim.Adam([
{‘params’: model.parameters()}, # Includes GaussianLikelihood parameters
], lr=0.1)
“Loss” for GPs - the marginal log likelihood
mll = gpytorch.mlls.ExactMarginalLogLikelihood(likelihood, model)
n_iter = 50
for i in range(n_iter):
optimizer.zero_grad()
output = model(train_x)
loss = -mll(output, train_y)
loss.backward()
print(‘Iter %d/%d - Loss: %.3f’ % (i + 1, n_iter, loss.item()))
optimizer.step()
RuntimeError Traceback (most recent call last)
in ()
16 optimizer.zero_grad()
17 output = model(train_x)
—> 18 loss = -mll(output, train_y)
19 loss.backward()
20 print(‘Iter %d/%d - Loss: %.3f’ % (i + 1, n_iter, loss.item()))
3 frames
/usr/local/lib/python3.6/dist-packages/gpytorch/distributions/multivariate_normal.py in log_prob(self, value)
112
113 mean, covar = self.loc, self.lazy_covariance_matrix
–> 114 diff = value - mean
115
116 # Repeat the covar to match the batch shape of diff
RuntimeError: expected backend CPU and dtype Double but got backend CPU and dtype Float
I am getting a runtime error i have two tensors one x and two y values and I am trying the multitask example.
| 1.0 | error in example - new data - Hi @dougalsutherland @jmhessel
Hi I am trying the below example with new data
https://github.com/cornellius-gp/gpytorch/blob/master/examples/03_Multitask_GP_Regression/Multitask_GP_Regression.ipynb
model.train()
likelihood.train()
Use the adam optimizer
optimizer = torch.optim.Adam([
{‘params’: model.parameters()}, # Includes GaussianLikelihood parameters
], lr=0.1)
“Loss” for GPs - the marginal log likelihood
mll = gpytorch.mlls.ExactMarginalLogLikelihood(likelihood, model)
n_iter = 50
for i in range(n_iter):
optimizer.zero_grad()
output = model(train_x)
loss = -mll(output, train_y)
loss.backward()
print(‘Iter %d/%d - Loss: %.3f’ % (i + 1, n_iter, loss.item()))
optimizer.step()
RuntimeError Traceback (most recent call last)
in ()
16 optimizer.zero_grad()
17 output = model(train_x)
—> 18 loss = -mll(output, train_y)
19 loss.backward()
20 print(‘Iter %d/%d - Loss: %.3f’ % (i + 1, n_iter, loss.item()))
3 frames
/usr/local/lib/python3.6/dist-packages/gpytorch/distributions/multivariate_normal.py in log_prob(self, value)
112
113 mean, covar = self.loc, self.lazy_covariance_matrix
–> 114 diff = value - mean
115
116 # Repeat the covar to match the batch shape of diff
RuntimeError: expected backend CPU and dtype Double but got backend CPU and dtype Float
I am getting a runtime error i have two tensors one x and two y values and I am trying the multitask example.
| non_defect | error in example new data hi dougalsutherland jmhessel hi i am trying the below example with new data model train likelihood train use the adam optimizer optimizer torch optim adam ‘params’ model parameters includes gaussianlikelihood parameters lr “loss” for gps the marginal log likelihood mll gpytorch mlls exactmarginalloglikelihood likelihood model n iter for i in range n iter optimizer zero grad output model train x loss mll output train y loss backward print ‘iter d d loss ’ i n iter loss item optimizer step runtimeerror traceback most recent call last in optimizer zero grad output model train x — loss mll output train y loss backward print ‘iter d d loss ’ i n iter loss item frames usr local lib dist packages gpytorch distributions multivariate normal py in log prob self value mean covar self loc self lazy covariance matrix – diff value mean repeat the covar to match the batch shape of diff runtimeerror expected backend cpu and dtype double but got backend cpu and dtype float i am getting a runtime error i have two tensors one x and two y values and i am trying the multitask example | 0 |
301,459 | 26,050,035,355 | IssuesEvent | 2022-12-22 17:42:00 | conda/conda-build | https://api.github.com/repos/conda/conda-build | opened | Flaky test `tests/test_api_build.py::test_failed_recipe_leaves_folders` | backlog source::anaconda type::testing | ### Checklist
- [X] I added a descriptive title
- [X] I searched open reports and couldn't find a duplicate
### What happened?
This test has proven very flaky on Windows recently, here are just some examples:
- [Python 3.9 tests](https://github.com/conda/conda-build/actions/runs/3716460346/jobs/6332948506#step:11:3099) https://github.com/conda/conda-build/pull/4624
- [Python 3.7 tests](https://github.com/conda/conda-build/actions/runs/3711735797/jobs/6332950346#step:11:3097) https://github.com/conda/conda-build/pull/4678
- [Python 3.10 tests](https://github.com/conda/conda-build/actions/runs/3727885677/jobs/6332969373#step:11:3084) https://github.com/conda/conda-build/pull/4677
- [Python 3.8 tests](https://github.com/conda/conda-build/actions/runs/3754014678/jobs/6379201621#step:7:2684) https://github.com/conda/conda-build/pull/4648
- [Python 3.8 tests](https://github.com/conda/conda-build/actions/runs/3738029828/jobs/6379202358#step:11:3099) https://github.com/conda/conda-build/pull/4624
- [Python 3.9 tests](https://github.com/conda/conda-build/actions/runs/3754962836/jobs/6379634873#step:11:3946) https://github.com/conda/conda-build/pull/4683
### Conda Info
_No response_
### Conda Config
_No response_
### Conda list
_No response_
### Additional Context
_No response_ | 1.0 | Flaky test `tests/test_api_build.py::test_failed_recipe_leaves_folders` - ### Checklist
- [X] I added a descriptive title
- [X] I searched open reports and couldn't find a duplicate
### What happened?
This test has proven very flaky on Windows recently, here are just some examples:
- [Python 3.9 tests](https://github.com/conda/conda-build/actions/runs/3716460346/jobs/6332948506#step:11:3099) https://github.com/conda/conda-build/pull/4624
- [Python 3.7 tests](https://github.com/conda/conda-build/actions/runs/3711735797/jobs/6332950346#step:11:3097) https://github.com/conda/conda-build/pull/4678
- [Python 3.10 tests](https://github.com/conda/conda-build/actions/runs/3727885677/jobs/6332969373#step:11:3084) https://github.com/conda/conda-build/pull/4677
- [Python 3.8 tests](https://github.com/conda/conda-build/actions/runs/3754014678/jobs/6379201621#step:7:2684) https://github.com/conda/conda-build/pull/4648
- [Python 3.8 tests](https://github.com/conda/conda-build/actions/runs/3738029828/jobs/6379202358#step:11:3099) https://github.com/conda/conda-build/pull/4624
- [Python 3.9 tests](https://github.com/conda/conda-build/actions/runs/3754962836/jobs/6379634873#step:11:3946) https://github.com/conda/conda-build/pull/4683
### Conda Info
_No response_
### Conda Config
_No response_
### Conda list
_No response_
### Additional Context
_No response_ | non_defect | flaky test tests test api build py test failed recipe leaves folders checklist i added a descriptive title i searched open reports and couldn t find a duplicate what happened this test has proven very flaky on windows recently here are just some examples conda info no response conda config no response conda list no response additional context no response | 0 |
55,107 | 14,219,320,060 | IssuesEvent | 2020-11-17 13:07:20 | PowerDNS/pdns | https://api.github.com/repos/PowerDNS/pdns | closed | auth mysql: mysql 8 defaults to utf8mb4 which is incompatible with our latin1 schema declaration | auth defect | <!-- Tell us what is issue is about -->
- Program: Authoritative 4.3.1 with PowerDNS-Admin
- Issue type: Bug report
### Short description
<!-- Explain in a few sentences what the issue/request is -->
Add a new record (or modify something as TTL) fail.
### Environment
<!-- Tell us about the environment -->
- Operating system: ubuntu 20.40
- Software version: 4.3.1 (fresh install) with mysql 8.0.22 (ubuntu repo) as backend
- Software source: PowerDNS repository
### Steps to reproduce
1. Try add new record, or change something like TTL.
### Actual behaviour
On logs I have this error:
`[webserver] 950d6d03-ca5e-44bc-9468-44f41b7064c8 HTTP ISE for "/api/v1/servers/localhost/zones/xxx.com": Exception: GSQLBackend unable to delete RRSet test.xxx.com|A: Could not execute mysql statement: delete from records where domain_id=? and name=? and type=?: Illegal mix of collations (latin1_swedish_ci,IMPLICIT) and (utf8mb4_0900_ai_ci,COERCIBLE) for operation '='`
| 1.0 | auth mysql: mysql 8 defaults to utf8mb4 which is incompatible with our latin1 schema declaration - <!-- Tell us what is issue is about -->
- Program: Authoritative 4.3.1 with PowerDNS-Admin
- Issue type: Bug report
### Short description
<!-- Explain in a few sentences what the issue/request is -->
Add a new record (or modify something as TTL) fail.
### Environment
<!-- Tell us about the environment -->
- Operating system: ubuntu 20.40
- Software version: 4.3.1 (fresh install) with mysql 8.0.22 (ubuntu repo) as backend
- Software source: PowerDNS repository
### Steps to reproduce
1. Try add new record, or change something like TTL.
### Actual behaviour
On logs I have this error:
`[webserver] 950d6d03-ca5e-44bc-9468-44f41b7064c8 HTTP ISE for "/api/v1/servers/localhost/zones/xxx.com": Exception: GSQLBackend unable to delete RRSet test.xxx.com|A: Could not execute mysql statement: delete from records where domain_id=? and name=? and type=?: Illegal mix of collations (latin1_swedish_ci,IMPLICIT) and (utf8mb4_0900_ai_ci,COERCIBLE) for operation '='`
| defect | auth mysql mysql defaults to which is incompatible with our schema declaration program authoritative with powerdns admin issue type bug report short description add a new record or modify something as ttl fail environment operating system ubuntu software version fresh install with mysql ubuntu repo as backend software source powerdns repository steps to reproduce try add new record or change something like ttl actual behaviour on logs i have this error http ise for api servers localhost zones xxx com exception gsqlbackend unable to delete rrset test xxx com a could not execute mysql statement delete from records where domain id and name and type illegal mix of collations swedish ci implicit and ai ci coercible for operation | 1 |
11,265 | 13,234,284,308 | IssuesEvent | 2020-08-18 16:02:46 | gudmdharalds-a8c/testing123 | https://api.github.com/repos/gudmdharalds-a8c/testing123 | closed | PHP Upgrade: Compatibility issues found in file1.php | PHP 7.4 Compatibility PHP Compatibility | The following issues were found when scanning for compatibility issues:
* <b>Warning</b>: Function ldap_sort() is deprecated since PHP 7.0 https://github.com/gudmdharalds-a8c/testing123/blob/b99a028e21f490f459f7095329fe4933e8643e79/file1.php#L3
* <b>Error</b>: Extension 'mcrypt' is deprecated since PHP 7.1 and removed since PHP 7.2; Use openssl (preferred) or pecl/mcrypt once available instead https://github.com/gudmdharalds-a8c/testing123/blob/b99a028e21f490f459f7095329fe4933e8643e79/file1.php#L5
* <b>Error</b>: Function mcrypt_cfb() is deprecated since PHP 5.5 and removed since PHP 7.0 https://github.com/gudmdharalds-a8c/testing123/blob/b99a028e21f490f459f7095329fe4933e8643e79/file1.php#L5
* <b>Error</b>: Extension 'mysql_' is deprecated since PHP 5.5 and removed since PHP 7.0; Use mysqli instead https://github.com/gudmdharalds-a8c/testing123/blob/b99a028e21f490f459f7095329fe4933e8643e79/file1.php#L7
Note that this is an automated report.
| True | PHP Upgrade: Compatibility issues found in file1.php - The following issues were found when scanning for compatibility issues:
* <b>Warning</b>: Function ldap_sort() is deprecated since PHP 7.0 https://github.com/gudmdharalds-a8c/testing123/blob/b99a028e21f490f459f7095329fe4933e8643e79/file1.php#L3
* <b>Error</b>: Extension 'mcrypt' is deprecated since PHP 7.1 and removed since PHP 7.2; Use openssl (preferred) or pecl/mcrypt once available instead https://github.com/gudmdharalds-a8c/testing123/blob/b99a028e21f490f459f7095329fe4933e8643e79/file1.php#L5
* <b>Error</b>: Function mcrypt_cfb() is deprecated since PHP 5.5 and removed since PHP 7.0 https://github.com/gudmdharalds-a8c/testing123/blob/b99a028e21f490f459f7095329fe4933e8643e79/file1.php#L5
* <b>Error</b>: Extension 'mysql_' is deprecated since PHP 5.5 and removed since PHP 7.0; Use mysqli instead https://github.com/gudmdharalds-a8c/testing123/blob/b99a028e21f490f459f7095329fe4933e8643e79/file1.php#L7
Note that this is an automated report.
| non_defect | php upgrade compatibility issues found in php the following issues were found when scanning for compatibility issues warning function ldap sort is deprecated since php error extension mcrypt is deprecated since php and removed since php use openssl preferred or pecl mcrypt once available instead error function mcrypt cfb is deprecated since php and removed since php error extension mysql is deprecated since php and removed since php use mysqli instead note that this is an automated report | 0 |
15,534 | 2,859,572,256 | IssuesEvent | 2015-06-03 11:28:44 | gama-platform/gama | https://api.github.com/repos/gama-platform/gama | closed | 'as_map' expects a pair as second argument - road traffic and maybe other model | Priority-Medium Type-Defect Usability | ```
What steps will reproduce the problem?
1.Model that are using as_map may have to be changed
For example road_traffic is not working anymore with the following use of as_maps
let weights_map type: map <- (list (road)) as_map [each:: each.destruction_coeff];
What is the expected output? What do you see instead?
What about the compatibility I think we should allow this notation to keep the compatibility
with older model?
Please use labels and text to provide additional information.
Mac OS - SVN Version
```
Original issue reported on code.google.com by `agrignard` on 2013-05-31 04:30:13 | 1.0 | 'as_map' expects a pair as second argument - road traffic and maybe other model - ```
What steps will reproduce the problem?
1.Model that are using as_map may have to be changed
For example road_traffic is not working anymore with the following use of as_maps
let weights_map type: map <- (list (road)) as_map [each:: each.destruction_coeff];
What is the expected output? What do you see instead?
What about the compatibility I think we should allow this notation to keep the compatibility
with older model?
Please use labels and text to provide additional information.
Mac OS - SVN Version
```
Original issue reported on code.google.com by `agrignard` on 2013-05-31 04:30:13 | defect | as map expects a pair as second argument road traffic and maybe other model what steps will reproduce the problem model that are using as map may have to be changed for example road traffic is not working anymore with the following use of as maps let weights map type map list road as map what is the expected output what do you see instead what about the compatibility i think we should allow this notation to keep the compatibility with older model please use labels and text to provide additional information mac os svn version original issue reported on code google com by agrignard on | 1 |
26,283 | 4,651,561,580 | IssuesEvent | 2016-10-03 10:37:36 | gbif/ipt | https://api.github.com/repos/gbif/ipt | closed | Resource gets added to RSS on PUBLIC, not on published | bug Component-Logic Priority-High Type-Defect | Steps:
1. I add a new resource to our IPT
2. The to-be-published version 1.0 of the resource will be public, so I set visibility to `PUBLIC`. Since I haven't finished writing the metadata or uploading the data yet, I'm not yet publishing the dataset.
3. The status `PUBLIC` triggers the IPT to include the resource in the RSS.
4. Via IFTTT a tweet is sent with the link to the not-yet-published dataset.
5. Visitors of the link see an `Not authorized` page
Solution: Setting a dataset to `PUBLIC` should not trigger the IPT to update the RSS. Only **publishing** a resource should trigger this.
Note: the not-yet-published resource is NOT included in the DCAT feed, which is the correct behaviour. | 1.0 | Resource gets added to RSS on PUBLIC, not on published - Steps:
1. I add a new resource to our IPT
2. The to-be-published version 1.0 of the resource will be public, so I set visibility to `PUBLIC`. Since I haven't finished writing the metadata or uploading the data yet, I'm not yet publishing the dataset.
3. The status `PUBLIC` triggers the IPT to include the resource in the RSS.
4. Via IFTTT a tweet is sent with the link to the not-yet-published dataset.
5. Visitors of the link see an `Not authorized` page
Solution: Setting a dataset to `PUBLIC` should not trigger the IPT to update the RSS. Only **publishing** a resource should trigger this.
Note: the not-yet-published resource is NOT included in the DCAT feed, which is the correct behaviour. | defect | resource gets added to rss on public not on published steps i add a new resource to our ipt the to be published version of the resource will be public so i set visibility to public since i haven t finished writing the metadata or uploading the data yet i m not yet publishing the dataset the status public triggers the ipt to include the resource in the rss via ifttt a tweet is sent with the link to the not yet published dataset visitors of the link see an not authorized page solution setting a dataset to public should not trigger the ipt to update the rss only publishing a resource should trigger this note the not yet published resource is not included in the dcat feed which is the correct behaviour | 1 |
68,770 | 3,292,575,261 | IssuesEvent | 2015-10-30 15:15:06 | thesgc/chembiohub_helpdesk | https://api.github.com/repos/thesgc/chembiohub_helpdesk | closed | Can't clear out the custom field values in the search.
It looks like I've deleted the 2 tagged cu | app: ChemReg bug name: Karen priority: Medium status: New | Can't clear out the custom field values in the search.
It looks like I've deleted the 2 tagged custom fields in my search, but the search isn't rerun and the URL isn't updated. If I change the functional group search that seems to be taken into account, but not custom fields. | 1.0 | Can't clear out the custom field values in the search.
It looks like I've deleted the 2 tagged cu - Can't clear out the custom field values in the search.
It looks like I've deleted the 2 tagged custom fields in my search, but the search isn't rerun and the URL isn't updated. If I change the functional group search that seems to be taken into account, but not custom fields. | non_defect | can t clear out the custom field values in the search it looks like i ve deleted the tagged cu can t clear out the custom field values in the search it looks like i ve deleted the tagged custom fields in my search but the search isn t rerun and the url isn t updated if i change the functional group search that seems to be taken into account but not custom fields | 0 |
343,808 | 10,336,590,439 | IssuesEvent | 2019-09-03 13:17:19 | wherebyus/general-tasks | https://api.github.com/repos/wherebyus/general-tasks | closed | Sponsored events should not also be able to be partner events | Priority: Medium Product: Events UX: Validated | ## Feature or problem
- [ ] In the EventDashboard__List component, disable these buttons.
## UX Validation
Validated
### Suggested priority
Medium
### Stakeholders
*Submitted:* michael
### Definition of done
How will we know when this feature is complete?
### Subtasks
A detailed list of changes that need to be made or subtasks. One checkbox per.
- [ ] Brew the coffee
## Developer estimate
To help the team accurately estimate the complexity of this task,
take a moment to walk through this list and estimate each item. At the end, you can total
the estimates and round to the nearest prime number.
If any of these are at a `5` or higher, or if the total is above a `5`, consider breaking
this issue into multiple smaller issues.
- [ ] Changes to the database ()
- [ ] Changes to the API ()
- [ ] Testing Changes to the API ()
- [ ] Changes to Application Code ()
- [ ] Adding or updating unit tests ()
- [ ] Local developer testing ()
### Total developer estimate: 0
## Additional estimate
- [ ] Code review ()
- [ ] QA Testing ()
- [ ] Stakeholder Sign-off ()
- [ ] Deploy to Production ()
### Total additional estimate:
## QA Notes
Detailed instructions for testing, one checkbox per test to be completed.
### Contextual tests
- [ ] Accessibility check
- [ ] Cross-browser check (Edge, Chrome, Firefox)
- [ ] Responsive check
| 1.0 | Sponsored events should not also be able to be partner events - ## Feature or problem
- [ ] In the EventDashboard__List component, disable these buttons.
## UX Validation
Validated
### Suggested priority
Medium
### Stakeholders
*Submitted:* michael
### Definition of done
How will we know when this feature is complete?
### Subtasks
A detailed list of changes that need to be made or subtasks. One checkbox per.
- [ ] Brew the coffee
## Developer estimate
To help the team accurately estimate the complexity of this task,
take a moment to walk through this list and estimate each item. At the end, you can total
the estimates and round to the nearest prime number.
If any of these are at a `5` or higher, or if the total is above a `5`, consider breaking
this issue into multiple smaller issues.
- [ ] Changes to the database ()
- [ ] Changes to the API ()
- [ ] Testing Changes to the API ()
- [ ] Changes to Application Code ()
- [ ] Adding or updating unit tests ()
- [ ] Local developer testing ()
### Total developer estimate: 0
## Additional estimate
- [ ] Code review ()
- [ ] QA Testing ()
- [ ] Stakeholder Sign-off ()
- [ ] Deploy to Production ()
### Total additional estimate:
## QA Notes
Detailed instructions for testing, one checkbox per test to be completed.
### Contextual tests
- [ ] Accessibility check
- [ ] Cross-browser check (Edge, Chrome, Firefox)
- [ ] Responsive check
| non_defect | sponsored events should not also be able to be partner events feature or problem in the eventdashboard list component disable these buttons ux validation validated suggested priority medium stakeholders submitted michael definition of done how will we know when this feature is complete subtasks a detailed list of changes that need to be made or subtasks one checkbox per brew the coffee developer estimate to help the team accurately estimate the complexity of this task take a moment to walk through this list and estimate each item at the end you can total the estimates and round to the nearest prime number if any of these are at a or higher or if the total is above a consider breaking this issue into multiple smaller issues changes to the database changes to the api testing changes to the api changes to application code adding or updating unit tests local developer testing total developer estimate additional estimate code review qa testing stakeholder sign off deploy to production total additional estimate qa notes detailed instructions for testing one checkbox per test to be completed contextual tests accessibility check cross browser check edge chrome firefox responsive check | 0 |
111,008 | 24,051,026,948 | IssuesEvent | 2022-09-16 12:49:57 | ices-eg/DIG | https://api.github.com/repos/ices-eg/DIG | closed | RDBES: RS_SpecimensState update of codes | vocab: CodeValue vocab: new vocab: concept change | Please see attached spreadsheet for the 8 code updates.
[RS_SpecimensState updates.xlsx](https://github.com/ices-eg/DIG/files/6931334/RS_SpecimensState.updates.xlsx)
| 1.0 | RDBES: RS_SpecimensState update of codes - Please see attached spreadsheet for the 8 code updates.
[RS_SpecimensState updates.xlsx](https://github.com/ices-eg/DIG/files/6931334/RS_SpecimensState.updates.xlsx)
| non_defect | rdbes rs specimensstate update of codes please see attached spreadsheet for the code updates | 0 |
334,241 | 29,829,095,643 | IssuesEvent | 2023-06-18 03:05:49 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | opened | Fix tensor.test_torch_special_neg | PyTorch Frontend Sub Task Failing Test | | | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5301229031/jobs/9595347278"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5301229031/jobs/9595347278"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5301229031/jobs/9595347278"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5301229031/jobs/9595347278"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5301229031/jobs/9595347278"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix tensor.test_torch_special_neg - | | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5301229031/jobs/9595347278"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5301229031/jobs/9595347278"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5301229031/jobs/9595347278"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5301229031/jobs/9595347278"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5301229031/jobs/9595347278"><img src=https://img.shields.io/badge/-success-success></a>
| non_defect | fix tensor test torch special neg numpy a href src torch a href src tensorflow a href src jax a href src paddle a href src | 0 |
73,612 | 7,346,071,215 | IssuesEvent | 2018-03-07 19:28:40 | SmartPhoenix/PWMod | https://api.github.com/repos/SmartPhoenix/PWMod | opened | Test Banner Dropping Again | need testing | ### Stuff such as:
- Does the banner start dropping when the person capping is isn't capable of capping it because of his class.
- Does the banner start dropping when you want to cap it when not in war.
- Does the banner start dropping when a commoner/outlaw tries to cap it.
| 1.0 | Test Banner Dropping Again - ### Stuff such as:
- Does the banner start dropping when the person capping is isn't capable of capping it because of his class.
- Does the banner start dropping when you want to cap it when not in war.
- Does the banner start dropping when a commoner/outlaw tries to cap it.
| non_defect | test banner dropping again stuff such as does the banner start dropping when the person capping is isn t capable of capping it because of his class does the banner start dropping when you want to cap it when not in war does the banner start dropping when a commoner outlaw tries to cap it | 0 |
63,477 | 3,195,662,486 | IssuesEvent | 2015-09-30 17:36:17 | fusioninventory/fusioninventory-for-glpi | https://api.github.com/repos/fusioninventory/fusioninventory-for-glpi | closed | Correctif des modèles SNMP 4767953.xml et 8157205.xml | Category: SNMP Component: For junior contributor Component: Found in version Priority: Normal Status: Rejected Tracker: Bug | ---
Author Name: **Christophe Pujol** (Christophe Pujol)
Original Redmine Issue: 2112, http://forge.fusioninventory.org/issues/2112
Original Date: 2013-05-11
Original Assignee: David Durieux
---
Bonjour,
Comme le bug perdure depuis le signalement dans le forum, je relance içi.
Il faut juste retirer le ZyWALL USG 200 dans le modèle 4767953.xml et l'ajouter dans le modèle 8157205.xml
Salutation,
| 1.0 | Correctif des modèles SNMP 4767953.xml et 8157205.xml - ---
Author Name: **Christophe Pujol** (Christophe Pujol)
Original Redmine Issue: 2112, http://forge.fusioninventory.org/issues/2112
Original Date: 2013-05-11
Original Assignee: David Durieux
---
Bonjour,
Comme le bug perdure depuis le signalement dans le forum, je relance içi.
Il faut juste retirer le ZyWALL USG 200 dans le modèle 4767953.xml et l'ajouter dans le modèle 8157205.xml
Salutation,
| non_defect | correctif des modèles snmp xml et xml author name christophe pujol christophe pujol original redmine issue original date original assignee david durieux bonjour comme le bug perdure depuis le signalement dans le forum je relance içi il faut juste retirer le zywall usg dans le modèle xml et l ajouter dans le modèle xml salutation | 0 |
63,161 | 17,399,059,776 | IssuesEvent | 2021-08-02 16:57:31 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Rooms where all messages have expired, are shown as unread | T-Defect | <!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->
<!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
#### Description
Rooms where all messages have expired on server side via m.room.retention and therefore are unavailable from the server, are always and continuously shown as unread in Element Web and Linux Desktop.
#### Steps to reproduce
- Create a room
- Set short expiration like 1 second using m.room.retention event
- Post something
- Wait for the message to expire
- Clear local cache (or login with new session which doesn't have the message cached)
- Now the room is shown as unread
- When something is posted to the room, and "seen" by the clients then it's marked as read, until this happens again.
I guess the logic is bad, it checks if the last event in room is seen, but what if the no last event isn't available? Why it then defaults to unread? If I try to jump to the last read message on desktop client, I'll just gives me an error: "Failed to load timeline position".
#### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: Web & Linux Desktop
For the web app:
- **Browser**: Firefox
- **OS**: Windows & Ubuntu
- **URL**: 1.7.30 (private server)
For the desktop app:
- **OS**: Ubuntu
- **Version**: 1.7.32
| 1.0 | Rooms where all messages have expired, are shown as unread - <!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->
<!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
#### Description
Rooms where all messages have expired on server side via m.room.retention and therefore are unavailable from the server, are always and continuously shown as unread in Element Web and Linux Desktop.
#### Steps to reproduce
- Create a room
- Set short expiration like 1 second using m.room.retention event
- Post something
- Wait for the message to expire
- Clear local cache (or login with new session which doesn't have the message cached)
- Now the room is shown as unread
- When something is posted to the room, and "seen" by the clients then it's marked as read, until this happens again.
I guess the logic is bad, it checks if the last event in room is seen, but what if the no last event isn't available? Why it then defaults to unread? If I try to jump to the last read message on desktop client, I'll just gives me an error: "Failed to load timeline position".
#### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: Web & Linux Desktop
For the web app:
- **Browser**: Firefox
- **OS**: Windows & Ubuntu
- **URL**: 1.7.30 (private server)
For the desktop app:
- **OS**: Ubuntu
- **Version**: 1.7.32
| defect | rooms where all messages have expired are shown as unread this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report description rooms where all messages have expired on server side via m room retention and therefore are unavailable from the server are always and continuously shown as unread in element web and linux desktop steps to reproduce create a room set short expiration like second using m room retention event post something wait for the message to expire clear local cache or login with new session which doesn t have the message cached now the room is shown as unread when something is posted to the room and seen by the clients then it s marked as read until this happens again i guess the logic is bad it checks if the last event in room is seen but what if the no last event isn t available why it then defaults to unread if i try to jump to the last read message on desktop client i ll just gives me an error failed to load timeline position version information platform web linux desktop for the web app browser firefox os windows ubuntu url private server for the desktop app os ubuntu version | 1 |
21,268 | 4,699,744,245 | IssuesEvent | 2016-10-12 16:30:48 | material-motion/material-motion-runtime-objc | https://api.github.com/repos/material-motion/material-motion-runtime-objc | closed | Write a "How to" guide for adding/removing named plans to a scheduler | Documentation | Should be added to the README.md. | 1.0 | Write a "How to" guide for adding/removing named plans to a scheduler - Should be added to the README.md. | non_defect | write a how to guide for adding removing named plans to a scheduler should be added to the readme md | 0 |
129,560 | 17,795,420,598 | IssuesEvent | 2021-08-31 21:26:41 | microsoft/responsible-ai-widgets | https://api.github.com/repos/microsoft/responsible-ai-widgets | closed | Zoom interaction on Tree map | Error Analysis Design | The zoom interaction is unavailable before you click the canvas area. Without clicking the background, the scroll reaction is a vertical scroll, not zooming.
<img width="1409" alt="Screen Shot 2020-12-10 at 9 29 14 AM" src="https://user-images.githubusercontent.com/1608824/101807948-91de4580-3aca-11eb-99e1-85b194103d92.png">
| 1.0 | Zoom interaction on Tree map - The zoom interaction is unavailable before you click the canvas area. Without clicking the background, the scroll reaction is a vertical scroll, not zooming.
<img width="1409" alt="Screen Shot 2020-12-10 at 9 29 14 AM" src="https://user-images.githubusercontent.com/1608824/101807948-91de4580-3aca-11eb-99e1-85b194103d92.png">
| non_defect | zoom interaction on tree map the zoom interaction is unavailable before you click the canvas area without clicking the background the scroll reaction is a vertical scroll not zooming img width alt screen shot at am src | 0 |
62,798 | 17,198,717,392 | IssuesEvent | 2021-07-16 22:14:43 | jccastillo0007/eFacturaT | https://api.github.com/repos/jccastillo0007/eFacturaT | closed | Carta Porte - no genera PDF cuando se incluye mas de 1 mercancía. | bug defect resolved | El XML si lo genera, pero con el PDF truena.
Es cuando existe mas de 1 mercancía. | 1.0 | Carta Porte - no genera PDF cuando se incluye mas de 1 mercancía. - El XML si lo genera, pero con el PDF truena.
Es cuando existe mas de 1 mercancía. | defect | carta porte no genera pdf cuando se incluye mas de mercancía el xml si lo genera pero con el pdf truena es cuando existe mas de mercancía | 1 |
58,113 | 16,342,489,598 | IssuesEvent | 2021-05-13 00:25:25 | darshan-hpc/darshan | https://api.github.com/repos/darshan-hpc/darshan | closed | darshan does not understand binded mount points | defect | In GitLab by @shanedsnyder on Mar 28, 2016, 11:18
When accessing a file path corresponding to a binded mount point, Darshan is unable to recognize the actual underlying mount point it references. This prevents Darshan from determining the underlying file system type, which is necessary for determining the FS block size (which is used to determine whether I/O access are aligned or not).
It would be nice if Darshan could recognize binded mount points and store mappings from these bind mount points to the true underlying mount point in its runtime data structures. When an application accesses files in the binded mount point, Darshan can map them to the actual mount points it stores in the log file and correctly determine the underlying FS type and block size.
| 1.0 | darshan does not understand binded mount points - In GitLab by @shanedsnyder on Mar 28, 2016, 11:18
When accessing a file path corresponding to a binded mount point, Darshan is unable to recognize the actual underlying mount point it references. This prevents Darshan from determining the underlying file system type, which is necessary for determining the FS block size (which is used to determine whether I/O access are aligned or not).
It would be nice if Darshan could recognize binded mount points and store mappings from these bind mount points to the true underlying mount point in its runtime data structures. When an application accesses files in the binded mount point, Darshan can map them to the actual mount points it stores in the log file and correctly determine the underlying FS type and block size.
| defect | darshan does not understand binded mount points in gitlab by shanedsnyder on mar when accessing a file path corresponding to a binded mount point darshan is unable to recognize the actual underlying mount point it references this prevents darshan from determining the underlying file system type which is necessary for determining the fs block size which is used to determine whether i o access are aligned or not it would be nice if darshan could recognize binded mount points and store mappings from these bind mount points to the true underlying mount point in its runtime data structures when an application accesses files in the binded mount point darshan can map them to the actual mount points it stores in the log file and correctly determine the underlying fs type and block size | 1 |
178,447 | 29,820,189,665 | IssuesEvent | 2023-06-17 01:06:06 | calcom/cal.com | https://api.github.com/repos/calcom/cal.com | closed | [CAL-1966] Top banner for verifying email doesn't match designs | ✅ good first issue 🎨 needs design ⚡ Quick Wins | Should be

Currently

<sub>From [SyncLinear.com](https://synclinear.com) | [CAL-1966](https://linear.app/calcom/issue/CAL-1966/top-banner-for-verifying-email-doesnt-match-designs)</sub> | 1.0 | [CAL-1966] Top banner for verifying email doesn't match designs - Should be

Currently

<sub>From [SyncLinear.com](https://synclinear.com) | [CAL-1966](https://linear.app/calcom/issue/CAL-1966/top-banner-for-verifying-email-doesnt-match-designs)</sub> | non_defect | top banner for verifying email doesn t match designs should be currently from | 0 |
126,528 | 4,997,114,921 | IssuesEvent | 2016-12-09 15:53:18 | sabino/ProjetoTeste | https://api.github.com/repos/sabino/ProjetoTeste | closed | asdfsadf | Ajuste base Priority - Critical Team - FK | # asdfsadf
> <b>Encantador:</b> Jéssica Santos
> <b>Case:</b> https://suporte.contaazul.com/agent/case/774745
> <b>Time:</b> Team - FK
> <b>Tipo:</b> Ajuste base
> <b>Prioridade:</b> Priority - Critical | 1.0 | asdfsadf - # asdfsadf
> <b>Encantador:</b> Jéssica Santos
> <b>Case:</b> https://suporte.contaazul.com/agent/case/774745
> <b>Time:</b> Team - FK
> <b>Tipo:</b> Ajuste base
> <b>Prioridade:</b> Priority - Critical | non_defect | asdfsadf asdfsadf encantador jéssica santos case time team fk tipo ajuste base prioridade priority critical | 0 |
16,021 | 2,870,251,864 | IssuesEvent | 2015-06-07 00:36:51 | pdelia/away3d | https://api.github.com/repos/pdelia/away3d | opened | Shaking Filters on MovieMaterial | auto-migrated Priority-Medium Type-Defect | #74 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:35Z
```
What steps will reproduce the problem?
1. Create a MovieClip with filters applied to it's nested clips
2. Supply that MovieClip to a MovieMaterial
3. Move mouse over MovieMaterial, watch MCs with filters applied shake
Using Away3D v3.4.0 (FP10).
DEMO:
http://www.milkisevil.com/dumpbox/away3d-shaking-filters-on-
moviematerial/index.html
SIMPLIFIED SOURCE:
http://www.milkisevil.com/dumpbox/away3d-shaking-filters-on-
moviematerial/source.rar
```
Original issue reported on code.google.com by `manimalc...@gmail.com` on 22 Oct 2009 at 10:01 | 1.0 | Shaking Filters on MovieMaterial - #74 Issue by __GoogleCodeExporter__, created on: 2015-04-24T07:51:35Z
```
What steps will reproduce the problem?
1. Create a MovieClip with filters applied to it's nested clips
2. Supply that MovieClip to a MovieMaterial
3. Move mouse over MovieMaterial, watch MCs with filters applied shake
Using Away3D v3.4.0 (FP10).
DEMO:
http://www.milkisevil.com/dumpbox/away3d-shaking-filters-on-
moviematerial/index.html
SIMPLIFIED SOURCE:
http://www.milkisevil.com/dumpbox/away3d-shaking-filters-on-
moviematerial/source.rar
```
Original issue reported on code.google.com by `manimalc...@gmail.com` on 22 Oct 2009 at 10:01 | defect | shaking filters on moviematerial issue by googlecodeexporter created on what steps will reproduce the problem create a movieclip with filters applied to it s nested clips supply that movieclip to a moviematerial move mouse over moviematerial watch mcs with filters applied shake using demo moviematerial index html simplified source moviematerial source rar original issue reported on code google com by manimalc gmail com on oct at | 1 |
29,079 | 5,530,672,001 | IssuesEvent | 2017-03-21 03:42:17 | ISWPOLI/poli-booking | https://api.github.com/repos/ISWPOLI/poli-booking | closed | invalidar la sesión en localStorage al cerrar sesión | defectos/bugs prioridad-deseable | Actualmente para cerrar sesión tenemos un link a /logout que se encarga de invalidar la sesión en el backend. Sin embargo, la sesión también la almacenamos en BackboneJS en localStorage y es necesario eliminarla de ahí, de lo contrario BackboneJS seguirá asumiendo que el usuario tiene una sesión activa. | 1.0 | invalidar la sesión en localStorage al cerrar sesión - Actualmente para cerrar sesión tenemos un link a /logout que se encarga de invalidar la sesión en el backend. Sin embargo, la sesión también la almacenamos en BackboneJS en localStorage y es necesario eliminarla de ahí, de lo contrario BackboneJS seguirá asumiendo que el usuario tiene una sesión activa. | defect | invalidar la sesión en localstorage al cerrar sesión actualmente para cerrar sesión tenemos un link a logout que se encarga de invalidar la sesión en el backend sin embargo la sesión también la almacenamos en backbonejs en localstorage y es necesario eliminarla de ahí de lo contrario backbonejs seguirá asumiendo que el usuario tiene una sesión activa | 1 |
15,960 | 2,870,102,216 | IssuesEvent | 2015-06-06 20:50:10 | dkpro/dkpro-tc | https://api.github.com/repos/dkpro/dkpro-tc | closed | Certain special characters cause problems when used as feature names. | Priority-Critical Release-0.5.0 Type-Defect | Originally reported on Google Code with ID 14
```
See test case.
Should be escaped or filtered.
```
Reported by `torsten.zesch` on 2013-05-20 16:04:18
| 1.0 | Certain special characters cause problems when used as feature names. - Originally reported on Google Code with ID 14
```
See test case.
Should be escaped or filtered.
```
Reported by `torsten.zesch` on 2013-05-20 16:04:18
| defect | certain special characters cause problems when used as feature names originally reported on google code with id see test case should be escaped or filtered reported by torsten zesch on | 1 |
3,972 | 2,610,085,056 | IssuesEvent | 2015-02-26 18:25:53 | chrsmith/dsdsdaadf | https://api.github.com/repos/chrsmith/dsdsdaadf | opened | 深圳彩光怎样治疗痘痘 | auto-migrated Priority-Medium Type-Defect | ```
深圳彩光怎样治疗痘痘【深圳韩方科颜全国热线400-869-1818,24
小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩��
�秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,�
��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹
”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内��
�业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上�
��痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:02 | 1.0 | 深圳彩光怎样治疗痘痘 - ```
深圳彩光怎样治疗痘痘【深圳韩方科颜全国热线400-869-1818,24
小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩��
�秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,�
��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹
”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内��
�业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上�
��痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:02 | defect | 深圳彩光怎样治疗痘痘 深圳彩光怎样治疗痘痘【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩�� �秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,� ��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹 ”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内�� �业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上� ��痘痘。 original issue reported on code google com by szft com on may at | 1 |
820,570 | 30,778,693,381 | IssuesEvent | 2023-07-31 08:30:42 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | Secondary network position changes when any network is selected | bug priority/P4 QA/Yes release-notes/exclude feature/web3/wallet OS/Android | <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description <!-- Provide a brief description of the issue -->
Secondary network position changes when any network is selected
## Steps to reproduce <!-- Please add a series of steps to reproduce the issue -->
1. Visit any page that triggers Web3 connection to shown wallet icon in URL bar
2. Open the panel and select network selector
3. Select any network and look at the position of the secondary networks
## Actual result <!-- Please add screenshots if needed -->
https://github.com/brave/brave-browser/assets/17010094/b86d0073-1a13-48fe-ac24-7bbf2a5f3d31
## Expected result
Should not change position
## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] -->
Easy
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current Play Store version? NA
- Can you reproduce this issue with the current Play Store Beta version? Yes
- Can you reproduce this issue with the current Play Store Nightly version? Not checked
## Device details
- Install type (ARM, x86): ARM
- Device type (Phone, Tablet, Phablet): Phone
- Android version: 13
## Brave version
1.57.1 x64 Beta build
### Website problems only
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards?NA
- Is the issue reproducible on the latest version of Chrome? NA
### Additional information
<!-- Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue -->
cc: @Pavneet-Sing @simoarpe | 1.0 | Secondary network position changes when any network is selected - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description <!-- Provide a brief description of the issue -->
Secondary network position changes when any network is selected
## Steps to reproduce <!-- Please add a series of steps to reproduce the issue -->
1. Visit any page that triggers Web3 connection to shown wallet icon in URL bar
2. Open the panel and select network selector
3. Select any network and look at the position of the secondary networks
## Actual result <!-- Please add screenshots if needed -->
https://github.com/brave/brave-browser/assets/17010094/b86d0073-1a13-48fe-ac24-7bbf2a5f3d31
## Expected result
Should not change position
## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] -->
Easy
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current Play Store version? NA
- Can you reproduce this issue with the current Play Store Beta version? Yes
- Can you reproduce this issue with the current Play Store Nightly version? Not checked
## Device details
- Install type (ARM, x86): ARM
- Device type (Phone, Tablet, Phablet): Phone
- Android version: 13
## Brave version
1.57.1 x64 Beta build
### Website problems only
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards?NA
- Is the issue reproducible on the latest version of Chrome? NA
### Additional information
<!-- Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue -->
cc: @Pavneet-Sing @simoarpe | non_defect | secondary network position changes when any network is selected have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description secondary network position changes when any network is selected steps to reproduce visit any page that triggers connection to shown wallet icon in url bar open the panel and select network selector select any network and look at the position of the secondary networks actual result expected result should not change position issue reproduces how often easy version channel information can you reproduce this issue with the current play store version na can you reproduce this issue with the current play store beta version yes can you reproduce this issue with the current play store nightly version not checked device details install type arm arm device type phone tablet phablet phone android version brave version beta build website problems only does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na additional information cc pavneet sing simoarpe | 0 |
73,850 | 24,833,746,900 | IssuesEvent | 2022-10-26 07:05:38 | vector-im/element-call | https://api.github.com/repos/vector-im/element-call | opened | display 4:3 cameras properly (currently crops them to 16:9) | T-Defect | ### Steps to reproduce
1. Where are you starting? What can you see?
2. What do you click?
3. More steps…
when you join the element-call call with a 4:3 camera, it crops it to 16:9 for the default "full view" of a participant. i think it would be better to make 4:3 cameras show the whole thing including the bottom. or if you like your camera getting chopped off the top and bottom maybe add a toggle for it, but i dont think this is optimal default behaviour imo.
### Outcome
#### What did you expect?
i expected the camera to show itself in its entirety 4:3
#### What happened instead?
the camera was cropped to 16:9, cutting off a significant amount of viewage
### Operating system
affects all of them currently
### Browser information
affects all of them currently
### URL for webapp
affects all of them currently, usually use netlify and/or call.element.io
### Will you send logs?
Yes | 1.0 | display 4:3 cameras properly (currently crops them to 16:9) - ### Steps to reproduce
1. Where are you starting? What can you see?
2. What do you click?
3. More steps…
when you join the element-call call with a 4:3 camera, it crops it to 16:9 for the default "full view" of a participant. i think it would be better to make 4:3 cameras show the whole thing including the bottom. or if you like your camera getting chopped off the top and bottom maybe add a toggle for it, but i dont think this is optimal default behaviour imo.
### Outcome
#### What did you expect?
i expected the camera to show itself in its entirety 4:3
#### What happened instead?
the camera was cropped to 16:9, cutting off a significant amount of viewage
### Operating system
affects all of them currently
### Browser information
affects all of them currently
### URL for webapp
affects all of them currently, usually use netlify and/or call.element.io
### Will you send logs?
Yes | defect | display cameras properly currently crops them to steps to reproduce where are you starting what can you see what do you click more steps… when you join the element call call with a camera it crops it to for the default full view of a participant i think it would be better to make cameras show the whole thing including the bottom or if you like your camera getting chopped off the top and bottom maybe add a toggle for it but i dont think this is optimal default behaviour imo outcome what did you expect i expected the camera to show itself in its entirety what happened instead the camera was cropped to cutting off a significant amount of viewage operating system affects all of them currently browser information affects all of them currently url for webapp affects all of them currently usually use netlify and or call element io will you send logs yes | 1 |
386,236 | 11,434,180,279 | IssuesEvent | 2020-02-04 16:55:17 | YangCatalog/bottle-yang-extractor-validator | https://api.github.com/repos/YangCatalog/bottle-yang-extractor-validator | closed | https://www.yangcatalog.org/yangvalidator/ => yumaworks | Priority: High enhancement | https://www.yangcatalog.org/yangvalidator/
I see pyang, confdc. yanglint. We need yumaworks as well.
On top of that, we need the same output as the compilation-result in the metadata.
That means: we need pyang --ietf
| 1.0 | https://www.yangcatalog.org/yangvalidator/ => yumaworks - https://www.yangcatalog.org/yangvalidator/
I see pyang, confdc. yanglint. We need yumaworks as well.
On top of that, we need the same output as the compilation-result in the metadata.
That means: we need pyang --ietf
| non_defect | yumaworks i see pyang confdc yanglint we need yumaworks as well on top of that we need the same output as the compilation result in the metadata that means we need pyang ietf | 0 |
32,075 | 12,061,780,616 | IssuesEvent | 2020-04-16 00:55:49 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Use correct OpenSSL libraries for FreeBSD | area-System.Security os-freebsd untriaged | src/libraries/Native/Unix/System.Security.Cryptography.Native/opensslshim.c
`OpenLibrary()` doesn't currently find the correct libraries on my FreeBSD 11.3 build box.
The version of OpenSSL included in the base 11.3 install is 1.0.2
```
[jason@freebsd11 ~/src/runtime]$ /usr/bin/openssl version
OpenSSL 1.0.2s-freebsd 28 May 2019
[jason@freebsd11 ~/src/runtime]$ ldd /usr/bin/openssl
/usr/bin/openssl:
libssl.so.8 => /usr/lib/libssl.so.8 (0x8008a4000)
libcrypto.so.8 => /lib/libcrypto.so.8 (0x800c00000)
libc.so.7 => /lib/libc.so.7 (0x801076000)
```
OpenSSL 1.1.1 can be installed with the FreeBSD package manager
```
[jason@freebsd11 ~/src/runtime]$ /usr/local/bin/openssl version
OpenSSL 1.1.1f 31 Mar 2020
[jason@freebsd11 ~/src/runtime]$ ldd /usr/local/bin/openssl
/usr/local/bin/openssl:
libssl.so.11 => /usr/local/lib/libssl.so.11 (0x8008b7000)
libcrypto.so.11 => /usr/local/lib/libcrypto.so.11 (0x800c00000)
libthr.so.3 => /lib/libthr.so.3 (0x8010ef000)
libc.so.7 => /lib/libc.so.7 (0x801317000)
```
`OpenLibrary()` needs to look for `libssl.so.11` and `libssl.so.8` to support these versions.
@wfurt | True | Use correct OpenSSL libraries for FreeBSD - src/libraries/Native/Unix/System.Security.Cryptography.Native/opensslshim.c
`OpenLibrary()` doesn't currently find the correct libraries on my FreeBSD 11.3 build box.
The version of OpenSSL included in the base 11.3 install is 1.0.2
```
[jason@freebsd11 ~/src/runtime]$ /usr/bin/openssl version
OpenSSL 1.0.2s-freebsd 28 May 2019
[jason@freebsd11 ~/src/runtime]$ ldd /usr/bin/openssl
/usr/bin/openssl:
libssl.so.8 => /usr/lib/libssl.so.8 (0x8008a4000)
libcrypto.so.8 => /lib/libcrypto.so.8 (0x800c00000)
libc.so.7 => /lib/libc.so.7 (0x801076000)
```
OpenSSL 1.1.1 can be installed with the FreeBSD package manager
```
[jason@freebsd11 ~/src/runtime]$ /usr/local/bin/openssl version
OpenSSL 1.1.1f 31 Mar 2020
[jason@freebsd11 ~/src/runtime]$ ldd /usr/local/bin/openssl
/usr/local/bin/openssl:
libssl.so.11 => /usr/local/lib/libssl.so.11 (0x8008b7000)
libcrypto.so.11 => /usr/local/lib/libcrypto.so.11 (0x800c00000)
libthr.so.3 => /lib/libthr.so.3 (0x8010ef000)
libc.so.7 => /lib/libc.so.7 (0x801317000)
```
`OpenLibrary()` needs to look for `libssl.so.11` and `libssl.so.8` to support these versions.
@wfurt | non_defect | use correct openssl libraries for freebsd src libraries native unix system security cryptography native opensslshim c openlibrary doesn t currently find the correct libraries on my freebsd build box the version of openssl included in the base install is usr bin openssl version openssl freebsd may ldd usr bin openssl usr bin openssl libssl so usr lib libssl so libcrypto so lib libcrypto so libc so lib libc so openssl can be installed with the freebsd package manager usr local bin openssl version openssl mar ldd usr local bin openssl usr local bin openssl libssl so usr local lib libssl so libcrypto so usr local lib libcrypto so libthr so lib libthr so libc so lib libc so openlibrary needs to look for libssl so and libssl so to support these versions wfurt | 0 |
28,397 | 5,247,706,475 | IssuesEvent | 2017-02-01 13:53:52 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | Boxed enum types are treated as regular numbers | defect | ### Expected
```
Boat
Boat
Boat
```
### Actual
```
Boat
2
2
```
### Steps To Reproduce
[Deck](http://deck.net/50c6936a439e6c4dcafdb53edfce9dee)
```cs
public enum VehicleType
{
Car,
Plane,
Boat
}
public class Program
{
public static void Main()
{
VehicleType vehicleType = VehicleType.Boat;
object box = vehicleType;
Console.WriteLine(vehicleType);
Console.WriteLine(box);
Console.WriteLine(Enum.Parse(typeof(VehicleType), "Boat"));
}
}
```
| 1.0 | Boxed enum types are treated as regular numbers - ### Expected
```
Boat
Boat
Boat
```
### Actual
```
Boat
2
2
```
### Steps To Reproduce
[Deck](http://deck.net/50c6936a439e6c4dcafdb53edfce9dee)
```cs
public enum VehicleType
{
Car,
Plane,
Boat
}
public class Program
{
public static void Main()
{
VehicleType vehicleType = VehicleType.Boat;
object box = vehicleType;
Console.WriteLine(vehicleType);
Console.WriteLine(box);
Console.WriteLine(Enum.Parse(typeof(VehicleType), "Boat"));
}
}
```
| defect | boxed enum types are treated as regular numbers expected boat boat boat actual boat steps to reproduce cs public enum vehicletype car plane boat public class program public static void main vehicletype vehicletype vehicletype boat object box vehicletype console writeline vehicletype console writeline box console writeline enum parse typeof vehicletype boat | 1 |
46,277 | 5,794,294,928 | IssuesEvent | 2017-05-02 14:37:21 | hashicorp/consul | https://api.github.com/repos/hashicorp/consul | closed | Track down TestCatalog_Nodes timeout failure | test-failure | ```
...
2016/10/31 21:10:00 [DEBUG] http: Request GET /v1/catalog/nodes (28.441µs) from=127.0.0.1:55097
2016/10/31 21:10:00 [DEBUG] http: Request GET /v1/catalog/nodes (33.41µs) from=127.0.0.1:55097
2016/10/31 21:10:00 [DEBUG] http: Request GET /v1/catalog/nodes (37.158µs) from=127.0.0.1:55097
--- FAIL: TestCatalog_Nodes (12.09s)
catalog_test.go:59: err: Bad: &{node1 127.0.0.1 map[]}
``` | 1.0 | Track down TestCatalog_Nodes timeout failure - ```
...
2016/10/31 21:10:00 [DEBUG] http: Request GET /v1/catalog/nodes (28.441µs) from=127.0.0.1:55097
2016/10/31 21:10:00 [DEBUG] http: Request GET /v1/catalog/nodes (33.41µs) from=127.0.0.1:55097
2016/10/31 21:10:00 [DEBUG] http: Request GET /v1/catalog/nodes (37.158µs) from=127.0.0.1:55097
--- FAIL: TestCatalog_Nodes (12.09s)
catalog_test.go:59: err: Bad: &{node1 127.0.0.1 map[]}
``` | non_defect | track down testcatalog nodes timeout failure http request get catalog nodes from http request get catalog nodes from http request get catalog nodes from fail testcatalog nodes catalog test go err bad map | 0 |
77,749 | 27,143,556,915 | IssuesEvent | 2023-02-16 18:04:49 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Composer doesn't refocus when "Unknown Command" dialog is quit | T-Defect S-Minor Help Wanted A-Composer A-Slash-Commands A11y O-Uncommon | ### Steps to reproduce
1. In a room, send an invalid command from the composer. e.g. `/>`
2. See the "Unknown Command" dialog appear
3. Press `Esc` on the keyboard to quit the dialog
### Outcome
#### What did you expect?
Dialog closes and composer is re-focused
#### What happened instead?
Dialog closes and composer lost focus
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
develop.element.io
### Application version
Element version: fc1db94c8030-react-6fb834200387-js-7a18991342cb Olm version: 3.2.8
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | Composer doesn't refocus when "Unknown Command" dialog is quit - ### Steps to reproduce
1. In a room, send an invalid command from the composer. e.g. `/>`
2. See the "Unknown Command" dialog appear
3. Press `Esc` on the keyboard to quit the dialog
### Outcome
#### What did you expect?
Dialog closes and composer is re-focused
#### What happened instead?
Dialog closes and composer lost focus
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
develop.element.io
### Application version
Element version: fc1db94c8030-react-6fb834200387-js-7a18991342cb Olm version: 3.2.8
### Homeserver
_No response_
### Will you send logs?
No | defect | composer doesn t refocus when unknown command dialog is quit steps to reproduce in a room send an invalid command from the composer e g see the unknown command dialog appear press esc on the keyboard to quit the dialog outcome what did you expect dialog closes and composer is re focused what happened instead dialog closes and composer lost focus operating system no response browser information no response url for webapp develop element io application version element version react js olm version homeserver no response will you send logs no | 1 |
26,489 | 4,729,397,480 | IssuesEvent | 2016-10-18 18:35:20 | edno/kleis | https://api.github.com/repos/edno/kleis | closed | Correction et amelioration des informations suggestives pour le formulaire de whitelist des domaines | defect | In gitlab by @ValentinGuerlesquin on Aug 10, 2016, 02:13
l’information suggestive pour ajouter un domaine a la liste blanche propose : `domaine.sousdomaine.ext` alors qu’il faudrait plutôt suggérer `sousdomaine.domaine.ext`
Accessoirement il serait mieux de mettre `domaine.ext ou sousdomaine.domaine.ext` pour aider l'utiisateur a comprendre que s'il ne précise pas le sous-domaine, alors l'ensemble du domaine sera ajouté a la whiteliste | 1.0 | Correction et amelioration des informations suggestives pour le formulaire de whitelist des domaines - In gitlab by @ValentinGuerlesquin on Aug 10, 2016, 02:13
l’information suggestive pour ajouter un domaine a la liste blanche propose : `domaine.sousdomaine.ext` alors qu’il faudrait plutôt suggérer `sousdomaine.domaine.ext`
Accessoirement il serait mieux de mettre `domaine.ext ou sousdomaine.domaine.ext` pour aider l'utiisateur a comprendre que s'il ne précise pas le sous-domaine, alors l'ensemble du domaine sera ajouté a la whiteliste | defect | correction et amelioration des informations suggestives pour le formulaire de whitelist des domaines in gitlab by valentinguerlesquin on aug l’information suggestive pour ajouter un domaine a la liste blanche propose domaine sousdomaine ext alors qu’il faudrait plutôt suggérer sousdomaine domaine ext accessoirement il serait mieux de mettre domaine ext ou sousdomaine domaine ext pour aider l utiisateur a comprendre que s il ne précise pas le sous domaine alors l ensemble du domaine sera ajouté a la whiteliste | 1 |
806,663 | 29,867,122,709 | IssuesEvent | 2023-06-20 05:26:30 | Fiserv/Support | https://api.github.com/repos/Fiserv/Support | closed | Download Postman Collection Option functionality issue, Server Error when downloading Postman Collection | bug Priority - Medium Severity - Medium client verification Download Postman | **Region (optional)**
EMEA and NA
**Is your feature request related to a problem? Please describe.**
In the page of every endpoint is a section called **Additional resources**, this section includes the "Download Postman collection" option. When this option is clicked, the download bar shows up, but the zip file containing the Postman Collection has the message "Failed - Server problem"
**Describe the solution you'd like**
After the download link is clicked, the Postman Collection link should download the json file relating to the postman Collection that the API endpoint is contained within.
For example: In the image below, the Postman Collection link should only download the Postman Collection for Account.yaml.
**Describe alternatives you've considered**
None
**Additional context**
<img width="1439" alt="Screenshot 2023-04-19 at 1 23 21 PM" src="https://user-images.githubusercontent.com/108832813/233152743-fa6ebf40-a3dd-4eef-9dad-5b2216b7c2be.png"> | 1.0 | Download Postman Collection Option functionality issue, Server Error when downloading Postman Collection - **Region (optional)**
EMEA and NA
**Is your feature request related to a problem? Please describe.**
In the page of every endpoint is a section called **Additional resources**, this section includes the "Download Postman collection" option. When this option is clicked, the download bar shows up, but the zip file containing the Postman Collection has the message "Failed - Server problem"
**Describe the solution you'd like**
After the download link is clicked, the Postman Collection link should download the json file relating to the postman Collection that the API endpoint is contained within.
For example: In the image below, the Postman Collection link should only download the Postman Collection for Account.yaml.
**Describe alternatives you've considered**
None
**Additional context**
<img width="1439" alt="Screenshot 2023-04-19 at 1 23 21 PM" src="https://user-images.githubusercontent.com/108832813/233152743-fa6ebf40-a3dd-4eef-9dad-5b2216b7c2be.png"> | non_defect | download postman collection option functionality issue server error when downloading postman collection region optional emea and na is your feature request related to a problem please describe in the page of every endpoint is a section called additional resources this section includes the download postman collection option when this option is clicked the download bar shows up but the zip file containing the postman collection has the message failed server problem describe the solution you d like after the download link is clicked the postman collection link should download the json file relating to the postman collection that the api endpoint is contained within for example in the image below the postman collection link should only download the postman collection for account yaml describe alternatives you ve considered none additional context img width alt screenshot at pm src | 0 |
498,892 | 14,435,135,646 | IssuesEvent | 2020-12-07 08:17:31 | aau-giraf/api_client | https://api.github.com/repos/aau-giraf/api_client | closed | As a developer I would like an interface for the offline features | Offline mode group 9 point: 1 prepare for handover priority: highest | **Is your feature request related to a problem? Please describe.**
At the moment, only one group can work on the offline mode at the same time.
**Describe the solution you'd like**
An interface would allow the work to be spread out to more groups
Related to #52 and #72 (And aau-giraf/weekplanner#9)
| 1.0 | As a developer I would like an interface for the offline features - **Is your feature request related to a problem? Please describe.**
At the moment, only one group can work on the offline mode at the same time.
**Describe the solution you'd like**
An interface would allow the work to be spread out to more groups
Related to #52 and #72 (And aau-giraf/weekplanner#9)
| non_defect | as a developer i would like an interface for the offline features is your feature request related to a problem please describe at the moment only one group can work on the offline mode at the same time describe the solution you d like an interface would allow the work to be spread out to more groups related to and and aau giraf weekplanner | 0 |
88,410 | 10,570,873,366 | IssuesEvent | 2019-10-07 04:50:40 | K-Desk/members | https://api.github.com/repos/K-Desk/members | opened | Tambah Nama Sendiri-Sendiri | documentation enhancement good first issue | Buat Pull Request untuk menambahkan namamu, harap perhatikan, nama harus lengkap dan ditambahkan nickname serta url akun github. Kemudian tambah sebuah snippet code tereserah yang menarik :sunglasses: | 1.0 | Tambah Nama Sendiri-Sendiri - Buat Pull Request untuk menambahkan namamu, harap perhatikan, nama harus lengkap dan ditambahkan nickname serta url akun github. Kemudian tambah sebuah snippet code tereserah yang menarik :sunglasses: | non_defect | tambah nama sendiri sendiri buat pull request untuk menambahkan namamu harap perhatikan nama harus lengkap dan ditambahkan nickname serta url akun github kemudian tambah sebuah snippet code tereserah yang menarik sunglasses | 0 |
25,365 | 18,543,958,225 | IssuesEvent | 2021-10-21 19:29:54 | wnbrb/wnb-rb-site | https://api.github.com/repos/wnbrb/wnb-rb-site | closed | Set up website deployment pipeline | infrastructure | This ticket involves a few steps:
- [x] Decide where to deploy (Heroku?)
- [x] Set up the initial deployment
- [x] Create a GitHub action that continuously deploys the website every time it is merged
I am happy to pair on this! | 1.0 | Set up website deployment pipeline - This ticket involves a few steps:
- [x] Decide where to deploy (Heroku?)
- [x] Set up the initial deployment
- [x] Create a GitHub action that continuously deploys the website every time it is merged
I am happy to pair on this! | non_defect | set up website deployment pipeline this ticket involves a few steps decide where to deploy heroku set up the initial deployment create a github action that continuously deploys the website every time it is merged i am happy to pair on this | 0 |
306,287 | 26,454,601,596 | IssuesEvent | 2023-01-16 13:39:10 | mautic/mautic | https://api.github.com/repos/mautic/mautic | closed | Create 5.x PR and increase test coverage of new Hubspot API changes | enhancement automated-tests | @kuzmany @escopecz this needs to go out in 4.4.5 but it is failing on test coverage. Anything that can be done to get this over the line?
_Originally posted by @RCheesley in https://github.com/mautic/mautic/issues/11470#issuecomment-1330898518_
This needs to be merged into the next 4.4.x release and also into the 5.x branch, creating the issue so that we do not lose sight of it as the 4.4.5 release needs this PR.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://app.bountysource.com/issues/114275233-create-5-x-pr-and-increase-test-coverage-of-new-hubspot-api-changes?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://app.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | Create 5.x PR and increase test coverage of new Hubspot API changes - @kuzmany @escopecz this needs to go out in 4.4.5 but it is failing on test coverage. Anything that can be done to get this over the line?
_Originally posted by @RCheesley in https://github.com/mautic/mautic/issues/11470#issuecomment-1330898518_
This needs to be merged into the next 4.4.x release and also into the 5.x branch, creating the issue so that we do not lose sight of it as the 4.4.5 release needs this PR.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://app.bountysource.com/issues/114275233-create-5-x-pr-and-increase-test-coverage-of-new-hubspot-api-changes?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://app.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F5355074&utm_medium=issues&utm_source=github).
</bountysource-plugin> | non_defect | create x pr and increase test coverage of new hubspot api changes kuzmany escopecz this needs to go out in but it is failing on test coverage anything that can be done to get this over the line originally posted by rcheesley in this needs to be merged into the next x release and also into the x branch creating the issue so that we do not lose sight of it as the release needs this pr want to back this issue we accept bounties via | 0 |
9,430 | 2,615,149,733 | IssuesEvent | 2015-03-01 06:26:50 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | closed | Alfa AWUS050NH support (rt2800usb) | auto-migrated Priority-Low Type-Defect | ```
add support to it. didnt work using backtrack 5 r1
```
Original issue reported on code.google.com by `wil.c.w...@gmail.com` on 12 Jan 2012 at 11:39 | 1.0 | Alfa AWUS050NH support (rt2800usb) - ```
add support to it. didnt work using backtrack 5 r1
```
Original issue reported on code.google.com by `wil.c.w...@gmail.com` on 12 Jan 2012 at 11:39 | defect | alfa support add support to it didnt work using backtrack original issue reported on code google com by wil c w gmail com on jan at | 1 |
4,733 | 2,610,153,856 | IssuesEvent | 2015-02-26 18:48:58 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | closed | CIS Patrol Frigate | auto-migrated Priority-Medium Type-Defect | ```
Select box scale needs to be increased
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:07 | 1.0 | CIS Patrol Frigate - ```
Select box scale needs to be increased
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:07 | defect | cis patrol frigate select box scale needs to be increased original issue reported on code google com by gmail com on jan at | 1 |
345,117 | 30,789,167,083 | IssuesEvent | 2023-07-31 15:02:21 | wazuh/wazuh-kibana-app | https://api.github.com/repos/wazuh/wazuh-kibana-app | closed | Compatibility with Wazuh 4.6.0 | type/test compatibility level/task | ## Description
We need to ensure the UI compatibility with the next version of Wazuh, and the latest version of wazuh-indexer and wazuh-dashboard
This update is still being discussed, but we need to be aware of potential issues.
For that, we need to:
- [x] Review wazuh-indexer and wazuh-dashboard latest stable changelog.
- [x] Identify improvements and potential impact on the UI.
- [x] Develop a testing environment to verify our components would work under this new build.
- [x] Update imposter with the changes made to Wazuh 4.6.0 for testing
- [x] Verify that the images for the testing environment works as expected
## Issues
- https://github.com/wazuh/wazuh-dashboard/issues/68
- https://github.com/wazuh/wazuh-dashboard/issues/62
| 1.0 | Compatibility with Wazuh 4.6.0 - ## Description
We need to ensure the UI compatibility with the next version of Wazuh, and the latest version of wazuh-indexer and wazuh-dashboard
This update is still being discussed, but we need to be aware of potential issues.
For that, we need to:
- [x] Review wazuh-indexer and wazuh-dashboard latest stable changelog.
- [x] Identify improvements and potential impact on the UI.
- [x] Develop a testing environment to verify our components would work under this new build.
- [x] Update imposter with the changes made to Wazuh 4.6.0 for testing
- [x] Verify that the images for the testing environment works as expected
## Issues
- https://github.com/wazuh/wazuh-dashboard/issues/68
- https://github.com/wazuh/wazuh-dashboard/issues/62
| non_defect | compatibility with wazuh description we need to ensure the ui compatibility with the next version of wazuh and the latest version of wazuh indexer and wazuh dashboard this update is still being discussed but we need to be aware of potential issues for that we need to review wazuh indexer and wazuh dashboard latest stable changelog identify improvements and potential impact on the ui develop a testing environment to verify our components would work under this new build update imposter with the changes made to wazuh for testing verify that the images for the testing environment works as expected issues | 0 |
584,956 | 17,467,660,083 | IssuesEvent | 2021-08-06 19:27:26 | googleapis/repo-automation-bots | https://api.github.com/repos/googleapis/repo-automation-bots | opened | OwlBot Failing on java-aiplatform, due to too many subsequent updates | priority: p1 bot: owl-bot | There's an error happening on OwlBot:
```
Error: too many OwlBot updates created in a row for googleapis/java-aiplatform at runPostProcessor (/workspace/build/src/owl-bot.js:267:15) at async Object.handlePullRequestLabeled (/workspace/build/src/owl-bot.js:173:5) at async /workspace/build/src/owl-bot.js:57:9 at async Promise.all (index 0) at async /workspace/node_modules/gcf-utils/build/src/gcf-utils.js:371:21
Error: too many OwlBot updates created in a row for googleapis/java-aiplatform at runPostProcessor (/workspace/build/src/owl-bot.js:267:15) at async Object.handlePullRequestLabeled (/workspace/build/src/owl-bot.js:173:5) at async /workspace/build/src/owl-bot.js:57:9 at async Promise.all (index 0) at async /workspace/node_modules/gcf-utils/build/src/gcf-utils.js:371:21
```
As a result of the PR:
https://github.com/googleapis/java-aiplatform/pull/449/commits
We have a safety in a place that doesn't allow three subsequent updates from OwlBot, I believe this safety is being incorrectly triggered on this PR, even though it's had new updates since OwlBot's changes.
CC: @Neenu1995 | 1.0 | OwlBot Failing on java-aiplatform, due to too many subsequent updates - There's an error happening on OwlBot:
```
Error: too many OwlBot updates created in a row for googleapis/java-aiplatform at runPostProcessor (/workspace/build/src/owl-bot.js:267:15) at async Object.handlePullRequestLabeled (/workspace/build/src/owl-bot.js:173:5) at async /workspace/build/src/owl-bot.js:57:9 at async Promise.all (index 0) at async /workspace/node_modules/gcf-utils/build/src/gcf-utils.js:371:21
Error: too many OwlBot updates created in a row for googleapis/java-aiplatform at runPostProcessor (/workspace/build/src/owl-bot.js:267:15) at async Object.handlePullRequestLabeled (/workspace/build/src/owl-bot.js:173:5) at async /workspace/build/src/owl-bot.js:57:9 at async Promise.all (index 0) at async /workspace/node_modules/gcf-utils/build/src/gcf-utils.js:371:21
```
As a result of the PR:
https://github.com/googleapis/java-aiplatform/pull/449/commits
We have a safety in a place that doesn't allow three subsequent updates from OwlBot, I believe this safety is being incorrectly triggered on this PR, even though it's had new updates since OwlBot's changes.
CC: @Neenu1995 | non_defect | owlbot failing on java aiplatform due to too many subsequent updates there s an error happening on owlbot error too many owlbot updates created in a row for googleapis java aiplatform at runpostprocessor workspace build src owl bot js at async object handlepullrequestlabeled workspace build src owl bot js at async workspace build src owl bot js at async promise all index at async workspace node modules gcf utils build src gcf utils js error too many owlbot updates created in a row for googleapis java aiplatform at runpostprocessor workspace build src owl bot js at async object handlepullrequestlabeled workspace build src owl bot js at async workspace build src owl bot js at async promise all index at async workspace node modules gcf utils build src gcf utils js as a result of the pr we have a safety in a place that doesn t allow three subsequent updates from owlbot i believe this safety is being incorrectly triggered on this pr even though it s had new updates since owlbot s changes cc | 0 |
69,548 | 7,147,492,300 | IssuesEvent | 2018-01-25 01:03:21 | fpco/stackage | https://api.github.com/repos/fpco/stackage | closed | zm-0.3 doctest build failure | failure: test-suite type: failure | ```haskell
> /tmp/stackage-build14/zm-0.3$ dist/build/zm-doctest/zm-doctest
src/ZM/Pretty.hs:33:1: error:
Ambiguous module name ‘Text.PrettyPrint.HughesPJClass’:
it was found in multiple packages:
pretty-1.1.3.3 pretty-class-1.0.1.1 prettyclass-1.0.0.0
|
33 | import Text.PrettyPrint.HughesPJClass
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
```
CC @tittoassini | 1.0 | zm-0.3 doctest build failure - ```haskell
> /tmp/stackage-build14/zm-0.3$ dist/build/zm-doctest/zm-doctest
src/ZM/Pretty.hs:33:1: error:
Ambiguous module name ‘Text.PrettyPrint.HughesPJClass’:
it was found in multiple packages:
pretty-1.1.3.3 pretty-class-1.0.1.1 prettyclass-1.0.0.0
|
33 | import Text.PrettyPrint.HughesPJClass
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
```
CC @tittoassini | non_defect | zm doctest build failure haskell tmp stackage zm dist build zm doctest zm doctest src zm pretty hs error ambiguous module name ‘text prettyprint hughespjclass’ it was found in multiple packages pretty pretty class prettyclass import text prettyprint hughespjclass cc tittoassini | 0 |
59,854 | 8,381,056,160 | IssuesEvent | 2018-10-07 20:52:26 | niamurrell/value-app | https://api.github.com/repos/niamurrell/value-app | reopened | DOCUMENTATION: Improve contribution guidelines | documentation hacktoberfest help wanted | ### Feature Request/Idea
Add clarifications or further prompts to improve the contribution guidelines for people new to open source. [These guidelines](https://github.com/thedaviddias/Front-End-Performance-Checklist/blob/master/.github/CONTRIBUTING.md) might have some good elements to include. | 1.0 | DOCUMENTATION: Improve contribution guidelines - ### Feature Request/Idea
Add clarifications or further prompts to improve the contribution guidelines for people new to open source. [These guidelines](https://github.com/thedaviddias/Front-End-Performance-Checklist/blob/master/.github/CONTRIBUTING.md) might have some good elements to include. | non_defect | documentation improve contribution guidelines feature request idea add clarifications or further prompts to improve the contribution guidelines for people new to open source might have some good elements to include | 0 |
3,144 | 2,607,985,506 | IssuesEvent | 2015-02-26 00:51:31 | chrsmithdemos/zen-coding | https://api.github.com/repos/chrsmithdemos/zen-coding | closed | Dreamweaver CS5.5: Wrapping with Abbreviation turns all '$' characters to '1''s | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Open Dreamweaver
2. Open New HTML Document
3. Add '$' Character
4. Highlight and either use the keyboard shortcut Control + H or Command > Zen
Coding > Wrap with Abbreviation
5. Wrap it in any HTML tag, a div for example
6. Your '$' character turns into a '1'
What is the expected output? What do you see instead?
One would expect it to wrap the '$', dollar sign, symbol, character in a
<div></div>, so one would see <div>$</div>; instead it changes the '$' to a '1'
and thus making <div>1</div>.
What version of the product are you using? On what operating system?
Dreamweaver CS5.5, Windows 7, Zen Coding 0.7.5 Extension
Please provide any additional information below.
This is a pretty straight forward issue to recreate, I only noticed it because
I have a WordPress PHP snippet with the line:
'items_wrap' => '%3$s'
and I noticed it kept changing it to:
'items_wrap' => '%31s'
...after I'd wrap it using Zen Coding.
I haven't seen this issue marked anywhere so if it is delete this report.
Cheers,
Frank
```
-----
Original issue reported on code.google.com by `IamFrankStallone` on 8 Mar 2012 at 12:51 | 1.0 | Dreamweaver CS5.5: Wrapping with Abbreviation turns all '$' characters to '1''s - ```
What steps will reproduce the problem?
1. Open Dreamweaver
2. Open New HTML Document
3. Add '$' Character
4. Highlight and either use the keyboard shortcut Control + H or Command > Zen
Coding > Wrap with Abbreviation
5. Wrap it in any HTML tag, a div for example
6. Your '$' character turns into a '1'
What is the expected output? What do you see instead?
One would expect it to wrap the '$', dollar sign, symbol, character in a
<div></div>, so one would see <div>$</div>; instead it changes the '$' to a '1'
and thus making <div>1</div>.
What version of the product are you using? On what operating system?
Dreamweaver CS5.5, Windows 7, Zen Coding 0.7.5 Extension
Please provide any additional information below.
This is a pretty straight forward issue to recreate, I only noticed it because
I have a WordPress PHP snippet with the line:
'items_wrap' => '%3$s'
and I noticed it kept changing it to:
'items_wrap' => '%31s'
...after I'd wrap it using Zen Coding.
I haven't seen this issue marked anywhere so if it is delete this report.
Cheers,
Frank
```
-----
Original issue reported on code.google.com by `IamFrankStallone` on 8 Mar 2012 at 12:51 | defect | dreamweaver wrapping with abbreviation turns all characters to s what steps will reproduce the problem open dreamweaver open new html document add character highlight and either use the keyboard shortcut control h or command zen coding wrap with abbreviation wrap it in any html tag a div for example your character turns into a what is the expected output what do you see instead one would expect it to wrap the dollar sign symbol character in a so one would see instead it changes the to a and thus making what version of the product are you using on what operating system dreamweaver windows zen coding extension please provide any additional information below this is a pretty straight forward issue to recreate i only noticed it because i have a wordpress php snippet with the line items wrap s and i noticed it kept changing it to items wrap after i d wrap it using zen coding i haven t seen this issue marked anywhere so if it is delete this report cheers frank original issue reported on code google com by iamfrankstallone on mar at | 1 |
12,462 | 2,700,551,744 | IssuesEvent | 2015-04-04 08:52:06 | lextm/obfuscar | https://api.github.com/repos/lextm/obfuscar | closed | Obfuscator crashes when trying to obfuscate two assemblies in a row using one config file | auto-migrated bug Type-Defect | ```
What steps will reproduce the problem?
1.
I have two assemblies, one class library, another an exe. I am trying to
obfuscate both at the same time using a config file similar to the following:
<?xml version='1.0'?>
<Obfuscator>
<Var name="InPath" value=".\Obfuscator_Input" />
<Var name="OutPath" value=".\Obfuscator_Output" />
<Var name="ReuseNames" value="true" />
<Module file="$(InPath)\myexe.exe"/>
<Module file="$(InPath)\somecontrols.dll"/>
</Obfuscator>
When I run the tool, it crashes with (full output):
Loading project...Loading assemblies...Done.
Renaming: fields...parameters...properties...events...methods...types...
Unhandled Exception: System.ArgumentException: An item with the same key
has already been added.
at System.ThrowHelper.ThrowArgumentException(ExceptionResource resource)
at System.Collections.Generic.Dictionary`2.Insert(TKey key, TValue
value, Boolean add)
at Obfuscar.Obfuscator.RenameTypes()
at Obfuscar.Program.Main(String[] args)
Note that running two separate config files for each assembly works just fine.
2.
3.
What is the expected output? What do you see instead?
No app crash.
What version of the product are you using? On what operating system?
1.5.4. Win XP with SP3
Please provide any additional information below.
```
Original issue reported on code.google.com by `radun...@gmail.com` on 28 Feb 2010 at 7:03 | 1.0 | Obfuscator crashes when trying to obfuscate two assemblies in a row using one config file - ```
What steps will reproduce the problem?
1.
I have two assemblies, one class library, another an exe. I am trying to
obfuscate both at the same time using a config file similar to the following:
<?xml version='1.0'?>
<Obfuscator>
<Var name="InPath" value=".\Obfuscator_Input" />
<Var name="OutPath" value=".\Obfuscator_Output" />
<Var name="ReuseNames" value="true" />
<Module file="$(InPath)\myexe.exe"/>
<Module file="$(InPath)\somecontrols.dll"/>
</Obfuscator>
When I run the tool, it crashes with (full output):
Loading project...Loading assemblies...Done.
Renaming: fields...parameters...properties...events...methods...types...
Unhandled Exception: System.ArgumentException: An item with the same key
has already been added.
at System.ThrowHelper.ThrowArgumentException(ExceptionResource resource)
at System.Collections.Generic.Dictionary`2.Insert(TKey key, TValue
value, Boolean add)
at Obfuscar.Obfuscator.RenameTypes()
at Obfuscar.Program.Main(String[] args)
Note that running two separate config files for each assembly works just fine.
2.
3.
What is the expected output? What do you see instead?
No app crash.
What version of the product are you using? On what operating system?
1.5.4. Win XP with SP3
Please provide any additional information below.
```
Original issue reported on code.google.com by `radun...@gmail.com` on 28 Feb 2010 at 7:03 | defect | obfuscator crashes when trying to obfuscate two assemblies in a row using one config file what steps will reproduce the problem i have two assemblies one class library another an exe i am trying to obfuscate both at the same time using a config file similar to the following when i run the tool it crashes with full output loading project loading assemblies done renaming fields parameters properties events methods types unhandled exception system argumentexception an item with the same key has already been added at system throwhelper throwargumentexception exceptionresource resource at system collections generic dictionary insert tkey key tvalue value boolean add at obfuscar obfuscator renametypes at obfuscar program main string args note that running two separate config files for each assembly works just fine what is the expected output what do you see instead no app crash what version of the product are you using on what operating system win xp with please provide any additional information below original issue reported on code google com by radun gmail com on feb at | 1 |
355,578 | 25,175,971,412 | IssuesEvent | 2022-11-11 09:17:45 | liviamil/pe | https://api.github.com/repos/liviamil/pe | opened | Inconvenienced UG flow as a new user | severity.VeryLow type.DocumentationBug |

When reading the UG from top to bottom, user is learning the app terms as they go. However, once user reached "sorting the task list", they are under the assumption of knowing what these criteria are. However, When executing the sort c/ command, nothing happens because user has yet to learn adding tags at this point of time.

*Although CRITERIA is explained under restrictions, the terms "priority", "deadline" are still very much unfamiliar for new users, hence deterring the explanation
Since criteria is apparently based on the tags added to tasks, it would have been better to introduce sort after introducing tags to prevent confusion and overall user experience for users trying out the app while concurrently using the user guide.
<!--session: 1668154738231-4d843099-6ea7-446e-ac40-75c85c67f836-->
<!--Version: Web v3.4.4--> | 1.0 | Inconvenienced UG flow as a new user -

When reading the UG from top to bottom, user is learning the app terms as they go. However, once user reached "sorting the task list", they are under the assumption of knowing what these criteria are. However, When executing the sort c/ command, nothing happens because user has yet to learn adding tags at this point of time.

*Although CRITERIA is explained under restrictions, the terms "priority", "deadline" are still very much unfamiliar for new users, hence deterring the explanation
Since criteria is apparently based on the tags added to tasks, it would have been better to introduce sort after introducing tags to prevent confusion and overall user experience for users trying out the app while concurrently using the user guide.
<!--session: 1668154738231-4d843099-6ea7-446e-ac40-75c85c67f836-->
<!--Version: Web v3.4.4--> | non_defect | inconvenienced ug flow as a new user when reading the ug from top to bottom user is learning the app terms as they go however once user reached sorting the task list they are under the assumption of knowing what these criteria are however when executing the sort c command nothing happens because user has yet to learn adding tags at this point of time although criteria is explained under restrictions the terms priority deadline are still very much unfamiliar for new users hence deterring the explanation since criteria is apparently based on the tags added to tasks it would have been better to introduce sort after introducing tags to prevent confusion and overall user experience for users trying out the app while concurrently using the user guide | 0 |
28,179 | 5,207,385,998 | IssuesEvent | 2017-01-24 23:21:57 | netty/netty | https://api.github.com/repos/netty/netty | closed | HTTP2 Encoder Failure on encode of oversized headers corrupts dynamic table consistency | defect | ### Problem Description
When the HTTP/2 encoder fails to encode headers due to headers that exceed the MAX_HEADER_LIST_SIZE during an encode where new headers are added, the dynamic on the sending side is updated, but the dynamic table on the receiving side is not. This can cause subsequent encodes of headers that already exist in the dynamic table to map to an incorrect index on the receiving side, in a production environment this can cause headers from one request/response to end up on another.
This bug is closely related to #6209
### Expected behavior
When oversized headers are encoded a GO_AWAY with an INTERNAL_ERROR should be sent to prevent the connection from becoming corrupted. Additionally, the failure to encode should be returned to the promise that represents the encode, right now this appears to fail silently.
### Steps to reproduce
See attached maven project zip
### Netty version
4.1.6, 4.1.7
[netty-debug.zip](https://github.com/netty/netty/files/705828/netty-debug.zip) | 1.0 | HTTP2 Encoder Failure on encode of oversized headers corrupts dynamic table consistency - ### Problem Description
When the HTTP/2 encoder fails to encode headers due to headers that exceed the MAX_HEADER_LIST_SIZE during an encode where new headers are added, the dynamic on the sending side is updated, but the dynamic table on the receiving side is not. This can cause subsequent encodes of headers that already exist in the dynamic table to map to an incorrect index on the receiving side, in a production environment this can cause headers from one request/response to end up on another.
This bug is closely related to #6209
### Expected behavior
When oversized headers are encoded a GO_AWAY with an INTERNAL_ERROR should be sent to prevent the connection from becoming corrupted. Additionally, the failure to encode should be returned to the promise that represents the encode, right now this appears to fail silently.
### Steps to reproduce
See attached maven project zip
### Netty version
4.1.6, 4.1.7
[netty-debug.zip](https://github.com/netty/netty/files/705828/netty-debug.zip) | defect | encoder failure on encode of oversized headers corrupts dynamic table consistency problem description when the http encoder fails to encode headers due to headers that exceed the max header list size during an encode where new headers are added the dynamic on the sending side is updated but the dynamic table on the receiving side is not this can cause subsequent encodes of headers that already exist in the dynamic table to map to an incorrect index on the receiving side in a production environment this can cause headers from one request response to end up on another this bug is closely related to expected behavior when oversized headers are encoded a go away with an internal error should be sent to prevent the connection from becoming corrupted additionally the failure to encode should be returned to the promise that represents the encode right now this appears to fail silently steps to reproduce see attached maven project zip netty version | 1 |
566,144 | 16,812,827,896 | IssuesEvent | 2021-06-17 01:36:30 | mozilla/fx-private-relay | https://api.github.com/repos/mozilla/fx-private-relay | closed | The banner section disappears after installing the dev version of the Relay addon | priority-P1 softvision | #### Platforms affected::
- Windows 10, Linux Ubuntu 18.04, MacOS 10.15;
#### Browsers tested on:
- Firefox;
#### Prerequisites:
- A new firefox profile has been created;
- The user has a Private Relay account without the Mozilla Privacy Pack purchased;
- Have the test addon downloaded, but not installed;
#### Steps to reproduce:
1. Open firefox in the new profile;
2. Go to https://dev.fxprivaterelay.nonprod.cloudops.mozgcp.net/accounts/profile/;
3. Log into the created profile and check the banner section just below the page header;
4. Install the test addon that points to the fxa stage;
5. Refresh the Relay page and check the banner section again;
#### Expected result:
- The promotional banner should stay displayed;
#### Actual result:
- The promotional banner disappears;
#### Notes:
- Uninstalling the extension and refreshing does not change anything;


| 1.0 | The banner section disappears after installing the dev version of the Relay addon - #### Platforms affected::
- Windows 10, Linux Ubuntu 18.04, MacOS 10.15;
#### Browsers tested on:
- Firefox;
#### Prerequisites:
- A new firefox profile has been created;
- The user has a Private Relay account without the Mozilla Privacy Pack purchased;
- Have the test addon downloaded, but not installed;
#### Steps to reproduce:
1. Open firefox in the new profile;
2. Go to https://dev.fxprivaterelay.nonprod.cloudops.mozgcp.net/accounts/profile/;
3. Log into the created profile and check the banner section just below the page header;
4. Install the test addon that points to the fxa stage;
5. Refresh the Relay page and check the banner section again;
#### Expected result:
- The promotional banner should stay displayed;
#### Actual result:
- The promotional banner disappears;
#### Notes:
- Uninstalling the extension and refreshing does not change anything;


| non_defect | the banner section disappears after installing the dev version of the relay addon platforms affected windows linux ubuntu macos browsers tested on firefox prerequisites a new firefox profile has been created the user has a private relay account without the mozilla privacy pack purchased have the test addon downloaded but not installed steps to reproduce open firefox in the new profile go to log into the created profile and check the banner section just below the page header install the test addon that points to the fxa stage refresh the relay page and check the banner section again expected result the promotional banner should stay displayed actual result the promotional banner disappears notes uninstalling the extension and refreshing does not change anything | 0 |
3,690 | 2,610,067,048 | IssuesEvent | 2015-02-26 18:19:41 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | 路桥治前列腺炎哪家权威 | auto-migrated Priority-Medium Type-Defect | ```
路桥治前列腺炎哪家权威【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:44 | 1.0 | 路桥治前列腺炎哪家权威 - ```
路桥治前列腺炎哪家权威【台州五洲生殖医院】24小时健康咨
询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州
市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108�
��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109
、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:44 | defect | 路桥治前列腺炎哪家权威 路桥治前列腺炎哪家权威【台州五洲生殖医院】 询热线 微信号tzwzszyy 医院地址 台州 (枫南大转盘旁)乘车线路 、 � �� 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at | 1 |
26,504 | 4,732,279,644 | IssuesEvent | 2016-10-19 07:09:27 | zotonic/zotonic | https://api.github.com/repos/zotonic/zotonic | closed | Blog site skeleton has problem on master. | defect | After starting a site made from the blog skeleton it returns this page.
I'm not sure if this is a problem with the skeleton or with zotonic itself.

| 1.0 | Blog site skeleton has problem on master. - After starting a site made from the blog skeleton it returns this page.
I'm not sure if this is a problem with the skeleton or with zotonic itself.

| defect | blog site skeleton has problem on master after starting a site made from the blog skeleton it returns this page i m not sure if this is a problem with the skeleton or with zotonic itself | 1 |
34,170 | 7,372,730,555 | IssuesEvent | 2018-03-13 15:26:59 | ONSdigital/eq-survey-runner | https://api.github.com/repos/ONSdigital/eq-survey-runner | opened | Highlight not padded correctly | design-defect | ### Expected behaviour
Text that is highlighted has 2px padding on left and right edges.
### Actual behaviour
Text that is highlighted is un-padded.
Use of `<em>` without the `class="highlight"` means that the appropriate padding is not applied.
### Steps to reproduce the behaviour
View a question with highlight (https://eq.onsdigital.uk/questionnaire/1/0005/ - e.g. total gross weekly pay)
### Technical information
#### Browser
Chrome 64
#### Operating System
Apple High Sierra
### Screenshot

| 1.0 | Highlight not padded correctly - ### Expected behaviour
Text that is highlighted has 2px padding on left and right edges.
### Actual behaviour
Text that is highlighted is un-padded.
Use of `<em>` without the `class="highlight"` means that the appropriate padding is not applied.
### Steps to reproduce the behaviour
View a question with highlight (https://eq.onsdigital.uk/questionnaire/1/0005/ - e.g. total gross weekly pay)
### Technical information
#### Browser
Chrome 64
#### Operating System
Apple High Sierra
### Screenshot

| defect | highlight not padded correctly expected behaviour text that is highlighted has padding on left and right edges actual behaviour text that is highlighted is un padded use of without the class highlight means that the appropriate padding is not applied steps to reproduce the behaviour view a question with highlight e g total gross weekly pay technical information browser chrome operating system apple high sierra screenshot | 1 |
387,307 | 11,459,726,762 | IssuesEvent | 2020-02-07 08:05:27 | kubernetes-sigs/cluster-api | https://api.github.com/repos/kubernetes-sigs/cluster-api | closed | Define clusterctl move process | area/clusterctl kind/feature lifecycle/active priority/important-soon | [2019-11-26 Updated] according to https://github.com/kubernetes-sigs/cluster-api/pull/1730#discussion_r346935816 Pivot is now rebranded into Move; issue updated accordingly
**User Story**
As an operator I would like to move cluster objects and all the associated resources (Machines, Machine Depolyments etc.) from the current management cluster to another management cluster for any reason
**Detailed Description**
The clusterctl CAEP currently in flight assumes the user should brig its own management cluster, so technically the sequence bootstrap cluster -> pivot to -> management cluster is not necessary anymore.
However, the same CAEP consider pivoting a possible answer to different operational needs, e.g because of maintenance or replacement of the management cluster, so pivoting is still in scope.
With v1alpha3 in flight and the new assumptions around clusterctl - one binary for rule all the providers -, the implementation detail should be re-validated, keeping in mind also https://github.com/kubernetes-sigs/cluster-api/pull/1730#discussion_r346935816 discussion that lead to transforming pivot into move.
***Goals***
- 1. To define a moving process that can work on any clusterctl generated management cluster (with any provider or any combinations of providers)
- To define how to move provider-specific objects (or eventually hierarchies of objects)
- 2. To define conventions/requirements the above process should rely on e.g.
- define a convention for forcing cluster-api controllers to not reconcile resources.
- [Auto labeling cluster resources](https://github.com/kubernetes-sigs/cluster-api/issues/1489)
- [Define a convention for namespaced deployments](https://github.com/kubernetes-sigs/cluster-api/issues/1490)
- 3. To define move specializations, like e.g. move only clusters in a given namespace or move only a specific cluster
- 4. To define eventual move side effects (e.g. If forcing cluster-api controllers to not reconcile resources will be implemented scaling down controllers, all the cluster objects stop to reconcile)
- 5. To define preflight checks to be executed before move, e.g.
- check all the required providers exists in the target cluster (how to identify required providers TBD)
- other checks e.g. objects already exist in the target cluster
***Non-Goals***
- 1. to support move for DIY management clusters (not created with clusterctl)
- 2. to support move for clusters < v1alpha3 (TO BE CONFIRMED)
**Anything else you would like to add:**
There is a lot of learning from past experiences on pivoting, so I'm pasting below some comments from different threads. Feel free to add more.
/kind feature
| 1.0 | Define clusterctl move process - [2019-11-26 Updated] according to https://github.com/kubernetes-sigs/cluster-api/pull/1730#discussion_r346935816 Pivot is now rebranded into Move; issue updated accordingly
**User Story**
As an operator I would like to move cluster objects and all the associated resources (Machines, Machine Depolyments etc.) from the current management cluster to another management cluster for any reason
**Detailed Description**
The clusterctl CAEP currently in flight assumes the user should brig its own management cluster, so technically the sequence bootstrap cluster -> pivot to -> management cluster is not necessary anymore.
However, the same CAEP consider pivoting a possible answer to different operational needs, e.g because of maintenance or replacement of the management cluster, so pivoting is still in scope.
With v1alpha3 in flight and the new assumptions around clusterctl - one binary for rule all the providers -, the implementation detail should be re-validated, keeping in mind also https://github.com/kubernetes-sigs/cluster-api/pull/1730#discussion_r346935816 discussion that lead to transforming pivot into move.
***Goals***
- 1. To define a moving process that can work on any clusterctl generated management cluster (with any provider or any combinations of providers)
- To define how to move provider-specific objects (or eventually hierarchies of objects)
- 2. To define conventions/requirements the above process should rely on e.g.
- define a convention for forcing cluster-api controllers to not reconcile resources.
- [Auto labeling cluster resources](https://github.com/kubernetes-sigs/cluster-api/issues/1489)
- [Define a convention for namespaced deployments](https://github.com/kubernetes-sigs/cluster-api/issues/1490)
- 3. To define move specializations, like e.g. move only clusters in a given namespace or move only a specific cluster
- 4. To define eventual move side effects (e.g. If forcing cluster-api controllers to not reconcile resources will be implemented scaling down controllers, all the cluster objects stop to reconcile)
- 5. To define preflight checks to be executed before move, e.g.
- check all the required providers exists in the target cluster (how to identify required providers TBD)
- other checks e.g. objects already exist in the target cluster
***Non-Goals***
- 1. to support move for DIY management clusters (not created with clusterctl)
- 2. to support move for clusters < v1alpha3 (TO BE CONFIRMED)
**Anything else you would like to add:**
There is a lot of learning from past experiences on pivoting, so I'm pasting below some comments from different threads. Feel free to add more.
/kind feature
| non_defect | define clusterctl move process according to pivot is now rebranded into move issue updated accordingly user story as an operator i would like to move cluster objects and all the associated resources machines machine depolyments etc from the current management cluster to another management cluster for any reason detailed description the clusterctl caep currently in flight assumes the user should brig its own management cluster so technically the sequence bootstrap cluster pivot to management cluster is not necessary anymore however the same caep consider pivoting a possible answer to different operational needs e g because of maintenance or replacement of the management cluster so pivoting is still in scope with in flight and the new assumptions around clusterctl one binary for rule all the providers the implementation detail should be re validated keeping in mind also discussion that lead to transforming pivot into move goals to define a moving process that can work on any clusterctl generated management cluster with any provider or any combinations of providers to define how to move provider specific objects or eventually hierarchies of objects to define conventions requirements the above process should rely on e g define a convention for forcing cluster api controllers to not reconcile resources to define move specializations like e g move only clusters in a given namespace or move only a specific cluster to define eventual move side effects e g if forcing cluster api controllers to not reconcile resources will be implemented scaling down controllers all the cluster objects stop to reconcile to define preflight checks to be executed before move e g check all the required providers exists in the target cluster how to identify required providers tbd other checks e g objects already exist in the target cluster non goals to support move for diy management clusters not created with clusterctl to support move for clusters to be confirmed anything else you would like to add there is a lot of learning from past experiences on pivoting so i m pasting below some comments from different threads feel free to add more kind feature | 0 |
36,397 | 7,923,515,583 | IssuesEvent | 2018-07-05 14:17:10 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | SerializationServiceBuilder ignores useNativeByteOrder | Team: Core Type: Defect | Reproducer:
```
@Test
public void test_useNativeByteOrder() {
ByteOrder nativeOrder = ByteOrder.nativeOrder();
SerializationServiceBuilder serializationServiceBuilder = new DefaultSerializationServiceBuilder();
InternalSerializationService serializationService = serializationServiceBuilder
.setUseNativeByteOrder(true).build();
assertEquals(nativeOrder, serializationService.getByteOrder());
}
```
Kudos @gokhanoner for catching this. | 1.0 | SerializationServiceBuilder ignores useNativeByteOrder - Reproducer:
```
@Test
public void test_useNativeByteOrder() {
ByteOrder nativeOrder = ByteOrder.nativeOrder();
SerializationServiceBuilder serializationServiceBuilder = new DefaultSerializationServiceBuilder();
InternalSerializationService serializationService = serializationServiceBuilder
.setUseNativeByteOrder(true).build();
assertEquals(nativeOrder, serializationService.getByteOrder());
}
```
Kudos @gokhanoner for catching this. | defect | serializationservicebuilder ignores usenativebyteorder reproducer test public void test usenativebyteorder byteorder nativeorder byteorder nativeorder serializationservicebuilder serializationservicebuilder new defaultserializationservicebuilder internalserializationservice serializationservice serializationservicebuilder setusenativebyteorder true build assertequals nativeorder serializationservice getbyteorder kudos gokhanoner for catching this | 1 |
18,075 | 3,022,982,088 | IssuesEvent | 2015-08-01 01:08:33 | GiraffaFS/giraffa | https://api.github.com/repos/GiraffaFS/giraffa | closed | Unnecessary Assertion in TestLeaseRecovery | bug Priority-High Type-Defect | I made a mistake in the fix for #188. On line 173 of TestLeaseManagement, INode.getINode may return null, which is supposed to be ok (it will retry 100 times). But I call INodeFile.valueOf on it, which will cause a NullPointerException unless the test succeeds on the first attempt. @weilintsaiWand, is this related to any of the race condition test failures you are seeing? I don't think it would be since we only committed this yesterday. | 1.0 | Unnecessary Assertion in TestLeaseRecovery - I made a mistake in the fix for #188. On line 173 of TestLeaseManagement, INode.getINode may return null, which is supposed to be ok (it will retry 100 times). But I call INodeFile.valueOf on it, which will cause a NullPointerException unless the test succeeds on the first attempt. @weilintsaiWand, is this related to any of the race condition test failures you are seeing? I don't think it would be since we only committed this yesterday. | defect | unnecessary assertion in testleaserecovery i made a mistake in the fix for on line of testleasemanagement inode getinode may return null which is supposed to be ok it will retry times but i call inodefile valueof on it which will cause a nullpointerexception unless the test succeeds on the first attempt weilintsaiwand is this related to any of the race condition test failures you are seeing i don t think it would be since we only committed this yesterday | 1 |
340,060 | 10,266,205,041 | IssuesEvent | 2019-08-22 20:49:01 | kubernetes/release | https://api.github.com/repos/kubernetes/release | closed | missing images for the ci/latest.txt label | area/release-eng kind/bug priority/critical-urgent | currently, it appears that the ci/latest.txt label points to images that are missing:
```
curl -L https://dl.k8s.io/ci/latest.txt
v1.17.0-alpha.0.400+b8e8130a9195d5
```
missing `400` tagged image for a control-plane component:
https://console.cloud.google.com/gcr/images/kubernetes-ci-images/GLOBAL/kube-apiserver?gcrImageListsize=30
at the time of writing this report the latest one is at `398`

ci-cross on the other hand returns `398`:
```
curl -L https://dl.k8s.io/ci-cross/latest.txt
v1.17.0-alpha.0.398+bfb69dbb8f0105
```
slack conversation:
https://kubernetes.slack.com/archives/C2C40FMNF/p1566398444348500
/kind bug
/priority critical-urgent
| 1.0 | missing images for the ci/latest.txt label - currently, it appears that the ci/latest.txt label points to images that are missing:
```
curl -L https://dl.k8s.io/ci/latest.txt
v1.17.0-alpha.0.400+b8e8130a9195d5
```
missing `400` tagged image for a control-plane component:
https://console.cloud.google.com/gcr/images/kubernetes-ci-images/GLOBAL/kube-apiserver?gcrImageListsize=30
at the time of writing this report the latest one is at `398`

ci-cross on the other hand returns `398`:
```
curl -L https://dl.k8s.io/ci-cross/latest.txt
v1.17.0-alpha.0.398+bfb69dbb8f0105
```
slack conversation:
https://kubernetes.slack.com/archives/C2C40FMNF/p1566398444348500
/kind bug
/priority critical-urgent
| non_defect | missing images for the ci latest txt label currently it appears that the ci latest txt label points to images that are missing curl l alpha missing tagged image for a control plane component at the time of writing this report the latest one is at ci cross on the other hand returns curl l alpha slack conversation kind bug priority critical urgent | 0 |
27,327 | 4,965,459,422 | IssuesEvent | 2016-12-04 09:42:48 | otros-systems/otroslogviewer | https://api.github.com/repos/otros-systems/otroslogviewer | closed | Save/Load log investigation is loosing information about markings | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Open log, add markings, notes
2. Save log investigation
3. Load log investigation
4. Events are not marked
What is the expected output? What do you see instead?
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `otros.sy...@gmail.com` on 18 Oct 2011 at 9:36
| 1.0 | Save/Load log investigation is loosing information about markings - ```
What steps will reproduce the problem?
1. Open log, add markings, notes
2. Save log investigation
3. Load log investigation
4. Events are not marked
What is the expected output? What do you see instead?
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `otros.sy...@gmail.com` on 18 Oct 2011 at 9:36
| defect | save load log investigation is loosing information about markings what steps will reproduce the problem open log add markings notes save log investigation load log investigation events are not marked what is the expected output what do you see instead please use labels and text to provide additional information original issue reported on code google com by otros sy gmail com on oct at | 1 |
344,317 | 30,734,218,904 | IssuesEvent | 2023-07-28 06:06:40 | SKT-AI-Research/backend | https://api.github.com/repos/SKT-AI-Research/backend | opened | Feat : 비밀번호 변경 | Back Test DB | ## 💡 기대 결과
- 사용자가 자신의 비밀번호를 변경할 수 있다.
## 🚨 주의 및 전달 사항
- 현재 비밀번호와 새 비밀번호, 새 비밀번호 확인이 필요하다.
## ✅ 작업 내용 겸 체크리스트
- [ ] UserController
- [ ] UserService
- [ ] UserRepository
- [ ] Test
| 1.0 | Feat : 비밀번호 변경 - ## 💡 기대 결과
- 사용자가 자신의 비밀번호를 변경할 수 있다.
## 🚨 주의 및 전달 사항
- 현재 비밀번호와 새 비밀번호, 새 비밀번호 확인이 필요하다.
## ✅ 작업 내용 겸 체크리스트
- [ ] UserController
- [ ] UserService
- [ ] UserRepository
- [ ] Test
| non_defect | feat 비밀번호 변경 💡 기대 결과 사용자가 자신의 비밀번호를 변경할 수 있다 🚨 주의 및 전달 사항 현재 비밀번호와 새 비밀번호 새 비밀번호 확인이 필요하다 ✅ 작업 내용 겸 체크리스트 usercontroller userservice userrepository test | 0 |
29,672 | 5,809,419,552 | IssuesEvent | 2017-05-04 13:23:05 | idaholab/moose | https://api.github.com/repos/idaholab/moose | closed | Saved State on Exodus viewer doesn't appear to work | C: Peacock P: normal T: defect | ### Description of the enhancement or error report
If you switch between different open files, the state of the check boxes and other controls on the Exodus tab doesn't reflect what's actually shown. It appears that state is saved when you switch between variables, just not files.
### Rationale for the enhancement or information for reproducing the error
Try opening a second file and checking various boxes. Then switch back and forth between the two output files.
### Identified impact
(i.e. Internal object changes, limited interface changes, public API change, or a list of specific applications impacted)
Peacock robustness | 1.0 | Saved State on Exodus viewer doesn't appear to work - ### Description of the enhancement or error report
If you switch between different open files, the state of the check boxes and other controls on the Exodus tab doesn't reflect what's actually shown. It appears that state is saved when you switch between variables, just not files.
### Rationale for the enhancement or information for reproducing the error
Try opening a second file and checking various boxes. Then switch back and forth between the two output files.
### Identified impact
(i.e. Internal object changes, limited interface changes, public API change, or a list of specific applications impacted)
Peacock robustness | defect | saved state on exodus viewer doesn t appear to work description of the enhancement or error report if you switch between different open files the state of the check boxes and other controls on the exodus tab doesn t reflect what s actually shown it appears that state is saved when you switch between variables just not files rationale for the enhancement or information for reproducing the error try opening a second file and checking various boxes then switch back and forth between the two output files identified impact i e internal object changes limited interface changes public api change or a list of specific applications impacted peacock robustness | 1 |
39,010 | 9,126,033,592 | IssuesEvent | 2019-02-24 18:27:25 | prettydiff/prettydiff | https://api.github.com/repos/prettydiff/prettydiff | closed | The "version" attribute must be the first attribute in the XML pragma | Defect | If version is not the first attribute Firefox will fail to render the document.
bad `<?xml encoding="UTF-8" version="1.0"?>`
good `<?xml version="1.0" encoding="UTF-8"?>` | 1.0 | The "version" attribute must be the first attribute in the XML pragma - If version is not the first attribute Firefox will fail to render the document.
bad `<?xml encoding="UTF-8" version="1.0"?>`
good `<?xml version="1.0" encoding="UTF-8"?>` | defect | the version attribute must be the first attribute in the xml pragma if version is not the first attribute firefox will fail to render the document bad good | 1 |
66,197 | 20,035,194,392 | IssuesEvent | 2022-02-02 11:06:44 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | fails to build on kernel 5.16 | Type: Defect | CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx2.o
CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx512f.o
CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx512bw.o
/var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.c: In function 'zvol_os_create_minor':
/var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.c:1075:3: error: ignoring return value of 'add_disk', declared with attribute warn_unused_result [-Werror=unused-result]
add_disk(zv->zv_zso->zvo_disk);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
cc1: all warnings being treated as errors
make[5]: *** [scripts/Makefile.build:287: /var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.o] Error 1
make[5]: *** Waiting for unfinished jobs....
LD [M] /var/lib/dkms/zfs/2.0.7/build/module/zstd/zzstd.o
make[4]: *** [scripts/Makefile.build:549: /var/lib/dkms/zfs/2.0.7/build/module/zfs] Error 2
make[3]: *** [Makefile:1846: /var/lib/dkms/zfs/2.0.7/build/module] Error 2
make[3]: Leaving directory '/usr/src/kernels/5.16.5-1.el8.elrepo.x86_64'
make[2]: *** [Makefile:49: modules-Linux] Error 2
make[2]: Leaving directory '/var/lib/dkms/zfs/2.0.7/build/module'
make[1]: *** [Makefile:898: all-recursive] Error 1
make[1]: Leaving directory '/var/lib/dkms/zfs/2.0.7/build'
make: *** [Makefile:759: all] Error 2 | 1.0 | fails to build on kernel 5.16 - CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx2.o
CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx512f.o
CC [M] /var/lib/dkms/zfs/2.0.7/build/module/zfs/vdev_raidz_math_avx512bw.o
/var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.c: In function 'zvol_os_create_minor':
/var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.c:1075:3: error: ignoring return value of 'add_disk', declared with attribute warn_unused_result [-Werror=unused-result]
add_disk(zv->zv_zso->zvo_disk);
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
cc1: all warnings being treated as errors
make[5]: *** [scripts/Makefile.build:287: /var/lib/dkms/zfs/2.0.7/build/module/zfs/../os/linux/zfs/zvol_os.o] Error 1
make[5]: *** Waiting for unfinished jobs....
LD [M] /var/lib/dkms/zfs/2.0.7/build/module/zstd/zzstd.o
make[4]: *** [scripts/Makefile.build:549: /var/lib/dkms/zfs/2.0.7/build/module/zfs] Error 2
make[3]: *** [Makefile:1846: /var/lib/dkms/zfs/2.0.7/build/module] Error 2
make[3]: Leaving directory '/usr/src/kernels/5.16.5-1.el8.elrepo.x86_64'
make[2]: *** [Makefile:49: modules-Linux] Error 2
make[2]: Leaving directory '/var/lib/dkms/zfs/2.0.7/build/module'
make[1]: *** [Makefile:898: all-recursive] Error 1
make[1]: Leaving directory '/var/lib/dkms/zfs/2.0.7/build'
make: *** [Makefile:759: all] Error 2 | defect | fails to build on kernel cc var lib dkms zfs build module zfs vdev raidz math o cc var lib dkms zfs build module zfs vdev raidz math o cc var lib dkms zfs build module zfs vdev raidz math o var lib dkms zfs build module zfs os linux zfs zvol os c in function zvol os create minor var lib dkms zfs build module zfs os linux zfs zvol os c error ignoring return value of add disk declared with attribute warn unused result add disk zv zv zso zvo disk all warnings being treated as errors make error make waiting for unfinished jobs ld var lib dkms zfs build module zstd zzstd o make error make error make leaving directory usr src kernels elrepo make error make leaving directory var lib dkms zfs build module make error make leaving directory var lib dkms zfs build make error | 1 |
62,396 | 17,023,914,234 | IssuesEvent | 2021-07-03 04:31:52 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Nominatim down over HTTPS | Component: nominatim Priority: major Resolution: wontfix Type: defect | **[Submitted to the original trac issue database at 7.17am, Sunday, 30th November 2014]**
OsmAnd+'s Nominatim integration has stopped working, with an error of:
"failed to connect to nominatim.openstreetmap.org/188.40.32.215 (port 443)".
I don't think anything changed in OsmAnd+ regarding this. Also, https://nominatim.openstreetmap.org/ does not work, but http://nominatim.openstreetmap.org/ does.
https://wiki.openstreetmap.org/wiki/Nominatim links to https://nominatim.openstreetmap.org/ (it's a protocol-relative link, so it uses the protocol you visit the wiki with), so if HTTPS is not intended to work (anymore), the documentation should be fixed. | 1.0 | Nominatim down over HTTPS - **[Submitted to the original trac issue database at 7.17am, Sunday, 30th November 2014]**
OsmAnd+'s Nominatim integration has stopped working, with an error of:
"failed to connect to nominatim.openstreetmap.org/188.40.32.215 (port 443)".
I don't think anything changed in OsmAnd+ regarding this. Also, https://nominatim.openstreetmap.org/ does not work, but http://nominatim.openstreetmap.org/ does.
https://wiki.openstreetmap.org/wiki/Nominatim links to https://nominatim.openstreetmap.org/ (it's a protocol-relative link, so it uses the protocol you visit the wiki with), so if HTTPS is not intended to work (anymore), the documentation should be fixed. | defect | nominatim down over https osmand s nominatim integration has stopped working with an error of failed to connect to nominatim openstreetmap org port i don t think anything changed in osmand regarding this also does not work but does links to it s a protocol relative link so it uses the protocol you visit the wiki with so if https is not intended to work anymore the documentation should be fixed | 1 |
55,857 | 14,711,439,878 | IssuesEvent | 2021-01-05 07:25:39 | mozilla-lockwise/lockwise-android | https://api.github.com/repos/mozilla-lockwise/lockwise-android | opened | Last update 4.0.3 not working on Android 9 anymore | defect | ## Steps to reproduce
Open application
Choose any stored item
When returning to the list app crashes
### Expected behavior
Correct return
### Actual behavior
App crashes
### Device & build information
* Device: Sony XZ1 compact G8441
* Build version: 47.2.A.11.228
### Notes
Attachments:
| 1.0 | Last update 4.0.3 not working on Android 9 anymore - ## Steps to reproduce
Open application
Choose any stored item
When returning to the list app crashes
### Expected behavior
Correct return
### Actual behavior
App crashes
### Device & build information
* Device: Sony XZ1 compact G8441
* Build version: 47.2.A.11.228
### Notes
Attachments:
| defect | last update not working on android anymore steps to reproduce open application choose any stored item when returning to the list app crashes expected behavior correct return actual behavior app crashes device build information device sony compact build version a notes attachments | 1 |
47,415 | 24,991,992,232 | IssuesEvent | 2022-11-02 19:39:11 | atomashpolskiy/bt | https://api.github.com/repos/atomashpolskiy/bt | closed | bt.torrent.messaging.ExchangedMetadata.getSha1Digest should not allocate larger buffer than necessary | enhancement performance good first issue | Instead of always allocating a fixed size buffer (1,000,000 bytes currently), it should leverage the upfront knowledge of total size of the metadata and allocate `Math.min(DEFAULT_BUFFER_LEN, metadata.length())`. | True | bt.torrent.messaging.ExchangedMetadata.getSha1Digest should not allocate larger buffer than necessary - Instead of always allocating a fixed size buffer (1,000,000 bytes currently), it should leverage the upfront knowledge of total size of the metadata and allocate `Math.min(DEFAULT_BUFFER_LEN, metadata.length())`. | non_defect | bt torrent messaging exchangedmetadata should not allocate larger buffer than necessary instead of always allocating a fixed size buffer bytes currently it should leverage the upfront knowledge of total size of the metadata and allocate math min default buffer len metadata length | 0 |
309,148 | 23,285,809,256 | IssuesEvent | 2022-08-05 16:21:41 | flutter-form-builder-ecosystem/flutter_form_builder | https://api.github.com/repos/flutter-form-builder-ecosystem/flutter_form_builder | closed | 7.4.0 has breaking change | documentation | ### Describe the change you would like to see
The "Changelog" on pub.dev for version 7.4.0 should indicate that there is a breaking change: it now requires Flutter SDK v3.0 or greater.
I think that the "Versions" info on pub.dev for version 7.4.0 should also indicate minimum Dart SDK v2.17
### How would the suggested change make the documentation more useful?
It is hard to know which versions of the package to install without this info on the underlying support requirements. I had to do trial-and-error to go back to a version that actually worked with my environment (Flutter 2.10.5) | 1.0 | 7.4.0 has breaking change - ### Describe the change you would like to see
The "Changelog" on pub.dev for version 7.4.0 should indicate that there is a breaking change: it now requires Flutter SDK v3.0 or greater.
I think that the "Versions" info on pub.dev for version 7.4.0 should also indicate minimum Dart SDK v2.17
### How would the suggested change make the documentation more useful?
It is hard to know which versions of the package to install without this info on the underlying support requirements. I had to do trial-and-error to go back to a version that actually worked with my environment (Flutter 2.10.5) | non_defect | has breaking change describe the change you would like to see the changelog on pub dev for version should indicate that there is a breaking change it now requires flutter sdk or greater i think that the versions info on pub dev for version should also indicate minimum dart sdk how would the suggested change make the documentation more useful it is hard to know which versions of the package to install without this info on the underlying support requirements i had to do trial and error to go back to a version that actually worked with my environment flutter | 0 |
12,803 | 2,717,707,710 | IssuesEvent | 2015-04-11 15:18:22 | codenameone/CodenameOne | https://api.github.com/repos/codenameone/CodenameOne | closed | RFE: Platform-version based theme overrides | Priority-Medium Type-Defect | Original [issue 226](https://code.google.com/p/codenameone/issues/detail?id=226) created by codenameone on 2012-06-21T18:48:49.000Z:
<b>What is the expected output? What do you see instead?</b>
Since the theme will change colors based on the device version it's installed on (example androidTheme.res on Gingerbread, android_holo_light.res on ICS), should "Override In Platform" include platform versions? I'm finding my custom theme looks good on ICS, but on Gingerbread, I have a mixture of blue and orange, and not sure how to properly handle that.
| 1.0 | RFE: Platform-version based theme overrides - Original [issue 226](https://code.google.com/p/codenameone/issues/detail?id=226) created by codenameone on 2012-06-21T18:48:49.000Z:
<b>What is the expected output? What do you see instead?</b>
Since the theme will change colors based on the device version it's installed on (example androidTheme.res on Gingerbread, android_holo_light.res on ICS), should "Override In Platform" include platform versions? I'm finding my custom theme looks good on ICS, but on Gingerbread, I have a mixture of blue and orange, and not sure how to properly handle that.
| defect | rfe platform version based theme overrides original created by codenameone on what is the expected output what do you see instead since the theme will change colors based on the device version it s installed on example androidtheme res on gingerbread android holo light res on ics should quot override in platform quot include platform versions i m finding my custom theme looks good on ics but on gingerbread i have a mixture of blue and orange and not sure how to properly handle that | 1 |
450,975 | 32,001,465,626 | IssuesEvent | 2023-09-21 12:33:05 | frostsg/inf2001-p11-2 | https://api.github.com/repos/frostsg/inf2001-p11-2 | closed | 1.7.3: Estimation of Effort and Project Duration | Documentation (Report) Calculation | Goal: Estimate the effort needed and the duration of the project
Success Criteria: The estimation should be done and it has to be realistic
Start Date: 19/09/23
End Date: 20/09/23
Owner: Soon Leung Isaac
Status: Done | 1.0 | 1.7.3: Estimation of Effort and Project Duration - Goal: Estimate the effort needed and the duration of the project
Success Criteria: The estimation should be done and it has to be realistic
Start Date: 19/09/23
End Date: 20/09/23
Owner: Soon Leung Isaac
Status: Done | non_defect | estimation of effort and project duration goal estimate the effort needed and the duration of the project success criteria the estimation should be done and it has to be realistic start date end date owner soon leung isaac status done | 0 |
503,051 | 14,578,367,171 | IssuesEvent | 2020-12-18 04:38:16 | numbersprotocol/capture-lite | https://api.github.com/repos/numbersprotocol/capture-lite | opened | 接收 postcapture 後再進行capture 拍照,速度變慢 | bug priority:critical | ## Description
1s機型 全新安裝跟全新帳號時 產生 capture 8~12s 但接收完 兩張 postcapture 後接著產生 11~15張 capture 21~25s
這個狀況比 note20 嚴重很多~~ 是這21~25s capture 頁面完全沒有新的東西出現是真的會以為當掉了
note20 機型
1. 拍攝到 50 張皆沒有明顯變慢顯示的情況 ( 拍照到顯示約在 6~8 s)
2. 但當 postcapture 接收到 超過 3 張後再進行拍照時 很明顯的拍照到顯示約在 15~20 s ( 拍攝張數在 30 張 ) 測試兩次
3. 但當 postcapture 接收到 10 張後再進行拍照時 很明顯的拍照到顯示約在 23~25 s ( 拍攝張數在 30 張 ) 測試兩次
4.capture 拍攝第一張 6s 接收 3 張 postcapture 後再進行拍照 5 張 11~13s ( 全新安裝跟全新帳號時 )
5. capture 拍攝第一張 7s 接收 5 張 postcapture 後再進行拍照 5 張 15 ~18s ( 全新安裝跟全新帳號時 )
6. capture 拍攝第一張 6s 接收 10 張 postcapture 後再進行拍照 5 張 18 ~21s ( 全新安裝跟全新帳號時 )
## Environment
- Version: 0.11.8
- OS: Android 8.1.0/Android 10
- Device: HTC Exodus 1s, note 20
| 1.0 | 接收 postcapture 後再進行capture 拍照,速度變慢 - ## Description
1s機型 全新安裝跟全新帳號時 產生 capture 8~12s 但接收完 兩張 postcapture 後接著產生 11~15張 capture 21~25s
這個狀況比 note20 嚴重很多~~ 是這21~25s capture 頁面完全沒有新的東西出現是真的會以為當掉了
note20 機型
1. 拍攝到 50 張皆沒有明顯變慢顯示的情況 ( 拍照到顯示約在 6~8 s)
2. 但當 postcapture 接收到 超過 3 張後再進行拍照時 很明顯的拍照到顯示約在 15~20 s ( 拍攝張數在 30 張 ) 測試兩次
3. 但當 postcapture 接收到 10 張後再進行拍照時 很明顯的拍照到顯示約在 23~25 s ( 拍攝張數在 30 張 ) 測試兩次
4.capture 拍攝第一張 6s 接收 3 張 postcapture 後再進行拍照 5 張 11~13s ( 全新安裝跟全新帳號時 )
5. capture 拍攝第一張 7s 接收 5 張 postcapture 後再進行拍照 5 張 15 ~18s ( 全新安裝跟全新帳號時 )
6. capture 拍攝第一張 6s 接收 10 張 postcapture 後再進行拍照 5 張 18 ~21s ( 全新安裝跟全新帳號時 )
## Environment
- Version: 0.11.8
- OS: Android 8.1.0/Android 10
- Device: HTC Exodus 1s, note 20
| non_defect | 接收 postcapture 後再進行capture 拍照,速度變慢 description 全新安裝跟全新帳號時 產生 capture 但接收完 兩張 postcapture 後接著產生 capture 這個狀況比 嚴重很多 capture 頁面完全沒有新的東西出現是真的會以為當掉了 機型 拍攝到 張皆沒有明顯變慢顯示的情況 拍照到顯示約在 s 但當 postcapture 接收到 超過 張後再進行拍照時 很明顯的拍照到顯示約在 s 拍攝張數在 張 測試兩次 但當 postcapture 接收到 張後再進行拍照時 很明顯的拍照到顯示約在 s 拍攝張數在 張 測試兩次 capture 拍攝第一張 接收 張 postcapture 後再進行拍照 張 全新安裝跟全新帳號時 capture 拍攝第一張 接收 張 postcapture 後再進行拍照 張 全新安裝跟全新帳號時 capture 拍攝第一張 接收 張 postcapture 後再進行拍照 張 全新安裝跟全新帳號時 environment version os android android device htc exodus note | 0 |
27,740 | 5,093,103,198 | IssuesEvent | 2017-01-03 02:48:01 | jquery/esprima | https://api.github.com/repos/jquery/esprima | opened | Generator parameters must not contain yield expressions | defect es6 | (As reported by @michaelficarra as case 17 in #1502)
Test case:
```js
esprima.parse('function *g(x = yield){}');
```
**Actual**: a valid syntax tree.
**Expected**: a thrown exception indicating a syntax error.
<hr>
Following [V8 error messages](https://chromium.googlesource.com/v8/v8/+/master/src/messages.h), the exception message should be _"Yield expression not allowed in formal parameter"_. | 1.0 | Generator parameters must not contain yield expressions - (As reported by @michaelficarra as case 17 in #1502)
Test case:
```js
esprima.parse('function *g(x = yield){}');
```
**Actual**: a valid syntax tree.
**Expected**: a thrown exception indicating a syntax error.
<hr>
Following [V8 error messages](https://chromium.googlesource.com/v8/v8/+/master/src/messages.h), the exception message should be _"Yield expression not allowed in formal parameter"_. | defect | generator parameters must not contain yield expressions as reported by michaelficarra as case in test case js esprima parse function g x yield actual a valid syntax tree expected a thrown exception indicating a syntax error following the exception message should be yield expression not allowed in formal parameter | 1 |
55,256 | 14,345,147,049 | IssuesEvent | 2020-11-28 17:39:54 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | New composer: Tiny text below compose is illegable and dissappears when you click it. | bug defect type:composer ui/ux | <!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->
### Description
### Steps to reproduce
When I mouse over and enable text entry in the compose, there seems to be a "Code" button that is highlighted - if it's like a lot of things, I can click to select/deselect to enable or disable code mode, or something.

I moved my mouse there and clicked and it went away. Success! I thought, I've done something so i won't be in Code mode.
I then started typing and it came back.
It appears in an area of the screen that is normally otherwise empty, and it comes back every time you type, maybe just show it all the time and save cycles making it show and hide?
Also it's unfeasably small - and there's no way to resize /just/ that part of the UI, nor can you select the text to copy it into another window (see above about how it goes away when you click it) so it's not very accessible to those with normal or worse than normal vision.
### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: web
For the web app:
- **Browser**: Firefox 68
- **OS**: linux
- **URL**: riot.im/develop
| 1.0 | New composer: Tiny text below compose is illegable and dissappears when you click it. - <!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P -->
### Description
### Steps to reproduce
When I mouse over and enable text entry in the compose, there seems to be a "Code" button that is highlighted - if it's like a lot of things, I can click to select/deselect to enable or disable code mode, or something.

I moved my mouse there and clicked and it went away. Success! I thought, I've done something so i won't be in Code mode.
I then started typing and it came back.
It appears in an area of the screen that is normally otherwise empty, and it comes back every time you type, maybe just show it all the time and save cycles making it show and hide?
Also it's unfeasably small - and there's no way to resize /just/ that part of the UI, nor can you select the text to copy it into another window (see above about how it goes away when you click it) so it's not very accessible to those with normal or worse than normal vision.
### Version information
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: web
For the web app:
- **Browser**: Firefox 68
- **OS**: linux
- **URL**: riot.im/develop
| defect | new composer tiny text below compose is illegable and dissappears when you click it description steps to reproduce when i mouse over and enable text entry in the compose there seems to be a code button that is highlighted if it s like a lot of things i can click to select deselect to enable or disable code mode or something i moved my mouse there and clicked and it went away success i thought i ve done something so i won t be in code mode i then started typing and it came back it appears in an area of the screen that is normally otherwise empty and it comes back every time you type maybe just show it all the time and save cycles making it show and hide also it s unfeasably small and there s no way to resize just that part of the ui nor can you select the text to copy it into another window see above about how it goes away when you click it so it s not very accessible to those with normal or worse than normal vision version information platform web for the web app browser firefox os linux url riot im develop | 1 |
65,114 | 19,099,265,291 | IssuesEvent | 2021-11-29 20:20:10 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Emoji picker has jumpy scroll jitter (x) | T-Defect S-Minor A-Emoji O-Uncommon | <!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
When filtering for the `:x:` ❌ emoji and scrolling to the bottom, the picker will jitter a couple pixels up and down continuously whenever you're touching the bottom.
### Steps to reproduce
1. Filter for `x` in the emoji picker
1. Scroll down to the bottom
1. Notice the jumpy scroll jitter whenever you are at the bottom

Log: not sent <!-- You can send us the app's logs via the 'Report bug'
link on the 'Settings' page. Very important for hard-to-reproduce bugs. Please
file a bug here too! -->
### Version information
- Browser: Chrome 87.0.4280.88
- OS: Windows 10
- URL: https://app.element.io/
I can't reproduce on macOS.
| 1.0 | Emoji picker has jumpy scroll jitter (x) - <!-- Please report security issues by email to security@matrix.org -->
<!-- This is a bug report template. By following the instructions below and
filling out the sections with your information, you will help the us to get all
the necessary data to fix your issue.
You can also preview your report before submitting it. You may remove sections
that aren't relevant to your particular case.
Text between <!-- and --> marks will be invisible in the report.
-->
### Description
When filtering for the `:x:` ❌ emoji and scrolling to the bottom, the picker will jitter a couple pixels up and down continuously whenever you're touching the bottom.
### Steps to reproduce
1. Filter for `x` in the emoji picker
1. Scroll down to the bottom
1. Notice the jumpy scroll jitter whenever you are at the bottom

Log: not sent <!-- You can send us the app's logs via the 'Report bug'
link on the 'Settings' page. Very important for hard-to-reproduce bugs. Please
file a bug here too! -->
### Version information
- Browser: Chrome 87.0.4280.88
- OS: Windows 10
- URL: https://app.element.io/
I can't reproduce on macOS.
| defect | emoji picker has jumpy scroll jitter x this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report description when filtering for the x ❌ emoji and scrolling to the bottom the picker will jitter a couple pixels up and down continuously whenever you re touching the bottom steps to reproduce filter for x in the emoji picker scroll down to the bottom notice the jumpy scroll jitter whenever you are at the bottom log not sent you can send us the app s logs via the report bug link on the settings page very important for hard to reproduce bugs please file a bug here too version information browser chrome os windows url i can t reproduce on macos | 1 |
18,880 | 3,090,946,461 | IssuesEvent | 2015-08-26 10:03:39 | gbif/ipt | https://api.github.com/repos/gbif/ipt | closed | Allow IPT to be installed without an organization | bug Component-i18n Milestone-Release2.3 Priority-Critical Translation Type-Defect Usability | The IPT should not need an organization to be installed.
Several people have come forward complaining about this, and have been using it to generate archives for exchange within networks but without any intention or need to publish to GBIF. Most recently the iDigBio folks have raised it as confusing and a regression.
The option to use the test mode is not ideal, as it both puts "TEST MODE" on the site which is a nuisance and will not use production level standards and vocabularies.
| 1.0 | Allow IPT to be installed without an organization - The IPT should not need an organization to be installed.
Several people have come forward complaining about this, and have been using it to generate archives for exchange within networks but without any intention or need to publish to GBIF. Most recently the iDigBio folks have raised it as confusing and a regression.
The option to use the test mode is not ideal, as it both puts "TEST MODE" on the site which is a nuisance and will not use production level standards and vocabularies.
| defect | allow ipt to be installed without an organization the ipt should not need an organization to be installed several people have come forward complaining about this and have been using it to generate archives for exchange within networks but without any intention or need to publish to gbif most recently the idigbio folks have raised it as confusing and a regression the option to use the test mode is not ideal as it both puts test mode on the site which is a nuisance and will not use production level standards and vocabularies | 1 |
29,559 | 5,720,790,716 | IssuesEvent | 2017-04-20 03:26:46 | pigfoot/formosa | https://api.github.com/repos/pigfoot/formosa | closed | [bug]百年蟲 | Defect Usability | src/cursor.c :93,94
sprintf(chdate, "%02d.%02d.%02d",tm->tm_year - 11, tm->tm_mon + 1, tm->tm_mday);
會導致無法進入看板
| 1.0 | [bug]百年蟲 - src/cursor.c :93,94
sprintf(chdate, "%02d.%02d.%02d",tm->tm_year - 11, tm->tm_mon + 1, tm->tm_mday);
會導致無法進入看板
| defect | 百年蟲 src cursor c sprintf chdate tm tm year tm tm mon tm tm mday 會導致無法進入看板 | 1 |
127,737 | 18,022,119,967 | IssuesEvent | 2021-09-16 20:55:11 | xinyuanwss/gradle-ghe | https://api.github.com/repos/xinyuanwss/gradle-ghe | opened | WS-2020-0408 (High) detected in netty-handler-4.1.39.Final.jar | security vulnerability | ## WS-2020-0408 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-handler-4.1.39.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p>
<p>Path to dependency file: gradle-ghe/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/io.netty/netty-handler/4.1.39.Final/4a63b56de071c1b10a56b5d90095e4201ea4098f/netty-handler-4.1.39.Final.jar</p>
<p>
Dependency Hierarchy:
- netty-codec-http-4.1.39.Final.jar (Root Library)
- :x: **netty-handler-4.1.39.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was found in all versions of io.netty:netty-all. Host verification in Netty is disabled by default. This can lead to MITM attack in which an attacker can forge valid SSL/TLS certificates for a different hostname in order to intercept traffic that doesn’t intend for him. This is an issue because the certificate is not matched with the host.
<p>Publish Date: 2020-06-22
<p>URL: <a href=https://github.com/netty/netty/issues/10362>WS-2020-0408</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-handler","packageVersion":"4.1.39.Final","packageFilePaths":["/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.netty:netty-codec-http:4.1.39.Final;io.netty:netty-handler:4.1.39.Final","isMinimumFixVersionAvailable":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2020-0408","vulnerabilityDetails":"An issue was found in all versions of io.netty:netty-all. Host verification in Netty is disabled by default. This can lead to MITM attack in which an attacker can forge valid SSL/TLS certificates for a different hostname in order to intercept traffic that doesn’t intend for him. This is an issue because the certificate is not matched with the host.","vulnerabilityUrl":"https://github.com/netty/netty/issues/10362","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | WS-2020-0408 (High) detected in netty-handler-4.1.39.Final.jar - ## WS-2020-0408 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-handler-4.1.39.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p>
<p>Path to dependency file: gradle-ghe/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/io.netty/netty-handler/4.1.39.Final/4a63b56de071c1b10a56b5d90095e4201ea4098f/netty-handler-4.1.39.Final.jar</p>
<p>
Dependency Hierarchy:
- netty-codec-http-4.1.39.Final.jar (Root Library)
- :x: **netty-handler-4.1.39.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was found in all versions of io.netty:netty-all. Host verification in Netty is disabled by default. This can lead to MITM attack in which an attacker can forge valid SSL/TLS certificates for a different hostname in order to intercept traffic that doesn’t intend for him. This is an issue because the certificate is not matched with the host.
<p>Publish Date: 2020-06-22
<p>URL: <a href=https://github.com/netty/netty/issues/10362>WS-2020-0408</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-handler","packageVersion":"4.1.39.Final","packageFilePaths":["/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.netty:netty-codec-http:4.1.39.Final;io.netty:netty-handler:4.1.39.Final","isMinimumFixVersionAvailable":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2020-0408","vulnerabilityDetails":"An issue was found in all versions of io.netty:netty-all. Host verification in Netty is disabled by default. This can lead to MITM attack in which an attacker can forge valid SSL/TLS certificates for a different hostname in order to intercept traffic that doesn’t intend for him. This is an issue because the certificate is not matched with the host.","vulnerabilityUrl":"https://github.com/netty/netty/issues/10362","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_defect | ws high detected in netty handler final jar ws high severity vulnerability vulnerable library netty handler final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file gradle ghe build gradle path to vulnerable library home wss scanner gradle caches modules files io netty netty handler final netty handler final jar dependency hierarchy netty codec http final jar root library x netty handler final jar vulnerable library found in base branch main vulnerability details an issue was found in all versions of io netty netty all host verification in netty is disabled by default this can lead to mitm attack in which an attacker can forge valid ssl tls certificates for a different hostname in order to intercept traffic that doesn’t intend for him this is an issue because the certificate is not matched with the host publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree io netty netty codec http final io netty netty handler final isminimumfixversionavailable false basebranches vulnerabilityidentifier ws vulnerabilitydetails an issue was found in all versions of io netty netty all host verification in netty is disabled by default this can lead to mitm attack in which an attacker can forge valid ssl tls certificates for a different hostname in order to intercept traffic that doesn’t intend for him this is an issue because the certificate is not matched with the host vulnerabilityurl | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.