Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 1 853 | labels stringlengths 4 898 | body stringlengths 2 262k | index stringclasses 13 values | text_combine stringlengths 96 262k | label stringclasses 2 values | text stringlengths 96 250k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
179,918 | 21,606,569,830 | IssuesEvent | 2022-05-04 04:25:43 | testdemo1227/Demo-WebGoat1 | https://api.github.com/repos/testdemo1227/Demo-WebGoat1 | opened | jstl-1.2.jar: 1 vulnerabilities (highest severity is: 7.3) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jstl-1.2.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /ository/javax/servlet/jstl/1.2/jstl-1.2.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2015-0254](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | jstl-1.2.jar | Direct | org.apache.taglibs:taglibs-standard-impl:1.2.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-0254</summary>
### Vulnerable Library - <b>jstl-1.2.jar</b></p>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /ository/javax/servlet/jstl/1.2/jstl-1.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jstl-1.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache Standard Taglibs before 1.2.3 allows remote attackers to execute arbitrary code or conduct external XML entity (XXE) attacks via a crafted XSLT extension in a (1) <x:parse> or (2) <x:transform> JSTL XML tag.
<p>Publish Date: 2015-03-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254>CVE-2015-0254</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tomcat.apache.org/taglibs/standard/">https://tomcat.apache.org/taglibs/standard/</a></p>
<p>Release Date: 2015-03-09</p>
<p>Fix Resolution: org.apache.taglibs:taglibs-standard-impl:1.2.3</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"javax.servlet","packageName":"jstl","packageVersion":"1.2","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"javax.servlet:jstl:1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.taglibs:taglibs-standard-impl:1.2.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-0254","vulnerabilityDetails":"Apache Standard Taglibs before 1.2.3 allows remote attackers to execute arbitrary code or conduct external XML entity (XXE) attacks via a crafted XSLT extension in a (1) \u003cx:parse\u003e or (2) \u003cx:transform\u003e JSTL XML tag.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> --> | True | jstl-1.2.jar: 1 vulnerabilities (highest severity is: 7.3) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jstl-1.2.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /ository/javax/servlet/jstl/1.2/jstl-1.2.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2015-0254](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | jstl-1.2.jar | Direct | org.apache.taglibs:taglibs-standard-impl:1.2.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-0254</summary>
### Vulnerable Library - <b>jstl-1.2.jar</b></p>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /ository/javax/servlet/jstl/1.2/jstl-1.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jstl-1.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache Standard Taglibs before 1.2.3 allows remote attackers to execute arbitrary code or conduct external XML entity (XXE) attacks via a crafted XSLT extension in a (1) <x:parse> or (2) <x:transform> JSTL XML tag.
<p>Publish Date: 2015-03-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254>CVE-2015-0254</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tomcat.apache.org/taglibs/standard/">https://tomcat.apache.org/taglibs/standard/</a></p>
<p>Release Date: 2015-03-09</p>
<p>Fix Resolution: org.apache.taglibs:taglibs-standard-impl:1.2.3</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"javax.servlet","packageName":"jstl","packageVersion":"1.2","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"javax.servlet:jstl:1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.taglibs:taglibs-standard-impl:1.2.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-0254","vulnerabilityDetails":"Apache Standard Taglibs before 1.2.3 allows remote attackers to execute arbitrary code or conduct external XML entity (XXE) attacks via a crafted XSLT extension in a (1) \u003cx:parse\u003e or (2) \u003cx:transform\u003e JSTL XML tag.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> --> | non_build | jstl jar vulnerabilities highest severity is vulnerable library jstl jar path to dependency file pom xml path to vulnerable library ository javax servlet jstl jstl jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high jstl jar direct org apache taglibs taglibs standard impl details cve vulnerable library jstl jar path to dependency file pom xml path to vulnerable library ository javax servlet jstl jstl jar dependency hierarchy x jstl jar vulnerable library found in head commit a href found in base branch master vulnerability details apache standard taglibs before allows remote attackers to execute arbitrary code or conduct external xml entity xxe attacks via a crafted xslt extension in a or jstl xml tag publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache taglibs taglibs standard impl step up your open source security game with whitesource istransitivedependency false dependencytree javax servlet jstl isminimumfixversionavailable true minimumfixversion org apache taglibs taglibs standard impl isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails apache standard taglibs before allows remote attackers to execute arbitrary code or conduct external xml entity xxe attacks via a crafted xslt extension in a parse or transform jstl xml tag vulnerabilityurl | 0 |
100,251 | 30,655,612,913 | IssuesEvent | 2023-07-25 11:56:29 | dotnet/fsharp | https://api.github.com/repos/dotnet/fsharp | closed | Generate deterministic MVID for reference assemblies based on contents of the assembly | Feature Improvement Area-Compiler-CodeGen Area-Build | Continuation of reference assembly work, initially done in https://github.com/dotnet/fsharp/pull/12334
Generate deterministic MVID for reference assemblies based on contents of the assembly where the content also includes signature and optimization data; basically SHA1 of the bytes of the entire assembly.
> **Note**
> We need to keep in mind, that we store some extra information in sigdata, which can affect MVID generation:
> - the metadata for "private" declarations in implementation files that don't have signature files
> - the metadata for "internal" declarations when there are no InternalsVisibleTo in the assembly
>
> Related issue: https://github.com/dotnet/fsharp/issues/12746 | 1.0 | Generate deterministic MVID for reference assemblies based on contents of the assembly - Continuation of reference assembly work, initially done in https://github.com/dotnet/fsharp/pull/12334
Generate deterministic MVID for reference assemblies based on contents of the assembly where the content also includes signature and optimization data; basically SHA1 of the bytes of the entire assembly.
> **Note**
> We need to keep in mind, that we store some extra information in sigdata, which can affect MVID generation:
> - the metadata for "private" declarations in implementation files that don't have signature files
> - the metadata for "internal" declarations when there are no InternalsVisibleTo in the assembly
>
> Related issue: https://github.com/dotnet/fsharp/issues/12746 | build | generate deterministic mvid for reference assemblies based on contents of the assembly continuation of reference assembly work initially done in generate deterministic mvid for reference assemblies based on contents of the assembly where the content also includes signature and optimization data basically of the bytes of the entire assembly note we need to keep in mind that we store some extra information in sigdata which can affect mvid generation the metadata for private declarations in implementation files that don t have signature files the metadata for internal declarations when there are no internalsvisibleto in the assembly related issue | 1 |
67,178 | 16,829,040,879 | IssuesEvent | 2021-06-17 23:45:26 | golang/go | https://api.github.com/repos/golang/go | reopened | x/build: new builder for cmd/compile's "unified IR" mode | Builders NeedsFix new-builder | I'd like to add a builder for testing the compiler's new "unified IR" mode. This has landed on dev.typeparams, and it took all of 1 commits for us to accidentally break it again.
Proposed build config (caveat: this is just copy/paste/tweaked based on some of the other compiler experiment configs; I don't feel strongly about any of these if other changes would be appropriate):
```
addBuilder(BuildConfig{
Name: "linux-amd64-unified-ir",
HostType: "host-linux-buster",
Notes: "builder with GO_GCFLAGS=-d=unified=1, see golang.org/issue/46786",
buildsRepo: func(repo, branch, goBranch string) bool {
return repo == "go" && branch == "dev.typeparams"
},
env: []string{
"GO_DISABLE_OUTBOUND_NETWORK=1",
"GO_GCFLAGS=-d=unified=1",
},
GoDeps: []string{
"cf1ae5fc364eb7f2ee5203e4c5e30411c3cfe01f", // dev.typeparams commit that added -d=unified
},
})
```
Caveat: It's expected that the `go tool dist test test:0_1` step will fail with unified IR currently. (There's a list of tests that are known to fail with the new types2 type checker, but test/run.go doesn't know -d=unified enables this too. I will fix this shortly.) | 2.0 | x/build: new builder for cmd/compile's "unified IR" mode - I'd like to add a builder for testing the compiler's new "unified IR" mode. This has landed on dev.typeparams, and it took all of 1 commits for us to accidentally break it again.
Proposed build config (caveat: this is just copy/paste/tweaked based on some of the other compiler experiment configs; I don't feel strongly about any of these if other changes would be appropriate):
```
addBuilder(BuildConfig{
Name: "linux-amd64-unified-ir",
HostType: "host-linux-buster",
Notes: "builder with GO_GCFLAGS=-d=unified=1, see golang.org/issue/46786",
buildsRepo: func(repo, branch, goBranch string) bool {
return repo == "go" && branch == "dev.typeparams"
},
env: []string{
"GO_DISABLE_OUTBOUND_NETWORK=1",
"GO_GCFLAGS=-d=unified=1",
},
GoDeps: []string{
"cf1ae5fc364eb7f2ee5203e4c5e30411c3cfe01f", // dev.typeparams commit that added -d=unified
},
})
```
Caveat: It's expected that the `go tool dist test test:0_1` step will fail with unified IR currently. (There's a list of tests that are known to fail with the new types2 type checker, but test/run.go doesn't know -d=unified enables this too. I will fix this shortly.) | build | x build new builder for cmd compile s unified ir mode i d like to add a builder for testing the compiler s new unified ir mode this has landed on dev typeparams and it took all of commits for us to accidentally break it again proposed build config caveat this is just copy paste tweaked based on some of the other compiler experiment configs i don t feel strongly about any of these if other changes would be appropriate addbuilder buildconfig name linux unified ir hosttype host linux buster notes builder with go gcflags d unified see golang org issue buildsrepo func repo branch gobranch string bool return repo go branch dev typeparams env string go disable outbound network go gcflags d unified godeps string dev typeparams commit that added d unified caveat it s expected that the go tool dist test test step will fail with unified ir currently there s a list of tests that are known to fail with the new type checker but test run go doesn t know d unified enables this too i will fix this shortly | 1 |
382,412 | 11,305,860,862 | IssuesEvent | 2020-01-18 09:27:26 | acidanthera/bugtracker | https://api.github.com/repos/acidanthera/bugtracker | closed | AppleSupportPkg\Platform\ApfsDriverLoader: warning C4244: '=': conversion from '...' to '...', possible loss of data | priority:normal project:apfs | ```
AppleSupportPkg\Platform\ApfsDriverLoader\ApfsDriverLoader.c(381): warning C4244: '=': conversion from 'EFI_LBA' to 'UINTN', possible loss of data
AppleSupportPkg\Platform\ApfsDriverLoader\ApfsDriverLoader.c(979): warning C4244: '=': conversion from 'UINT64' to 'UINTN', possible loss of data
``` | 1.0 | AppleSupportPkg\Platform\ApfsDriverLoader: warning C4244: '=': conversion from '...' to '...', possible loss of data - ```
AppleSupportPkg\Platform\ApfsDriverLoader\ApfsDriverLoader.c(381): warning C4244: '=': conversion from 'EFI_LBA' to 'UINTN', possible loss of data
AppleSupportPkg\Platform\ApfsDriverLoader\ApfsDriverLoader.c(979): warning C4244: '=': conversion from 'UINT64' to 'UINTN', possible loss of data
``` | non_build | applesupportpkg platform apfsdriverloader warning conversion from to possible loss of data applesupportpkg platform apfsdriverloader apfsdriverloader c warning conversion from efi lba to uintn possible loss of data applesupportpkg platform apfsdriverloader apfsdriverloader c warning conversion from to uintn possible loss of data | 0 |
374,320 | 11,083,931,556 | IssuesEvent | 2019-12-13 15:29:06 | googleapis/google-cloud-cpp-spanner | https://api.github.com/repos/googleapis/google-cloud-cpp-spanner | closed | Properly handle Session NOT_FOUND | priority: p2 type: bug | @coryan reported that one of his tests failed with the following error:
`Exception caught in HTTP handler: Permanent failure in RunTransactionWithPolicies: Session not found: <XXX> [NOT FOUND]`
We should handle this error internally and avoid exposing it to the user. From https://cloud.google.com/spanner/docs/sessions
```
Attempts to use a deleted session result in NOT_FOUND. If you encounter this error, create and use a new session, add the new session to the cache, and remove the deleted session from the cache.
```
However, I'm not entirely sure how to accomplish this given that a `Transaction` is bound to a `Session`. For RW transactions using `RunTransaction` we should be able to treat it similarly to an `ABORTED` and retry the whole transaction with a new session.
For RO transactions or non-idempotent mutations, it's less clear what we should do. The page linked-to above gives some general advice, but nothing specifically about this case. | 1.0 | Properly handle Session NOT_FOUND - @coryan reported that one of his tests failed with the following error:
`Exception caught in HTTP handler: Permanent failure in RunTransactionWithPolicies: Session not found: <XXX> [NOT FOUND]`
We should handle this error internally and avoid exposing it to the user. From https://cloud.google.com/spanner/docs/sessions
```
Attempts to use a deleted session result in NOT_FOUND. If you encounter this error, create and use a new session, add the new session to the cache, and remove the deleted session from the cache.
```
However, I'm not entirely sure how to accomplish this given that a `Transaction` is bound to a `Session`. For RW transactions using `RunTransaction` we should be able to treat it similarly to an `ABORTED` and retry the whole transaction with a new session.
For RO transactions or non-idempotent mutations, it's less clear what we should do. The page linked-to above gives some general advice, but nothing specifically about this case. | non_build | properly handle session not found coryan reported that one of his tests failed with the following error exception caught in http handler permanent failure in runtransactionwithpolicies session not found we should handle this error internally and avoid exposing it to the user from attempts to use a deleted session result in not found if you encounter this error create and use a new session add the new session to the cache and remove the deleted session from the cache however i m not entirely sure how to accomplish this given that a transaction is bound to a session for rw transactions using runtransaction we should be able to treat it similarly to an aborted and retry the whole transaction with a new session for ro transactions or non idempotent mutations it s less clear what we should do the page linked to above gives some general advice but nothing specifically about this case | 0 |
27,496 | 7,970,704,837 | IssuesEvent | 2018-07-16 13:35:47 | ninia/jep | https://api.github.com/repos/ninia/jep | closed | pip install : no such file or directory for classlist_7.txt | build | When installing jep with pip i get the error
error: [Errno 2] No such file or directory: 'src/main/resources/jep/classlist_7.txt'
- OS Platform, Distribution, and Version: Windows 7
- Python Distribution and Version: Python 3.6
- Java Distribution and Version: Java 1.8
| 1.0 | pip install : no such file or directory for classlist_7.txt - When installing jep with pip i get the error
error: [Errno 2] No such file or directory: 'src/main/resources/jep/classlist_7.txt'
- OS Platform, Distribution, and Version: Windows 7
- Python Distribution and Version: Python 3.6
- Java Distribution and Version: Java 1.8
| build | pip install no such file or directory for classlist txt when installing jep with pip i get the error error no such file or directory src main resources jep classlist txt os platform distribution and version windows python distribution and version python java distribution and version java | 1 |
19,809 | 3,263,407,907 | IssuesEvent | 2015-10-22 04:26:56 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | Is the Amber netcdf writer fully compliant with the official spec? | Component-Writers defect Format-Amber Priority-High | I have had a problem reading AMBER .ncdf files into the AMBER analysis tool cpptraj:
```
CPPTRAJ: Trajectory Analysis. V14.22
___ ___ ___ ___
| \/ | \/ | \/ |
_|_/\_|_/\_|_/\_|_
Reading 'penta.top' as Amber Topology
> trajin mdanalout.ncdf
Reading 'mdanalout.ncdf' as Amber NetCDF
NETCDF Error: NetCDF: Attribute not found
Error: Getting length for attribute title
NETCDF Error: NetCDF: Variable not found
Error: Getting spatial VID
Error: Could not set up mdanalout.ncdf for reading.
Error: Could not set up input trajectory 'mdanalout.ncdf'.
```
At first I thought this was an issue with cpptraj and raised it as an issue there (https://github.com/Amber-MD/cpptraj/issues/123) , but it seems it may be a MDAnalysis issue. | 1.0 | Is the Amber netcdf writer fully compliant with the official spec? - I have had a problem reading AMBER .ncdf files into the AMBER analysis tool cpptraj:
```
CPPTRAJ: Trajectory Analysis. V14.22
___ ___ ___ ___
| \/ | \/ | \/ |
_|_/\_|_/\_|_/\_|_
Reading 'penta.top' as Amber Topology
> trajin mdanalout.ncdf
Reading 'mdanalout.ncdf' as Amber NetCDF
NETCDF Error: NetCDF: Attribute not found
Error: Getting length for attribute title
NETCDF Error: NetCDF: Variable not found
Error: Getting spatial VID
Error: Could not set up mdanalout.ncdf for reading.
Error: Could not set up input trajectory 'mdanalout.ncdf'.
```
At first I thought this was an issue with cpptraj and raised it as an issue there (https://github.com/Amber-MD/cpptraj/issues/123) , but it seems it may be a MDAnalysis issue. | non_build | is the amber netcdf writer fully compliant with the official spec i have had a problem reading amber ncdf files into the amber analysis tool cpptraj cpptraj trajectory analysis reading penta top as amber topology trajin mdanalout ncdf reading mdanalout ncdf as amber netcdf netcdf error netcdf attribute not found error getting length for attribute title netcdf error netcdf variable not found error getting spatial vid error could not set up mdanalout ncdf for reading error could not set up input trajectory mdanalout ncdf at first i thought this was an issue with cpptraj and raised it as an issue there but it seems it may be a mdanalysis issue | 0 |
283,930 | 21,337,808,079 | IssuesEvent | 2022-04-18 16:35:26 | pyrsia/pyrsia | https://api.github.com/repos/pyrsia/pyrsia | opened | Build on "Good PR" for Reviewing Suggestions | documentation | There were a few items noted during the reviewers policy change, https://github.com/pyrsia/.github/pull/6#issuecomment-1098362917
- Try to avoid requesting changes
- How to change review (in other words, dismiss your own approval)
Other ideas are welcomed! | 1.0 | Build on "Good PR" for Reviewing Suggestions - There were a few items noted during the reviewers policy change, https://github.com/pyrsia/.github/pull/6#issuecomment-1098362917
- Try to avoid requesting changes
- How to change review (in other words, dismiss your own approval)
Other ideas are welcomed! | non_build | build on good pr for reviewing suggestions there were a few items noted during the reviewers policy change try to avoid requesting changes how to change review in other words dismiss your own approval other ideas are welcomed | 0 |
669 | 3,964,688,323 | IssuesEvent | 2016-05-03 02:43:27 | comozilla/Parapara-Canvas-Editor | https://api.github.com/repos/comozilla/Parapara-Canvas-Editor | closed | Viewクラスたちを継承できるようにしたい! | architecture code enhancement wontfix | `ColorPickerView`、`LineWidthPickerView`は、どちらも、
コンストラクタで引数として element、drawingConfig を受け取っていて、
それらをメンバにしているところが共通であるため、
継承したい!(jsで継承とかってできるのかなあ) | 1.0 | Viewクラスたちを継承できるようにしたい! - `ColorPickerView`、`LineWidthPickerView`は、どちらも、
コンストラクタで引数として element、drawingConfig を受け取っていて、
それらをメンバにしているところが共通であるため、
継承したい!(jsで継承とかってできるのかなあ) | non_build | viewクラスたちを継承できるようにしたい! colorpickerview 、 linewidthpickerview は、どちらも、 コンストラクタで引数として element、drawingconfig を受け取っていて、 それらをメンバにしているところが共通であるため、 継承したい!(jsで継承とかってできるのかなあ) | 0 |
70,986 | 18,362,452,703 | IssuesEvent | 2021-10-09 13:07:31 | beaverbuilder/feature-requests | https://api.github.com/repos/beaverbuilder/feature-requests | opened | Module slider | Beaver Builder | Imagine a slider in which you could put different kinds of modules. One 'pane' might be an image, the other a paragraph, and yet the other a callout. It would allow to build up a 'story' or 'action' set that gets the user to ultimately do something. There are plenty of other benefits, such as alternating between two modules. Imagine a photo, then details of that photo, then a photo again, and so on... | 1.0 | Module slider - Imagine a slider in which you could put different kinds of modules. One 'pane' might be an image, the other a paragraph, and yet the other a callout. It would allow to build up a 'story' or 'action' set that gets the user to ultimately do something. There are plenty of other benefits, such as alternating between two modules. Imagine a photo, then details of that photo, then a photo again, and so on... | build | module slider imagine a slider in which you could put different kinds of modules one pane might be an image the other a paragraph and yet the other a callout it would allow to build up a story or action set that gets the user to ultimately do something there are plenty of other benefits such as alternating between two modules imagine a photo then details of that photo then a photo again and so on | 1 |
275,953 | 30,310,126,927 | IssuesEvent | 2023-07-10 12:17:17 | openraven/magpie | https://api.github.com/repos/openraven/magpie | closed | Assets tracking rule with AWS Tags | aws security rules | People are wanting to ensure tags are on assets. In general security teams are focused on having asset owners and manual data / application tagging.
It would be great to have a rule that returns all ARN's that don't have tags. Ideally the rule would allow the user to read in an array of expected tags as well so they can add missing tags as well as the case where there are no tags. This is likely related to #247 the policy rule actions ticket.
┆Issue is synchronized with this [Jira Task](https://openraven.atlassian.net/browse/RAD-248) by [Unito](https://www.unito.io)
| True | Assets tracking rule with AWS Tags - People are wanting to ensure tags are on assets. In general security teams are focused on having asset owners and manual data / application tagging.
It would be great to have a rule that returns all ARN's that don't have tags. Ideally the rule would allow the user to read in an array of expected tags as well so they can add missing tags as well as the case where there are no tags. This is likely related to #247 the policy rule actions ticket.
┆Issue is synchronized with this [Jira Task](https://openraven.atlassian.net/browse/RAD-248) by [Unito](https://www.unito.io)
| non_build | assets tracking rule with aws tags people are wanting to ensure tags are on assets in general security teams are focused on having asset owners and manual data application tagging it would be great to have a rule that returns all arn s that don t have tags ideally the rule would allow the user to read in an array of expected tags as well so they can add missing tags as well as the case where there are no tags this is likely related to the policy rule actions ticket ┆issue is synchronized with this by | 0 |
169,144 | 20,828,057,698 | IssuesEvent | 2022-03-19 01:27:20 | Seagate/cortx-re | https://api.github.com/repos/Seagate/cortx-re | opened | CVE-2022-24302 (Medium) detected in paramiko-2.7.1-py2.py3-none-any.whl | security vulnerability | ## CVE-2022-24302 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-2.7.1-py2.py3-none-any.whl</b></p></summary>
<p>SSH2 protocol library</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /docker/cortx-deploy/python_requirements.txt</p>
<p>Path to vulnerable library: /docker/cortx-deploy/python_requirements.txt,/scripts/third-party-rpm/python_requirements.txt,/docker/cortx-deploy/cortx-rgw/python_requirements.txt,/docker/cortx-deploy/cortx-control/python_requirements.txt,/docker/cortx-deploy/cortx-data/python_requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **paramiko-2.7.1-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302>CVE-2022-24302</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.paramiko.org/changelog.html">https://www.paramiko.org/changelog.html</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: paramiko - 2.10.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"paramiko","packageVersion":"2.7.1","packageFilePaths":["/docker/cortx-deploy/python_requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"paramiko:2.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"paramiko - 2.10.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-24302","vulnerabilityDetails":"In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2022-24302 (Medium) detected in paramiko-2.7.1-py2.py3-none-any.whl - ## CVE-2022-24302 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-2.7.1-py2.py3-none-any.whl</b></p></summary>
<p>SSH2 protocol library</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /docker/cortx-deploy/python_requirements.txt</p>
<p>Path to vulnerable library: /docker/cortx-deploy/python_requirements.txt,/scripts/third-party-rpm/python_requirements.txt,/docker/cortx-deploy/cortx-rgw/python_requirements.txt,/docker/cortx-deploy/cortx-control/python_requirements.txt,/docker/cortx-deploy/cortx-data/python_requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **paramiko-2.7.1-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302>CVE-2022-24302</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.paramiko.org/changelog.html">https://www.paramiko.org/changelog.html</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: paramiko - 2.10.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"paramiko","packageVersion":"2.7.1","packageFilePaths":["/docker/cortx-deploy/python_requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"paramiko:2.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"paramiko - 2.10.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-24302","vulnerabilityDetails":"In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> --> | non_build | cve medium detected in paramiko none any whl cve medium severity vulnerability vulnerable library paramiko none any whl protocol library library home page a href path to dependency file docker cortx deploy python requirements txt path to vulnerable library docker cortx deploy python requirements txt scripts third party rpm python requirements txt docker cortx deploy cortx rgw python requirements txt docker cortx deploy cortx control python requirements txt docker cortx deploy cortx data python requirements txt dependency hierarchy x paramiko none any whl vulnerable library found in base branch main vulnerability details in paramiko before a race condition between creation and chmod in the write private key file function could allow unauthorized information disclosure publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution paramiko rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree paramiko isminimumfixversionavailable true minimumfixversion paramiko isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in paramiko before a race condition between creation and chmod in the write private key file function could allow unauthorized information disclosure vulnerabilityurl | 0 |
42,374 | 10,964,126,942 | IssuesEvent | 2019-11-27 21:36:07 | pytorch/vision | https://api.github.com/repos/pytorch/vision | closed | torchvision build from source failed | module: io needs reproduction topic: build topic: video | Environment:
python3.7
cuda-10.0
pytorch successfully build from source on latest version
ffmpeg
```
ffmpeg version 4.1.3-0york1~16.04 Copyright (c) 2000-2019 the FFmpeg developers
built with gcc 5.4.0 (Ubuntu 5.4.0-6ubuntu1~16.04.11) 20160609
configuration: --prefix=/usr --extra-version='0york1~16.04' --toolchain=hardened --libdir=/usr/lib/x86_64-linux-gnu --incdir=/usr/include/x86_64-linux-gnu --arch=amd64 --enable-gpl --disable-stripping --enable-avresample --disable-filter=resample --enable-avisynth --enable-gnutls --enable-ladspa --enable-libaom --enable-libass --enable-libbluray --enable-libbs2b --enable-libcaca --enable-libcdio --enable-libcodec2 --enable-libflite --enable-libfontconfig --enable-libfreetype --enable-libfribidi --enable-libgme --enable-libgsm --enable-libjack --enable-libmp3lame --enable-libmysofa --enable-libopenjpeg --enable-libopenmpt --enable-libopus --enable-libpulse --enable-librsvg --enable-librubberband --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libspeex --enable-libssh --enable-libtheora --enable-libtwolame --enable-libvidstab --enable-libvorbis --enable-libvpx --enable-libwavpack --enable-libwebp --enable-libx265 --enable-libxml2 --enable-libxvid --enable-libzmq --enable-libzvbi --enable-lv2 --enable-omx --enable-openal --enable-opengl --enable-sdl2 --enable-nonfree --enable-libfdk-aac --enable-libdc1394 --enable-libdrm --enable-libiec61883 --enable-chromaprint --enable-frei0r --enable-libx264 --enable-shared
libavutil 56. 22.100 / 56. 22.100
libavcodec 58. 35.100 / 58. 35.100
libavformat 58. 20.100 / 58. 20.100
libavdevice 58. 5.100 / 58. 5.100
libavfilter 7. 40.101 / 7. 40.101
libavresample 4. 0. 0 / 4. 0. 0
libswscale 5. 3.100 / 5. 3.100
libswresample 3. 3.100 / 3. 3.100
libpostproc 55. 3.100 / 55. 3.100
Hyper fast Audio and Video encoder
```
when run `python setup.py install` in `vision` directory:
```
cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::openCodecContext()’:
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:31:23: error: ‘AVStream {aka struct AVStream}’ has no member named ‘codecpar’
auto codec_id = st->codecpar->codec_id;
^
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:48:59: error: ‘AVStream {aka struct AVStream}’ has no member named ‘codecpar’
if ((ret = avcodec_parameters_to_context(codecCtx_, st->codecpar)) < 0) {
^
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:48:67: error: ‘avcodec_parameters_to_context’ was not declared in this scope
if ((ret = avcodec_parameters_to_context(codecCtx_, st->codecpar)) < 0) {
^
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::sendPacket(const AVPacket*)’:
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:146:47: error: ‘avcodec_send_packet’ was not declared in this scope
return avcodec_send_packet(codecCtx_, packet);
^
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::receiveFrame()’:
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:150:52: error: ‘avcodec_receive_frame’ was not declared in this scope
int ret = avcodec_receive_frame(codecCtx_, frame_);
^
error: command 'gcc' failed with exit status 1
``` | 1.0 | torchvision build from source failed - Environment:
python3.7
cuda-10.0
pytorch successfully build from source on latest version
ffmpeg
```
ffmpeg version 4.1.3-0york1~16.04 Copyright (c) 2000-2019 the FFmpeg developers
built with gcc 5.4.0 (Ubuntu 5.4.0-6ubuntu1~16.04.11) 20160609
configuration: --prefix=/usr --extra-version='0york1~16.04' --toolchain=hardened --libdir=/usr/lib/x86_64-linux-gnu --incdir=/usr/include/x86_64-linux-gnu --arch=amd64 --enable-gpl --disable-stripping --enable-avresample --disable-filter=resample --enable-avisynth --enable-gnutls --enable-ladspa --enable-libaom --enable-libass --enable-libbluray --enable-libbs2b --enable-libcaca --enable-libcdio --enable-libcodec2 --enable-libflite --enable-libfontconfig --enable-libfreetype --enable-libfribidi --enable-libgme --enable-libgsm --enable-libjack --enable-libmp3lame --enable-libmysofa --enable-libopenjpeg --enable-libopenmpt --enable-libopus --enable-libpulse --enable-librsvg --enable-librubberband --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libspeex --enable-libssh --enable-libtheora --enable-libtwolame --enable-libvidstab --enable-libvorbis --enable-libvpx --enable-libwavpack --enable-libwebp --enable-libx265 --enable-libxml2 --enable-libxvid --enable-libzmq --enable-libzvbi --enable-lv2 --enable-omx --enable-openal --enable-opengl --enable-sdl2 --enable-nonfree --enable-libfdk-aac --enable-libdc1394 --enable-libdrm --enable-libiec61883 --enable-chromaprint --enable-frei0r --enable-libx264 --enable-shared
libavutil 56. 22.100 / 56. 22.100
libavcodec 58. 35.100 / 58. 35.100
libavformat 58. 20.100 / 58. 20.100
libavdevice 58. 5.100 / 58. 5.100
libavfilter 7. 40.101 / 7. 40.101
libavresample 4. 0. 0 / 4. 0. 0
libswscale 5. 3.100 / 5. 3.100
libswresample 3. 3.100 / 3. 3.100
libpostproc 55. 3.100 / 55. 3.100
Hyper fast Audio and Video encoder
```
when run `python setup.py install` in `vision` directory:
```
cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::openCodecContext()’:
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:31:23: error: ‘AVStream {aka struct AVStream}’ has no member named ‘codecpar’
auto codec_id = st->codecpar->codec_id;
^
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:48:59: error: ‘AVStream {aka struct AVStream}’ has no member named ‘codecpar’
if ((ret = avcodec_parameters_to_context(codecCtx_, st->codecpar)) < 0) {
^
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:48:67: error: ‘avcodec_parameters_to_context’ was not declared in this scope
if ((ret = avcodec_parameters_to_context(codecCtx_, st->codecpar)) < 0) {
^
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::sendPacket(const AVPacket*)’:
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:146:47: error: ‘avcodec_send_packet’ was not declared in this scope
return avcodec_send_packet(codecCtx_, packet);
^
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::receiveFrame()’:
/home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:150:52: error: ‘avcodec_receive_frame’ was not declared in this scope
int ret = avcodec_receive_frame(codecCtx_, frame_);
^
error: command 'gcc' failed with exit status 1
``` | build | torchvision build from source failed environment cuda pytorch successfully build from source on latest version ffmpeg ffmpeg version copyright c the ffmpeg developers built with gcc ubuntu configuration prefix usr extra version toolchain hardened libdir usr lib linux gnu incdir usr include linux gnu arch enable gpl disable stripping enable avresample disable filter resample enable avisynth enable gnutls enable ladspa enable libaom enable libass enable libbluray enable enable libcaca enable libcdio enable enable libflite enable libfontconfig enable libfreetype enable libfribidi enable libgme enable libgsm enable libjack enable enable libmysofa enable libopenjpeg enable libopenmpt enable libopus enable libpulse enable librsvg enable librubberband enable libshine enable libsnappy enable libsoxr enable libspeex enable libssh enable libtheora enable libtwolame enable libvidstab enable libvorbis enable libvpx enable libwavpack enable libwebp enable enable enable libxvid enable libzmq enable libzvbi enable enable omx enable openal enable opengl enable enable nonfree enable libfdk aac enable enable libdrm enable enable chromaprint enable enable enable shared libavutil libavcodec libavformat libavdevice libavfilter libavresample libswscale libswresample libpostproc hyper fast audio and video encoder when run python setup py install in vision directory warning command line option ‘ wstrict prototypes’ is valid for c objc but not for c home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp in member function ‘int ffmpegstream opencodeccontext ’ home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp error ‘avstream aka struct avstream ’ has no member named ‘codecpar’ auto codec id st codecpar codec id home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp error ‘avstream aka struct avstream ’ has no member named ‘codecpar’ if ret avcodec parameters to context codecctx st codecpar home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp error ‘avcodec parameters to context’ was not declared in this scope if ret avcodec parameters to context codecctx st codecpar home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp in member function ‘int ffmpegstream sendpacket const avpacket ’ home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp error ‘avcodec send packet’ was not declared in this scope return avcodec send packet codecctx packet home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp in member function ‘int ffmpegstream receiveframe ’ home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp error ‘avcodec receive frame’ was not declared in this scope int ret avcodec receive frame codecctx frame error command gcc failed with exit status | 1 |
107,765 | 23,479,858,358 | IssuesEvent | 2022-08-17 09:35:56 | zond/dipact | https://api.github.com/repos/zond/dipact | closed | Remove Color.js component | code quality | This component is not necessary as there is a color input type built in to browsers. | 1.0 | Remove Color.js component - This component is not necessary as there is a color input type built in to browsers. | non_build | remove color js component this component is not necessary as there is a color input type built in to browsers | 0 |
637,404 | 20,627,326,806 | IssuesEvent | 2022-03-08 00:35:21 | NRCan/GSC-Field-Application | https://api.github.com/repos/NRCan/GSC-Field-Application | closed | Location - edit fails on saving | bug low priority | For testing purposes I took some stations with a really bad location, actually pretty much nothing.
Then I went into the location form to edit it and clicked saved without modifying anything. Now I can't get out of the location form or even edit it to fix the problem. There is a warning message that pops on screen and prevents user from closing it or going elsewhere.

| 1.0 | Location - edit fails on saving - For testing purposes I took some stations with a really bad location, actually pretty much nothing.
Then I went into the location form to edit it and clicked saved without modifying anything. Now I can't get out of the location form or even edit it to fix the problem. There is a warning message that pops on screen and prevents user from closing it or going elsewhere.

| non_build | location edit fails on saving for testing purposes i took some stations with a really bad location actually pretty much nothing then i went into the location form to edit it and clicked saved without modifying anything now i can t get out of the location form or even edit it to fix the problem there is a warning message that pops on screen and prevents user from closing it or going elsewhere | 0 |
89,905 | 25,920,949,408 | IssuesEvent | 2022-12-15 21:53:30 | brtnfld/cgnsjira | https://api.github.com/repos/brtnfld/cgnsjira | opened | [CGNS-164] Limit on size of writing a single array for parallel library: 2Gb limit and 4Gb windows (no limit in serial library) | bug To Do Major Build |
> This issue has been migrated from the Forge. Read the [original ticket here](https://cgnsorg.atlassian.net/browse/CGNS-164).
- _**Reporter:**_ None
- _**Created at:**_ Thu, 6 Jun 2019 07:35:56 -0500
<p>(1) There are no issues writing large files on either Windows or Linux when serial CGNS used<br/>
(2) When parallel CGNS is used, there appears to be a limit on the size of a single array that can be written. For Windows the limit appears to be 4Gb. For Linux the limit appears to be 2Gb. The overall file size can be larger than 4Gb, provided no single array written is >4Gb (windows) or 2Gb (Linux)</p>
<p>So, in summary, the issue is with writing large single arrays on both platforms when using the parallel library. This is happening even when using cgp_open with a serial communicator. </p>
<p>Since the CFD solution we are writing is written across a number of separate arrays (one for velocity, another the mesh etc etc) we have not hit this restriction yet, but it could happen with very large meshes, so this needs to be fixed. </p>
<p>How did I come to this conclusion?</p>
<p>I used the attached Large.cpp file to write >4Gb files on windows and Linux. This only uses the serial cg_open call and writes a single large array. This works fine on both platforms.<br/>
Our existing Forte simulator is using parallel CGNS 3.3.0 and I used it to write a 20Gb file on windows when running a real engine simulation. However, although the file is 20Gb, no single array is >=4Gb.<br/>
To determine the limit for writing single arrays in parallel, I modified my test code to include an option to write multiple arrays (attached). The first arg is the size of each array and the second the number of arrays. Running with 4 MPI processes and args 2 10 - i.e. write 10 arrays each of size 2Gb, leading to a 20Gb file, works on Windows. This code writes serially using cgp_open and a serial communicator and then repeats the same write in parallel. However, as soon as each single array is >=4Gb, the write fails. On Linux, the limit is 2Gb.</p>
<p>Windows:<br/>
mpiexec -localonly -n 2 CGNSLarge_sections.exe 4 2<br/>
Large CGNS file test program, size=4Gb, written in 2 slices<br/>
Open in serial<br/>
Opening test.cgns...<br/>
Now write the data......1......2......done<br/>
cgio_write_all_data:H5Dwrite:write to node data failed</p>
<p>Linux:<br/>
mpirun -n 2 CGNSLarge_sections_new 2 1<br/>
Large CGNS file test program, size=2Gb, written in 1 slices <br/>
Open in serial <br/>
Opening test.cgns...<br/>
Now write the data......1......done<br/>
File written successfully<br/>
File closed successfully<br/>
Re-open the file.....done<br/>
...1...268434944 2.68435e+08 <> 0</p>
| 1.0 | [CGNS-164] Limit on size of writing a single array for parallel library: 2Gb limit and 4Gb windows (no limit in serial library) -
> This issue has been migrated from the Forge. Read the [original ticket here](https://cgnsorg.atlassian.net/browse/CGNS-164).
- _**Reporter:**_ None
- _**Created at:**_ Thu, 6 Jun 2019 07:35:56 -0500
<p>(1) There are no issues writing large files on either Windows or Linux when serial CGNS used<br/>
(2) When parallel CGNS is used, there appears to be a limit on the size of a single array that can be written. For Windows the limit appears to be 4Gb. For Linux the limit appears to be 2Gb. The overall file size can be larger than 4Gb, provided no single array written is >4Gb (windows) or 2Gb (Linux)</p>
<p>So, in summary, the issue is with writing large single arrays on both platforms when using the parallel library. This is happening even when using cgp_open with a serial communicator. </p>
<p>Since the CFD solution we are writing is written across a number of separate arrays (one for velocity, another the mesh etc etc) we have not hit this restriction yet, but it could happen with very large meshes, so this needs to be fixed. </p>
<p>How did I come to this conclusion?</p>
<p>I used the attached Large.cpp file to write >4Gb files on windows and Linux. This only uses the serial cg_open call and writes a single large array. This works fine on both platforms.<br/>
Our existing Forte simulator is using parallel CGNS 3.3.0 and I used it to write a 20Gb file on windows when running a real engine simulation. However, although the file is 20Gb, no single array is >=4Gb.<br/>
To determine the limit for writing single arrays in parallel, I modified my test code to include an option to write multiple arrays (attached). The first arg is the size of each array and the second the number of arrays. Running with 4 MPI processes and args 2 10 - i.e. write 10 arrays each of size 2Gb, leading to a 20Gb file, works on Windows. This code writes serially using cgp_open and a serial communicator and then repeats the same write in parallel. However, as soon as each single array is >=4Gb, the write fails. On Linux, the limit is 2Gb.</p>
<p>Windows:<br/>
mpiexec -localonly -n 2 CGNSLarge_sections.exe 4 2<br/>
Large CGNS file test program, size=4Gb, written in 2 slices<br/>
Open in serial<br/>
Opening test.cgns...<br/>
Now write the data......1......2......done<br/>
cgio_write_all_data:H5Dwrite:write to node data failed</p>
<p>Linux:<br/>
mpirun -n 2 CGNSLarge_sections_new 2 1<br/>
Large CGNS file test program, size=2Gb, written in 1 slices <br/>
Open in serial <br/>
Opening test.cgns...<br/>
Now write the data......1......done<br/>
File written successfully<br/>
File closed successfully<br/>
Re-open the file.....done<br/>
...1...268434944 2.68435e+08 <> 0</p>
| build | limit on size of writing a single array for parallel library limit and windows no limit in serial library this issue has been migrated from the forge read the reporter none created at thu jun there are no issues writing large files on either windows or linux when serial cgns used when parallel cgns is used there appears to be a limit on the size of a single array that can be written for windows the limit appears to be for linux the limit appears to be the overall file size can be larger than provided no single array written is windows or linux so in summary the issue is with writing large single arrays on both platforms when using the parallel library this is happening even when using cgp open with a serial communicator since the cfd solution we are writing is written across a number of separate arrays one for velocity another the mesh etc etc we have not hit this restriction yet but it could happen with very large meshes so this needs to be fixed how did i come to this conclusion i used the attached large cpp file to write files on windows and linux this only uses the serial cg open call and writes a single large array this works fine on both platforms our existing forte simulator is using parallel cgns and i used it to write a file on windows when running a real engine simulation however although the file is no single array is to determine the limit for writing single arrays in parallel i modified my test code to include an option to write multiple arrays attached the first arg is the size of each array and the second the number of arrays running with mpi processes and args i e write arrays each of size leading to a file works on windows this code writes serially using cgp open and a serial communicator and then repeats the same write in parallel however as soon as each single array is the write fails on linux the limit is windows mpiexec localonly n cgnslarge sections exe large cgns file test program size written in slices open in serial opening test cgns now write the data done cgio write all data write to node data failed linux mpirun n cgnslarge sections new large cgns file test program size written in slices open in serial opening test cgns now write the data done file written successfully file closed successfully re open the file done | 1 |
8,563 | 4,273,866,807 | IssuesEvent | 2016-07-13 18:39:27 | haskell/cabal | https://api.github.com/repos/haskell/cabal | closed | new-repl with other ghci | component: nix-local-build | I'd like to use `cabal new-repl --with-ghc=intero`, this used to work with `cabal repl`. But now I get the following error:
```
jacco@xw4600:~/projects/leksah/leksah$ cabal new-repl --with-ghc intero
cabal: '/home/jacco/.local/bin/intero' exited with an error:
intero: UnhelpfulSpan "on the commandline": cannot use `--interactive' with
`--supported-languages'
Usage: For basic information, try the `--help' option.
```
This also happens with any other ghci (e.g. `cabal new-repl --with-ghc=ghci`) | 1.0 | new-repl with other ghci - I'd like to use `cabal new-repl --with-ghc=intero`, this used to work with `cabal repl`. But now I get the following error:
```
jacco@xw4600:~/projects/leksah/leksah$ cabal new-repl --with-ghc intero
cabal: '/home/jacco/.local/bin/intero' exited with an error:
intero: UnhelpfulSpan "on the commandline": cannot use `--interactive' with
`--supported-languages'
Usage: For basic information, try the `--help' option.
```
This also happens with any other ghci (e.g. `cabal new-repl --with-ghc=ghci`) | build | new repl with other ghci i d like to use cabal new repl with ghc intero this used to work with cabal repl but now i get the following error jacco projects leksah leksah cabal new repl with ghc intero cabal home jacco local bin intero exited with an error intero unhelpfulspan on the commandline cannot use interactive with supported languages usage for basic information try the help option this also happens with any other ghci e g cabal new repl with ghc ghci | 1 |
76,452 | 21,427,646,769 | IssuesEvent | 2022-04-23 00:00:43 | PowerShell/PowerShell | https://api.github.com/repos/PowerShell/PowerShell | closed | PowerShell v7.2.1 - Windows Update Loop | Issue-Question Area-Maintainers-Build Resolution-Answered | ### Prerequisites
- [X] Write a descriptive title.
- [X] Make sure you are able to repro it on the [latest released version](https://github.com/PowerShell/PowerShell/releases)
- [X] Search the existing issues.
- [X] Refer to the [FAQ](https://github.com/PowerShell/PowerShell/blob/master/docs/FAQ.md).
- [X] Refer to [Differences between Windows PowerShell 5.1 and PowerShell](https://docs.microsoft.com/powershell/scripting/whats-new/differences-from-windows-powershell).
### Steps to reproduce
Run Windows Update with PowerShell installed and check for updates.
### Expected behavior
```console
PowerShell v7.2.1 installs and completes.
```
### Actual behavior
```console
Same version of PowerShell is installed and again and again.
```
### Environment data
```powershell
Name Value
---- -----
PSVersion 7.2.1
PSEdition Core
GitCommitId 7.2.1
OS Microsoft Windows 10.0.19044
Platform Win32NT
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…}
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
WSManStackVersion 3.0
```
### OS Data
```
Version Caption
------- -------
10.0.19044 Microsoft Windows 10 Pro
```
### Windows update log
Please see [this gist](https://gist.githubusercontent.com/hl2guide/dffed20e92f4346395eb2ae0cc7317cf/raw/3ff41f46d544fc72b4507bfe55819a546c726018/gistfile1.txt).
### Visuals
_No response_ | 1.0 | PowerShell v7.2.1 - Windows Update Loop - ### Prerequisites
- [X] Write a descriptive title.
- [X] Make sure you are able to repro it on the [latest released version](https://github.com/PowerShell/PowerShell/releases)
- [X] Search the existing issues.
- [X] Refer to the [FAQ](https://github.com/PowerShell/PowerShell/blob/master/docs/FAQ.md).
- [X] Refer to [Differences between Windows PowerShell 5.1 and PowerShell](https://docs.microsoft.com/powershell/scripting/whats-new/differences-from-windows-powershell).
### Steps to reproduce
Run Windows Update with PowerShell installed and check for updates.
### Expected behavior
```console
PowerShell v7.2.1 installs and completes.
```
### Actual behavior
```console
Same version of PowerShell is installed and again and again.
```
### Environment data
```powershell
Name Value
---- -----
PSVersion 7.2.1
PSEdition Core
GitCommitId 7.2.1
OS Microsoft Windows 10.0.19044
Platform Win32NT
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…}
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
WSManStackVersion 3.0
```
### OS Data
```
Version Caption
------- -------
10.0.19044 Microsoft Windows 10 Pro
```
### Windows update log
Please see [this gist](https://gist.githubusercontent.com/hl2guide/dffed20e92f4346395eb2ae0cc7317cf/raw/3ff41f46d544fc72b4507bfe55819a546c726018/gistfile1.txt).
### Visuals
_No response_ | build | powershell windows update loop prerequisites write a descriptive title make sure you are able to repro it on the search the existing issues refer to the refer to steps to reproduce run windows update with powershell installed and check for updates expected behavior console powershell installs and completes actual behavior console same version of powershell is installed and again and again environment data powershell name value psversion psedition core gitcommitid os microsoft windows platform pscompatibleversions … psremotingprotocolversion serializationversion wsmanstackversion os data version caption microsoft windows pro windows update log please see visuals no response | 1 |
424,784 | 12,323,250,532 | IssuesEvent | 2020-05-13 11:50:24 | Azure/ARO-RP | https://api.github.com/repos/Azure/ARO-RP | opened | APIServer status to metrics | priority-medium size-medium | APIserver status has very good insights on install failures and modes:
```
oc get kubeapiservers cluster -o yaml
...
- lastTransitionTime: "2020-05-13T11:40:42Z"
message: 3 nodes are active; 2 nodes are at revision 2; 1 nodes are at revision
3; 0 nodes have achieved new revision 4
status: "True"
type: StaticPodsAvailable
- lastTransitionTime: "2020-05-13T11:38:11Z"
message: 2 nodes are at revision 2; 1 nodes are at revision 3; 0 nodes have achieved
new revision 4
status: "True"
type: NodeInstallerProgressing
- lastTransitionTime: "2020-05-13T11:47:09Z"
message: |-
nodes/mjudeikis3-8bcbz-master-1 pods/kube-apiserver-mjudeikis3-8bcbz-master-1 container="kube-apiserver" is not ready
nodes/mjudeikis3-8bcbz-master-1 pods/kube-apiserver-mjudeikis3-8bcbz-master-1 container="kube-apiserver" is waiting: "CrashLoopBackOff" - "back-off 20s restarting failed container=kube-apiserver pod=kube-apiserver-mjudeikis3-8bcbz-master-1_openshift-kube-apiserver(458b630d98792d8c21612b955384b7e7)"
reason: Error
status: "True"
type: StaticPodsDegraded
```
We might want to consider those as metrics/logs for failed cluster install debugging when we don't have access to API server to use geneva.
| 1.0 | APIServer status to metrics - APIserver status has very good insights on install failures and modes:
```
oc get kubeapiservers cluster -o yaml
...
- lastTransitionTime: "2020-05-13T11:40:42Z"
message: 3 nodes are active; 2 nodes are at revision 2; 1 nodes are at revision
3; 0 nodes have achieved new revision 4
status: "True"
type: StaticPodsAvailable
- lastTransitionTime: "2020-05-13T11:38:11Z"
message: 2 nodes are at revision 2; 1 nodes are at revision 3; 0 nodes have achieved
new revision 4
status: "True"
type: NodeInstallerProgressing
- lastTransitionTime: "2020-05-13T11:47:09Z"
message: |-
nodes/mjudeikis3-8bcbz-master-1 pods/kube-apiserver-mjudeikis3-8bcbz-master-1 container="kube-apiserver" is not ready
nodes/mjudeikis3-8bcbz-master-1 pods/kube-apiserver-mjudeikis3-8bcbz-master-1 container="kube-apiserver" is waiting: "CrashLoopBackOff" - "back-off 20s restarting failed container=kube-apiserver pod=kube-apiserver-mjudeikis3-8bcbz-master-1_openshift-kube-apiserver(458b630d98792d8c21612b955384b7e7)"
reason: Error
status: "True"
type: StaticPodsDegraded
```
We might want to consider those as metrics/logs for failed cluster install debugging when we don't have access to API server to use geneva.
| non_build | apiserver status to metrics apiserver status has very good insights on install failures and modes oc get kubeapiservers cluster o yaml lasttransitiontime message nodes are active nodes are at revision nodes are at revision nodes have achieved new revision status true type staticpodsavailable lasttransitiontime message nodes are at revision nodes are at revision nodes have achieved new revision status true type nodeinstallerprogressing lasttransitiontime message nodes master pods kube apiserver master container kube apiserver is not ready nodes master pods kube apiserver master container kube apiserver is waiting crashloopbackoff back off restarting failed container kube apiserver pod kube apiserver master openshift kube apiserver reason error status true type staticpodsdegraded we might want to consider those as metrics logs for failed cluster install debugging when we don t have access to api server to use geneva | 0 |
653,135 | 21,572,876,372 | IssuesEvent | 2022-05-02 10:23:22 | clarin-eric/standards | https://api.github.com/repos/clarin-eric/standards | closed | add a schema for domain data | priority SIS:schemas | We need a schema to constrain the domain data (including domain groupings). It's not a critical addition, but ... | 1.0 | add a schema for domain data - We need a schema to constrain the domain data (including domain groupings). It's not a critical addition, but ... | non_build | add a schema for domain data we need a schema to constrain the domain data including domain groupings it s not a critical addition but | 0 |
17,162 | 22,739,148,923 | IssuesEvent | 2022-07-07 00:44:42 | MPMG-DCC-UFMG/C01 | https://api.github.com/repos/MPMG-DCC-UFMG/C01 | closed | Erro código 1006 ao coletar algumas páginas | [1] Bug [2] Baixa Prioridade [0] Desenvolvimento [3] Processamento Dinâmico | # Comportamento Esperado
Realizar as coletas sem erro algum.
# Comportamento Atual
As coletas apresentam o erro code = 1006. Na maioria das vezes este erro não impede o funcionamento do coletor.
Segue log exemplo para este problema:
`2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - event = eof_received()
2021-08-03 19:55:04 [websockets.protocol] DEBUG: client ! failing CLOSING WebSocket connection with code 1006
2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - event = connection_lost(None)
2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - state = CLOSED
2021-08-03 19:55:04 [websockets.protocol] DEBUG: client x code = 1006, reason = [no reason]
2021-08-03 19:55:04 [scrapy.statscollectors] INFO: Dumping Scrapy stats:
{'elapsed_time_seconds': 0.0047,
'finish_reason': 'finished',
'finish_time': datetime.datetime(2021, 8, 3, 19, 54, 57, 215496),
'log_count/DEBUG': 116,
'log_count/ERROR': 1,
'log_count/INFO': 10,
'log_count/WARNING': 3,
'memusage/max': 302215168,
'memusage/startup': 302215168,
'start_time': datetime.datetime(2021, 8, 3, 19, 54, 57, 210796)}
2021-08-03 19:55:04 [scrapy.core.engine] INFO: Spider closed (finished)`
# Passos para reproduzir o erro
Para recriar o problema, basta criar um coletor para a fonte "http://www.congonhas.mg.gov.br/index.php/licitacao-publica-prefeitura/", com um passo dinâmico de espera e um de salvar página.
Pode ter alguma relação com o seguinte problema: https://github.com/miyakogi/pyppeteer/issues/62 | 1.0 | Erro código 1006 ao coletar algumas páginas - # Comportamento Esperado
Realizar as coletas sem erro algum.
# Comportamento Atual
As coletas apresentam o erro code = 1006. Na maioria das vezes este erro não impede o funcionamento do coletor.
Segue log exemplo para este problema:
`2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - event = eof_received()
2021-08-03 19:55:04 [websockets.protocol] DEBUG: client ! failing CLOSING WebSocket connection with code 1006
2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - event = connection_lost(None)
2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - state = CLOSED
2021-08-03 19:55:04 [websockets.protocol] DEBUG: client x code = 1006, reason = [no reason]
2021-08-03 19:55:04 [scrapy.statscollectors] INFO: Dumping Scrapy stats:
{'elapsed_time_seconds': 0.0047,
'finish_reason': 'finished',
'finish_time': datetime.datetime(2021, 8, 3, 19, 54, 57, 215496),
'log_count/DEBUG': 116,
'log_count/ERROR': 1,
'log_count/INFO': 10,
'log_count/WARNING': 3,
'memusage/max': 302215168,
'memusage/startup': 302215168,
'start_time': datetime.datetime(2021, 8, 3, 19, 54, 57, 210796)}
2021-08-03 19:55:04 [scrapy.core.engine] INFO: Spider closed (finished)`
# Passos para reproduzir o erro
Para recriar o problema, basta criar um coletor para a fonte "http://www.congonhas.mg.gov.br/index.php/licitacao-publica-prefeitura/", com um passo dinâmico de espera e um de salvar página.
Pode ter alguma relação com o seguinte problema: https://github.com/miyakogi/pyppeteer/issues/62 | non_build | erro código ao coletar algumas páginas comportamento esperado realizar as coletas sem erro algum comportamento atual as coletas apresentam o erro code na maioria das vezes este erro não impede o funcionamento do coletor segue log exemplo para este problema debug client event eof received debug client failing closing websocket connection with code debug client event connection lost none debug client state closed debug client x code reason info dumping scrapy stats elapsed time seconds finish reason finished finish time datetime datetime log count debug log count error log count info log count warning memusage max memusage startup start time datetime datetime info spider closed finished passos para reproduzir o erro para recriar o problema basta criar um coletor para a fonte com um passo dinâmico de espera e um de salvar página pode ter alguma relação com o seguinte problema | 0 |
90,394 | 26,080,597,426 | IssuesEvent | 2022-12-25 09:23:59 | OpenShot/openshot-qt | https://api.github.com/repos/OpenShot/openshot-qt | closed | Audio is not synced after exporting video | media-handling 📅 Daily Build | **Describe the bug:**
I created a video with the latest daily build. The preview of the video is working just fine, that is, the video and the audio are in sync. Exporting the video with:
* Target: MP4 (h.264)
* Video Profile (HD 720p 30 fps (1280x720)
* Quality (High)
Result: the audio and video are out of sync
Tried the same with 60 fps and then it is even worse. The audio starts later than the video.
**Expected behavior:**
Exported video and audio are in sync.
**System Details:**
- OpenShot Version [e.g. 2.4.3]: v2.6.1-dev-daily-9975-b890417a-46255e46, build date: 2022-10-20 20:23
- Operating System / Distro: Ubuntu 22.04.1 LTS
**Log Files:**
If you are experiencing a crash, please collect and attach logs of the problem. Log files can be found in the `.openshot_qt` folder in your user home directory. Log files over 2 MB in size will need to be compressed, please attach a ZIP file instead of the raw logs.
**Exception / Stacktrace:**
No crash or exception
| 1.0 | Audio is not synced after exporting video - **Describe the bug:**
I created a video with the latest daily build. The preview of the video is working just fine, that is, the video and the audio are in sync. Exporting the video with:
* Target: MP4 (h.264)
* Video Profile (HD 720p 30 fps (1280x720)
* Quality (High)
Result: the audio and video are out of sync
Tried the same with 60 fps and then it is even worse. The audio starts later than the video.
**Expected behavior:**
Exported video and audio are in sync.
**System Details:**
- OpenShot Version [e.g. 2.4.3]: v2.6.1-dev-daily-9975-b890417a-46255e46, build date: 2022-10-20 20:23
- Operating System / Distro: Ubuntu 22.04.1 LTS
**Log Files:**
If you are experiencing a crash, please collect and attach logs of the problem. Log files can be found in the `.openshot_qt` folder in your user home directory. Log files over 2 MB in size will need to be compressed, please attach a ZIP file instead of the raw logs.
**Exception / Stacktrace:**
No crash or exception
| build | audio is not synced after exporting video describe the bug i created a video with the latest daily build the preview of the video is working just fine that is the video and the audio are in sync exporting the video with target h video profile hd fps quality high result the audio and video are out of sync tried the same with fps and then it is even worse the audio starts later than the video expected behavior exported video and audio are in sync system details openshot version dev daily build date operating system distro ubuntu lts log files if you are experiencing a crash please collect and attach logs of the problem log files can be found in the openshot qt folder in your user home directory log files over mb in size will need to be compressed please attach a zip file instead of the raw logs exception stacktrace no crash or exception | 1 |
692,461 | 23,735,440,669 | IssuesEvent | 2022-08-31 07:41:12 | magento/magento2 | https://api.github.com/repos/magento/magento2 | closed | [Issue] Fixed the welcome message contains single quote character | Issue: Confirmed Component: Theme Reproduced on 2.4.x Progress: PR in progress Priority: P1 Area: Account | This issue is automatically created based on existing pull request: magento/magento2#34061: Fixed the welcome message contains single quote character
---------
<!---
Thank you for contributing to Magento.
To help us process this pull request we recommend that you add the following information:
- Summary of the pull request,
- Issue(s) related to the changes made,
- Manual testing scenarios
Fields marked with (*) are required. Please don't remove the template.
-->
<!--- Please provide a general summary of the Pull Request in the Title above -->
### Description (*)
This PR is fixed for the issue, that if we put the single quote character to the header welcome message config, it will lead the frontend to have a bunch of console errors:
> Uncaught SyntaxError: Unable to process binding "ifnot: function(){return customer().fullname }"
Message: Unable to parse bindings.
### Manual testing scenarios (*)
<!---
Please provide a set of unambiguous steps to test the proposed code change.
Giving us manual testing scenarios will help with the processing and validation process.
-->
1. From Admin, navigate to Content -> Design -> Configuration
2. Select a theme and Edit
3. In the edit page, expand the Header setting, put the `Welcome Text`, like "Welcome Khoi's store!"
4. Clear cache, then go to frontend, you'll see a JS error, and some JS functions not working because of this issue.
### Questions or comments
Maybe we need to bind the message as a text, instead of HTML binding, like:
> data-bind="text: '<?= $escaper->escapeHtmlAttr($welcomeMessage) ?>'"
### Contribution checklist (*)
- [ ] Pull request has a meaningful description of its purpose
- [ ] All commits are accompanied by meaningful commit messages
- [ ] All new or changed code is covered with unit/integration tests (if applicable)
- [ ] README.md files for modified modules are updated and included in the pull request if any [README.md predefined sections](https://github.com/magento/devdocs/wiki/Magento-module-README.md) require an update
- [ ] All automated tests passed successfully (all builds are green)
| 1.0 | [Issue] Fixed the welcome message contains single quote character - This issue is automatically created based on existing pull request: magento/magento2#34061: Fixed the welcome message contains single quote character
---------
<!---
Thank you for contributing to Magento.
To help us process this pull request we recommend that you add the following information:
- Summary of the pull request,
- Issue(s) related to the changes made,
- Manual testing scenarios
Fields marked with (*) are required. Please don't remove the template.
-->
<!--- Please provide a general summary of the Pull Request in the Title above -->
### Description (*)
This PR is fixed for the issue, that if we put the single quote character to the header welcome message config, it will lead the frontend to have a bunch of console errors:
> Uncaught SyntaxError: Unable to process binding "ifnot: function(){return customer().fullname }"
Message: Unable to parse bindings.
### Manual testing scenarios (*)
<!---
Please provide a set of unambiguous steps to test the proposed code change.
Giving us manual testing scenarios will help with the processing and validation process.
-->
1. From Admin, navigate to Content -> Design -> Configuration
2. Select a theme and Edit
3. In the edit page, expand the Header setting, put the `Welcome Text`, like "Welcome Khoi's store!"
4. Clear cache, then go to frontend, you'll see a JS error, and some JS functions not working because of this issue.
### Questions or comments
Maybe we need to bind the message as a text, instead of HTML binding, like:
> data-bind="text: '<?= $escaper->escapeHtmlAttr($welcomeMessage) ?>'"
### Contribution checklist (*)
- [ ] Pull request has a meaningful description of its purpose
- [ ] All commits are accompanied by meaningful commit messages
- [ ] All new or changed code is covered with unit/integration tests (if applicable)
- [ ] README.md files for modified modules are updated and included in the pull request if any [README.md predefined sections](https://github.com/magento/devdocs/wiki/Magento-module-README.md) require an update
- [ ] All automated tests passed successfully (all builds are green)
| non_build | fixed the welcome message contains single quote character this issue is automatically created based on existing pull request magento fixed the welcome message contains single quote character thank you for contributing to magento to help us process this pull request we recommend that you add the following information summary of the pull request issue s related to the changes made manual testing scenarios fields marked with are required please don t remove the template description this pr is fixed for the issue that if we put the single quote character to the header welcome message config it will lead the frontend to have a bunch of console errors uncaught syntaxerror unable to process binding ifnot function return customer fullname message unable to parse bindings manual testing scenarios please provide a set of unambiguous steps to test the proposed code change giving us manual testing scenarios will help with the processing and validation process from admin navigate to content design configuration select a theme and edit in the edit page expand the header setting put the welcome text like welcome khoi s store clear cache then go to frontend you ll see a js error and some js functions not working because of this issue questions or comments maybe we need to bind the message as a text instead of html binding like data bind text escapehtmlattr welcomemessage contribution checklist pull request has a meaningful description of its purpose all commits are accompanied by meaningful commit messages all new or changed code is covered with unit integration tests if applicable readme md files for modified modules are updated and included in the pull request if any require an update all automated tests passed successfully all builds are green | 0 |
66,519 | 16,627,201,458 | IssuesEvent | 2021-06-03 11:08:22 | zalando/restful-api-guidelines | https://api.github.com/repos/zalando/restful-api-guidelines | closed | Change build process to not need a personal access token | build | Currently, the build process to create the web side from the repository is [running on Travis CI](https://travis-ci.org/zalando/restful-api-guidelines), and publishes the result by pushing to the gh_pages branch.
For that, it's using a `GH_TOKEN` variable configured in the [project settings](https://travis-ci.org/zalando/restful-api-guidelines/settings), which is a [personal access token](https://help.github.com/en/github/authenticating-to-github/creating-a-personal-access-token-for-the-command-line).
This token was previously created by @maxim-tschumak, and after he left the organization, and thus lost the write rights, a new one was created today by me (@ePaul).
This token gives read and write rights to all repositories I have access to, in my name. Unfortunately there seems to be no way of creating a token which has access just to a single repository.
This induces those risks:
* If I ever leave the organization, my token automatically loses write access to the repository, and thus breaks the publishing of new versions.
* If something goes wrong and some attacker gets hold of the token, that person can impersonate me on all repositories – both my personal ones, and ones in organizations where I have access.
We should think about setting up the build process in a way that a personal token is not needed. A possible solution might be using a repository-specific [deployment key](https://developer.github.com/v3/guides/managing-deploy-keys/#deploy-keys). | 1.0 | Change build process to not need a personal access token - Currently, the build process to create the web side from the repository is [running on Travis CI](https://travis-ci.org/zalando/restful-api-guidelines), and publishes the result by pushing to the gh_pages branch.
For that, it's using a `GH_TOKEN` variable configured in the [project settings](https://travis-ci.org/zalando/restful-api-guidelines/settings), which is a [personal access token](https://help.github.com/en/github/authenticating-to-github/creating-a-personal-access-token-for-the-command-line).
This token was previously created by @maxim-tschumak, and after he left the organization, and thus lost the write rights, a new one was created today by me (@ePaul).
This token gives read and write rights to all repositories I have access to, in my name. Unfortunately there seems to be no way of creating a token which has access just to a single repository.
This induces those risks:
* If I ever leave the organization, my token automatically loses write access to the repository, and thus breaks the publishing of new versions.
* If something goes wrong and some attacker gets hold of the token, that person can impersonate me on all repositories – both my personal ones, and ones in organizations where I have access.
We should think about setting up the build process in a way that a personal token is not needed. A possible solution might be using a repository-specific [deployment key](https://developer.github.com/v3/guides/managing-deploy-keys/#deploy-keys). | build | change build process to not need a personal access token currently the build process to create the web side from the repository is and publishes the result by pushing to the gh pages branch for that it s using a gh token variable configured in the which is a this token was previously created by maxim tschumak and after he left the organization and thus lost the write rights a new one was created today by me epaul this token gives read and write rights to all repositories i have access to in my name unfortunately there seems to be no way of creating a token which has access just to a single repository this induces those risks if i ever leave the organization my token automatically loses write access to the repository and thus breaks the publishing of new versions if something goes wrong and some attacker gets hold of the token that person can impersonate me on all repositories – both my personal ones and ones in organizations where i have access we should think about setting up the build process in a way that a personal token is not needed a possible solution might be using a repository specific | 1 |
89,741 | 25,894,424,002 | IssuesEvent | 2022-12-14 20:57:48 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | Build 470 for 7.17 with status FAILURE | automation ci-reported Team:Elastic-Agent-Data-Plane build-failures |
## :broken_heart: Tests Failed
<!-- BUILD BADGES-->
> _the below badges are clickable and redirect to their specific view in the CI or DOCS_
[](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//pipeline) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//tests) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//changes) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//artifacts) [](http://beats_null.docs-preview.app.elstc.co/diff) [](https://ci-stats.elastic.co/app/apm/services/beats-ci/transactions/view?rangeFrom=2022-08-25T12:53:35.690Z&rangeTo=2022-08-25T13:13:35.690Z&transactionName=BUILD+Beats%2Fbeats%2F7.17&transactionType=job&latencyAggregationType=avg&traceId=5f1bada60f5f914aa4f76d955c272a38&transactionId=80fe1f1eb38b0665)
<!-- BUILD SUMMARY-->
<details><summary>Expand to view the summary</summary>
<p>
#### Build stats
* Start Time: 2022-08-25T13:03:35.690+0000
* Duration: 113 min 7 sec
#### Test stats :test_tube:
| Test | Results |
| ------------ | :-----------------------------: |
| Failed | 2 |
| Passed | 25709 |
| Skipped | 1884 |
| Total | 27595 |
</p>
</details>
<!-- TEST RESULTS IF ANY-->
### Test errors [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//tests)
<details><summary>Expand to view the tests failures</summary><p>
##### `Build&Test / metricbeat-goIntegTest / TestFetch – github.com/elastic/beats/v7/metricbeat/module/jolokia/jmx`
<ul>
<details><summary>Expand to view the error details</summary><p>
```
Failed
```
</p></details>
<details><summary>Expand to view the stacktrace</summary><p>
```
=== RUN TestFetch
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:33: failed to start service "jolokia: exit status 1
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:33: failed to start service "jolokia: exit status 1
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:33: failed to start service "jolokia: exit status 1
jmx_integration_test.go:33: getting host for jolokia: no container running for service
--- FAIL: TestFetch (6.64s)
```
</p></details>
</ul>
##### `Build&Test / metricbeat-goIntegTest / TestData – github.com/elastic/beats/v7/metricbeat/module/jolokia/jmx`
<ul>
<details><summary>Expand to view the error details</summary><p>
```
Failed
```
</p></details>
<details><summary>Expand to view the stacktrace</summary><p>
```
=== RUN TestData
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:45: failed to start service "jolokia: exit status 1
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:45: failed to start service "jolokia: exit status 1
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:45: failed to start service "jolokia: exit status 1
jmx_integration_test.go:45: getting host for jolokia: no container running for service
--- FAIL: TestData (6.54s)
```
</p></details>
</ul>
</p></details>
<!-- STEPS ERRORS IF ANY -->
### Steps errors [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//pipeline)
<details><summary>Expand to view the steps failures</summary>
<p>
##### `filebeat-unitTest - mage build unitTest`
<ul>
<li>Took 6 min 51 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14102/log/?start=0">here</a></li>
<li>Description: <code>mage build unitTest</code></l1>
</ul>
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 28 min 46 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15344/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 19 min 50 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/21860/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 19 min 55 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/22367/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `metricbeat-pythonIntegTest - mage pythonIntegTest`
<ul>
<li>Took 2 min 34 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14743/log/?start=0">here</a></li>
<li>Description: <code>mage pythonIntegTest</code></l1>
</ul>
##### `metricbeat-pythonIntegTest - mage pythonIntegTest`
<ul>
<li>Took 0 min 23 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15511/log/?start=0">here</a></li>
<li>Description: <code>mage pythonIntegTest</code></l1>
</ul>
##### `metricbeat-pythonIntegTest - mage pythonIntegTest`
<ul>
<li>Took 0 min 22 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15716/log/?start=0">here</a></li>
<li>Description: <code>mage pythonIntegTest</code></l1>
</ul>
##### `x-pack/filebeat-arm-arm - mage build unitTest`
<ul>
<li>Took 8 min 2 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14923/log/?start=0">here</a></li>
<li>Description: <code>mage build unitTest</code></l1>
</ul>
##### `Error signal`
<ul>
<li>Took 0 min 0 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/22383/log/?start=0">here</a></li>
<li>Description: <code>Error "hudson.AbortException: script returned exit code 1"</code></l1>
</ul>
</p>
</details>
| 1.0 | Build 470 for 7.17 with status FAILURE -
## :broken_heart: Tests Failed
<!-- BUILD BADGES-->
> _the below badges are clickable and redirect to their specific view in the CI or DOCS_
[](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//pipeline) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//tests) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//changes) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//artifacts) [](http://beats_null.docs-preview.app.elstc.co/diff) [](https://ci-stats.elastic.co/app/apm/services/beats-ci/transactions/view?rangeFrom=2022-08-25T12:53:35.690Z&rangeTo=2022-08-25T13:13:35.690Z&transactionName=BUILD+Beats%2Fbeats%2F7.17&transactionType=job&latencyAggregationType=avg&traceId=5f1bada60f5f914aa4f76d955c272a38&transactionId=80fe1f1eb38b0665)
<!-- BUILD SUMMARY-->
<details><summary>Expand to view the summary</summary>
<p>
#### Build stats
* Start Time: 2022-08-25T13:03:35.690+0000
* Duration: 113 min 7 sec
#### Test stats :test_tube:
| Test | Results |
| ------------ | :-----------------------------: |
| Failed | 2 |
| Passed | 25709 |
| Skipped | 1884 |
| Total | 27595 |
</p>
</details>
<!-- TEST RESULTS IF ANY-->
### Test errors [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//tests)
<details><summary>Expand to view the tests failures</summary><p>
##### `Build&Test / metricbeat-goIntegTest / TestFetch – github.com/elastic/beats/v7/metricbeat/module/jolokia/jmx`
<ul>
<details><summary>Expand to view the error details</summary><p>
```
Failed
```
</p></details>
<details><summary>Expand to view the stacktrace</summary><p>
```
=== RUN TestFetch
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:33: failed to start service "jolokia: exit status 1
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:33: failed to start service "jolokia: exit status 1
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:33: failed to start service "jolokia: exit status 1
jmx_integration_test.go:33: getting host for jolokia: no container running for service
--- FAIL: TestFetch (6.64s)
```
</p></details>
</ul>
##### `Build&Test / metricbeat-goIntegTest / TestData – github.com/elastic/beats/v7/metricbeat/module/jolokia/jmx`
<ul>
<details><summary>Expand to view the error details</summary><p>
```
Failed
```
</p></details>
<details><summary>Expand to view the stacktrace</summary><p>
```
=== RUN TestData
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:45: failed to start service "jolokia: exit status 1
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:45: failed to start service "jolokia: exit status 1
Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up.
Building jolokia
Step 1/14 : FROM java:8-jdk-alpine
Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown
jmx_integration_test.go:45: failed to start service "jolokia: exit status 1
jmx_integration_test.go:45: getting host for jolokia: no container running for service
--- FAIL: TestData (6.54s)
```
</p></details>
</ul>
</p></details>
<!-- STEPS ERRORS IF ANY -->
### Steps errors [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//pipeline)
<details><summary>Expand to view the steps failures</summary>
<p>
##### `filebeat-unitTest - mage build unitTest`
<ul>
<li>Took 6 min 51 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14102/log/?start=0">here</a></li>
<li>Description: <code>mage build unitTest</code></l1>
</ul>
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 28 min 46 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15344/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 19 min 50 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/21860/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 19 min 55 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/22367/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `metricbeat-pythonIntegTest - mage pythonIntegTest`
<ul>
<li>Took 2 min 34 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14743/log/?start=0">here</a></li>
<li>Description: <code>mage pythonIntegTest</code></l1>
</ul>
##### `metricbeat-pythonIntegTest - mage pythonIntegTest`
<ul>
<li>Took 0 min 23 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15511/log/?start=0">here</a></li>
<li>Description: <code>mage pythonIntegTest</code></l1>
</ul>
##### `metricbeat-pythonIntegTest - mage pythonIntegTest`
<ul>
<li>Took 0 min 22 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15716/log/?start=0">here</a></li>
<li>Description: <code>mage pythonIntegTest</code></l1>
</ul>
##### `x-pack/filebeat-arm-arm - mage build unitTest`
<ul>
<li>Took 8 min 2 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14923/log/?start=0">here</a></li>
<li>Description: <code>mage build unitTest</code></l1>
</ul>
##### `Error signal`
<ul>
<li>Took 0 min 0 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/22383/log/?start=0">here</a></li>
<li>Description: <code>Error "hudson.AbortException: script returned exit code 1"</code></l1>
</ul>
</p>
</details>
| build | build for with status failure broken heart tests failed the below badges are clickable and redirect to their specific view in the ci or docs expand to view the summary build stats start time duration min sec test stats test tube test results failed passed skipped total test errors expand to view the tests failures build test metricbeat gointegtest testfetch – github com elastic beats metricbeat module jolokia jmx expand to view the error details failed expand to view the stacktrace run testfetch found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status jmx integration test go getting host for jolokia no container running for service fail testfetch build test metricbeat gointegtest testdata – github com elastic beats metricbeat module jolokia jmx expand to view the error details failed expand to view the stacktrace run testdata found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status jmx integration test go getting host for jolokia no container running for service fail testdata steps errors expand to view the steps failures filebeat unittest mage build unittest took min sec view more details a href description mage build unittest metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest metricbeat pythonintegtest mage pythonintegtest took min sec view more details a href description mage pythonintegtest metricbeat pythonintegtest mage pythonintegtest took min sec view more details a href description mage pythonintegtest metricbeat pythonintegtest mage pythonintegtest took min sec view more details a href description mage pythonintegtest x pack filebeat arm arm mage build unittest took min sec view more details a href description mage build unittest error signal took min sec view more details a href description error hudson abortexception script returned exit code | 1 |
72,707 | 31,769,020,249 | IssuesEvent | 2023-09-12 10:32:16 | gauravrs18/issue_onboarding | https://api.github.com/repos/gauravrs18/issue_onboarding | closed | dev-angular-code-account-services-new-connection-component-approve-component
-consumer-details-component
-connect-component
-reject-button-component | CX-account-services | dev-angular-code-account-services-new-connection-component-approve-component
-consumer-details-component
-connect-component
-reject-button-component | 1.0 | dev-angular-code-account-services-new-connection-component-approve-component
-consumer-details-component
-connect-component
-reject-button-component - dev-angular-code-account-services-new-connection-component-approve-component
-consumer-details-component
-connect-component
-reject-button-component | non_build | dev angular code account services new connection component approve component consumer details component connect component reject button component dev angular code account services new connection component approve component consumer details component connect component reject button component | 0 |
62,332 | 15,219,448,375 | IssuesEvent | 2021-02-17 19:13:07 | spring-projects/spring-security | https://api.github.com/repos/spring-projects/spring-security | closed | Update Git workflows to require JDK 11 | in: build type: enhancement | As part of updating the build to require JDK 11, the Git workflows need to change. | 1.0 | Update Git workflows to require JDK 11 - As part of updating the build to require JDK 11, the Git workflows need to change. | build | update git workflows to require jdk as part of updating the build to require jdk the git workflows need to change | 1 |
14,467 | 17,571,238,762 | IssuesEvent | 2021-08-14 18:43:02 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | Add glob or recursive option for processor `drop_fields` | enhancement libbeat :Processors Team:Integrations Stalled | **Describe the enhancement:**
Add glob or recursive option for processor `drop_fields` which would be able to check the entire document for the field to remove.
**Describe a specific use case for the enhancement or feature:**
The processor works only with statically defined list of the fields and also only with fields at root level. If user needs to check the entire document for a field and does not know exact position, the only option is a script.
Thank you | 1.0 | Add glob or recursive option for processor `drop_fields` - **Describe the enhancement:**
Add glob or recursive option for processor `drop_fields` which would be able to check the entire document for the field to remove.
**Describe a specific use case for the enhancement or feature:**
The processor works only with statically defined list of the fields and also only with fields at root level. If user needs to check the entire document for a field and does not know exact position, the only option is a script.
Thank you | non_build | add glob or recursive option for processor drop fields describe the enhancement add glob or recursive option for processor drop fields which would be able to check the entire document for the field to remove describe a specific use case for the enhancement or feature the processor works only with statically defined list of the fields and also only with fields at root level if user needs to check the entire document for a field and does not know exact position the only option is a script thank you | 0 |
55,690 | 13,650,581,899 | IssuesEvent | 2020-09-26 19:52:51 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | closed | Add `x.py setup` or similar | A-contributor-roadblock A-rustbuild C-feature-request | It has been [impressed](https://github.com/rust-lang/rust/issues/76371#issuecomment-687618541) [into](https://github.com/rust-lang/rust/issues/76165#issuecomment-687634489) [me](https://github.com/rust-lang/rust/pull/73964#issuecomment-653134345) [several](https://github.com/rust-lang/rust/pull/76446#issuecomment-688417692) [times](https://github.com/rust-lang/rust/pull/76417#issuecomment-687881451) that there are no good defaults for everyone. Instead, what if people could choose their own defaults, but in a guided way? I'm imagining something like this:
```
$ x.py setup
Welcome to the Rust project! What do you want to do with x.py?
a) Contribute to the standard library
b) Contribute to the compiler
c) Install Rust from source
Please choose one (a/b/c): b
Changed the following settings:
- `incremental = true`
- `debug = true`
Do you plan to modify codegen/LLVM? (y/[n]): n
# maybe some magic with $(which llvm-config) here; but won't be necessary after https://github.com/rust-lang/rust/pull/76349
To get started, try one of the following commands:
- `x.py check`
- `x.py build library/std`
- `x.py test src/test/ui`
```
This would be entirely opt-in, if you don't like it, just don't run the command and modify `config.toml.example` as before. This would work best in conjunction with #76165, but does not require it.
cc @Lokathor, @thomcc | 1.0 | Add `x.py setup` or similar - It has been [impressed](https://github.com/rust-lang/rust/issues/76371#issuecomment-687618541) [into](https://github.com/rust-lang/rust/issues/76165#issuecomment-687634489) [me](https://github.com/rust-lang/rust/pull/73964#issuecomment-653134345) [several](https://github.com/rust-lang/rust/pull/76446#issuecomment-688417692) [times](https://github.com/rust-lang/rust/pull/76417#issuecomment-687881451) that there are no good defaults for everyone. Instead, what if people could choose their own defaults, but in a guided way? I'm imagining something like this:
```
$ x.py setup
Welcome to the Rust project! What do you want to do with x.py?
a) Contribute to the standard library
b) Contribute to the compiler
c) Install Rust from source
Please choose one (a/b/c): b
Changed the following settings:
- `incremental = true`
- `debug = true`
Do you plan to modify codegen/LLVM? (y/[n]): n
# maybe some magic with $(which llvm-config) here; but won't be necessary after https://github.com/rust-lang/rust/pull/76349
To get started, try one of the following commands:
- `x.py check`
- `x.py build library/std`
- `x.py test src/test/ui`
```
This would be entirely opt-in, if you don't like it, just don't run the command and modify `config.toml.example` as before. This would work best in conjunction with #76165, but does not require it.
cc @Lokathor, @thomcc | build | add x py setup or similar it has been that there are no good defaults for everyone instead what if people could choose their own defaults but in a guided way i m imagining something like this x py setup welcome to the rust project what do you want to do with x py a contribute to the standard library b contribute to the compiler c install rust from source please choose one a b c b changed the following settings incremental true debug true do you plan to modify codegen llvm y n maybe some magic with which llvm config here but won t be necessary after to get started try one of the following commands x py check x py build library std x py test src test ui this would be entirely opt in if you don t like it just don t run the command and modify config toml example as before this would work best in conjunction with but does not require it cc lokathor thomcc | 1 |
86,871 | 3,734,396,324 | IssuesEvent | 2016-03-08 06:36:53 | notsecure/uTox | https://api.github.com/repos/notsecure/uTox | closed | Avatars should look better | enhancement Friends mid_priority user_interface | Two things:
- They should be resized to 40x40 using cubic filtering(maybe this can be stolen from GIMP's source code?) so they look better(and look the same for every OS, currently Windows' built-in filtering looks worse)
- They should use rounded corners as in the mockup: https://github.com/ItsDuke/Tox-UI | 1.0 | Avatars should look better - Two things:
- They should be resized to 40x40 using cubic filtering(maybe this can be stolen from GIMP's source code?) so they look better(and look the same for every OS, currently Windows' built-in filtering looks worse)
- They should use rounded corners as in the mockup: https://github.com/ItsDuke/Tox-UI | non_build | avatars should look better two things they should be resized to using cubic filtering maybe this can be stolen from gimp s source code so they look better and look the same for every os currently windows built in filtering looks worse they should use rounded corners as in the mockup | 0 |
127,451 | 18,010,479,620 | IssuesEvent | 2021-09-16 08:01:32 | maddyCode23/linux-4.1.15 | https://api.github.com/repos/maddyCode23/linux-4.1.15 | opened | CVE-2019-19927 (Medium) detected in linux-yocto-4.1v4.1.17 | security vulnerability | ## CVE-2019-19927 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.1v4.1.17</b></p></summary>
<p>
<p>[no description]</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.1>https://git.yoctoproject.org/git/linux-yocto-4.1</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/gpu/drm/ttm/ttm_page_alloc.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/gpu/drm/ttm/ttm_page_alloc.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel 5.0.0-rc7 (as distributed in ubuntu/linux.git on kernel.ubuntu.com), mounting a crafted f2fs filesystem image and performing some operations can lead to slab-out-of-bounds read access in ttm_put_pages in drivers/gpu/drm/ttm/ttm_page_alloc.c. This is related to the vmwgfx or ttm module.
<p>Publish Date: 2019-12-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19927>CVE-2019-19927</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/torvalds/linux/tree/v5.1-rc6">https://github.com/torvalds/linux/tree/v5.1-rc6</a></p>
<p>Release Date: 2019-12-31</p>
<p>Fix Resolution: v5.1-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-19927 (Medium) detected in linux-yocto-4.1v4.1.17 - ## CVE-2019-19927 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.1v4.1.17</b></p></summary>
<p>
<p>[no description]</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.1>https://git.yoctoproject.org/git/linux-yocto-4.1</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/gpu/drm/ttm/ttm_page_alloc.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/gpu/drm/ttm/ttm_page_alloc.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel 5.0.0-rc7 (as distributed in ubuntu/linux.git on kernel.ubuntu.com), mounting a crafted f2fs filesystem image and performing some operations can lead to slab-out-of-bounds read access in ttm_put_pages in drivers/gpu/drm/ttm/ttm_page_alloc.c. This is related to the vmwgfx or ttm module.
<p>Publish Date: 2019-12-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19927>CVE-2019-19927</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/torvalds/linux/tree/v5.1-rc6">https://github.com/torvalds/linux/tree/v5.1-rc6</a></p>
<p>Release Date: 2019-12-31</p>
<p>Fix Resolution: v5.1-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_build | cve medium detected in linux yocto cve medium severity vulnerability vulnerable library linux yocto library home page a href found in head commit a href vulnerable source files drivers gpu drm ttm ttm page alloc c drivers gpu drm ttm ttm page alloc c vulnerability details in the linux kernel as distributed in ubuntu linux git on kernel ubuntu com mounting a crafted filesystem image and performing some operations can lead to slab out of bounds read access in ttm put pages in drivers gpu drm ttm ttm page alloc c this is related to the vmwgfx or ttm module publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
137,179 | 30,643,463,371 | IssuesEvent | 2023-07-25 01:15:26 | libai8723/front-end-dev-notes-bignerdbook | https://api.github.com/repos/libai8723/front-end-dev-notes-bignerdbook | closed | LeetCode. 1218. Longest Arithmetic Subsequence of Given Difference | DP LeetCode | 拥有给定公差的最长算术子序列。
https://leetcode.com/problems/longest-arithmetic-subsequence-of-given-difference/description/
这个题算是做完了,耗费了我前前后后一周的时间,主要还是想法不对。一开始没有想到DP的做法,而是直接想到了排序,二分,外加缓存的优化,所以虽然还是accept了,但是整体的性能差距很大。
### 1. 朴素的想法
1. 首先把数组排序了,排序按照数字的大小排序,其次按照数字本身在arr的序号排序,这样相同值的数字,最小序号的都排在前面了。
2. 对于同样取值的数字,如果是以这类数字作为起点的话,那么肯定是找序号最好的那个开始就可以了,因为序号大的同数值类型的作为起点,肯定长度不会超过序号小的。
3. 如果一个数字在前序的搜索中被用过了,那么就要考虑排除了。因为要么之前已经搜索过了,要么就是搜索到这个数字的时候,对应的长度肯定没办法超过之前的了。这就是我的缓存的原理。
### 2. DP的搞法
1. DP的核心都是空间换时间,那么这道题的想法就是,对于已经看过的arr[0]~arr[i],定义一个dp函数:dp(number) = max_length
2. 这里的这个公式的含义比较特殊,这里的number是arr[0~i]这个区间内的一个数字,dp(number)的含义是,以number作为结束的,以difference为公差的,最长子数组的长度。
3. 这里我想过一些问题,那么如果number在arr[0~i]区间有重复怎么办?好像也不影响,这个函数只是表达,存在一个number结尾的,公差是difference的数组,长度是max_length
4. 解决了这个公式的定义之后,现在看一下怎么存储,因为入参只有一个number,取值也是一个整数max_length,这样搞一个map就可以了,比较高效。
5. 然后再看当i扩展到i+1的时候怎么办呢?number = arr[i+1],这时候要回看,已有的map(其实就是dp函数的定义)中是否有一个dp(number - difference)的定义,如果有的话,那么dp(number) = dp(number -difference) + 1, 同时判断一下,dp(number)是否大于accumulated_max_length,如果大于的话,就更新accumulated_max_length,如果dp(number - difference)不存在,就默认dp(number)为1.
6. 这样一路推导下去,就搞定了,map的最大的大小就是arr中distinct value count的大小,在空间上也是非常高效的。
| 1.0 | LeetCode. 1218. Longest Arithmetic Subsequence of Given Difference - 拥有给定公差的最长算术子序列。
https://leetcode.com/problems/longest-arithmetic-subsequence-of-given-difference/description/
这个题算是做完了,耗费了我前前后后一周的时间,主要还是想法不对。一开始没有想到DP的做法,而是直接想到了排序,二分,外加缓存的优化,所以虽然还是accept了,但是整体的性能差距很大。
### 1. 朴素的想法
1. 首先把数组排序了,排序按照数字的大小排序,其次按照数字本身在arr的序号排序,这样相同值的数字,最小序号的都排在前面了。
2. 对于同样取值的数字,如果是以这类数字作为起点的话,那么肯定是找序号最好的那个开始就可以了,因为序号大的同数值类型的作为起点,肯定长度不会超过序号小的。
3. 如果一个数字在前序的搜索中被用过了,那么就要考虑排除了。因为要么之前已经搜索过了,要么就是搜索到这个数字的时候,对应的长度肯定没办法超过之前的了。这就是我的缓存的原理。
### 2. DP的搞法
1. DP的核心都是空间换时间,那么这道题的想法就是,对于已经看过的arr[0]~arr[i],定义一个dp函数:dp(number) = max_length
2. 这里的这个公式的含义比较特殊,这里的number是arr[0~i]这个区间内的一个数字,dp(number)的含义是,以number作为结束的,以difference为公差的,最长子数组的长度。
3. 这里我想过一些问题,那么如果number在arr[0~i]区间有重复怎么办?好像也不影响,这个函数只是表达,存在一个number结尾的,公差是difference的数组,长度是max_length
4. 解决了这个公式的定义之后,现在看一下怎么存储,因为入参只有一个number,取值也是一个整数max_length,这样搞一个map就可以了,比较高效。
5. 然后再看当i扩展到i+1的时候怎么办呢?number = arr[i+1],这时候要回看,已有的map(其实就是dp函数的定义)中是否有一个dp(number - difference)的定义,如果有的话,那么dp(number) = dp(number -difference) + 1, 同时判断一下,dp(number)是否大于accumulated_max_length,如果大于的话,就更新accumulated_max_length,如果dp(number - difference)不存在,就默认dp(number)为1.
6. 这样一路推导下去,就搞定了,map的最大的大小就是arr中distinct value count的大小,在空间上也是非常高效的。
| non_build | leetcode longest arithmetic subsequence of given difference 拥有给定公差的最长算术子序列。 这个题算是做完了,耗费了我前前后后一周的时间,主要还是想法不对。一开始没有想到dp的做法,而是直接想到了排序,二分,外加缓存的优化,所以虽然还是accept了,但是整体的性能差距很大。 朴素的想法 首先把数组排序了,排序按照数字的大小排序,其次按照数字本身在arr的序号排序,这样相同值的数字,最小序号的都排在前面了。 对于同样取值的数字,如果是以这类数字作为起点的话,那么肯定是找序号最好的那个开始就可以了,因为序号大的同数值类型的作为起点,肯定长度不会超过序号小的。 如果一个数字在前序的搜索中被用过了,那么就要考虑排除了。因为要么之前已经搜索过了,要么就是搜索到这个数字的时候,对应的长度肯定没办法超过之前的了。这就是我的缓存的原理。 dp的搞法 dp的核心都是空间换时间,那么这道题的想法就是,对于已经看过的arr arr ,定义一个dp函数:dp number max length 这里的这个公式的含义比较特殊,这里的number是arr 这个区间内的一个数字,dp number 的含义是,以number作为结束的,以difference为公差的,最长子数组的长度。 这里我想过一些问题,那么如果number在arr 区间有重复怎么办?好像也不影响,这个函数只是表达,存在一个number结尾的,公差是difference的数组,长度是max length 解决了这个公式的定义之后,现在看一下怎么存储,因为入参只有一个number,取值也是一个整数max length,这样搞一个map就可以了,比较高效。 然后再看当i扩展到i ?number arr ,这时候要回看,已有的map(其实就是dp函数的定义)中是否有一个dp number difference 的定义,如果有的话,那么dp number dp number difference , 同时判断一下,dp number 是否大于accumulated max length,如果大于的话,就更新accumulated max length,如果dp number difference 不存在,就默认dp number 这样一路推导下去,就搞定了,map的最大的大小就是arr中distinct value count的大小,在空间上也是非常高效的。 | 0 |
36,442 | 17,692,740,580 | IssuesEvent | 2021-08-24 12:06:52 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | closed | Global sound issues | Performance Sound Bug | So for a while now there's been a number off issues around admin loaded sounds.
The first is that some sounds, such as emitters firing, will interrupt the sound being played, rather like there aren't enough channels to play both sounds at the same time. Its particularly noticeable when multiple emitters are set up and it will make it stutter, due to their short sound length and regular triggering of said effect.
There also seems to be some major issue of lag being created when they are uploaded. It will cause the server to hang for at least 1 second during a full 1mb song with a population of around 60 or higher (havent tried it on low pop).
Its hard to tell if its just a population increase thats effecting it, but Im fairly certain it didnt used to be this bad in terms of lag.
I realise this issue is relatively vague but its more a call to arms to do some further testing and seeing if anyone might know of what might have changed to the code of late (never noticed these issues before number value sound levels became a thing (i.e. when it was just on or off as a pref) | True | Global sound issues - So for a while now there's been a number off issues around admin loaded sounds.
The first is that some sounds, such as emitters firing, will interrupt the sound being played, rather like there aren't enough channels to play both sounds at the same time. Its particularly noticeable when multiple emitters are set up and it will make it stutter, due to their short sound length and regular triggering of said effect.
There also seems to be some major issue of lag being created when they are uploaded. It will cause the server to hang for at least 1 second during a full 1mb song with a population of around 60 or higher (havent tried it on low pop).
Its hard to tell if its just a population increase thats effecting it, but Im fairly certain it didnt used to be this bad in terms of lag.
I realise this issue is relatively vague but its more a call to arms to do some further testing and seeing if anyone might know of what might have changed to the code of late (never noticed these issues before number value sound levels became a thing (i.e. when it was just on or off as a pref) | non_build | global sound issues so for a while now there s been a number off issues around admin loaded sounds the first is that some sounds such as emitters firing will interrupt the sound being played rather like there aren t enough channels to play both sounds at the same time its particularly noticeable when multiple emitters are set up and it will make it stutter due to their short sound length and regular triggering of said effect there also seems to be some major issue of lag being created when they are uploaded it will cause the server to hang for at least second during a full song with a population of around or higher havent tried it on low pop its hard to tell if its just a population increase thats effecting it but im fairly certain it didnt used to be this bad in terms of lag i realise this issue is relatively vague but its more a call to arms to do some further testing and seeing if anyone might know of what might have changed to the code of late never noticed these issues before number value sound levels became a thing i e when it was just on or off as a pref | 0 |
35,206 | 16,985,382,652 | IssuesEvent | 2021-06-30 13:52:30 | supaero-aircraft-design/FAST-GA | https://api.github.com/repos/supaero-aircraft-design/FAST-GA | closed | Calculer et utiliser au bon endroit la trainée des trains d'atterrissage | aerodynamics methodology performances | Principalement au décollage.
On part du principe que pour un train d'atterrissage non rétractable on paye la pénalité de traînée à tout moment.
Pour un train rétractable, le cd0 du aux trains n'est additionné que sur le phases basse vitesse. Sachant que la montée et la descente sont des phases en haute-vitesse, ça reste juste. | True | Calculer et utiliser au bon endroit la trainée des trains d'atterrissage - Principalement au décollage.
On part du principe que pour un train d'atterrissage non rétractable on paye la pénalité de traînée à tout moment.
Pour un train rétractable, le cd0 du aux trains n'est additionné que sur le phases basse vitesse. Sachant que la montée et la descente sont des phases en haute-vitesse, ça reste juste. | non_build | calculer et utiliser au bon endroit la trainée des trains d atterrissage principalement au décollage on part du principe que pour un train d atterrissage non rétractable on paye la pénalité de traînée à tout moment pour un train rétractable le du aux trains n est additionné que sur le phases basse vitesse sachant que la montée et la descente sont des phases en haute vitesse ça reste juste | 0 |
653,597 | 21,608,147,096 | IssuesEvent | 2022-05-04 07:09:46 | cyrusae/highlighter-public | https://api.github.com/repos/cyrusae/highlighter-public | opened | Clean paragraphs on the go | bug database frontend Priority: ++ | Doesn't appear to be reliably saving, so may want to revisit how that's working | 1.0 | Clean paragraphs on the go - Doesn't appear to be reliably saving, so may want to revisit how that's working | non_build | clean paragraphs on the go doesn t appear to be reliably saving so may want to revisit how that s working | 0 |
334,046 | 29,820,341,007 | IssuesEvent | 2023-06-17 01:30:45 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix device.test_as_ivy_dev | Sub Task Failing Test | | | |
|---|---|
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix device.test_as_ivy_dev - | | |
|---|---|
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a>
| non_build | fix device test as ivy dev jax a href src torch a href src tensorflow a href src numpy a href src paddle a href src | 0 |
238,387 | 18,240,125,839 | IssuesEvent | 2021-10-01 11:55:28 | hashicorp/terraform-ls | https://api.github.com/repos/hashicorp/terraform-ls | opened | docs: Reflect tfvars support in USAGE (client setup guides) | documentation | Currently we do not actively encourage clients to detect and send tfvars files to the language server here:
https://github.com/hashicorp/terraform-ls/blob/main/docs/USAGE.md
The only mention so far is in the dev-oriented docs:
https://github.com/hashicorp/terraform-ls/blob/main/docs/language-clients.md
| 1.0 | docs: Reflect tfvars support in USAGE (client setup guides) - Currently we do not actively encourage clients to detect and send tfvars files to the language server here:
https://github.com/hashicorp/terraform-ls/blob/main/docs/USAGE.md
The only mention so far is in the dev-oriented docs:
https://github.com/hashicorp/terraform-ls/blob/main/docs/language-clients.md
| non_build | docs reflect tfvars support in usage client setup guides currently we do not actively encourage clients to detect and send tfvars files to the language server here the only mention so far is in the dev oriented docs | 0 |
32,849 | 8,960,682,081 | IssuesEvent | 2019-01-28 07:04:54 | akka/akka | https://api.github.com/repos/akka/akka | closed | sbt 1.2.7 broke unidoc/javadoc with jdk9 | t:build | The update to sbt 1.2.7 in https://github.com/akka/akka/commit/6ba4c341f5d2a090c28e9ec92d0ea05f7188de61 broke unidoc/javadoc
Reproduce with sbt 1.2.7:
```
sbt -Dakka.genjavadoc.enabled=true
> clean
> unidoc
...
javadoc exited with exit code 1
```
I have verified that it was not the `classpath 4.4.12` update that caused it. | 1.0 | sbt 1.2.7 broke unidoc/javadoc with jdk9 - The update to sbt 1.2.7 in https://github.com/akka/akka/commit/6ba4c341f5d2a090c28e9ec92d0ea05f7188de61 broke unidoc/javadoc
Reproduce with sbt 1.2.7:
```
sbt -Dakka.genjavadoc.enabled=true
> clean
> unidoc
...
javadoc exited with exit code 1
```
I have verified that it was not the `classpath 4.4.12` update that caused it. | build | sbt broke unidoc javadoc with the update to sbt in broke unidoc javadoc reproduce with sbt sbt dakka genjavadoc enabled true clean unidoc javadoc exited with exit code i have verified that it was not the classpath update that caused it | 1 |
286,793 | 21,609,480,041 | IssuesEvent | 2022-05-04 08:36:49 | WordPress/Documentation-Issue-Tracker | https://api.github.com/repos/WordPress/Documentation-Issue-Tracker | closed | [HelpHub] Paragraph Block | user documentation 5.7 5.9 block editor | Article: https://wordpress.org/support/article/paragraph-block/
## Updating for 5.7
- [ ] Add Keyboard Input inline format - [26801](https://github.com/WordPress/gutenberg/pull/26801)
## General
- [ ] make sure all screenshots are relevant for current version (5.9)
- [ ] Video updates
- [ ] update changelog at the end of the article
Issue migrated from Trello: https://trello.com/c/mWOy1uu2/1-paragraph-block | 1.0 | [HelpHub] Paragraph Block - Article: https://wordpress.org/support/article/paragraph-block/
## Updating for 5.7
- [ ] Add Keyboard Input inline format - [26801](https://github.com/WordPress/gutenberg/pull/26801)
## General
- [ ] make sure all screenshots are relevant for current version (5.9)
- [ ] Video updates
- [ ] update changelog at the end of the article
Issue migrated from Trello: https://trello.com/c/mWOy1uu2/1-paragraph-block | non_build | paragraph block article updating for add keyboard input inline format general make sure all screenshots are relevant for current version video updates update changelog at the end of the article issue migrated from trello | 0 |
24,178 | 2,666,730,519 | IssuesEvent | 2015-03-21 21:25:16 | Templarian/MaterialDesign | https://api.github.com/repos/Templarian/MaterialDesign | closed | H(eading)1 ... H(eading)6 and P(aragraph) icons for editor | High Priority Icon Request | These are often used in embedded editors and would be nice to have
- [x] format-header-pound
- [x] format-header-1
- [x] format-header-2
- [x] format-header-3
- [x] format-header-4
- [x] format-header-5
- [x] format-header-6
- [x] format-paragraph | 1.0 | H(eading)1 ... H(eading)6 and P(aragraph) icons for editor - These are often used in embedded editors and would be nice to have
- [x] format-header-pound
- [x] format-header-1
- [x] format-header-2
- [x] format-header-3
- [x] format-header-4
- [x] format-header-5
- [x] format-header-6
- [x] format-paragraph | non_build | h eading h eading and p aragraph icons for editor these are often used in embedded editors and would be nice to have format header pound format header format header format header format header format header format header format paragraph | 0 |
56,283 | 13,784,651,070 | IssuesEvent | 2020-10-08 21:14:47 | microsoft/vscode-cpptools | https://api.github.com/repos/microsoft/vscode-cpptools | closed | ${vcpkgRoot} variable is not expanded in tasks.json | Visual Studio Code build/tasks duplicate question | ${vcpkgRoot} variable is expanded in c_cpp_properties.json (for example in includePath) but not in tasks.json.
I need this expansion to write a cmake task command. Something like this:
```
{
"label": "cmake",
"type": "shell",
"command": "cmake",
"args": [
"${workspaceFolder}",
"-B",
"${workspaceFolder}/build",
"-DCMAKE_TOOLCHAIN_FILE=${vcpkgRoot}/scripts/buildsystems/vcpkg.cmake"
],
"group": "build",
"problemMatcher": []
}
```
Thanks | 1.0 | ${vcpkgRoot} variable is not expanded in tasks.json - ${vcpkgRoot} variable is expanded in c_cpp_properties.json (for example in includePath) but not in tasks.json.
I need this expansion to write a cmake task command. Something like this:
```
{
"label": "cmake",
"type": "shell",
"command": "cmake",
"args": [
"${workspaceFolder}",
"-B",
"${workspaceFolder}/build",
"-DCMAKE_TOOLCHAIN_FILE=${vcpkgRoot}/scripts/buildsystems/vcpkg.cmake"
],
"group": "build",
"problemMatcher": []
}
```
Thanks | build | vcpkgroot variable is not expanded in tasks json vcpkgroot variable is expanded in c cpp properties json for example in includepath but not in tasks json i need this expansion to write a cmake task command something like this label cmake type shell command cmake args workspacefolder b workspacefolder build dcmake toolchain file vcpkgroot scripts buildsystems vcpkg cmake group build problemmatcher thanks | 1 |
143,371 | 19,177,931,077 | IssuesEvent | 2021-12-04 00:07:11 | samq-ghdemo/js-monorepo | https://api.github.com/repos/samq-ghdemo/js-monorepo | opened | CVE-2020-7608 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-2.4.1.tgz</b>, <b>yargs-parser-8.1.0.tgz</b>, <b>yargs-parser-9.0.2.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-2.4.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz</a></p>
<p>Path to dependency file: js-monorepo/NodeGoat/package.json</p>
<p>Path to vulnerable library: js-monorepo/NodeGoat/node_modules/nyc/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- realize-package-specifier-3.0.3.tgz (Root Library)
- grunt-contrib-nodeunit-1.0.0.tgz
- nodeunit-0.9.5.tgz
- tap-7.1.2.tgz
- nyc-7.1.0.tgz
- :x: **yargs-parser-2.4.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-8.1.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz</a></p>
<p>Path to dependency file: js-monorepo/nodejs-goof/package.json</p>
<p>Path to vulnerable library: js-monorepo/nodejs-goof/node_modules/nyc/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- tap-11.1.5.tgz (Root Library)
- nyc-11.9.0.tgz
- :x: **yargs-parser-8.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-9.0.2.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz</a></p>
<p>Path to dependency file: js-monorepo/nodejs-goof/package.json</p>
<p>Path to vulnerable library: js-monorepo/nodejs-goof/node_modules/nyc/node_modules/yargs/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- tap-11.1.5.tgz (Root Library)
- nyc-11.9.0.tgz
- yargs-11.1.0.tgz
- :x: **yargs-parser-9.0.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/js-monorepo/commit/f3701923c18333c1e4e49bf595dd36b3f186812f">f3701923c18333c1e4e49bf595dd36b3f186812f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution: 5.0.1;13.1.2;15.0.1;18.1.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"2.4.1","packageFilePaths":["/NodeGoat/package.json"],"isTransitiveDependency":true,"dependencyTree":"realize-package-specifier:3.0.3;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;yargs-parser:2.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"8.1.0","packageFilePaths":["/nodejs-goof/package.json"],"isTransitiveDependency":true,"dependencyTree":"tap:11.1.5;nyc:11.9.0;yargs-parser:8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"9.0.2","packageFilePaths":["/nodejs-goof/package.json"],"isTransitiveDependency":true,"dependencyTree":"tap:11.1.5;nyc:11.9.0;yargs:11.1.0;yargs-parser:9.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-7608 (Medium) detected in multiple libraries - ## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-2.4.1.tgz</b>, <b>yargs-parser-8.1.0.tgz</b>, <b>yargs-parser-9.0.2.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-2.4.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz</a></p>
<p>Path to dependency file: js-monorepo/NodeGoat/package.json</p>
<p>Path to vulnerable library: js-monorepo/NodeGoat/node_modules/nyc/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- realize-package-specifier-3.0.3.tgz (Root Library)
- grunt-contrib-nodeunit-1.0.0.tgz
- nodeunit-0.9.5.tgz
- tap-7.1.2.tgz
- nyc-7.1.0.tgz
- :x: **yargs-parser-2.4.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-8.1.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz</a></p>
<p>Path to dependency file: js-monorepo/nodejs-goof/package.json</p>
<p>Path to vulnerable library: js-monorepo/nodejs-goof/node_modules/nyc/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- tap-11.1.5.tgz (Root Library)
- nyc-11.9.0.tgz
- :x: **yargs-parser-8.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-9.0.2.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz</a></p>
<p>Path to dependency file: js-monorepo/nodejs-goof/package.json</p>
<p>Path to vulnerable library: js-monorepo/nodejs-goof/node_modules/nyc/node_modules/yargs/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- tap-11.1.5.tgz (Root Library)
- nyc-11.9.0.tgz
- yargs-11.1.0.tgz
- :x: **yargs-parser-9.0.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/js-monorepo/commit/f3701923c18333c1e4e49bf595dd36b3f186812f">f3701923c18333c1e4e49bf595dd36b3f186812f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution: 5.0.1;13.1.2;15.0.1;18.1.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"2.4.1","packageFilePaths":["/NodeGoat/package.json"],"isTransitiveDependency":true,"dependencyTree":"realize-package-specifier:3.0.3;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;yargs-parser:2.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"8.1.0","packageFilePaths":["/nodejs-goof/package.json"],"isTransitiveDependency":true,"dependencyTree":"tap:11.1.5;nyc:11.9.0;yargs-parser:8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"9.0.2","packageFilePaths":["/nodejs-goof/package.json"],"isTransitiveDependency":true,"dependencyTree":"tap:11.1.5;nyc:11.9.0;yargs:11.1.0;yargs-parser:9.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_build | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries yargs parser tgz yargs parser tgz yargs parser tgz yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file js monorepo nodegoat package json path to vulnerable library js monorepo nodegoat node modules nyc node modules yargs parser package json dependency hierarchy realize package specifier tgz root library grunt contrib nodeunit tgz nodeunit tgz tap tgz nyc tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file js monorepo nodejs goof package json path to vulnerable library js monorepo nodejs goof node modules nyc node modules yargs parser package json dependency hierarchy tap tgz root library nyc tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file js monorepo nodejs goof package json path to vulnerable library js monorepo nodejs goof node modules nyc node modules yargs node modules yargs parser package json dependency hierarchy tap tgz root library nyc tgz yargs tgz x yargs parser tgz vulnerable library found in head commit a href found in base branch main vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree realize package specifier grunt contrib nodeunit nodeunit tap nyc yargs parser isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename yargs parser packageversion packagefilepaths istransitivedependency true dependencytree tap nyc yargs parser isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename yargs parser packageversion packagefilepaths istransitivedependency true dependencytree tap nyc yargs yargs parser isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload vulnerabilityurl | 0 |
349,292 | 31,791,005,868 | IssuesEvent | 2023-09-13 03:18:36 | TencentBlueKing/bk-cmdb | https://api.github.com/repos/TencentBlueKing/bk-cmdb | closed | 【CMDB+3.11.2-alpha1】业务拓扑下导出的主机数据前提提示成功,实际上Excel表格里面没有数据 | for test priority: Urgent | 问题描述:
业务拓扑下导出的主机数据前提提示成功,实际上Excel表格里面没有塑化剂
一,前置条件:
业务拓扑下存在主机,并在业务拓扑页面
二,操作步骤:
1.勾选多条主机
2.点击更多--导出选中

3.点击下一步-->开始导出

三,预期结果:
数据导出成功,Excel表格正常查看到导出的数据
四,实际结果:
前端提示导出成功,Excel表格里面没有数据

| 1.0 | 【CMDB+3.11.2-alpha1】业务拓扑下导出的主机数据前提提示成功,实际上Excel表格里面没有数据 - 问题描述:
业务拓扑下导出的主机数据前提提示成功,实际上Excel表格里面没有塑化剂
一,前置条件:
业务拓扑下存在主机,并在业务拓扑页面
二,操作步骤:
1.勾选多条主机
2.点击更多--导出选中

3.点击下一步-->开始导出

三,预期结果:
数据导出成功,Excel表格正常查看到导出的数据
四,实际结果:
前端提示导出成功,Excel表格里面没有数据

| non_build | 【cmdb 】业务拓扑下导出的主机数据前提提示成功,实际上excel表格里面没有数据 问题描述: 业务拓扑下导出的主机数据前提提示成功,实际上excel表格里面没有塑化剂 一,前置条件: 业务拓扑下存在主机,并在业务拓扑页面 二,操作步骤: 勾选多条主机 点击更多 导出选中 点击下一步 开始导出 三,预期结果: 数据导出成功,excel表格正常查看到导出的数据 四,实际结果: 前端提示导出成功,excel表格里面没有数据 | 0 |
568,873 | 16,989,623,224 | IssuesEvent | 2021-06-30 18:36:41 | crispy-computing-machine/Winbinder | https://api.github.com/repos/crispy-computing-machine/Winbinder | closed | Allocate memory dynamically in various places | Low priority enhancement help wanted | Allocate memory dynamically instead of limited static memory
- In wbtemp_create_menu(): allocate accel[] dynamically like pitem
- Allocate tab controls and pages dynamically: TABDATA should allocate items dynamically; now it is carrying a somewhat large data structure | 1.0 | Allocate memory dynamically in various places - Allocate memory dynamically instead of limited static memory
- In wbtemp_create_menu(): allocate accel[] dynamically like pitem
- Allocate tab controls and pages dynamically: TABDATA should allocate items dynamically; now it is carrying a somewhat large data structure | non_build | allocate memory dynamically in various places allocate memory dynamically instead of limited static memory in wbtemp create menu allocate accel dynamically like pitem allocate tab controls and pages dynamically tabdata should allocate items dynamically now it is carrying a somewhat large data structure | 0 |
105,395 | 13,183,326,087 | IssuesEvent | 2020-08-12 17:18:41 | COVID19Tracking/website | https://api.github.com/repos/COVID19Tracking/website | closed | Remove Florida exception made in #1250 | DESIGN DEV | See #1250 -- we should no longer need this exception since FL has been reporting hosp. data for over 30 days now. | 1.0 | Remove Florida exception made in #1250 - See #1250 -- we should no longer need this exception since FL has been reporting hosp. data for over 30 days now. | non_build | remove florida exception made in see we should no longer need this exception since fl has been reporting hosp data for over days now | 0 |
180,534 | 6,650,521,737 | IssuesEvent | 2017-09-28 16:36:51 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.yam.com - see bug description | browser-firefox priority-normal status-invalid | <!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: web -->
**URL**: https://www.yam.com/
**Browser / Version**: Firefox 58.0
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: The connection is unsafe.
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2017/9/f13f8200-7ae0-4f1b-96dc-018e374e39ed.jpeg)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.yam.com - see bug description - <!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: web -->
**URL**: https://www.yam.com/
**Browser / Version**: Firefox 58.0
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Something else
**Description**: The connection is unsafe.
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2017/9/f13f8200-7ae0-4f1b-96dc-018e374e39ed.jpeg)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_build | see bug description url browser version firefox operating system windows tested another browser no problem type something else description the connection is unsafe steps to reproduce from with ❤️ | 0 |
4,906 | 3,477,900,140 | IssuesEvent | 2015-12-28 07:16:21 | rubinius/rubinius | https://api.github.com/repos/rubinius/rubinius | closed | Rubinius fails to compile: '_setjmp' was not declared in this scope | build Solaris | Hi I'm trying to compile Rubinius pulled from the git repo as of tonight within a Solaris zone and I keep getting the same error. I've tried the current git repo and the last two releases. I first attempted this with Ruby 1.9.3-xxx installed, but after reading on the Rubinius website, I restarted the template with a fresh Ruby 2.0.0-xxx install. They all error out in the same place:
```
Running 191 tasks using 1 parallel threads
2: CXX vm/accessor_primitives.cpp
2: CXX vm/agent.cpp
2: CXX vm/agent_components.cpp
2: CXX vm/arguments.cpp
2: CXX vm/auxiliary_threads.cpp
2: CXX vm/builtin/access_variable.cpp
2: CXX vm/builtin/alias.cpp
2: CXX vm/builtin/array.cpp
2: CXX vm/builtin/atomic.cpp
2: CXX vm/builtin/autoload.cpp
2: CXX vm/builtin/bignum.cpp
2: CXX vm/builtin/block_as_method.cpp
2: CXX vm/builtin/block_environment.cpp
2: CXX vm/builtin/byte_array.cpp
2: CXX vm/builtin/call_custom_cache.cpp
2: CXX vm/builtin/call_site.cpp
2: CXX vm/builtin/call_unit.cpp
2: CXX vm/builtin/call_unit_adapter.cpp
2: CXX vm/builtin/channel.cpp
2: CXX vm/builtin/character.cpp
2: CXX vm/builtin/class.cpp
2: CXX vm/builtin/compact_lookup_table.cpp
2: CXX vm/builtin/compiled_code.cpp
2: CXX vm/builtin/constant_cache.cpp
2: CXX vm/builtin/constant_scope.cpp
2: CXX vm/builtin/constant_table.cpp
2: CXX vm/builtin/data.cpp
2: CXX vm/builtin/dir.cpp
2: CXX vm/builtin/encoding.cpp
2: CXX vm/builtin/exception.cpp
2: CXX vm/builtin/executable.cpp
2: CXX vm/builtin/ffi_pointer.cpp
2: CXX vm/builtin/fiber.cpp
2: CXX vm/builtin/find_object.cpp
2: CXX vm/builtin/fixnum.cpp
2: CXX vm/builtin/float.cpp
2: CXX vm/builtin/heap_dump.cpp
2: CXX vm/builtin/immediates.cpp
2: CXX vm/builtin/integer.cpp
2: CXX vm/builtin/io.cpp
2: CXX vm/builtin/iseq.cpp
2: CXX vm/builtin/list.cpp
2: CXX vm/builtin/location.cpp
2: CXX vm/builtin/lookup_table.cpp
2: CXX vm/builtin/method_table.cpp
2: CXX vm/builtin/module.cpp
2: CXX vm/builtin/mono_inline_cache.cpp
2: CXX vm/builtin/native_function.cpp
2: CXX vm/builtin/native_method.cpp
vm/builtin/native_method.cpp: In static member function 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&)':
vm/builtin/native_method.cpp:691:9: error: there are no arguments to '_setjmp' that depend on a template parameter, so a declaration of '_setjmp' must be available [-fpermissive]
vm/builtin/native_method.cpp:691:9: note: (if you use '-fpermissive', G++ will accept your code, but allowing the use of an undeclared name is deprecated)
vm/builtin/native_method.cpp:702:9: error: there are no arguments to '_setjmp' that depend on a template parameter, so a declaration of '_setjmp' must be available [-fpermissive]
vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::ZeroArguments; rubinius::Executable = rubinius::Executable]':
vm/builtin/native_method.cpp:776:82: required from here
vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:691:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:702:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::OneArgument; rubinius::Executable = rubinius::Executable]':
vm/builtin/native_method.cpp:780:80: required from here
vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:691:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:702:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::TwoArguments; rubinius::Executable = rubinius::Executable]':
vm/builtin/native_method.cpp:784:81: required from here
vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:691:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:702:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::ThreeArguments; rubinius::Executable = rubinius::Executable]':
vm/builtin/native_method.cpp:788:83: required from here
vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:691:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:702:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::GenericArguments; rubinius::Executable = rubinius::Executable]':
vm/builtin/native_method.cpp:792:85: required from here
vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:691:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:702:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
Error: g++ -I/usr/local/src/rubinius/vm -I/usr/local/src/rubinius/vm/include -I/usr/local/src/rubinius/vm/builtin -I. -Ivm/test/cxxtest -I/usr/local/src/rubinius/vendor/udis86 -I/usr/local/src/rubinius/vendor/libffi/include -Ivendor/double-conversion/src -DHAVE_CONFIG_H -I/usr/local/src/rubinius/vm/include/capi -I/usr/local/src/rubinius/vendor/oniguruma -I/usr/local/src/rubinius/vendor/libtommath -pipe -Wall -fno-omit-frame-pointer -g -I/usr/local/include -I/opt/local/include -fPIC -Wno-strict-aliasing -O2 -DHAS_EXECINFO -DHAVE_CLOCK_GETTIME -DHAVE_NL_LANGINFO -DHAVE_POSIX_FADVISE -DHAVE_STRNLEN -DHAVE_TIMEZONE -DHAVE_TZNAME -DHAVE_DAYLIGHT -DHAVE_ALLOCA_H -DHAVE_STRING_H -DHAVE_SYS_TIME_H -DHAVE_SYS_TIMES_H -DHAVE_SYS_TYPES_H -DHAVE_UNISTD_H -DHAVE_STDARG_H -I/usr/local/include -D_DEBUG -include llvm/Support/Solaris.h -D_GNU_SOURCE -fPIC -DENABLE_LLVM -Wno-unused-function -Werror -DRBX_PROFILER -D__STDC_LIMIT_MACROS -D__STDC_CONSTANT_MACROS -D_LARGEFILE_SOURCE -D_FILE_OFFSET_BITS=64 -fno-rtti -fvisibility-inlines-hidden -c -o vm/builtin/artifacts/native_method.cpp.o vm/builtin/native_method.cpp
rake aborted!
Error compiling
Tasks: TOP => default => spec => build => build:build => vm/vm
(See full trace by running task with --trace)
```
Installed packages:
I built LLVM from source as it wasn't in the repos, using LLVM-3.3
```
binutils-2.23.2nb1 GNU binary utilities
bison-3.0 GNU yacc(1) replacement
bmake-20110606nb1 Portable (autoconf) version of NetBSD 'make' utility
bootstrap-mk-files-20130912 *.mk files for the bootstrap bmake utility
bsdinstall-20130905 Portable version of the BSD install(1) program
bzip2-1.0.6nb1 Block-sorting file compressor
changepass-1.3.3 Simple tool to set user passwords non-interactively
cloog-0.18.0nb1 Code generator for loop optimization (used by gcc)
coreutils-8.13nb6 GNU basic file, shell and text manipulation utilities
curl-7.32.0 Client that groks URLs
cyrus-sasl-2.1.26nb3 Simple Authentication and Security Layer
db4-4.8.30 Berkeley DB version 4 from Oracle
diffutils-3.3 GNU diff utilities - find the differences between files
duo-unix-1.9.5 2 Factor Authentication provided by Duo Security
expat-2.1.0 XML parser library written in C
findutils-4.2.33nb3 The GNU find, locate, updatedb, and xargs utilities
fontconfig-2.10.95 Library for configuring and customizing font access
freetype2-2.4.12 Font rendering engine and library API
gawk-4.1.0 GNU awk
gcc47-4.7.3nb1 The GNU Compiler Collection (GCC) - 4.7 Release Series
gcc47-libs-4.7.3nb3 The GNU Compiler Collection (GCC) support shared libraries
genbfpw-0 Simple tool to generate Blowfish crypts of passwords
gettext-lib-0.18.3.1 Internationalized Message Handling Library (libintl)
ghostscript-9.05nb6 Meta-package for installing the preferred ghostscript version
ghostscript-fonts-8.11nb3 Postscript fonts for Aladdin Ghostscript
ghostscript-gpl-9.05nb6 Postscript interpreter
git-base-1.8.4nb1 GIT Tree History Storage Tool (base package)
gmake-3.82nb7 GNU version of 'make' utility
gmp-5.1.2 Library for arbitrary precision arithmetic
grep-2.14nb2 GNU grep
groff-1.20.1nb12 GNU roff text processing suite
gsed-4.2.2nb4 GNU implementation of sed, the POSIX stream editor
gtar-base-1.26nb2 The GNU tape archiver with remote magnetic tape support
guardtime-1.0.12 Command line client for GuardTime keyless signature service
isl-0.11.2 Integer set library required by gcc graphite
jasper-1.900.1nb7 Software-based reference implementation of the JPEG-2000 codec
jbigkit-2.0 JBIG-KIT lossless image compression library
jpeg-9 IJG's jpeg compression utilities
json-c-0.10 JSON library in C
less-458 Pager similar to more and pg
libICE-1.0.8 Inter Client Exchange (ICE) library for X
libSM-1.2.2 X Session Management Library
libX11-1.6.2 Base X libraries from modular Xorg X11
libXau-1.0.8 Authorization Protocol for X from X.org
libXaw-1.0.12 X Athena Widgets Library from modular Xorg X11
libXdmcp-1.1.1 X Display Manager Control Protocol library from X.org
libXext-1.3.2 X Extension library
libXmu-1.1.2 X Miscellaneous Utilities library
libXpm-3.5.11 X PixMap Library from modular Xorg X11
libXt-1.1.4 X Toolkit Intrinsics library
libarchive-2.8.4nb2 Library to read/create different archive formats
libee-0.4.1 Event Expression Library inspired by CEE
libestr-0.1.5 Library for some string essentials
libffi-3.0.13 Foreign function interface
libgcrypt-1.5.3 GNU cryptographic library
libgetopt-1.4.4 Library for handling --long options
libgpg-error-1.12 Definitions of common error values for all GnuPG components
libguardtime-0.3.11 GuardTime Client C SDK
libiconv-1.14nb2 Character set conversion library
libidn-1.28 Internationalized Domain Names command line tool
liblognorm-0.3.7 Tool to normalize log data
libpaper-1.1.24 Paper size handling library
libssh2-1.2.2nb2 SSH2 protocol library
libuuid-2.19.1 Generate unique identifiers for objects
libxcb-1.9.1 X protocol C-language Binding
libxml2-2.9.1 XML parser library from the GNOME project
libyaml-0.1.4 YAML 1.1 parser and emitter written in C
m4-1.4.16nb3 GNU version of UNIX m4 macro language processor
mit-krb5-1.10.6 MIT Kerberos 5 authentication system
mozilla-rootcerts-1.0.20121229nb1 Root CA certificates from the Mozilla Project
mpfr-3.1.2 GMP-based library for multiple-precision floating-point computations
nawk-20050424nb3 Brian Kernighan's pattern-directed scanning and processing language
nbsed-20120308 NetBSD-current's sed(1)
ncurses-5.9nb1 CRT screen handling and optimization package
netpbm-10.35.80nb7 Toolkit for conversion of images between different formats
nodejs-0.10.22 V8 JavaScript for clients and servers
openldap-client-2.4.36 Lightweight Directory Access Protocol libraries and client programs
openssl-1.0.1enb2 Secure Socket Layer and cryptographic library
p5-Authen-SASL-2.16nb1 Perl module to handle SASL authentication
p5-Digest-HMAC-1.03nb2 Perl5 module for HMAC
p5-Digest-MD5-2.53 Perl5 extension interface for MD5 algorithm
p5-Digest-SHA-5.85 Perl module for SHA-1/224/256/384/512 algorithms
p5-Email-Valid-1.192 Perl5 module for testing validity of an email address
p5-Error-0.17021 Perl extension module for try/throw/catch exception handling
p5-GSSAPI-0.28nb4 Perl extension providing access to the GSSAPIv2 library
p5-IO-Socket-INET6-2.71 Perl object interface for AF_INET|AF_INET6 domain sockets
p5-IO-Socket-SSL-1.953 Perl5 SSL socket interface class
p5-MIME-Base64-3.14 Perl5 module for Base64 and Quoted-Printable encodings
p5-MailTools-2.12nb1 Perl5 modules related to mail applications
p5-Net-DNS-0.72 Perl5 module for DNS resolution
p5-Net-IP-1.26nb1 Perl extension for manipulating IPv4/IPv6 addresses
p5-Net-LibIDN-0.12nb5 Perl bindings for GNU Libidn
p5-Net-SMTP-SSL-1.01nb4 Perl5 module providing SSL support for Net::SMTP
p5-Net-SSLeay-1.55 Perl5 module for using OpenSSL
p5-Socket6-0.23nb4 Perl5 module to support getaddrinfo() and getnameinfo()
p5-TimeDate-2.30 Perl5 TimeDate distribution
patch-2.5.9nb2 Patch files using diff output
pcre-8.33 Perl Compatible Regular Expressions library
perl-5.18.1nb1 Practical Extraction and Report Language
pkg_install-20130902 Package management and administration tools for pkgsrc
pkg_install-info-4.5nb3 Standalone GNU info file installation utility
pkgin-0.6.4nb1 Apt / yum like tool for managing pkgsrc binary packages
png-1.6.6 Library for manipulating PNG images
postfix-2.9.8 Postfix SMTP server and tools
ppl-0.11.2nb2 The Parma Polyhedra Library. Used by gcc for loop optimization
psutils-1.17nb4 Utilities for manipulating PostScript documents
python27-2.7.5nb3 Interpreted, interactive, object-oriented programming language
readline-6.2 GNU library that can recall and edit previous input
rsyslog-7.4.4nb2 The enhanced syslogd for Unix
ruby-2.0.0p353 Wrapper package for Ruby programming language
ruby200-base-2.0.0p353 Ruby 2.0.0 release minimum base package
ruby200-bundler-1.3.5 Manage your application's dependencies
ruby200-rake-10.0.4 Ruby Make
ruby200-rubygems-2.0.10 Ruby standard for publishing and managing third party libraries
sdc-manta-1.2.2 Node.js SDK for Manta
sdc-node-0.10.22 V8 JavaScript for clients and servers
sdc-smartdc-7.1.1 Client SDK and CLI for the Joyent SmartDataCenter API
smtools-20131031 Joyent tools relevant to SmartOS and SmartMachines
sqlite3-3.8.0.2 SQL Database Engine in a C Library
sudo-1.7.10p7 Allow others to run commands as root
tcp_wrappers-7.6.4 Monitor and filter incoming requests for network services
tiff-4.0.3nb6 Library and tools for reading and writing TIFF data files
xmlcatmgr-2.2nb1 XML and SGML catalog manager
xz-5.0.5 XZ utilities
zlib-1.2.8 General purpose data compression library
zoneinit-1.6.5 Joyent Instance initialization toolchain
``` | 1.0 | Rubinius fails to compile: '_setjmp' was not declared in this scope - Hi I'm trying to compile Rubinius pulled from the git repo as of tonight within a Solaris zone and I keep getting the same error. I've tried the current git repo and the last two releases. I first attempted this with Ruby 1.9.3-xxx installed, but after reading on the Rubinius website, I restarted the template with a fresh Ruby 2.0.0-xxx install. They all error out in the same place:
```
Running 191 tasks using 1 parallel threads
2: CXX vm/accessor_primitives.cpp
2: CXX vm/agent.cpp
2: CXX vm/agent_components.cpp
2: CXX vm/arguments.cpp
2: CXX vm/auxiliary_threads.cpp
2: CXX vm/builtin/access_variable.cpp
2: CXX vm/builtin/alias.cpp
2: CXX vm/builtin/array.cpp
2: CXX vm/builtin/atomic.cpp
2: CXX vm/builtin/autoload.cpp
2: CXX vm/builtin/bignum.cpp
2: CXX vm/builtin/block_as_method.cpp
2: CXX vm/builtin/block_environment.cpp
2: CXX vm/builtin/byte_array.cpp
2: CXX vm/builtin/call_custom_cache.cpp
2: CXX vm/builtin/call_site.cpp
2: CXX vm/builtin/call_unit.cpp
2: CXX vm/builtin/call_unit_adapter.cpp
2: CXX vm/builtin/channel.cpp
2: CXX vm/builtin/character.cpp
2: CXX vm/builtin/class.cpp
2: CXX vm/builtin/compact_lookup_table.cpp
2: CXX vm/builtin/compiled_code.cpp
2: CXX vm/builtin/constant_cache.cpp
2: CXX vm/builtin/constant_scope.cpp
2: CXX vm/builtin/constant_table.cpp
2: CXX vm/builtin/data.cpp
2: CXX vm/builtin/dir.cpp
2: CXX vm/builtin/encoding.cpp
2: CXX vm/builtin/exception.cpp
2: CXX vm/builtin/executable.cpp
2: CXX vm/builtin/ffi_pointer.cpp
2: CXX vm/builtin/fiber.cpp
2: CXX vm/builtin/find_object.cpp
2: CXX vm/builtin/fixnum.cpp
2: CXX vm/builtin/float.cpp
2: CXX vm/builtin/heap_dump.cpp
2: CXX vm/builtin/immediates.cpp
2: CXX vm/builtin/integer.cpp
2: CXX vm/builtin/io.cpp
2: CXX vm/builtin/iseq.cpp
2: CXX vm/builtin/list.cpp
2: CXX vm/builtin/location.cpp
2: CXX vm/builtin/lookup_table.cpp
2: CXX vm/builtin/method_table.cpp
2: CXX vm/builtin/module.cpp
2: CXX vm/builtin/mono_inline_cache.cpp
2: CXX vm/builtin/native_function.cpp
2: CXX vm/builtin/native_method.cpp
vm/builtin/native_method.cpp: In static member function 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&)':
vm/builtin/native_method.cpp:691:9: error: there are no arguments to '_setjmp' that depend on a template parameter, so a declaration of '_setjmp' must be available [-fpermissive]
vm/builtin/native_method.cpp:691:9: note: (if you use '-fpermissive', G++ will accept your code, but allowing the use of an undeclared name is deprecated)
vm/builtin/native_method.cpp:702:9: error: there are no arguments to '_setjmp' that depend on a template parameter, so a declaration of '_setjmp' must be available [-fpermissive]
vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::ZeroArguments; rubinius::Executable = rubinius::Executable]':
vm/builtin/native_method.cpp:776:82: required from here
vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:691:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:702:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::OneArgument; rubinius::Executable = rubinius::Executable]':
vm/builtin/native_method.cpp:780:80: required from here
vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:691:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:702:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::TwoArguments; rubinius::Executable = rubinius::Executable]':
vm/builtin/native_method.cpp:784:81: required from here
vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:691:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:702:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::ThreeArguments; rubinius::Executable = rubinius::Executable]':
vm/builtin/native_method.cpp:788:83: required from here
vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:691:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:702:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::GenericArguments; rubinius::Executable = rubinius::Executable]':
vm/builtin/native_method.cpp:792:85: required from here
vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:691:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope
vm/builtin/native_method.cpp:702:9: note: suggested alternative:
In file included from /usr/include/setjmp.h:34:0,
from /usr/local/src/rubinius/vm/vm.hpp:25,
from /usr/local/src/rubinius/vm/builtin/object.hpp:6,
from /usr/local/src/rubinius/vm/builtin/array.hpp:4,
from /usr/local/src/rubinius/vm/arguments.hpp:5,
from vm/builtin/native_method.cpp:1:
/usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp'
Error: g++ -I/usr/local/src/rubinius/vm -I/usr/local/src/rubinius/vm/include -I/usr/local/src/rubinius/vm/builtin -I. -Ivm/test/cxxtest -I/usr/local/src/rubinius/vendor/udis86 -I/usr/local/src/rubinius/vendor/libffi/include -Ivendor/double-conversion/src -DHAVE_CONFIG_H -I/usr/local/src/rubinius/vm/include/capi -I/usr/local/src/rubinius/vendor/oniguruma -I/usr/local/src/rubinius/vendor/libtommath -pipe -Wall -fno-omit-frame-pointer -g -I/usr/local/include -I/opt/local/include -fPIC -Wno-strict-aliasing -O2 -DHAS_EXECINFO -DHAVE_CLOCK_GETTIME -DHAVE_NL_LANGINFO -DHAVE_POSIX_FADVISE -DHAVE_STRNLEN -DHAVE_TIMEZONE -DHAVE_TZNAME -DHAVE_DAYLIGHT -DHAVE_ALLOCA_H -DHAVE_STRING_H -DHAVE_SYS_TIME_H -DHAVE_SYS_TIMES_H -DHAVE_SYS_TYPES_H -DHAVE_UNISTD_H -DHAVE_STDARG_H -I/usr/local/include -D_DEBUG -include llvm/Support/Solaris.h -D_GNU_SOURCE -fPIC -DENABLE_LLVM -Wno-unused-function -Werror -DRBX_PROFILER -D__STDC_LIMIT_MACROS -D__STDC_CONSTANT_MACROS -D_LARGEFILE_SOURCE -D_FILE_OFFSET_BITS=64 -fno-rtti -fvisibility-inlines-hidden -c -o vm/builtin/artifacts/native_method.cpp.o vm/builtin/native_method.cpp
rake aborted!
Error compiling
Tasks: TOP => default => spec => build => build:build => vm/vm
(See full trace by running task with --trace)
```
Installed packages:
I built LLVM from source as it wasn't in the repos, using LLVM-3.3
```
binutils-2.23.2nb1 GNU binary utilities
bison-3.0 GNU yacc(1) replacement
bmake-20110606nb1 Portable (autoconf) version of NetBSD 'make' utility
bootstrap-mk-files-20130912 *.mk files for the bootstrap bmake utility
bsdinstall-20130905 Portable version of the BSD install(1) program
bzip2-1.0.6nb1 Block-sorting file compressor
changepass-1.3.3 Simple tool to set user passwords non-interactively
cloog-0.18.0nb1 Code generator for loop optimization (used by gcc)
coreutils-8.13nb6 GNU basic file, shell and text manipulation utilities
curl-7.32.0 Client that groks URLs
cyrus-sasl-2.1.26nb3 Simple Authentication and Security Layer
db4-4.8.30 Berkeley DB version 4 from Oracle
diffutils-3.3 GNU diff utilities - find the differences between files
duo-unix-1.9.5 2 Factor Authentication provided by Duo Security
expat-2.1.0 XML parser library written in C
findutils-4.2.33nb3 The GNU find, locate, updatedb, and xargs utilities
fontconfig-2.10.95 Library for configuring and customizing font access
freetype2-2.4.12 Font rendering engine and library API
gawk-4.1.0 GNU awk
gcc47-4.7.3nb1 The GNU Compiler Collection (GCC) - 4.7 Release Series
gcc47-libs-4.7.3nb3 The GNU Compiler Collection (GCC) support shared libraries
genbfpw-0 Simple tool to generate Blowfish crypts of passwords
gettext-lib-0.18.3.1 Internationalized Message Handling Library (libintl)
ghostscript-9.05nb6 Meta-package for installing the preferred ghostscript version
ghostscript-fonts-8.11nb3 Postscript fonts for Aladdin Ghostscript
ghostscript-gpl-9.05nb6 Postscript interpreter
git-base-1.8.4nb1 GIT Tree History Storage Tool (base package)
gmake-3.82nb7 GNU version of 'make' utility
gmp-5.1.2 Library for arbitrary precision arithmetic
grep-2.14nb2 GNU grep
groff-1.20.1nb12 GNU roff text processing suite
gsed-4.2.2nb4 GNU implementation of sed, the POSIX stream editor
gtar-base-1.26nb2 The GNU tape archiver with remote magnetic tape support
guardtime-1.0.12 Command line client for GuardTime keyless signature service
isl-0.11.2 Integer set library required by gcc graphite
jasper-1.900.1nb7 Software-based reference implementation of the JPEG-2000 codec
jbigkit-2.0 JBIG-KIT lossless image compression library
jpeg-9 IJG's jpeg compression utilities
json-c-0.10 JSON library in C
less-458 Pager similar to more and pg
libICE-1.0.8 Inter Client Exchange (ICE) library for X
libSM-1.2.2 X Session Management Library
libX11-1.6.2 Base X libraries from modular Xorg X11
libXau-1.0.8 Authorization Protocol for X from X.org
libXaw-1.0.12 X Athena Widgets Library from modular Xorg X11
libXdmcp-1.1.1 X Display Manager Control Protocol library from X.org
libXext-1.3.2 X Extension library
libXmu-1.1.2 X Miscellaneous Utilities library
libXpm-3.5.11 X PixMap Library from modular Xorg X11
libXt-1.1.4 X Toolkit Intrinsics library
libarchive-2.8.4nb2 Library to read/create different archive formats
libee-0.4.1 Event Expression Library inspired by CEE
libestr-0.1.5 Library for some string essentials
libffi-3.0.13 Foreign function interface
libgcrypt-1.5.3 GNU cryptographic library
libgetopt-1.4.4 Library for handling --long options
libgpg-error-1.12 Definitions of common error values for all GnuPG components
libguardtime-0.3.11 GuardTime Client C SDK
libiconv-1.14nb2 Character set conversion library
libidn-1.28 Internationalized Domain Names command line tool
liblognorm-0.3.7 Tool to normalize log data
libpaper-1.1.24 Paper size handling library
libssh2-1.2.2nb2 SSH2 protocol library
libuuid-2.19.1 Generate unique identifiers for objects
libxcb-1.9.1 X protocol C-language Binding
libxml2-2.9.1 XML parser library from the GNOME project
libyaml-0.1.4 YAML 1.1 parser and emitter written in C
m4-1.4.16nb3 GNU version of UNIX m4 macro language processor
mit-krb5-1.10.6 MIT Kerberos 5 authentication system
mozilla-rootcerts-1.0.20121229nb1 Root CA certificates from the Mozilla Project
mpfr-3.1.2 GMP-based library for multiple-precision floating-point computations
nawk-20050424nb3 Brian Kernighan's pattern-directed scanning and processing language
nbsed-20120308 NetBSD-current's sed(1)
ncurses-5.9nb1 CRT screen handling and optimization package
netpbm-10.35.80nb7 Toolkit for conversion of images between different formats
nodejs-0.10.22 V8 JavaScript for clients and servers
openldap-client-2.4.36 Lightweight Directory Access Protocol libraries and client programs
openssl-1.0.1enb2 Secure Socket Layer and cryptographic library
p5-Authen-SASL-2.16nb1 Perl module to handle SASL authentication
p5-Digest-HMAC-1.03nb2 Perl5 module for HMAC
p5-Digest-MD5-2.53 Perl5 extension interface for MD5 algorithm
p5-Digest-SHA-5.85 Perl module for SHA-1/224/256/384/512 algorithms
p5-Email-Valid-1.192 Perl5 module for testing validity of an email address
p5-Error-0.17021 Perl extension module for try/throw/catch exception handling
p5-GSSAPI-0.28nb4 Perl extension providing access to the GSSAPIv2 library
p5-IO-Socket-INET6-2.71 Perl object interface for AF_INET|AF_INET6 domain sockets
p5-IO-Socket-SSL-1.953 Perl5 SSL socket interface class
p5-MIME-Base64-3.14 Perl5 module for Base64 and Quoted-Printable encodings
p5-MailTools-2.12nb1 Perl5 modules related to mail applications
p5-Net-DNS-0.72 Perl5 module for DNS resolution
p5-Net-IP-1.26nb1 Perl extension for manipulating IPv4/IPv6 addresses
p5-Net-LibIDN-0.12nb5 Perl bindings for GNU Libidn
p5-Net-SMTP-SSL-1.01nb4 Perl5 module providing SSL support for Net::SMTP
p5-Net-SSLeay-1.55 Perl5 module for using OpenSSL
p5-Socket6-0.23nb4 Perl5 module to support getaddrinfo() and getnameinfo()
p5-TimeDate-2.30 Perl5 TimeDate distribution
patch-2.5.9nb2 Patch files using diff output
pcre-8.33 Perl Compatible Regular Expressions library
perl-5.18.1nb1 Practical Extraction and Report Language
pkg_install-20130902 Package management and administration tools for pkgsrc
pkg_install-info-4.5nb3 Standalone GNU info file installation utility
pkgin-0.6.4nb1 Apt / yum like tool for managing pkgsrc binary packages
png-1.6.6 Library for manipulating PNG images
postfix-2.9.8 Postfix SMTP server and tools
ppl-0.11.2nb2 The Parma Polyhedra Library. Used by gcc for loop optimization
psutils-1.17nb4 Utilities for manipulating PostScript documents
python27-2.7.5nb3 Interpreted, interactive, object-oriented programming language
readline-6.2 GNU library that can recall and edit previous input
rsyslog-7.4.4nb2 The enhanced syslogd for Unix
ruby-2.0.0p353 Wrapper package for Ruby programming language
ruby200-base-2.0.0p353 Ruby 2.0.0 release minimum base package
ruby200-bundler-1.3.5 Manage your application's dependencies
ruby200-rake-10.0.4 Ruby Make
ruby200-rubygems-2.0.10 Ruby standard for publishing and managing third party libraries
sdc-manta-1.2.2 Node.js SDK for Manta
sdc-node-0.10.22 V8 JavaScript for clients and servers
sdc-smartdc-7.1.1 Client SDK and CLI for the Joyent SmartDataCenter API
smtools-20131031 Joyent tools relevant to SmartOS and SmartMachines
sqlite3-3.8.0.2 SQL Database Engine in a C Library
sudo-1.7.10p7 Allow others to run commands as root
tcp_wrappers-7.6.4 Monitor and filter incoming requests for network services
tiff-4.0.3nb6 Library and tools for reading and writing TIFF data files
xmlcatmgr-2.2nb1 XML and SGML catalog manager
xz-5.0.5 XZ utilities
zlib-1.2.8 General purpose data compression library
zoneinit-1.6.5 Joyent Instance initialization toolchain
``` | build | rubinius fails to compile setjmp was not declared in this scope hi i m trying to compile rubinius pulled from the git repo as of tonight within a solaris zone and i keep getting the same error i ve tried the current git repo and the last two releases i first attempted this with ruby xxx installed but after reading on the rubinius website i restarted the template with a fresh ruby xxx install they all error out in the same place running tasks using parallel threads cxx vm accessor primitives cpp cxx vm agent cpp cxx vm agent components cpp cxx vm arguments cpp cxx vm auxiliary threads cpp cxx vm builtin access variable cpp cxx vm builtin alias cpp cxx vm builtin array cpp cxx vm builtin atomic cpp cxx vm builtin autoload cpp cxx vm builtin bignum cpp cxx vm builtin block as method cpp cxx vm builtin block environment cpp cxx vm builtin byte array cpp cxx vm builtin call custom cache cpp cxx vm builtin call site cpp cxx vm builtin call unit cpp cxx vm builtin call unit adapter cpp cxx vm builtin channel cpp cxx vm builtin character cpp cxx vm builtin class cpp cxx vm builtin compact lookup table cpp cxx vm builtin compiled code cpp cxx vm builtin constant cache cpp cxx vm builtin constant scope cpp cxx vm builtin constant table cpp cxx vm builtin data cpp cxx vm builtin dir cpp cxx vm builtin encoding cpp cxx vm builtin exception cpp cxx vm builtin executable cpp cxx vm builtin ffi pointer cpp cxx vm builtin fiber cpp cxx vm builtin find object cpp cxx vm builtin fixnum cpp cxx vm builtin float cpp cxx vm builtin heap dump cpp cxx vm builtin immediates cpp cxx vm builtin integer cpp cxx vm builtin io cpp cxx vm builtin iseq cpp cxx vm builtin list cpp cxx vm builtin location cpp cxx vm builtin lookup table cpp cxx vm builtin method table cpp cxx vm builtin module cpp cxx vm builtin mono inline cache cpp cxx vm builtin native function cpp cxx vm builtin native method cpp vm builtin native method cpp in static member function static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp error there are no arguments to setjmp that depend on a template parameter so a declaration of setjmp must be available vm builtin native method cpp note if you use fpermissive g will accept your code but allowing the use of an undeclared name is deprecated vm builtin native method cpp error there are no arguments to setjmp that depend on a template parameter so a declaration of setjmp must be available vm builtin native method cpp in instantiation of static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp required from here vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp in instantiation of static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp required from here vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp in instantiation of static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp required from here vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp in instantiation of static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp required from here vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp in instantiation of static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp required from here vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp error g i usr local src rubinius vm i usr local src rubinius vm include i usr local src rubinius vm builtin i ivm test cxxtest i usr local src rubinius vendor i usr local src rubinius vendor libffi include ivendor double conversion src dhave config h i usr local src rubinius vm include capi i usr local src rubinius vendor oniguruma i usr local src rubinius vendor libtommath pipe wall fno omit frame pointer g i usr local include i opt local include fpic wno strict aliasing dhas execinfo dhave clock gettime dhave nl langinfo dhave posix fadvise dhave strnlen dhave timezone dhave tzname dhave daylight dhave alloca h dhave string h dhave sys time h dhave sys times h dhave sys types h dhave unistd h dhave stdarg h i usr local include d debug include llvm support solaris h d gnu source fpic denable llvm wno unused function werror drbx profiler d stdc limit macros d stdc constant macros d largefile source d file offset bits fno rtti fvisibility inlines hidden c o vm builtin artifacts native method cpp o vm builtin native method cpp rake aborted error compiling tasks top default spec build build build vm vm see full trace by running task with trace installed packages i built llvm from source as it wasn t in the repos using llvm binutils gnu binary utilities bison gnu yacc replacement bmake portable autoconf version of netbsd make utility bootstrap mk files mk files for the bootstrap bmake utility bsdinstall portable version of the bsd install program block sorting file compressor changepass simple tool to set user passwords non interactively cloog code generator for loop optimization used by gcc coreutils gnu basic file shell and text manipulation utilities curl client that groks urls cyrus sasl simple authentication and security layer berkeley db version from oracle diffutils gnu diff utilities find the differences between files duo unix factor authentication provided by duo security expat xml parser library written in c findutils the gnu find locate updatedb and xargs utilities fontconfig library for configuring and customizing font access font rendering engine and library api gawk gnu awk the gnu compiler collection gcc release series libs the gnu compiler collection gcc support shared libraries genbfpw simple tool to generate blowfish crypts of passwords gettext lib internationalized message handling library libintl ghostscript meta package for installing the preferred ghostscript version ghostscript fonts postscript fonts for aladdin ghostscript ghostscript gpl postscript interpreter git base git tree history storage tool base package gmake gnu version of make utility gmp library for arbitrary precision arithmetic grep gnu grep groff gnu roff text processing suite gsed gnu implementation of sed the posix stream editor gtar base the gnu tape archiver with remote magnetic tape support guardtime command line client for guardtime keyless signature service isl integer set library required by gcc graphite jasper software based reference implementation of the jpeg codec jbigkit jbig kit lossless image compression library jpeg ijg s jpeg compression utilities json c json library in c less pager similar to more and pg libice inter client exchange ice library for x libsm x session management library base x libraries from modular xorg libxau authorization protocol for x from x org libxaw x athena widgets library from modular xorg libxdmcp x display manager control protocol library from x org libxext x extension library libxmu x miscellaneous utilities library libxpm x pixmap library from modular xorg libxt x toolkit intrinsics library libarchive library to read create different archive formats libee event expression library inspired by cee libestr library for some string essentials libffi foreign function interface libgcrypt gnu cryptographic library libgetopt library for handling long options libgpg error definitions of common error values for all gnupg components libguardtime guardtime client c sdk libiconv character set conversion library libidn internationalized domain names command line tool liblognorm tool to normalize log data libpaper paper size handling library protocol library libuuid generate unique identifiers for objects libxcb x protocol c language binding xml parser library from the gnome project libyaml yaml parser and emitter written in c gnu version of unix macro language processor mit mit kerberos authentication system mozilla rootcerts root ca certificates from the mozilla project mpfr gmp based library for multiple precision floating point computations nawk brian kernighan s pattern directed scanning and processing language nbsed netbsd current s sed ncurses crt screen handling and optimization package netpbm toolkit for conversion of images between different formats nodejs javascript for clients and servers openldap client lightweight directory access protocol libraries and client programs openssl secure socket layer and cryptographic library authen sasl perl module to handle sasl authentication digest hmac module for hmac digest extension interface for algorithm digest sha perl module for sha algorithms email valid module for testing validity of an email address error perl extension module for try throw catch exception handling gssapi perl extension providing access to the library io socket perl object interface for af inet af domain sockets io socket ssl ssl socket interface class mime module for and quoted printable encodings mailtools modules related to mail applications net dns module for dns resolution net ip perl extension for manipulating addresses net libidn perl bindings for gnu libidn net smtp ssl module providing ssl support for net smtp net ssleay module for using openssl module to support getaddrinfo and getnameinfo timedate timedate distribution patch patch files using diff output pcre perl compatible regular expressions library perl practical extraction and report language pkg install package management and administration tools for pkgsrc pkg install info standalone gnu info file installation utility pkgin apt yum like tool for managing pkgsrc binary packages png library for manipulating png images postfix postfix smtp server and tools ppl the parma polyhedra library used by gcc for loop optimization psutils utilities for manipulating postscript documents interpreted interactive object oriented programming language readline gnu library that can recall and edit previous input rsyslog the enhanced syslogd for unix ruby wrapper package for ruby programming language base ruby release minimum base package bundler manage your application s dependencies rake ruby make rubygems ruby standard for publishing and managing third party libraries sdc manta node js sdk for manta sdc node javascript for clients and servers sdc smartdc client sdk and cli for the joyent smartdatacenter api smtools joyent tools relevant to smartos and smartmachines sql database engine in a c library sudo allow others to run commands as root tcp wrappers monitor and filter incoming requests for network services tiff library and tools for reading and writing tiff data files xmlcatmgr xml and sgml catalog manager xz xz utilities zlib general purpose data compression library zoneinit joyent instance initialization toolchain | 1 |
167,764 | 13,041,626,133 | IssuesEvent | 2020-07-28 20:45:09 | nextras/orm | https://api.github.com/repos/nextras/orm | closed | Removing entities in foreach | bug failing-testcase-needed | I get error `Property Smartlook\Domain\Project\ProjectPermission::$project is not nullable.`, when in 2nd foreach run call remove method.
Code:
```php
// un-assign user from projects
foreach ($user->projectPermissions as $projectPermission) {
$this->projectPermissionRepository->remove($projectPermission);
}
$this->userRepository->flush();
```
Entity:
```php
/**
* @property-read int $id {primary-proxy}
* @property Project $project {m:1 Project::$userPermissions} {primary}
* @property User $user {m:1 User::$projectPermissions} {primary}
* @property bool $isOwner
* @property-read int $permission
*/
class ProjectPermission extends Entity
```
**Versions::**
- Database: MySQL 5.5
- Orm: master
- Dbal: master
| 1.0 | Removing entities in foreach - I get error `Property Smartlook\Domain\Project\ProjectPermission::$project is not nullable.`, when in 2nd foreach run call remove method.
Code:
```php
// un-assign user from projects
foreach ($user->projectPermissions as $projectPermission) {
$this->projectPermissionRepository->remove($projectPermission);
}
$this->userRepository->flush();
```
Entity:
```php
/**
* @property-read int $id {primary-proxy}
* @property Project $project {m:1 Project::$userPermissions} {primary}
* @property User $user {m:1 User::$projectPermissions} {primary}
* @property bool $isOwner
* @property-read int $permission
*/
class ProjectPermission extends Entity
```
**Versions::**
- Database: MySQL 5.5
- Orm: master
- Dbal: master
| non_build | removing entities in foreach i get error property smartlook domain project projectpermission project is not nullable when in foreach run call remove method code php un assign user from projects foreach user projectpermissions as projectpermission this projectpermissionrepository remove projectpermission this userrepository flush entity php property read int id primary proxy property project project m project userpermissions primary property user user m user projectpermissions primary property bool isowner property read int permission class projectpermission extends entity versions database mysql orm master dbal master | 0 |
127,235 | 12,310,004,304 | IssuesEvent | 2020-05-12 09:53:01 | jstockwin/py-pdf-parser | https://api.github.com/repos/jstockwin/py-pdf-parser | closed | Add some examples to the documentation | Component - Documentation Difficulty - Easy Priority - Low Type - Enhancement | It would be good to have some example use cases in the documentation somewhere. | 1.0 | Add some examples to the documentation - It would be good to have some example use cases in the documentation somewhere. | non_build | add some examples to the documentation it would be good to have some example use cases in the documentation somewhere | 0 |
48,541 | 12,214,707,488 | IssuesEvent | 2020-05-01 10:45:33 | woocommerce/woocommerce-admin | https://api.github.com/repos/woocommerce/woocommerce-admin | closed | Webpack builds files in project root | Build [Type] Bug | **Describe the bug**
Occassionally, the new optimized files introduced in https://github.com/woocommerce/woocommerce-admin/pull/4094 get built directly in the wc-admin root instead of the `dist` folder.
This seems to happen on branch change. This may just be a point of frustration temporarily and not long-term since it's unlikely we'll be switching back and forth between branches with this change for too long. We may decide that this is an edge case and does not need a fix.
If we do decide to fix this, my best guess is that this is stemming from the introduction of the `path` property in webpack `output`.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to an older branch or commit.
1. Run `npm start`.
1. Switch to the current master.
1. Note the files are built in the root directory.
**Expected behavior**
Expected to only have those files built into `dist` folder.
**Screenshots**
<img width="518" alt="Screen Shot 2020-04-30 at 3 01 56 PM" src="https://user-images.githubusercontent.com/10561050/80713600-6ccd3180-8afc-11ea-9b79-f78b11521090.png">
| 1.0 | Webpack builds files in project root - **Describe the bug**
Occassionally, the new optimized files introduced in https://github.com/woocommerce/woocommerce-admin/pull/4094 get built directly in the wc-admin root instead of the `dist` folder.
This seems to happen on branch change. This may just be a point of frustration temporarily and not long-term since it's unlikely we'll be switching back and forth between branches with this change for too long. We may decide that this is an edge case and does not need a fix.
If we do decide to fix this, my best guess is that this is stemming from the introduction of the `path` property in webpack `output`.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to an older branch or commit.
1. Run `npm start`.
1. Switch to the current master.
1. Note the files are built in the root directory.
**Expected behavior**
Expected to only have those files built into `dist` folder.
**Screenshots**
<img width="518" alt="Screen Shot 2020-04-30 at 3 01 56 PM" src="https://user-images.githubusercontent.com/10561050/80713600-6ccd3180-8afc-11ea-9b79-f78b11521090.png">
| build | webpack builds files in project root describe the bug occassionally the new optimized files introduced in get built directly in the wc admin root instead of the dist folder this seems to happen on branch change this may just be a point of frustration temporarily and not long term since it s unlikely we ll be switching back and forth between branches with this change for too long we may decide that this is an edge case and does not need a fix if we do decide to fix this my best guess is that this is stemming from the introduction of the path property in webpack output to reproduce steps to reproduce the behavior go to an older branch or commit run npm start switch to the current master note the files are built in the root directory expected behavior expected to only have those files built into dist folder screenshots img width alt screen shot at pm src | 1 |
411,081 | 27,812,515,618 | IssuesEvent | 2023-03-18 09:43:22 | caodaion/caodaion.github.io | https://api.github.com/repos/caodaion/caodaion.github.io | closed | DOCUMENTATION | PROJECT DOCUMENTATION | Requirement specification | documentation | If you are leading a software development or system implementation project – such as an ERP rollout, then this is a must have document that is required to be prepared.
A requirement specification is a detailed document that serves as a guide to the development team on the various features that need to be present in the system. Usually, a requirement specification follows a modular approach which means that you identify various modules (or processes) and then drill down/detail out the requirements at a detailed level.
To describe each feature, a diagrammatic approach along with a detailed logic of the feature is usually the best way to go about. Examples of diagrams that may be included are:
Context diagram: A bird’s eye level view of the module/feature showing the various users that interact with the system
Process flow diagram: usually a flow chart or swim lane diagram defining the proces
When you have had a business analyst prepare the above, what you have is the functional requirement – meaning the various features/functionality to be present in the system. However, it has been my experience that many times, people forget to write out non-functional requirements.
A non functional requirement is, in short, a system requirement but it cannot be directly connected with any module/feature/functionality. Example of non-functional requirements are:
- Number of concurrent users, of the system
- Any particular performance criteria that is present e.g. a transaction should be approved/rejected in 20 seconds
The next phase of the project after analysis is the design (or implementation) phase and in order to do a good design, non-functional requirements come in handy. The point that you have to remember is that if you have not catered to the non-functional requirements, then even if during the final user acceptance testing, it is seen that you have catered to all the functional requirements, even then your project will not get passed.
In short, non-functional requirements are usually a very small part of the requirement, but they are perhaps the most neglected/least documented, and as a good project manager it is your job to document the non-functional requirements and see to it that the system design caters to these. (Check out this page from California Polytechnic State University for an [overview of non-functional requirements](http://users.csc.calpoly.edu/~jdalbey/SWE/QA/nonfunctional.html)).
If you are looking for a Requirement Specification template, check out [my requirement specification template](https://www.tacticalprojectmanager.com/requirement-specification-template-requirement-numbering/) which supports automatic numbering of requirements and matching of customer & product requirements (to ensure traceability). | 1.0 | DOCUMENTATION | PROJECT DOCUMENTATION | Requirement specification - If you are leading a software development or system implementation project – such as an ERP rollout, then this is a must have document that is required to be prepared.
A requirement specification is a detailed document that serves as a guide to the development team on the various features that need to be present in the system. Usually, a requirement specification follows a modular approach which means that you identify various modules (or processes) and then drill down/detail out the requirements at a detailed level.
To describe each feature, a diagrammatic approach along with a detailed logic of the feature is usually the best way to go about. Examples of diagrams that may be included are:
Context diagram: A bird’s eye level view of the module/feature showing the various users that interact with the system
Process flow diagram: usually a flow chart or swim lane diagram defining the proces
When you have had a business analyst prepare the above, what you have is the functional requirement – meaning the various features/functionality to be present in the system. However, it has been my experience that many times, people forget to write out non-functional requirements.
A non functional requirement is, in short, a system requirement but it cannot be directly connected with any module/feature/functionality. Example of non-functional requirements are:
- Number of concurrent users, of the system
- Any particular performance criteria that is present e.g. a transaction should be approved/rejected in 20 seconds
The next phase of the project after analysis is the design (or implementation) phase and in order to do a good design, non-functional requirements come in handy. The point that you have to remember is that if you have not catered to the non-functional requirements, then even if during the final user acceptance testing, it is seen that you have catered to all the functional requirements, even then your project will not get passed.
In short, non-functional requirements are usually a very small part of the requirement, but they are perhaps the most neglected/least documented, and as a good project manager it is your job to document the non-functional requirements and see to it that the system design caters to these. (Check out this page from California Polytechnic State University for an [overview of non-functional requirements](http://users.csc.calpoly.edu/~jdalbey/SWE/QA/nonfunctional.html)).
If you are looking for a Requirement Specification template, check out [my requirement specification template](https://www.tacticalprojectmanager.com/requirement-specification-template-requirement-numbering/) which supports automatic numbering of requirements and matching of customer & product requirements (to ensure traceability). | non_build | documentation project documentation requirement specification if you are leading a software development or system implementation project – such as an erp rollout then this is a must have document that is required to be prepared a requirement specification is a detailed document that serves as a guide to the development team on the various features that need to be present in the system usually a requirement specification follows a modular approach which means that you identify various modules or processes and then drill down detail out the requirements at a detailed level to describe each feature a diagrammatic approach along with a detailed logic of the feature is usually the best way to go about examples of diagrams that may be included are context diagram a bird’s eye level view of the module feature showing the various users that interact with the system process flow diagram usually a flow chart or swim lane diagram defining the proces when you have had a business analyst prepare the above what you have is the functional requirement – meaning the various features functionality to be present in the system however it has been my experience that many times people forget to write out non functional requirements a non functional requirement is in short a system requirement but it cannot be directly connected with any module feature functionality example of non functional requirements are number of concurrent users of the system any particular performance criteria that is present e g a transaction should be approved rejected in seconds the next phase of the project after analysis is the design or implementation phase and in order to do a good design non functional requirements come in handy the point that you have to remember is that if you have not catered to the non functional requirements then even if during the final user acceptance testing it is seen that you have catered to all the functional requirements even then your project will not get passed in short non functional requirements are usually a very small part of the requirement but they are perhaps the most neglected least documented and as a good project manager it is your job to document the non functional requirements and see to it that the system design caters to these check out this page from california polytechnic state university for an if you are looking for a requirement specification template check out which supports automatic numbering of requirements and matching of customer product requirements to ensure traceability | 0 |
23,531 | 22,146,650,047 | IssuesEvent | 2022-06-03 12:46:28 | WDscholia/scholia | https://api.github.com/repos/WDscholia/scholia | closed | Link co-occurring topics to topic comparison aspect | enhancement usability | Example screenshot from https://tools.wmflabs.org/scholia/topic/Q5227350 :
<img width="1156" alt="Screen Shot 2020-05-29 at 22 58 18" src="https://user-images.githubusercontent.com/465923/83318149-e6d9fe80-a1ff-11ea-84c0-5909982827b8.png">
Probably best to link it via the count value, i.e. the "383" would become [383](https://tools.wmflabs.org/scholia/topics/Q5227350,Q45933174). | True | Link co-occurring topics to topic comparison aspect - Example screenshot from https://tools.wmflabs.org/scholia/topic/Q5227350 :
<img width="1156" alt="Screen Shot 2020-05-29 at 22 58 18" src="https://user-images.githubusercontent.com/465923/83318149-e6d9fe80-a1ff-11ea-84c0-5909982827b8.png">
Probably best to link it via the count value, i.e. the "383" would become [383](https://tools.wmflabs.org/scholia/topics/Q5227350,Q45933174). | non_build | link co occurring topics to topic comparison aspect example screenshot from img width alt screen shot at src probably best to link it via the count value i e the would become | 0 |
21,923 | 30,446,460,024 | IssuesEvent | 2023-07-15 18:31:24 | h4sh5/pypi-auto-scanner | https://api.github.com/repos/h4sh5/pypi-auto-scanner | opened | pyutils 0.0.1b18 has 2 GuardDog issues | guarddog typosquatting silent-process-execution | https://pypi.org/project/pyutils
https://inspector.pypi.io/project/pyutils
```{
"dependency": "pyutils",
"version": "0.0.1b18",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils",
"silent-process-execution": [
{
"location": "pyutils/exec_utils.py/pyutils/exec_utils.py:214",
"code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpooezcz_t/pyutils"
}
}``` | 1.0 | pyutils 0.0.1b18 has 2 GuardDog issues - https://pypi.org/project/pyutils
https://inspector.pypi.io/project/pyutils
```{
"dependency": "pyutils",
"version": "0.0.1b18",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils",
"silent-process-execution": [
{
"location": "pyutils/exec_utils.py/pyutils/exec_utils.py:214",
"code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpooezcz_t/pyutils"
}
}``` | non_build | pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt pytils python utils silent process execution location pyutils exec utils py pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmpooezcz t pyutils | 0 |
32,866 | 8,966,831,388 | IssuesEvent | 2019-01-29 00:34:57 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | Issue with training of object detection api (ssd_mobilenet_v2_coco.config) | comp:apis stat:awaiting response type:build/install | Command Used:
python train.py --logtostderr --train_dir=training/ --pipeline_config_path=training/ssd_mobilenet_v2_coco.config
Please guide me if i am doing something wrong.
Error Trace:
Traceback (most recent call last):
File "train.py", line 163, in <module>
tf.app.run()
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/tensorflow/python/platform/app.py", line 125, in run
_sys.exit(main(argv))
File "train.py", line 91, in main
FLAGS.pipeline_config_path)
File "/home/tayyab/Desktop/models/object_detection/utils/config_util.py", line 43, in get_configs_from_pipeline_file
text_format.Merge(proto_str, pipeline_config)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 536, in Merge
descriptor_pool=descriptor_pool)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 590, in MergeLines
return parser.MergeLines(lines, message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 623, in MergeLines
self._ParseOrMerge(lines, message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 638, in _ParseOrMerge
self._MergeField(tokenizer, message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField
merger(tokenizer, message, field)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField
self._MergeField(tokenizer, sub_message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField
merger(tokenizer, message, field)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField
self._MergeField(tokenizer, sub_message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField
merger(tokenizer, message, field)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField
self._MergeField(tokenizer, sub_message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 730, in _MergeField
(message_descriptor.full_name, name))
google.protobuf.text_format.ParseError: 86:7 : Message type "object_detection.protos.SsdFeatureExtractor" has no field named "use_depthwise". | 1.0 | Issue with training of object detection api (ssd_mobilenet_v2_coco.config) - Command Used:
python train.py --logtostderr --train_dir=training/ --pipeline_config_path=training/ssd_mobilenet_v2_coco.config
Please guide me if i am doing something wrong.
Error Trace:
Traceback (most recent call last):
File "train.py", line 163, in <module>
tf.app.run()
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/tensorflow/python/platform/app.py", line 125, in run
_sys.exit(main(argv))
File "train.py", line 91, in main
FLAGS.pipeline_config_path)
File "/home/tayyab/Desktop/models/object_detection/utils/config_util.py", line 43, in get_configs_from_pipeline_file
text_format.Merge(proto_str, pipeline_config)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 536, in Merge
descriptor_pool=descriptor_pool)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 590, in MergeLines
return parser.MergeLines(lines, message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 623, in MergeLines
self._ParseOrMerge(lines, message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 638, in _ParseOrMerge
self._MergeField(tokenizer, message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField
merger(tokenizer, message, field)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField
self._MergeField(tokenizer, sub_message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField
merger(tokenizer, message, field)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField
self._MergeField(tokenizer, sub_message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField
merger(tokenizer, message, field)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField
self._MergeField(tokenizer, sub_message)
File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 730, in _MergeField
(message_descriptor.full_name, name))
google.protobuf.text_format.ParseError: 86:7 : Message type "object_detection.protos.SsdFeatureExtractor" has no field named "use_depthwise". | build | issue with training of object detection api ssd mobilenet coco config command used python train py logtostderr train dir training pipeline config path training ssd mobilenet coco config please guide me if i am doing something wrong error trace traceback most recent call last file train py line in tf app run file home tayyab desktop tensor android local lib site packages tensorflow python platform app py line in run sys exit main argv file train py line in main flags pipeline config path file home tayyab desktop models object detection utils config util py line in get configs from pipeline file text format merge proto str pipeline config file home tayyab desktop tensor android local lib site packages google protobuf text format py line in merge descriptor pool descriptor pool file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergelines return parser mergelines lines message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergelines self parseormerge lines message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in parseormerge self mergefield tokenizer message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergefield merger tokenizer message field file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergemessagefield self mergefield tokenizer sub message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergefield merger tokenizer message field file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergemessagefield self mergefield tokenizer sub message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergefield merger tokenizer message field file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergemessagefield self mergefield tokenizer sub message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergefield message descriptor full name name google protobuf text format parseerror message type object detection protos ssdfeatureextractor has no field named use depthwise | 1 |
456,821 | 13,151,008,267 | IssuesEvent | 2020-08-09 14:38:24 | chrisjsewell/docutils | https://api.github.com/repos/chrisjsewell/docutils | opened | support unicode box characters for table markup [SF:feature-requests:6] | feature-requests open priority-5 |
author: trentbuck
created: 2006-03-05 23:41:02
assigned: None
SF_url: https://sourceforge.net/p/docutils/feature-requests/6
See first attachment.
---
commenter: trentbuck
posted: 2006-03-05 23:41:02
title: #6 support unicode box characters for table markup
attachments:
- https://sourceforge.net/p/docutils/feature-requests/_discuss/thread/03996de8/4305/attachment/tables.txt
issue description
---
commenter: goodger
posted: 2006-03-06 20:39:33
title: #6 support unicode box characters for table markup
Logged In: YES
user\_id=7733
Unicode contains many characters that would be suitable for
reST markup. Few are currently supported. Adding this
feature would not be exceedingly difficult, but I'd classify
it as low-priority. We depend on developers to implement
features that interest them; patches are welcome\!
Not a bug; changed to "feature request".
---
commenter: fwiemann
posted: 2006-05-03 21:42:09
title: #6 support unicode box characters for table markup
Logged In: YES
user\_id=1374215
This is not \*quite\* easy to implement because we'd need to
distinguish different types of intersection characters: not
only "+" for everything, but "┌", "┬", "┐", etc.
---
commenter: milde
posted: 2015-02-19 13:01:05.637000
title: #6 support unicode box characters for table markup
It should suffice to map all intersection characters to "+".
| 1.0 | support unicode box characters for table markup [SF:feature-requests:6] -
author: trentbuck
created: 2006-03-05 23:41:02
assigned: None
SF_url: https://sourceforge.net/p/docutils/feature-requests/6
See first attachment.
---
commenter: trentbuck
posted: 2006-03-05 23:41:02
title: #6 support unicode box characters for table markup
attachments:
- https://sourceforge.net/p/docutils/feature-requests/_discuss/thread/03996de8/4305/attachment/tables.txt
issue description
---
commenter: goodger
posted: 2006-03-06 20:39:33
title: #6 support unicode box characters for table markup
Logged In: YES
user\_id=7733
Unicode contains many characters that would be suitable for
reST markup. Few are currently supported. Adding this
feature would not be exceedingly difficult, but I'd classify
it as low-priority. We depend on developers to implement
features that interest them; patches are welcome\!
Not a bug; changed to "feature request".
---
commenter: fwiemann
posted: 2006-05-03 21:42:09
title: #6 support unicode box characters for table markup
Logged In: YES
user\_id=1374215
This is not \*quite\* easy to implement because we'd need to
distinguish different types of intersection characters: not
only "+" for everything, but "┌", "┬", "┐", etc.
---
commenter: milde
posted: 2015-02-19 13:01:05.637000
title: #6 support unicode box characters for table markup
It should suffice to map all intersection characters to "+".
| non_build | support unicode box characters for table markup author trentbuck created assigned none sf url see first attachment commenter trentbuck posted title support unicode box characters for table markup attachments issue description commenter goodger posted title support unicode box characters for table markup logged in yes user id unicode contains many characters that would be suitable for rest markup few are currently supported adding this feature would not be exceedingly difficult but i d classify it as low priority we depend on developers to implement features that interest them patches are welcome not a bug changed to feature request commenter fwiemann posted title support unicode box characters for table markup logged in yes user id this is not quite easy to implement because we d need to distinguish different types of intersection characters not only for everything but ┌ ┬ ┐ etc commenter milde posted title support unicode box characters for table markup it should suffice to map all intersection characters to | 0 |
21,738 | 7,065,112,867 | IssuesEvent | 2018-01-06 16:07:39 | jupyterlab/jupyterlab | https://api.github.com/repos/jupyterlab/jupyterlab | closed | Rogue intransigent extension | status:Needs Discussion tag:Build System | #I was working with @ian-r-rose on the jupyterlab-latex extension, and in the course of attempting to debug it, we attempted to link to the extension in a number of ways.
Most notably, we used a number of methods including
- `jlpm install && jlpm run build && jupyterlab labextension install`
- `jupyterlab labextension link`
- `jlpm run add:sibling ../jupyterlab-latex` (from inside the jupyterlab repo)
We were running `jupyterlab` both with the public packages (`jupyter lab --watch` with `jlpm run watch` in `jupyterlab-latex`) and (`jupyter lab --dev-mode --watch` after running `jlpm run add:sibling ../jupyterlab-latex`).
Unfortunately, I've not been able to make the extension go away after many attempts in a variety of methods:
- `jlpm run clean` (in both jupyterlab-latex and jupyterlab)
- `jlpm run clean:slate` (in jupyterlab)
- `git clean -xfd` (in jupyterlab-latex)
- `git clean -xfd && jlpm install && jlpm run build` (in jupyterlab)
- `jupyterlab labextension uninstall @jupyterlab/latex`
- `jupyterlab labextension unlink @jupyterlab/latex`
- `jupyter_conf_search latex` → leading to me removing the `"jupyterlab_latex": true,` line
Even after all this I'm still getting the claim `@jupyterlab/latex` exists when I run `jupyter labextension list`
```
Build recommended:
@jupyterlab/latex needs to be removed
@jupyterlab/latex needs to be removed
```
and with `import ipdb; ipdb.set_trace()` inside the `jupyterlab/commands.py` `build_check` I'm still seeing both
```
'extensions': {'@jupyterlab/application-extension': '',
⋮
'@jupyterlab/latex': 'lib/latex',
⋮
```
```
'linkedPackages': {'@jupyterlab/latex': '/Users/mpacer/jupyter/jupyterlab-latex'},
```
and
```
'mimeExtensions': {'@jupyterlab/json-extension': '',
'@jupyterlab/latex': 'lib/pdf',
'@jupyterlab/pdf-extension': '',
⋮
```
I'm going to keep trying to debug this but this is still way too complicated in order to figure this out. Happy to help figure out how to make this nicer.
| 1.0 | Rogue intransigent extension - #I was working with @ian-r-rose on the jupyterlab-latex extension, and in the course of attempting to debug it, we attempted to link to the extension in a number of ways.
Most notably, we used a number of methods including
- `jlpm install && jlpm run build && jupyterlab labextension install`
- `jupyterlab labextension link`
- `jlpm run add:sibling ../jupyterlab-latex` (from inside the jupyterlab repo)
We were running `jupyterlab` both with the public packages (`jupyter lab --watch` with `jlpm run watch` in `jupyterlab-latex`) and (`jupyter lab --dev-mode --watch` after running `jlpm run add:sibling ../jupyterlab-latex`).
Unfortunately, I've not been able to make the extension go away after many attempts in a variety of methods:
- `jlpm run clean` (in both jupyterlab-latex and jupyterlab)
- `jlpm run clean:slate` (in jupyterlab)
- `git clean -xfd` (in jupyterlab-latex)
- `git clean -xfd && jlpm install && jlpm run build` (in jupyterlab)
- `jupyterlab labextension uninstall @jupyterlab/latex`
- `jupyterlab labextension unlink @jupyterlab/latex`
- `jupyter_conf_search latex` → leading to me removing the `"jupyterlab_latex": true,` line
Even after all this I'm still getting the claim `@jupyterlab/latex` exists when I run `jupyter labextension list`
```
Build recommended:
@jupyterlab/latex needs to be removed
@jupyterlab/latex needs to be removed
```
and with `import ipdb; ipdb.set_trace()` inside the `jupyterlab/commands.py` `build_check` I'm still seeing both
```
'extensions': {'@jupyterlab/application-extension': '',
⋮
'@jupyterlab/latex': 'lib/latex',
⋮
```
```
'linkedPackages': {'@jupyterlab/latex': '/Users/mpacer/jupyter/jupyterlab-latex'},
```
and
```
'mimeExtensions': {'@jupyterlab/json-extension': '',
'@jupyterlab/latex': 'lib/pdf',
'@jupyterlab/pdf-extension': '',
⋮
```
I'm going to keep trying to debug this but this is still way too complicated in order to figure this out. Happy to help figure out how to make this nicer.
| build | rogue intransigent extension i was working with ian r rose on the jupyterlab latex extension and in the course of attempting to debug it we attempted to link to the extension in a number of ways most notably we used a number of methods including jlpm install jlpm run build jupyterlab labextension install jupyterlab labextension link jlpm run add sibling jupyterlab latex from inside the jupyterlab repo we were running jupyterlab both with the public packages jupyter lab watch with jlpm run watch in jupyterlab latex and jupyter lab dev mode watch after running jlpm run add sibling jupyterlab latex unfortunately i ve not been able to make the extension go away after many attempts in a variety of methods jlpm run clean in both jupyterlab latex and jupyterlab jlpm run clean slate in jupyterlab git clean xfd in jupyterlab latex git clean xfd jlpm install jlpm run build in jupyterlab jupyterlab labextension uninstall jupyterlab latex jupyterlab labextension unlink jupyterlab latex jupyter conf search latex → leading to me removing the jupyterlab latex true line even after all this i m still getting the claim jupyterlab latex exists when i run jupyter labextension list build recommended jupyterlab latex needs to be removed jupyterlab latex needs to be removed and with import ipdb ipdb set trace inside the jupyterlab commands py build check i m still seeing both extensions jupyterlab application extension ⋮ jupyterlab latex lib latex ⋮ linkedpackages jupyterlab latex users mpacer jupyter jupyterlab latex and mimeextensions jupyterlab json extension jupyterlab latex lib pdf jupyterlab pdf extension ⋮ i m going to keep trying to debug this but this is still way too complicated in order to figure this out happy to help figure out how to make this nicer | 1 |
289,741 | 21,790,654,232 | IssuesEvent | 2022-05-14 21:10:23 | sirfuzzalot/textual-inputs | https://api.github.com/repos/sirfuzzalot/textual-inputs | closed | Add Contributing Guide | documentation | **USER STORY**
As a developer I want clear guidance on how I may contribute my work to this project
**REQUIREMENTS**
- add a contributing guide | 1.0 | Add Contributing Guide - **USER STORY**
As a developer I want clear guidance on how I may contribute my work to this project
**REQUIREMENTS**
- add a contributing guide | non_build | add contributing guide user story as a developer i want clear guidance on how i may contribute my work to this project requirements add a contributing guide | 0 |
84,844 | 10,568,085,964 | IssuesEvent | 2019-10-06 10:21:55 | elementary/files | https://api.github.com/repos/elementary/files | opened | Give aggregate info in overlay in ListView | Needs Design Priority: Wishlist | ## Prerequisites
- [* ] I have searched open and closed issues for duplicates.
## Feature
In ListView the information overlay is suppressed. This makes sense for single files as the same information is available in the view. However, it would be useful, and more consistent to show aggregate information when multiple files are selected, as happens in the other views.
**Describe the solution you'd like**
Do not suppress the information overlay when more than one file is selected when in ListView.
**Existing work**
All the required code exists in Files - it should just require tweaking the suppression logic.
| 1.0 | Give aggregate info in overlay in ListView - ## Prerequisites
- [* ] I have searched open and closed issues for duplicates.
## Feature
In ListView the information overlay is suppressed. This makes sense for single files as the same information is available in the view. However, it would be useful, and more consistent to show aggregate information when multiple files are selected, as happens in the other views.
**Describe the solution you'd like**
Do not suppress the information overlay when more than one file is selected when in ListView.
**Existing work**
All the required code exists in Files - it should just require tweaking the suppression logic.
| non_build | give aggregate info in overlay in listview prerequisites i have searched open and closed issues for duplicates feature in listview the information overlay is suppressed this makes sense for single files as the same information is available in the view however it would be useful and more consistent to show aggregate information when multiple files are selected as happens in the other views describe the solution you d like do not suppress the information overlay when more than one file is selected when in listview existing work all the required code exists in files it should just require tweaking the suppression logic | 0 |
37,403 | 15,285,064,554 | IssuesEvent | 2021-02-23 13:05:04 | alexwonet/challenge-new | https://api.github.com/repos/alexwonet/challenge-new | closed | Your Own IOT Service | Your Own IOT Service | # Cloud deployment
Your challenge is to develop and deploy your own IoT service to the SAP Cloud Platform CloudFoundry environment.
## Step 1 - Prepare service
First we must make sure that your new service implements the Open Service Broker API.
### The Open Service Broker API spec and what it's good for
https://ict.swisscom.ch/2019/12/open-service-broker-api/
https://github.com/openservicebrokerapi/servicebroker/blob/master/spec.md
### Endpoints to implement:
- Catalog (GET)
- Provision (PUT)
- Bind (GET, PUT)
---
### Requirements
1.1. Your service should allow to create a service instance that match an active device on http://localhost:9090/device.
1.2. The device binding should return the full device credentials as
```
{
"credentials": {
"connection_string": "device://438bcdca-c714-4c4a-8ba0-da975dec49f1/fNhptfPn",
"device_type": "ROUTER",
"id": "438bcdca-c714-4c4a-8ba0-da975dec49f1",
"state": "Online",
}
}
```
1.3. Your service should support basic authentication and allow access for admin:admin.
https://docs.iris-go.com/iris/security/basicauth
1.4. The create instance endpoint should support the device_type service instance param (ROUTER,MODEM,SWITCH)
See provisioning parameters:
https://github.com/openservicebrokerapi/servicebroker/blob/master/spec.md#provisioning
### Recommended Golang libs
https://github.com/pmorie/osb-broker-lib
This library is most useful if you want to build your own broker from scratch
Golang server library:
https://www.iris-go.com/docs/#
Connecting to the /device endpoint
https://github.com/go-resty/resty
osb-broker-lib with Golang server example:
```go
package main
import (
"github.com/fatih/structs"
"github.com/gavv/httpexpect"
"github.com/kataras/iris/v12"
"github.com/onsi/ginkgo"
"github.wdf.sap.corp/cloudenablers/challenger/pkg/types"
"github.wdf.sap.corp/cloudenablers/challenger/test/utils"
"gopkg.in/yaml.v3"
"net"
"net/http"
"time"
osb "sigs.k8s.io/go-open-service-broker-client/v2"
)
func main() {
app := iris.New()
brokerAPIs := app.Party("/v2")
{
brokerAPIs.Get("/catalog", GetCatalog)
brokerAPIs.Get("/service_instances/{instance_id:string}/service_bindings/{binding_id:string}", GetServiceBinding)
brokerAPIs.Put("/service_instances/{instance_id:string}", CreateServiceInstance)
brokerAPIs.Put("/service_instances/{instance_id:string}/service_bindings/{binding_id:string}", CreateServiceBinding)
}
serverPort := ":" + os.Getenv("PORT")
app.Listen(serverPort)
}
func CreateServiceInstance(ctx iris.Context) {
provisionDetails := &osb.ProvisionRequest{}
instanceID = ctx.Params().Get("instance_id")
deviceType := provisionDetails.Parameters["device_type"]
iotBackendURL := "http://localhost:9090"
apiCookieKey := "YOUR-API_COOKIE"
reqBody, _ := ctx.GetBody()
json.Unmarshal(reqBody, provisionDetails)
//find device that matches thew provided deviceType in the provisionDetails
deviceType := provisionDetails.Parameters["device_type"]
client := getClient(iotBackendURL, apiCookieKey)
pages := utils.GetDevicesPages(client) // Replace with your client impl (resty)
for _, item := range pages {} // go over the device service pages result
ctx.JSON(&osb.ProvisionResponse{})
}
func GetServiceBinding(ctx iris.Context) {
ctx.JSON(&osb.GetBindingResponse{
Credentials: structs.Map(foundItem),
})
}
func CreateServiceBinding(ctx iris.Context) {
ctx.JSON(&osb.BindResponse{
Credentials: structs.Map(foundItem),
})
}
func GetCatalog(ctx iris.Context) {
response := &osb.CatalogResponse{}
data := `
---
services:
- name: your-iot-service
id: 4f6e6cf6-ffdd-425f-a2c7-3c9258ad246e-alexwonet
description: The example service!
bindable: true
metadata:
displayName: "Example service"
imageUrl: https://avatars2.githubusercontent.com/u/19862012?s=200&v=4
plans:
- name: iot-service
id: iot-service-alexwonet
description: The default plan for the service
free: true
`
yaml.Unmarshal([]byte(data), &response)
ctx.JSON(response)
}
```
### Running Golang apps on cloudfoundry:
Define a correct port for the server to listen on:
```
serverPort := ":" + os.Getenv("PORT")
app.Listen(serverPort)
```
### define your cloudfoundry manifest.yml
```
applications:
- name: alexwonet-challenge-app
instances: 1
buildpacks:
- go_buildpack
```
## Step 2 - Deploy Your Service
Create a new [trial account](https://www.sap.com/products/cloud-platform.html) on SAP CF - (Hint) [tutorial](https://developers.sap.com/tutorials/hcp-create-trial-account.html).
[Deploy](https://docs.cloudfoundry.org/devguide/deploy-apps/manifest.html) the service and create a [service broker](https://docs.cloudfoundry.org/services/managing-service-brokers.html).
Lastly, create a [service instance](https://docs.cloudfoundry.org/devguide/services/managing-services.html) and [bind](http://cli.cloudfoundry.org/en-US/v6/bind-service.html) the app!
If all works well go back to http://localhost:9090 and try out your service.
### Try out will try to confirm that
1. We can create a service instance -> select a real device from the "/device" list api
2. Bind to it -> get the credentials to use the device
3. Confirm the device actually matches an active & cool device (not above 29°C)
## Step 3 - Validate & pass CI
In your selected "challenge-new" repository create:
`.github/workflows/golangci-lint.yml`
```
GET /device/{id}/status
name: golangci-lint
on:
push:
tags:
- v*
branches:
- master
- main
pull_request:
release:
types: # This configuration does not affect the page_build event above
- created
jobs:
golangci:
name: lint
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- name: golangci-lint
uses: golangci/golangci-lint-action@v2
with:
# Required: the version of golangci-lint is required and must be specified without patch version: we always use the latest patch version.
version: v1.33
# Optional: working directory, useful for monorepos
# working-directory: somedir
# Optional: golangci-lint command line arguments.
# args: --issues-exit-code=0
# Optional: show only new issues if it's a pull request. The default value is `false`.
# only-new-issues: true
```
3.1 Make sure that your go code for this challenge is in the repository master branch.
3.2 Fix all lint issues that golangci will report (we want our code to be clean)
3.3 Check that the CI pass (In github actions) & Create a new release
If all works well go back to http://localhost:9090 and validate your release pass CI!
### Use full Links
https://github.com/golangci/golangci-lint
https://github.com/features/actions
Good luck!!!
| 1.0 | Your Own IOT Service - # Cloud deployment
Your challenge is to develop and deploy your own IoT service to the SAP Cloud Platform CloudFoundry environment.
## Step 1 - Prepare service
First we must make sure that your new service implements the Open Service Broker API.
### The Open Service Broker API spec and what it's good for
https://ict.swisscom.ch/2019/12/open-service-broker-api/
https://github.com/openservicebrokerapi/servicebroker/blob/master/spec.md
### Endpoints to implement:
- Catalog (GET)
- Provision (PUT)
- Bind (GET, PUT)
---
### Requirements
1.1. Your service should allow to create a service instance that match an active device on http://localhost:9090/device.
1.2. The device binding should return the full device credentials as
```
{
"credentials": {
"connection_string": "device://438bcdca-c714-4c4a-8ba0-da975dec49f1/fNhptfPn",
"device_type": "ROUTER",
"id": "438bcdca-c714-4c4a-8ba0-da975dec49f1",
"state": "Online",
}
}
```
1.3. Your service should support basic authentication and allow access for admin:admin.
https://docs.iris-go.com/iris/security/basicauth
1.4. The create instance endpoint should support the device_type service instance param (ROUTER,MODEM,SWITCH)
See provisioning parameters:
https://github.com/openservicebrokerapi/servicebroker/blob/master/spec.md#provisioning
### Recommended Golang libs
https://github.com/pmorie/osb-broker-lib
This library is most useful if you want to build your own broker from scratch
Golang server library:
https://www.iris-go.com/docs/#
Connecting to the /device endpoint
https://github.com/go-resty/resty
osb-broker-lib with Golang server example:
```go
package main
import (
"github.com/fatih/structs"
"github.com/gavv/httpexpect"
"github.com/kataras/iris/v12"
"github.com/onsi/ginkgo"
"github.wdf.sap.corp/cloudenablers/challenger/pkg/types"
"github.wdf.sap.corp/cloudenablers/challenger/test/utils"
"gopkg.in/yaml.v3"
"net"
"net/http"
"time"
osb "sigs.k8s.io/go-open-service-broker-client/v2"
)
func main() {
app := iris.New()
brokerAPIs := app.Party("/v2")
{
brokerAPIs.Get("/catalog", GetCatalog)
brokerAPIs.Get("/service_instances/{instance_id:string}/service_bindings/{binding_id:string}", GetServiceBinding)
brokerAPIs.Put("/service_instances/{instance_id:string}", CreateServiceInstance)
brokerAPIs.Put("/service_instances/{instance_id:string}/service_bindings/{binding_id:string}", CreateServiceBinding)
}
serverPort := ":" + os.Getenv("PORT")
app.Listen(serverPort)
}
func CreateServiceInstance(ctx iris.Context) {
provisionDetails := &osb.ProvisionRequest{}
instanceID = ctx.Params().Get("instance_id")
deviceType := provisionDetails.Parameters["device_type"]
iotBackendURL := "http://localhost:9090"
apiCookieKey := "YOUR-API_COOKIE"
reqBody, _ := ctx.GetBody()
json.Unmarshal(reqBody, provisionDetails)
//find device that matches thew provided deviceType in the provisionDetails
deviceType := provisionDetails.Parameters["device_type"]
client := getClient(iotBackendURL, apiCookieKey)
pages := utils.GetDevicesPages(client) // Replace with your client impl (resty)
for _, item := range pages {} // go over the device service pages result
ctx.JSON(&osb.ProvisionResponse{})
}
func GetServiceBinding(ctx iris.Context) {
ctx.JSON(&osb.GetBindingResponse{
Credentials: structs.Map(foundItem),
})
}
func CreateServiceBinding(ctx iris.Context) {
ctx.JSON(&osb.BindResponse{
Credentials: structs.Map(foundItem),
})
}
func GetCatalog(ctx iris.Context) {
response := &osb.CatalogResponse{}
data := `
---
services:
- name: your-iot-service
id: 4f6e6cf6-ffdd-425f-a2c7-3c9258ad246e-alexwonet
description: The example service!
bindable: true
metadata:
displayName: "Example service"
imageUrl: https://avatars2.githubusercontent.com/u/19862012?s=200&v=4
plans:
- name: iot-service
id: iot-service-alexwonet
description: The default plan for the service
free: true
`
yaml.Unmarshal([]byte(data), &response)
ctx.JSON(response)
}
```
### Running Golang apps on cloudfoundry:
Define a correct port for the server to listen on:
```
serverPort := ":" + os.Getenv("PORT")
app.Listen(serverPort)
```
### define your cloudfoundry manifest.yml
```
applications:
- name: alexwonet-challenge-app
instances: 1
buildpacks:
- go_buildpack
```
## Step 2 - Deploy Your Service
Create a new [trial account](https://www.sap.com/products/cloud-platform.html) on SAP CF - (Hint) [tutorial](https://developers.sap.com/tutorials/hcp-create-trial-account.html).
[Deploy](https://docs.cloudfoundry.org/devguide/deploy-apps/manifest.html) the service and create a [service broker](https://docs.cloudfoundry.org/services/managing-service-brokers.html).
Lastly, create a [service instance](https://docs.cloudfoundry.org/devguide/services/managing-services.html) and [bind](http://cli.cloudfoundry.org/en-US/v6/bind-service.html) the app!
If all works well go back to http://localhost:9090 and try out your service.
### Try out will try to confirm that
1. We can create a service instance -> select a real device from the "/device" list api
2. Bind to it -> get the credentials to use the device
3. Confirm the device actually matches an active & cool device (not above 29°C)
## Step 3 - Validate & pass CI
In your selected "challenge-new" repository create:
`.github/workflows/golangci-lint.yml`
```
GET /device/{id}/status
name: golangci-lint
on:
push:
tags:
- v*
branches:
- master
- main
pull_request:
release:
types: # This configuration does not affect the page_build event above
- created
jobs:
golangci:
name: lint
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- name: golangci-lint
uses: golangci/golangci-lint-action@v2
with:
# Required: the version of golangci-lint is required and must be specified without patch version: we always use the latest patch version.
version: v1.33
# Optional: working directory, useful for monorepos
# working-directory: somedir
# Optional: golangci-lint command line arguments.
# args: --issues-exit-code=0
# Optional: show only new issues if it's a pull request. The default value is `false`.
# only-new-issues: true
```
3.1 Make sure that your go code for this challenge is in the repository master branch.
3.2 Fix all lint issues that golangci will report (we want our code to be clean)
3.3 Check that the CI pass (In github actions) & Create a new release
If all works well go back to http://localhost:9090 and validate your release pass CI!
### Use full Links
https://github.com/golangci/golangci-lint
https://github.com/features/actions
Good luck!!!
| non_build | your own iot service cloud deployment your challenge is to develop and deploy your own iot service to the sap cloud platform cloudfoundry environment step prepare service first we must make sure that your new service implements the open service broker api the open service broker api spec and what it s good for endpoints to implement catalog get provision put bind get put requirements your service should allow to create a service instance that match an active device on the device binding should return the full device credentials as credentials connection string device fnhptfpn device type router id state online your service should support basic authentication and allow access for admin admin the create instance endpoint should support the device type service instance param router modem switch see provisioning parameters recommended golang libs this library is most useful if you want to build your own broker from scratch golang server library connecting to the device endpoint osb broker lib with golang server example go package main import github com fatih structs github com gavv httpexpect github com kataras iris github com onsi ginkgo github wdf sap corp cloudenablers challenger pkg types github wdf sap corp cloudenablers challenger test utils gopkg in yaml net net http time osb sigs io go open service broker client func main app iris new brokerapis app party brokerapis get catalog getcatalog brokerapis get service instances instance id string service bindings binding id string getservicebinding brokerapis put service instances instance id string createserviceinstance brokerapis put service instances instance id string service bindings binding id string createservicebinding serverport os getenv port app listen serverport func createserviceinstance ctx iris context provisiondetails osb provisionrequest instanceid ctx params get instance id devicetype provisiondetails parameters iotbackendurl apicookiekey your api cookie reqbody ctx getbody json unmarshal reqbody provisiondetails find device that matches thew provided devicetype in the provisiondetails devicetype provisiondetails parameters client getclient iotbackendurl apicookiekey pages utils getdevicespages client replace with your client impl resty for item range pages go over the device service pages result ctx json osb provisionresponse func getservicebinding ctx iris context ctx json osb getbindingresponse credentials structs map founditem func createservicebinding ctx iris context ctx json osb bindresponse credentials structs map founditem func getcatalog ctx iris context response osb catalogresponse data services name your iot service id ffdd alexwonet description the example service bindable true metadata displayname example service imageurl plans name iot service id iot service alexwonet description the default plan for the service free true yaml unmarshal byte data response ctx json response running golang apps on cloudfoundry define a correct port for the server to listen on serverport os getenv port app listen serverport define your cloudfoundry manifest yml applications name alexwonet challenge app instances buildpacks go buildpack step deploy your service create a new on sap cf hint the service and create a lastly create a and the app if all works well go back to and try out your service try out will try to confirm that we can create a service instance select a real device from the device list api bind to it get the credentials to use the device confirm the device actually matches an active cool device not above °c step validate pass ci in your selected challenge new repository create github workflows golangci lint yml get device id status name golangci lint on push tags v branches master main pull request release types this configuration does not affect the page build event above created jobs golangci name lint runs on ubuntu latest steps uses actions checkout name golangci lint uses golangci golangci lint action with required the version of golangci lint is required and must be specified without patch version we always use the latest patch version version optional working directory useful for monorepos working directory somedir optional golangci lint command line arguments args issues exit code optional show only new issues if it s a pull request the default value is false only new issues true make sure that your go code for this challenge is in the repository master branch fix all lint issues that golangci will report we want our code to be clean check that the ci pass in github actions create a new release if all works well go back to and validate your release pass ci use full links good luck | 0 |
93,383 | 26,935,230,333 | IssuesEvent | 2023-02-07 20:03:40 | openhwgroup/cva6 | https://api.github.com/repos/openhwgroup/cva6 | closed | Unable to view waveform of variable in Questasim | Component:Tool-and-build Type:Question Status:Invalid | In corev_apu/fpga/src, the axi2apb directory contains the axi2apb_64_32.sv. axi2apb_64_32.sv has two variables CS, NS of type enum logic. I am trying to simulate a testbench for this module using Questasim. As shown in the attached screenshot, these two variables CS, NS don't appear in the simulation window. I wish to view the waveforms for these two variables. Can you please advise how to view them in Questasim?

| 1.0 | Unable to view waveform of variable in Questasim - In corev_apu/fpga/src, the axi2apb directory contains the axi2apb_64_32.sv. axi2apb_64_32.sv has two variables CS, NS of type enum logic. I am trying to simulate a testbench for this module using Questasim. As shown in the attached screenshot, these two variables CS, NS don't appear in the simulation window. I wish to view the waveforms for these two variables. Can you please advise how to view them in Questasim?

| build | unable to view waveform of variable in questasim in corev apu fpga src the directory contains the sv sv has two variables cs ns of type enum logic i am trying to simulate a testbench for this module using questasim as shown in the attached screenshot these two variables cs ns don t appear in the simulation window i wish to view the waveforms for these two variables can you please advise how to view them in questasim | 1 |
538,880 | 15,780,282,852 | IssuesEvent | 2021-04-01 09:45:44 | sopra-fs21-group-03/Server | https://api.github.com/repos/sopra-fs21-group-03/Server | closed | Input in fields is handled correctly and wehen registering/loging in the correct api call is made | medium priority task | Time estimate: 2h
This task is part of user story #1 | 1.0 | Input in fields is handled correctly and wehen registering/loging in the correct api call is made - Time estimate: 2h
This task is part of user story #1 | non_build | input in fields is handled correctly and wehen registering loging in the correct api call is made time estimate this task is part of user story | 0 |
821,440 | 30,823,411,723 | IssuesEvent | 2023-08-01 18:03:55 | googleapis/google-cloud-go | https://api.github.com/repos/googleapis/google-cloud-go | closed | profiler: TestAgentIntegration failed | type: bug priority: p1 api: cloudprofiler flakybot: issue | Note: #7152 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: b2964a92d093a3d751e8b43aa70471bd989b1244
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f532917b-d222-41c3-89ee-274229ab20e1), [Sponge](http://sponge2/f532917b-d222-41c3-89ee-274229ab20e1)
status: failed
<details><summary>Test output</summary><br><pre> proftest.go:481: Serial port output for profiler-test-gomaster-2023-07-20-15-47-00-115169-pdt:
\033[2J\033[01;01H\033[=3h\033[2J\033[01;01H\033[2J\033[01;01H\033[=3h\033[2J\033[01;01HBdsDxe: loading Boot0001 "UEFI Google PersistentDisk " from PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0)
BdsDxe: starting Boot0001 "UEFI Google PersistentDisk " from PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0)
UEFI: Attempting to start image.
Description: UEFI Google PersistentDisk
FilePath: PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0)
OptionNumber: 1.
\033[0m\033[30m\033[47mWelcome to GRUB!
\033[0m\033[37m\033[40m\033[0m\033[30m\033[40m\033[2J\033[01;01H\033[0m\033[37m\033[40m\033[0m\033[30m\033[40m\033[2J\033[01;01H\033[0m\033[37m\033[40m Booting `Debian GNU/Linux'
Loading Linux 5.10.0-23-cloud-amd64 ...
Loading initial ramdisk ...
error: no suitable video mode found.
Booting in blind mode
Thu Jul 20 22:47:26 UTC 2023: + retry apt-get update
Thu Jul 20 22:47:26 UTC 2023: + for i in {1..3}
Thu Jul 20 22:47:26 UTC 2023: + '[' 1 == 1 ']'
Thu Jul 20 22:47:26 UTC 2023: + apt-get update
Thu Jul 20 22:47:28 UTC 2023: + return 0
Thu Jul 20 22:47:28 UTC 2023: + retry apt-get -y -q install git
Thu Jul 20 22:47:28 UTC 2023: + for i in {1..3}
Thu Jul 20 22:47:28 UTC 2023: + '[' 1 == 1 ']'
Thu Jul 20 22:47:28 UTC 2023: + apt-get -y -q install git
Thu Jul 20 22:47:29 UTC 2023: debconf: unable to initialize frontend: Dialog
Thu Jul 20 22:47:29 UTC 2023: debconf: (TERM is not set, so the dialog frontend is not usable.)
Thu Jul 20 22:47:29 UTC 2023: debconf: falling back to frontend: Readline
Thu Jul 20 22:47:29 UTC 2023: debconf: unable to initialize frontend: Readline
Thu Jul 20 22:47:29 UTC 2023: debconf: (Can't locate Term/ReadLine.pm in @INC (you may need to install the Term::ReadLine module) (@INC contains: /etc/perl /usr/local/lib/x86_64-linux-gnu/perl/5.32.1 /usr/local/share/perl/5.32.1 /usr/lib/x86_64-linux-gnu/perl5/5.32 /usr/share/perl5 /usr/lib/x86_64-linux-gnu/perl-base /usr/lib/x86_64-linux-gnu/perl/5.32 /usr/share/perl/5.32 /usr/local/lib/site_perl) at /usr/share/perl5/Debconf/FrontEnd/Readline.pm line 7, <> line 8.)
Thu Jul 20 22:47:29 UTC 2023: debconf: falling back to frontend: Teletype
Thu Jul 20 22:47:29 UTC 2023: dpkg-preconfigure: unable to re-open stdin:
Thu Jul 20 22:47:39 UTC 2023: + return 0
Thu Jul 20 22:47:39 UTC 2023: + mkdir -p /tmp/gocache
Thu Jul 20 22:47:39 UTC 2023: + export GOCACHE=/tmp/gocache
Thu Jul 20 22:47:39 UTC 2023: + GOCACHE=/tmp/gocache
Thu Jul 20 22:47:39 UTC 2023: + '[' master = master ']'
Thu Jul 20 22:47:39 UTC 2023: + retry apt-get -y -q install gcc
Thu Jul 20 22:47:39 UTC 2023: + for i in {1..3}
Thu Jul 20 22:47:39 UTC 2023: + '[' 1 == 1 ']'
Thu Jul 20 22:47:39 UTC 2023: + apt-get -y -q install gcc
Thu Jul 20 22:47:40 UTC 2023: debconf: unable to initialize frontend: Dialog
Thu Jul 20 22:47:40 UTC 2023: debconf: (TERM is not set, so the dialog frontend is not usable.)
Thu Jul 20 22:47:40 UTC 2023: debconf: falling back to frontend: Readline
Thu Jul 20 22:47:40 UTC 2023: debconf: unable to initialize frontend: Readline
Thu Jul 20 22:47:40 UTC 2023: debconf: (This frontend requires a controlling tty.)
Thu Jul 20 22:47:40 UTC 2023: debconf: falling back to frontend: Teletype
Thu Jul 20 22:47:40 UTC 2023: dpkg-preconfigure: unable to re-open stdin:
Thu Jul 20 22:48:13 UTC 2023: + return 0
Thu Jul 20 22:48:13 UTC 2023: + mkdir -p /tmp/bin
Thu Jul 20 22:48:13 UTC 2023: + retry curl -sL -o /tmp/bin/gimme https://raw.githubusercontent.com/travis-ci/gimme/master/gimme
Thu Jul 20 22:48:13 UTC 2023: + for i in {1..3}
Thu Jul 20 22:48:13 UTC 2023: + '[' 1 == 1 ']'
Thu Jul 20 22:48:13 UTC 2023: + curl -sL -o /tmp/bin/gimme https://raw.githubusercontent.com/travis-ci/gimme/master/gimme
Thu Jul 20 22:48:13 UTC 2023: + return 0
Thu Jul 20 22:48:13 UTC 2023: + chmod +x /tmp/bin/gimme
Thu Jul 20 22:48:14 UTC 2023: + export PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/tmp/bin
Thu Jul 20 22:48:14 UTC 2023: + PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/tmp/bin
Thu Jul 20 22:48:14 UTC 2023: + retry gimme_retrier
Thu Jul 20 22:48:14 UTC 2023: + for i in {1..3}
Thu Jul 20 22:48:14 UTC 2023: + '[' 1 == 1 ']'
Thu Jul 20 22:48:14 UTC 2023: + gimme_retrier
Thu Jul 20 22:48:14 UTC 2023: ++ gimme master
Thu Jul 20 22:49:18 UTC 2023: I don't have any idea what to do with 'master'.
Thu Jul 20 22:49:18 UTC 2023: (using download type 'auto')
Thu Jul 20 22:49:18 UTC 2023: + eval ''
Thu Jul 20 22:49:18 UTC 2023: + which go
Thu Jul 20 22:49:18 UTC 2023: + for i in {1..3}
Thu Jul 20 22:49:18 UTC 2023: + '[' 2 == 1 ']'
Thu Jul 20 22:49:18 UTC 2023: + sleep 10
Thu Jul 20 22:49:28 UTC 2023: + gimme_retrier
Thu Jul 20 22:49:28 UTC 2023: ++ gimme master
Thu Jul 20 22:49:29 UTC 2023: I don't have any idea what to do with 'master'.
Thu Jul 20 22:49:29 UTC 2023: (using download type 'auto')
Thu Jul 20 22:49:29 UTC 2023: + eval ''
Thu Jul 20 22:49:29 UTC 2023: + which go
Thu Jul 20 22:49:29 UTC 2023: + for i in {1..3}
Thu Jul 20 22:49:29 UTC 2023: + '[' 3 == 1 ']'
Thu Jul 20 22:49:29 UTC 2023: + sleep 10
Thu Jul 20 22:49:39 UTC 2023: + gimme_retrier
Thu Jul 20 22:49:39 UTC 2023: ++ gimme master
Thu Jul 20 22:49:40 UTC 2023: I don't have any idea what to do with 'master'.
Thu Jul 20 22:49:40 UTC 2023: (using download type 'auto')
Thu Jul 20 22:49:40 UTC 2023: + eval ''
Thu Jul 20 22:49:40 UTC 2023: + which go
Thu Jul 20 22:49:40 UTC 2023: + return 1
Thu Jul 20 22:49:40 UTC 2023: ++ echo 'failed to set up or run the benchmark'
Thu Jul 20 22:49:40 UTC 2023: failed to set up or run the benchmark
Thu Jul 20 22:49:40 UTC 2023: ++ sleep 300
integration_test.go:364: PollAndLogSerialPort() got error: failed to execute the prober benchmark script</pre></details> | 1.0 | profiler: TestAgentIntegration failed - Note: #7152 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: b2964a92d093a3d751e8b43aa70471bd989b1244
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f532917b-d222-41c3-89ee-274229ab20e1), [Sponge](http://sponge2/f532917b-d222-41c3-89ee-274229ab20e1)
status: failed
<details><summary>Test output</summary><br><pre> proftest.go:481: Serial port output for profiler-test-gomaster-2023-07-20-15-47-00-115169-pdt:
\033[2J\033[01;01H\033[=3h\033[2J\033[01;01H\033[2J\033[01;01H\033[=3h\033[2J\033[01;01HBdsDxe: loading Boot0001 "UEFI Google PersistentDisk " from PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0)
BdsDxe: starting Boot0001 "UEFI Google PersistentDisk " from PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0)
UEFI: Attempting to start image.
Description: UEFI Google PersistentDisk
FilePath: PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0)
OptionNumber: 1.
\033[0m\033[30m\033[47mWelcome to GRUB!
\033[0m\033[37m\033[40m\033[0m\033[30m\033[40m\033[2J\033[01;01H\033[0m\033[37m\033[40m\033[0m\033[30m\033[40m\033[2J\033[01;01H\033[0m\033[37m\033[40m Booting `Debian GNU/Linux'
Loading Linux 5.10.0-23-cloud-amd64 ...
Loading initial ramdisk ...
error: no suitable video mode found.
Booting in blind mode
Thu Jul 20 22:47:26 UTC 2023: + retry apt-get update
Thu Jul 20 22:47:26 UTC 2023: + for i in {1..3}
Thu Jul 20 22:47:26 UTC 2023: + '[' 1 == 1 ']'
Thu Jul 20 22:47:26 UTC 2023: + apt-get update
Thu Jul 20 22:47:28 UTC 2023: + return 0
Thu Jul 20 22:47:28 UTC 2023: + retry apt-get -y -q install git
Thu Jul 20 22:47:28 UTC 2023: + for i in {1..3}
Thu Jul 20 22:47:28 UTC 2023: + '[' 1 == 1 ']'
Thu Jul 20 22:47:28 UTC 2023: + apt-get -y -q install git
Thu Jul 20 22:47:29 UTC 2023: debconf: unable to initialize frontend: Dialog
Thu Jul 20 22:47:29 UTC 2023: debconf: (TERM is not set, so the dialog frontend is not usable.)
Thu Jul 20 22:47:29 UTC 2023: debconf: falling back to frontend: Readline
Thu Jul 20 22:47:29 UTC 2023: debconf: unable to initialize frontend: Readline
Thu Jul 20 22:47:29 UTC 2023: debconf: (Can't locate Term/ReadLine.pm in @INC (you may need to install the Term::ReadLine module) (@INC contains: /etc/perl /usr/local/lib/x86_64-linux-gnu/perl/5.32.1 /usr/local/share/perl/5.32.1 /usr/lib/x86_64-linux-gnu/perl5/5.32 /usr/share/perl5 /usr/lib/x86_64-linux-gnu/perl-base /usr/lib/x86_64-linux-gnu/perl/5.32 /usr/share/perl/5.32 /usr/local/lib/site_perl) at /usr/share/perl5/Debconf/FrontEnd/Readline.pm line 7, <> line 8.)
Thu Jul 20 22:47:29 UTC 2023: debconf: falling back to frontend: Teletype
Thu Jul 20 22:47:29 UTC 2023: dpkg-preconfigure: unable to re-open stdin:
Thu Jul 20 22:47:39 UTC 2023: + return 0
Thu Jul 20 22:47:39 UTC 2023: + mkdir -p /tmp/gocache
Thu Jul 20 22:47:39 UTC 2023: + export GOCACHE=/tmp/gocache
Thu Jul 20 22:47:39 UTC 2023: + GOCACHE=/tmp/gocache
Thu Jul 20 22:47:39 UTC 2023: + '[' master = master ']'
Thu Jul 20 22:47:39 UTC 2023: + retry apt-get -y -q install gcc
Thu Jul 20 22:47:39 UTC 2023: + for i in {1..3}
Thu Jul 20 22:47:39 UTC 2023: + '[' 1 == 1 ']'
Thu Jul 20 22:47:39 UTC 2023: + apt-get -y -q install gcc
Thu Jul 20 22:47:40 UTC 2023: debconf: unable to initialize frontend: Dialog
Thu Jul 20 22:47:40 UTC 2023: debconf: (TERM is not set, so the dialog frontend is not usable.)
Thu Jul 20 22:47:40 UTC 2023: debconf: falling back to frontend: Readline
Thu Jul 20 22:47:40 UTC 2023: debconf: unable to initialize frontend: Readline
Thu Jul 20 22:47:40 UTC 2023: debconf: (This frontend requires a controlling tty.)
Thu Jul 20 22:47:40 UTC 2023: debconf: falling back to frontend: Teletype
Thu Jul 20 22:47:40 UTC 2023: dpkg-preconfigure: unable to re-open stdin:
Thu Jul 20 22:48:13 UTC 2023: + return 0
Thu Jul 20 22:48:13 UTC 2023: + mkdir -p /tmp/bin
Thu Jul 20 22:48:13 UTC 2023: + retry curl -sL -o /tmp/bin/gimme https://raw.githubusercontent.com/travis-ci/gimme/master/gimme
Thu Jul 20 22:48:13 UTC 2023: + for i in {1..3}
Thu Jul 20 22:48:13 UTC 2023: + '[' 1 == 1 ']'
Thu Jul 20 22:48:13 UTC 2023: + curl -sL -o /tmp/bin/gimme https://raw.githubusercontent.com/travis-ci/gimme/master/gimme
Thu Jul 20 22:48:13 UTC 2023: + return 0
Thu Jul 20 22:48:13 UTC 2023: + chmod +x /tmp/bin/gimme
Thu Jul 20 22:48:14 UTC 2023: + export PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/tmp/bin
Thu Jul 20 22:48:14 UTC 2023: + PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/tmp/bin
Thu Jul 20 22:48:14 UTC 2023: + retry gimme_retrier
Thu Jul 20 22:48:14 UTC 2023: + for i in {1..3}
Thu Jul 20 22:48:14 UTC 2023: + '[' 1 == 1 ']'
Thu Jul 20 22:48:14 UTC 2023: + gimme_retrier
Thu Jul 20 22:48:14 UTC 2023: ++ gimme master
Thu Jul 20 22:49:18 UTC 2023: I don't have any idea what to do with 'master'.
Thu Jul 20 22:49:18 UTC 2023: (using download type 'auto')
Thu Jul 20 22:49:18 UTC 2023: + eval ''
Thu Jul 20 22:49:18 UTC 2023: + which go
Thu Jul 20 22:49:18 UTC 2023: + for i in {1..3}
Thu Jul 20 22:49:18 UTC 2023: + '[' 2 == 1 ']'
Thu Jul 20 22:49:18 UTC 2023: + sleep 10
Thu Jul 20 22:49:28 UTC 2023: + gimme_retrier
Thu Jul 20 22:49:28 UTC 2023: ++ gimme master
Thu Jul 20 22:49:29 UTC 2023: I don't have any idea what to do with 'master'.
Thu Jul 20 22:49:29 UTC 2023: (using download type 'auto')
Thu Jul 20 22:49:29 UTC 2023: + eval ''
Thu Jul 20 22:49:29 UTC 2023: + which go
Thu Jul 20 22:49:29 UTC 2023: + for i in {1..3}
Thu Jul 20 22:49:29 UTC 2023: + '[' 3 == 1 ']'
Thu Jul 20 22:49:29 UTC 2023: + sleep 10
Thu Jul 20 22:49:39 UTC 2023: + gimme_retrier
Thu Jul 20 22:49:39 UTC 2023: ++ gimme master
Thu Jul 20 22:49:40 UTC 2023: I don't have any idea what to do with 'master'.
Thu Jul 20 22:49:40 UTC 2023: (using download type 'auto')
Thu Jul 20 22:49:40 UTC 2023: + eval ''
Thu Jul 20 22:49:40 UTC 2023: + which go
Thu Jul 20 22:49:40 UTC 2023: + return 1
Thu Jul 20 22:49:40 UTC 2023: ++ echo 'failed to set up or run the benchmark'
Thu Jul 20 22:49:40 UTC 2023: failed to set up or run the benchmark
Thu Jul 20 22:49:40 UTC 2023: ++ sleep 300
integration_test.go:364: PollAndLogSerialPort() got error: failed to execute the prober benchmark script</pre></details> | non_build | profiler testagentintegration failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output proftest go serial port output for profiler test gomaster pdt loading uefi google persistentdisk from pciroot pci scsi bdsdxe starting uefi google persistentdisk from pciroot pci scsi uefi attempting to start image description uefi google persistentdisk filepath pciroot pci scsi optionnumber to grub booting debian gnu linux loading linux cloud loading initial ramdisk error no suitable video mode found booting in blind mode thu jul utc retry apt get update thu jul utc for i in thu jul utc thu jul utc apt get update thu jul utc return thu jul utc retry apt get y q install git thu jul utc for i in thu jul utc thu jul utc apt get y q install git thu jul utc debconf unable to initialize frontend dialog thu jul utc debconf term is not set so the dialog frontend is not usable thu jul utc debconf falling back to frontend readline thu jul utc debconf unable to initialize frontend readline thu jul utc debconf can t locate term readline pm in inc you may need to install the term readline module inc contains etc perl usr local lib linux gnu perl usr local share perl usr lib linux gnu usr share usr lib linux gnu perl base usr lib linux gnu perl usr share perl usr local lib site perl at usr share debconf frontend readline pm line line thu jul utc debconf falling back to frontend teletype thu jul utc dpkg preconfigure unable to re open stdin thu jul utc return thu jul utc mkdir p tmp gocache thu jul utc export gocache tmp gocache thu jul utc gocache tmp gocache thu jul utc thu jul utc retry apt get y q install gcc thu jul utc for i in thu jul utc thu jul utc apt get y q install gcc thu jul utc debconf unable to initialize frontend dialog thu jul utc debconf term is not set so the dialog frontend is not usable thu jul utc debconf falling back to frontend readline thu jul utc debconf unable to initialize frontend readline thu jul utc debconf this frontend requires a controlling tty thu jul utc debconf falling back to frontend teletype thu jul utc dpkg preconfigure unable to re open stdin thu jul utc return thu jul utc mkdir p tmp bin thu jul utc retry curl sl o tmp bin gimme thu jul utc for i in thu jul utc thu jul utc curl sl o tmp bin gimme thu jul utc return thu jul utc chmod x tmp bin gimme thu jul utc export path usr local sbin usr local bin usr sbin usr bin sbin bin tmp bin thu jul utc path usr local sbin usr local bin usr sbin usr bin sbin bin tmp bin thu jul utc retry gimme retrier thu jul utc for i in thu jul utc thu jul utc gimme retrier thu jul utc gimme master thu jul utc i don t have any idea what to do with master thu jul utc using download type auto thu jul utc eval thu jul utc which go thu jul utc for i in thu jul utc thu jul utc sleep thu jul utc gimme retrier thu jul utc gimme master thu jul utc i don t have any idea what to do with master thu jul utc using download type auto thu jul utc eval thu jul utc which go thu jul utc for i in thu jul utc thu jul utc sleep thu jul utc gimme retrier thu jul utc gimme master thu jul utc i don t have any idea what to do with master thu jul utc using download type auto thu jul utc eval thu jul utc which go thu jul utc return thu jul utc echo failed to set up or run the benchmark thu jul utc failed to set up or run the benchmark thu jul utc sleep integration test go pollandlogserialport got error failed to execute the prober benchmark script | 0 |
186,548 | 14,397,483,867 | IssuesEvent | 2020-12-03 08:14:30 | zeek/spicy | https://api.github.com/repos/zeek/spicy | closed | Add unit tests for IntrusivePtr | Runtime Library Testing / CI | With #491 came a new IntrusivePtr class (adapted from Zeek). We need more test coverage for that. | 1.0 | Add unit tests for IntrusivePtr - With #491 came a new IntrusivePtr class (adapted from Zeek). We need more test coverage for that. | non_build | add unit tests for intrusiveptr with came a new intrusiveptr class adapted from zeek we need more test coverage for that | 0 |
68,213 | 14,914,662,904 | IssuesEvent | 2021-01-22 15:42:37 | finos/secref-data | https://api.github.com/repos/finos/secref-data | closed | Nov 10th 2020 - SecRef Meeting Minutes | Securities Reference Data meeting | # Security Reference Data Project Meeting
## Date
Nov 10th 2020 - 11am EST / 4pm BST
## Untracked attendees
- Lodovico Gavotti / Refinitiv
- ...
## Agenda
- [x] Convene & roll call (5mins)
- [x] Introduce new joiners & volunteers
- [x] Approve [Oct 27th](https://github.com/finos/secref-data/issues/40) meeting minutes - _meeting minutes were approved_
- [ ] Review submitted use cases [#33](https://github.com/finos/secref-data/issues/33) - (use cases around data ingestion from new providers and duplicates had been scheduled to be submitted by Oct 27th, 2020)
- [ ] Review workflow [#26](https://github.com/finos/secref-data/issues/26)
- [x] AOB, Q&A & Adjourn (5mins)
## Decisions Made
- [ ] Confirmed that access to the mapping tool/data will be via API calls and that a GUI is a secondary priority.
- [ ] Decision 2
- [ ] ...
## Action Items
- [ ] @toshaellison to add the use cases discussed during the call to [https://github.com/finos/secref-data/issues/33](https://github.com/finos/secref-data/issues/33)
- [ ] Discuss how new data sources will be mapped into the tool.
- [ ] Discuss the process by which licenses can be validated when API calls are made.
### WebEx info
- [WebEx Meeting URL](https://finos.webex.com/finos/j.php?MTID=mf43ec72e7ce8e771da12093c940c7ab9)
- Meeting Number: 668 805 739
- Join by phone: +1.415.655.0003 USA Toll | +44-20319-88141 UK Toll | Access code: 668 805 739
- [Global call-in numbers](https://finos.webex.com/finos/globalcallin.php?serviceType=MC&ED=582460637&tollFree=0)
| True | Nov 10th 2020 - SecRef Meeting Minutes - # Security Reference Data Project Meeting
## Date
Nov 10th 2020 - 11am EST / 4pm BST
## Untracked attendees
- Lodovico Gavotti / Refinitiv
- ...
## Agenda
- [x] Convene & roll call (5mins)
- [x] Introduce new joiners & volunteers
- [x] Approve [Oct 27th](https://github.com/finos/secref-data/issues/40) meeting minutes - _meeting minutes were approved_
- [ ] Review submitted use cases [#33](https://github.com/finos/secref-data/issues/33) - (use cases around data ingestion from new providers and duplicates had been scheduled to be submitted by Oct 27th, 2020)
- [ ] Review workflow [#26](https://github.com/finos/secref-data/issues/26)
- [x] AOB, Q&A & Adjourn (5mins)
## Decisions Made
- [ ] Confirmed that access to the mapping tool/data will be via API calls and that a GUI is a secondary priority.
- [ ] Decision 2
- [ ] ...
## Action Items
- [ ] @toshaellison to add the use cases discussed during the call to [https://github.com/finos/secref-data/issues/33](https://github.com/finos/secref-data/issues/33)
- [ ] Discuss how new data sources will be mapped into the tool.
- [ ] Discuss the process by which licenses can be validated when API calls are made.
### WebEx info
- [WebEx Meeting URL](https://finos.webex.com/finos/j.php?MTID=mf43ec72e7ce8e771da12093c940c7ab9)
- Meeting Number: 668 805 739
- Join by phone: +1.415.655.0003 USA Toll | +44-20319-88141 UK Toll | Access code: 668 805 739
- [Global call-in numbers](https://finos.webex.com/finos/globalcallin.php?serviceType=MC&ED=582460637&tollFree=0)
| non_build | nov secref meeting minutes security reference data project meeting date nov est bst untracked attendees lodovico gavotti refinitiv agenda convene roll call introduce new joiners volunteers approve meeting minutes meeting minutes were approved review submitted use cases use cases around data ingestion from new providers and duplicates had been scheduled to be submitted by oct review workflow aob q a adjourn decisions made confirmed that access to the mapping tool data will be via api calls and that a gui is a secondary priority decision action items toshaellison to add the use cases discussed during the call to discuss how new data sources will be mapped into the tool discuss the process by which licenses can be validated when api calls are made webex info meeting number join by phone usa toll uk toll access code | 0 |
320,370 | 23,808,797,170 | IssuesEvent | 2022-09-04 12:56:11 | WeCanWebserv/Webserv | https://api.github.com/repos/WeCanWebserv/Webserv | closed | [FEAT] HTTP Request message의 Format(token) 파악하기 | documentation | ## Description
- HTTP Request message의 token들 파악하기
## TO DO
- [x] request-line 파악하기
- [x] header 파악하기
- [x] body 파악하기
| 1.0 | [FEAT] HTTP Request message의 Format(token) 파악하기 - ## Description
- HTTP Request message의 token들 파악하기
## TO DO
- [x] request-line 파악하기
- [x] header 파악하기
- [x] body 파악하기
| non_build | http request message의 format token 파악하기 description http request message의 token들 파악하기 to do request line 파악하기 header 파악하기 body 파악하기 | 0 |
30,916 | 8,621,951,994 | IssuesEvent | 2018-11-20 18:50:58 | hashicorp/packer | https://api.github.com/repos/hashicorp/packer | closed | Vsphere-iso validation is failing for RHEL7.5 VM deployment | builder/vmware-esxi question | I am trying to build a vm with vsphere-iso Plugin ..
Below is my rhel75.Json file ...
```
{
"builders":[
{
"type": "vsphere-iso",
"vcenter_server": "nahq-rvapl00072",
"insecure_connection": "true",
"username": "srv_vra_lab",
"password": "XXXXXXXXXXX",
"datacenter": "LAB-NUTANIX",
"vm_name": "RHEL75_SB",
"folder": "_Packer",
"host": "HQLABINF002/nahq-rpehv00644",
"cluster": "HQLABINF002",
"resource_pool": "vRA-LAB-Site2-APP",
"datastore": "lab-nutanix02-ds01",
"CPUs": 8,
"RAM": 10240,
"disk_size": 20480,
"guest_os_type": "rhel7_64Guest",
"cdrom_adapter_type": "ide",
"disk_adapter_type": "lsiLogic",
"disk_thin_provisioned": "true",
"network": "Site1-LAB_930",
"network_adapter": "vmxnet3",
"iso_datastore": "lab-nutanix01-ds01",
"iso": "[LABX_TEMPLATES_NFS] ISOs/rhel75.iso",
"communicator": "ssh",
"ssh_username": "root",
"ssh_password": "packer",
"convert_to_template": "true"
}
],
"provisioners": [
{
"type": "shell",
"inline": [ "echo hello" ]
}
]
}
```
Validation is failing with below errors.
[bodduls@packer]$ ./packer validate rhel75.json
Template validation failed. Errors are shown below.
Errors validating build 'vsphere-iso'. 3 errors occurred:
* unknown configuration key: "cdrom_adapter_type"
* unknown configuration key: "disk_adapter_type"
* unknown configuration key: "disk_thin_provisioned"
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
When I remove these optional parameters from the file .. build is failing with the below message and from V-center logs I can see this is related to VirtualCdrom which is not mounting the specified ISO.
[bodduls@ packer]$ ./packer build rhel75.json
vsphere-iso output will be in this color.
==> vsphere-iso: Creating VM...
Build 'vsphere-iso' errored: Invalid configuration for device '2'.
==> Some builds didn't complete successfully and had errors:
--> vsphere-iso: Invalid configuration for device '2'.
==> Builds finished but no artifacts were created.
[bodduls@packer]$
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
I am using Packer v1.3.2
[bodduls@ packer]$ ./packer version
Packer v1.3.2
[bodduls@ packer]$
[bodduls@labd-rclap00142 packer]$ ls -ltr
total 151572
-rwxr-xr-x 1 bodduls bodduls 86309745 Oct 29 07:40 packer
-rw-rw-r-- 1 bodduls bodduls 27459221 Oct 29 07:47 packer_1.3.2_linux_amd64.zip
-rwxr-xr-x 1 bodduls bodduls 41392224 Nov 7 11:19 packer-builder-vsphere-iso.linux
-rw-rw-r-- 1 bodduls bodduls 19466 Nov 7 14:52 crash.log
-rw-rw-r-- 1 bodduls bodduls 1033 Nov 13 09:40 rhel75.json
[bodduls@labd-rclap00142 packer]$
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Any suggestion to overcome this ?
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
| 1.0 | Vsphere-iso validation is failing for RHEL7.5 VM deployment - I am trying to build a vm with vsphere-iso Plugin ..
Below is my rhel75.Json file ...
```
{
"builders":[
{
"type": "vsphere-iso",
"vcenter_server": "nahq-rvapl00072",
"insecure_connection": "true",
"username": "srv_vra_lab",
"password": "XXXXXXXXXXX",
"datacenter": "LAB-NUTANIX",
"vm_name": "RHEL75_SB",
"folder": "_Packer",
"host": "HQLABINF002/nahq-rpehv00644",
"cluster": "HQLABINF002",
"resource_pool": "vRA-LAB-Site2-APP",
"datastore": "lab-nutanix02-ds01",
"CPUs": 8,
"RAM": 10240,
"disk_size": 20480,
"guest_os_type": "rhel7_64Guest",
"cdrom_adapter_type": "ide",
"disk_adapter_type": "lsiLogic",
"disk_thin_provisioned": "true",
"network": "Site1-LAB_930",
"network_adapter": "vmxnet3",
"iso_datastore": "lab-nutanix01-ds01",
"iso": "[LABX_TEMPLATES_NFS] ISOs/rhel75.iso",
"communicator": "ssh",
"ssh_username": "root",
"ssh_password": "packer",
"convert_to_template": "true"
}
],
"provisioners": [
{
"type": "shell",
"inline": [ "echo hello" ]
}
]
}
```
Validation is failing with below errors.
[bodduls@packer]$ ./packer validate rhel75.json
Template validation failed. Errors are shown below.
Errors validating build 'vsphere-iso'. 3 errors occurred:
* unknown configuration key: "cdrom_adapter_type"
* unknown configuration key: "disk_adapter_type"
* unknown configuration key: "disk_thin_provisioned"
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
When I remove these optional parameters from the file .. build is failing with the below message and from V-center logs I can see this is related to VirtualCdrom which is not mounting the specified ISO.
[bodduls@ packer]$ ./packer build rhel75.json
vsphere-iso output will be in this color.
==> vsphere-iso: Creating VM...
Build 'vsphere-iso' errored: Invalid configuration for device '2'.
==> Some builds didn't complete successfully and had errors:
--> vsphere-iso: Invalid configuration for device '2'.
==> Builds finished but no artifacts were created.
[bodduls@packer]$
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
I am using Packer v1.3.2
[bodduls@ packer]$ ./packer version
Packer v1.3.2
[bodduls@ packer]$
[bodduls@labd-rclap00142 packer]$ ls -ltr
total 151572
-rwxr-xr-x 1 bodduls bodduls 86309745 Oct 29 07:40 packer
-rw-rw-r-- 1 bodduls bodduls 27459221 Oct 29 07:47 packer_1.3.2_linux_amd64.zip
-rwxr-xr-x 1 bodduls bodduls 41392224 Nov 7 11:19 packer-builder-vsphere-iso.linux
-rw-rw-r-- 1 bodduls bodduls 19466 Nov 7 14:52 crash.log
-rw-rw-r-- 1 bodduls bodduls 1033 Nov 13 09:40 rhel75.json
[bodduls@labd-rclap00142 packer]$
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Any suggestion to overcome this ?
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
| build | vsphere iso validation is failing for vm deployment i am trying to build a vm with vsphere iso plugin below is my json file builders type vsphere iso vcenter server nahq insecure connection true username srv vra lab password xxxxxxxxxxx datacenter lab nutanix vm name sb folder packer host nahq cluster resource pool vra lab app datastore lab cpus ram disk size guest os type cdrom adapter type ide disk adapter type lsilogic disk thin provisioned true network lab network adapter iso datastore lab iso isos iso communicator ssh ssh username root ssh password packer convert to template true provisioners type shell inline validation is failing with below errors packer validate json template validation failed errors are shown below errors validating build vsphere iso errors occurred unknown configuration key cdrom adapter type unknown configuration key disk adapter type unknown configuration key disk thin provisioned when i remove these optional parameters from the file build is failing with the below message and from v center logs i can see this is related to virtualcdrom which is not mounting the specified iso packer build json vsphere iso output will be in this color vsphere iso creating vm build vsphere iso errored invalid configuration for device some builds didn t complete successfully and had errors vsphere iso invalid configuration for device builds finished but no artifacts were created i am using packer packer version packer ls ltr total rwxr xr x bodduls bodduls oct packer rw rw r bodduls bodduls oct packer linux zip rwxr xr x bodduls bodduls nov packer builder vsphere iso linux rw rw r bodduls bodduls nov crash log rw rw r bodduls bodduls nov json any suggestion to overcome this | 1 |
13,514 | 5,392,625,845 | IssuesEvent | 2017-02-26 13:03:13 | junit-team/junit5 | https://api.github.com/repos/junit-team/junit5 | closed | Upgrade to Gradle 3.4 | build enhancement up-for-grabs | ## Overview
The build fails after upgrading to Gradle 3.4 with following stacktrace:
```
...
:junit-platform-gradle-plugin:spotlessCheck
:junit-platform-gradle-plugin:compileTestJava NO-SOURCE
:junit-platform-gradle-plugin:compileTestGroovy
:junit-platform-gradle-plugin:processTestResources NO-SOURCE
:junit-platform-gradle-plugin:testClasses
:junit-platform-gradle-plugin:test
org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec > adds dependencies to configuration FAILED
java.lang.UnsupportedOperationException
at org.gradle.api.internal.CompositeDomainObjectSet$DomainObjectCompositeCollection.containsAll(CompositeDomainObjectSet.java:228)
at org.gradle.api.internal.DefaultDomainObjectCollection.containsAll(DefaultDomainObjectCollection.java:241)
at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98)
at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98)
at org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec.adds dependencies to configuration(JUnitPlatformPluginSpec.groovy:251)
10 tests completed, 1 failed
:junit-platform-gradle-plugin:test FAILED
```
## Deliverables
- [ ] Solve `UnsupportedOperationException`
- ~Make use of https://docs.gradle.org/current/userguide/java_library_plugin.html~ Moved to #691
| 1.0 | Upgrade to Gradle 3.4 - ## Overview
The build fails after upgrading to Gradle 3.4 with following stacktrace:
```
...
:junit-platform-gradle-plugin:spotlessCheck
:junit-platform-gradle-plugin:compileTestJava NO-SOURCE
:junit-platform-gradle-plugin:compileTestGroovy
:junit-platform-gradle-plugin:processTestResources NO-SOURCE
:junit-platform-gradle-plugin:testClasses
:junit-platform-gradle-plugin:test
org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec > adds dependencies to configuration FAILED
java.lang.UnsupportedOperationException
at org.gradle.api.internal.CompositeDomainObjectSet$DomainObjectCompositeCollection.containsAll(CompositeDomainObjectSet.java:228)
at org.gradle.api.internal.DefaultDomainObjectCollection.containsAll(DefaultDomainObjectCollection.java:241)
at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98)
at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98)
at org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec.adds dependencies to configuration(JUnitPlatformPluginSpec.groovy:251)
10 tests completed, 1 failed
:junit-platform-gradle-plugin:test FAILED
```
## Deliverables
- [ ] Solve `UnsupportedOperationException`
- ~Make use of https://docs.gradle.org/current/userguide/java_library_plugin.html~ Moved to #691
| build | upgrade to gradle overview the build fails after upgrading to gradle with following stacktrace junit platform gradle plugin spotlesscheck junit platform gradle plugin compiletestjava no source junit platform gradle plugin compiletestgroovy junit platform gradle plugin processtestresources no source junit platform gradle plugin testclasses junit platform gradle plugin test org junit platform gradle plugin junitplatformpluginspec adds dependencies to configuration failed java lang unsupportedoperationexception at org gradle api internal compositedomainobjectset domainobjectcompositecollection containsall compositedomainobjectset java at org gradle api internal defaultdomainobjectcollection containsall defaultdomainobjectcollection java at org gradle api internal delegatingdomainobjectset containsall delegatingdomainobjectset java at org gradle api internal delegatingdomainobjectset containsall delegatingdomainobjectset java at org junit platform gradle plugin junitplatformpluginspec adds dependencies to configuration junitplatformpluginspec groovy tests completed failed junit platform gradle plugin test failed deliverables solve unsupportedoperationexception make use of moved to | 1 |
24,432 | 12,103,737,745 | IssuesEvent | 2020-04-20 18:56:28 | Azure/azure-sdk-for-net | https://api.github.com/repos/Azure/azure-sdk-for-net | closed | Add ReceiptTypeConfidence field to USReceipt | Client Cognitive Services FormRecognizer |
`ReceiptType` comes back as a field in `DocumentResults`, without a bounding box, but with a confidence value. Add a `float? ReceiptTypeConfidence` to `USReceipt` to capture the confidence. | 1.0 | Add ReceiptTypeConfidence field to USReceipt -
`ReceiptType` comes back as a field in `DocumentResults`, without a bounding box, but with a confidence value. Add a `float? ReceiptTypeConfidence` to `USReceipt` to capture the confidence. | non_build | add receipttypeconfidence field to usreceipt receipttype comes back as a field in documentresults without a bounding box but with a confidence value add a float receipttypeconfidence to usreceipt to capture the confidence | 0 |
563,731 | 16,704,702,966 | IssuesEvent | 2021-06-09 08:36:40 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.ebay.com - design is broken | browser-firefox engine-gecko priority-critical | <!-- @browser: Firefox 89.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:89.0) Gecko/20100101 Firefox/89.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/76109 -->
**URL**: https://www.ebay.com/
**Browser / Version**: Firefox 89.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Images not loaded
**Steps to Reproduce**:
The Page not visible correctly, no image, no order
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/6/9f2b29a9-412b-4998-9d31-8c548ac8aaa9.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.ebay.com - design is broken - <!-- @browser: Firefox 89.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:89.0) Gecko/20100101 Firefox/89.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/76109 -->
**URL**: https://www.ebay.com/
**Browser / Version**: Firefox 89.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Design is broken
**Description**: Images not loaded
**Steps to Reproduce**:
The Page not visible correctly, no image, no order
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/6/9f2b29a9-412b-4998-9d31-8c548ac8aaa9.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_build | design is broken url browser version firefox operating system windows tested another browser yes chrome problem type design is broken description images not loaded steps to reproduce the page not visible correctly no image no order view the screenshot img alt screenshot src browser configuration none from with ❤️ | 0 |
196,568 | 6,935,334,656 | IssuesEvent | 2017-12-03 07:24:04 | manrajsingh960/GetTogether | https://api.github.com/repos/manrajsingh960/GetTogether | closed | Joined Events (When data is deleted) | functional problem priority | The joined events list is cleared when you clear the data | 1.0 | Joined Events (When data is deleted) - The joined events list is cleared when you clear the data | non_build | joined events when data is deleted the joined events list is cleared when you clear the data | 0 |
265,283 | 23,158,358,956 | IssuesEvent | 2022-07-29 15:03:18 | redpanda-data/redpanda | https://api.github.com/repos/redpanda-data/redpanda | closed | Failure in `PartitionMovementTest`.`test_availability_when_one_node_down` (ECONNREFUSED) | kind/bug area/redpanda area/tests ci-failure | PartitionMovementTest.test_availability_when_one_node_down fails with ECONNREFUSED trying to query the `partitions/kafka/<topic>/0/replicas` endpoint, which is very similar to #5189. Same error, same query, same test class, but different test case--thus the new bug instead of reopening #5189.
```
PartitionMovementTest.test_availability_when_one_node_down (3/9 runs)
failure at 2022-07-27T21:40:32.761Z: ConnectionError(MaxRetryError("HTTPConnectionPool(host='docker-rp-7', port=9644):
Max retries exceeded with url: /v1/partitions/kafka/topic-ezddcdbxtc/0/replicas (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f6d6d258280>:
Failed to establish a new connection: [Errno 111] Connection refused'))"))
```
from [build 13204](https://buildkite.com/redpanda/redpanda/builds/13204#01824125-c0c1-4155-a6c5-914fc29a724f) | 1.0 | Failure in `PartitionMovementTest`.`test_availability_when_one_node_down` (ECONNREFUSED) - PartitionMovementTest.test_availability_when_one_node_down fails with ECONNREFUSED trying to query the `partitions/kafka/<topic>/0/replicas` endpoint, which is very similar to #5189. Same error, same query, same test class, but different test case--thus the new bug instead of reopening #5189.
```
PartitionMovementTest.test_availability_when_one_node_down (3/9 runs)
failure at 2022-07-27T21:40:32.761Z: ConnectionError(MaxRetryError("HTTPConnectionPool(host='docker-rp-7', port=9644):
Max retries exceeded with url: /v1/partitions/kafka/topic-ezddcdbxtc/0/replicas (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f6d6d258280>:
Failed to establish a new connection: [Errno 111] Connection refused'))"))
```
from [build 13204](https://buildkite.com/redpanda/redpanda/builds/13204#01824125-c0c1-4155-a6c5-914fc29a724f) | non_build | failure in partitionmovementtest test availability when one node down econnrefused partitionmovementtest test availability when one node down fails with econnrefused trying to query the partitions kafka replicas endpoint which is very similar to same error same query same test class but different test case thus the new bug instead of reopening partitionmovementtest test availability when one node down runs failure at connectionerror maxretryerror httpconnectionpool host docker rp port max retries exceeded with url partitions kafka topic ezddcdbxtc replicas caused by newconnectionerror failed to establish a new connection connection refused from | 0 |
14,107 | 5,556,494,506 | IssuesEvent | 2017-03-24 09:20:31 | ng-bootstrap/ng-bootstrap | https://api.github.com/repos/ng-bootstrap/ng-bootstrap | closed | angular 4 compatibility | type: build | ### Bug description:
See *how to reproduce* below. The issue is that currently the generated package.json has:
"peerDependencies": {
"@angular/common": "^2.0.0",
"@angular/compiler": "^2.0.0",
"@angular/core": "^2.0.0",
"@angular/forms": "^2.0.0"
},
which is too restrictive - according to the [official docs](https://nodejs.org/en/blog/npm/peer-dependencies/) :
> One piece of advice: peer dependency requirements, unlike those for regular dependencies, should be lenient
*Note*: this is a follow-up to https://github.com/ng-bootstrap/ng-bootstrap/pull/1390 which is indeed wrong - I didn't know the package.json was autogenerated and changed the actual dependencies (because I was looking at using the fork directly as a dependency of my project), but the issue remains.
*Note2*: by using `"^2.0.0 || >=4.0.0-beta <5.0.0",` you would still be supporting 2.x.
### How to reproduce the issue:
```bash
npm install -g @angular/cli
ng new ngcli-ng4 --ng4
cd ngcli-ng4
npm install --save ng-bootstrap
```
output
```
ngcli-ng4@0.0.0 /ything/projects/tmk/tmp/ngcli-ng4
├── UNMET PEER DEPENDENCY @angular/common@4.0.0-rc.5
├── UNMET PEER DEPENDENCY @angular/compiler@4.0.0-rc.5
├── UNMET PEER DEPENDENCY @angular/core@4.0.0-rc.5
├── UNMET PEER DEPENDENCY @angular/forms@4.0.0-rc.5
├─┬ ng-bootstrap@1.1.16-1
│ └── moment@2.15.1
└── UNMET PEER DEPENDENCY zone.js@0.7.8
npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@^1.0.0 (node_modules/chokidar/node_modules/fsevents):
npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.1.1: wanted {"os":"darwin","arch":"any"} (current: {"os":"linux","arch":"x64"})
npm WARN @angular/core@4.0.0-rc.5 requires a peer of zone.js@^0.8.4 but none was installed.
npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/common@^2.0.0 but none was installed.
npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/compiler@^2.0.0 but none was installed.
npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/core@^2.0.0 but none was installed.
npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/forms@^2.0.0 but none was installed.
```
### Versions
* npm 3.10.10
* Angular: 4.0.0-rc.5
* ng-bootstrap: 1.1.16 | 1.0 | angular 4 compatibility - ### Bug description:
See *how to reproduce* below. The issue is that currently the generated package.json has:
"peerDependencies": {
"@angular/common": "^2.0.0",
"@angular/compiler": "^2.0.0",
"@angular/core": "^2.0.0",
"@angular/forms": "^2.0.0"
},
which is too restrictive - according to the [official docs](https://nodejs.org/en/blog/npm/peer-dependencies/) :
> One piece of advice: peer dependency requirements, unlike those for regular dependencies, should be lenient
*Note*: this is a follow-up to https://github.com/ng-bootstrap/ng-bootstrap/pull/1390 which is indeed wrong - I didn't know the package.json was autogenerated and changed the actual dependencies (because I was looking at using the fork directly as a dependency of my project), but the issue remains.
*Note2*: by using `"^2.0.0 || >=4.0.0-beta <5.0.0",` you would still be supporting 2.x.
### How to reproduce the issue:
```bash
npm install -g @angular/cli
ng new ngcli-ng4 --ng4
cd ngcli-ng4
npm install --save ng-bootstrap
```
output
```
ngcli-ng4@0.0.0 /ything/projects/tmk/tmp/ngcli-ng4
├── UNMET PEER DEPENDENCY @angular/common@4.0.0-rc.5
├── UNMET PEER DEPENDENCY @angular/compiler@4.0.0-rc.5
├── UNMET PEER DEPENDENCY @angular/core@4.0.0-rc.5
├── UNMET PEER DEPENDENCY @angular/forms@4.0.0-rc.5
├─┬ ng-bootstrap@1.1.16-1
│ └── moment@2.15.1
└── UNMET PEER DEPENDENCY zone.js@0.7.8
npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@^1.0.0 (node_modules/chokidar/node_modules/fsevents):
npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.1.1: wanted {"os":"darwin","arch":"any"} (current: {"os":"linux","arch":"x64"})
npm WARN @angular/core@4.0.0-rc.5 requires a peer of zone.js@^0.8.4 but none was installed.
npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/common@^2.0.0 but none was installed.
npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/compiler@^2.0.0 but none was installed.
npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/core@^2.0.0 but none was installed.
npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/forms@^2.0.0 but none was installed.
```
### Versions
* npm 3.10.10
* Angular: 4.0.0-rc.5
* ng-bootstrap: 1.1.16 | build | angular compatibility bug description see how to reproduce below the issue is that currently the generated package json has peerdependencies angular common angular compiler angular core angular forms which is too restrictive according to the one piece of advice peer dependency requirements unlike those for regular dependencies should be lenient note this is a follow up to which is indeed wrong i didn t know the package json was autogenerated and changed the actual dependencies because i was looking at using the fork directly as a dependency of my project but the issue remains by using beta you would still be supporting x how to reproduce the issue bash npm install g angular cli ng new ngcli cd ngcli npm install save ng bootstrap output ngcli ything projects tmk tmp ngcli ├── unmet peer dependency angular common rc ├── unmet peer dependency angular compiler rc ├── unmet peer dependency angular core rc ├── unmet peer dependency angular forms rc ├─┬ ng bootstrap │ └── moment └── unmet peer dependency zone js npm warn optional skipping optional dependency fsevents node modules chokidar node modules fsevents npm warn notsup skipping optional dependency unsupported platform for fsevents wanted os darwin arch any current os linux arch npm warn angular core rc requires a peer of zone js but none was installed npm warn ng bootstrap requires a peer of angular common but none was installed npm warn ng bootstrap requires a peer of angular compiler but none was installed npm warn ng bootstrap requires a peer of angular core but none was installed npm warn ng bootstrap requires a peer of angular forms but none was installed versions npm angular rc ng bootstrap | 1 |
147,408 | 19,522,528,587 | IssuesEvent | 2021-12-29 21:42:44 | barreljan/netlog | https://api.github.com/repos/barreljan/netlog | closed | Sanitizing $_POST inputs to SQL | bug security | As there is a way to for instance cURL your way into a http POST, any key or value could be a risk.
A potential issue is there with the new host in the settings.php page. If you could craft a few headers, it will hit line 128 (foreach loop) and then line 167 to start the seq. of adding a new host. The hostip or hosttype fields are not prepared, escaped of by whatever means so with knowlegde of the code, you could drop a table there or worse.
Must investigate the index.php also however the search input is already in order with a prepared statement. Other lines must be investigated as well. | True | Sanitizing $_POST inputs to SQL - As there is a way to for instance cURL your way into a http POST, any key or value could be a risk.
A potential issue is there with the new host in the settings.php page. If you could craft a few headers, it will hit line 128 (foreach loop) and then line 167 to start the seq. of adding a new host. The hostip or hosttype fields are not prepared, escaped of by whatever means so with knowlegde of the code, you could drop a table there or worse.
Must investigate the index.php also however the search input is already in order with a prepared statement. Other lines must be investigated as well. | non_build | sanitizing post inputs to sql as there is a way to for instance curl your way into a http post any key or value could be a risk a potential issue is there with the new host in the settings php page if you could craft a few headers it will hit line foreach loop and then line to start the seq of adding a new host the hostip or hosttype fields are not prepared escaped of by whatever means so with knowlegde of the code you could drop a table there or worse must investigate the index php also however the search input is already in order with a prepared statement other lines must be investigated as well | 0 |
5,056 | 3,497,535,023 | IssuesEvent | 2016-01-06 01:50:17 | mxe/mxe | https://api.github.com/repos/mxe/mxe | closed | Create Travis CI example of using MXE binary distribution | build-pkg cmake | Current repo for this task: https://github.com/starius/mxe-travis-example (empty)
Can we move it to github.com/mxe/mxe-travis-example when it's finished? | 1.0 | Create Travis CI example of using MXE binary distribution - Current repo for this task: https://github.com/starius/mxe-travis-example (empty)
Can we move it to github.com/mxe/mxe-travis-example when it's finished? | build | create travis ci example of using mxe binary distribution current repo for this task empty can we move it to github com mxe mxe travis example when it s finished | 1 |
74,295 | 20,109,360,022 | IssuesEvent | 2022-02-07 13:45:07 | gakonst/foundry | https://api.github.com/repos/gakonst/foundry | closed | Seeing bad cache invalidation in makerdao/spells-mainnet | T-bug C-forge P-normal Cmd-forge-build | Steps to reproduce:
1. Fork https://github.com/makerdao/spells-mainnet and checkout commit be8c6e946a5fa7768d8af0f6e200712f397ad587 .
2. Remove the `--force` arg from `test-dssspell-forge.sh`.
3. Run `make test-forge`.
4. Tests should correctly mostly error (4 success, 9 failing) and report that the spell has already been cast. Cache has been built.
5. Edit `src/DssSpell.t.base.sol` line 294 and set the `deployed_spell` to `address(0)`.
6. Run `make test-forge`.
Expected output:
Most tests should start passing (11 succeed, 2 failing) with no mention of spell already cast.
Actual output:
Identical from the first test which is incorrect. | 1.0 | Seeing bad cache invalidation in makerdao/spells-mainnet - Steps to reproduce:
1. Fork https://github.com/makerdao/spells-mainnet and checkout commit be8c6e946a5fa7768d8af0f6e200712f397ad587 .
2. Remove the `--force` arg from `test-dssspell-forge.sh`.
3. Run `make test-forge`.
4. Tests should correctly mostly error (4 success, 9 failing) and report that the spell has already been cast. Cache has been built.
5. Edit `src/DssSpell.t.base.sol` line 294 and set the `deployed_spell` to `address(0)`.
6. Run `make test-forge`.
Expected output:
Most tests should start passing (11 succeed, 2 failing) with no mention of spell already cast.
Actual output:
Identical from the first test which is incorrect. | build | seeing bad cache invalidation in makerdao spells mainnet steps to reproduce fork and checkout commit remove the force arg from test dssspell forge sh run make test forge tests should correctly mostly error success failing and report that the spell has already been cast cache has been built edit src dssspell t base sol line and set the deployed spell to address run make test forge expected output most tests should start passing succeed failing with no mention of spell already cast actual output identical from the first test which is incorrect | 1 |
74,683 | 9,104,687,630 | IssuesEvent | 2019-02-20 18:49:50 | envoyproxy/envoy | https://api.github.com/repos/envoyproxy/envoy | closed | healthcheck: Allow non-200 status codes as healthy HTTP healthchecks | design proposal help wanted | *Title*: *healthcheck: Allow non-200 status codes as HTTP healthchecks*
*Description*:
We have certain services that do not return a 200 on a successful healthcheck. The most common one I've seen is returning an HTTP 204. Envoy marks these as unhealthy because it is checking [explicitly](https://github.com/envoyproxy/envoy/blob/master/source/common/upstream/health_checker_impl.cc#L180) for a 200
I think a good default would be 2xx or 3xx is considered healthy. Whether a given status code should be considered healthy is up for debate, but i think this is pretty standard for proxies right now, and so i think it makes sense to do this.
For extra credit, we can have configuration for what constitutes healthy. e.g. regex check on the body of the response.
| 1.0 | healthcheck: Allow non-200 status codes as healthy HTTP healthchecks - *Title*: *healthcheck: Allow non-200 status codes as HTTP healthchecks*
*Description*:
We have certain services that do not return a 200 on a successful healthcheck. The most common one I've seen is returning an HTTP 204. Envoy marks these as unhealthy because it is checking [explicitly](https://github.com/envoyproxy/envoy/blob/master/source/common/upstream/health_checker_impl.cc#L180) for a 200
I think a good default would be 2xx or 3xx is considered healthy. Whether a given status code should be considered healthy is up for debate, but i think this is pretty standard for proxies right now, and so i think it makes sense to do this.
For extra credit, we can have configuration for what constitutes healthy. e.g. regex check on the body of the response.
| non_build | healthcheck allow non status codes as healthy http healthchecks title healthcheck allow non status codes as http healthchecks description we have certain services that do not return a on a successful healthcheck the most common one i ve seen is returning an http envoy marks these as unhealthy because it is checking for a i think a good default would be or is considered healthy whether a given status code should be considered healthy is up for debate but i think this is pretty standard for proxies right now and so i think it makes sense to do this for extra credit we can have configuration for what constitutes healthy e g regex check on the body of the response | 0 |
34,040 | 9,254,581,667 | IssuesEvent | 2019-03-16 00:41:24 | golang/go | https://api.github.com/repos/golang/go | opened | x/build/app: build dashboard for subrepos displays wrong columns | Builders | For example, on x/tools, the Go revision 64b1889 line on freebsd-10_3/386 shows the following:
freebsd-386-12_0 at 64b1889e2d98e336160cad337a7781c720696290 building tools at c74ccfb859c02fdeda7e69caf9e7fdccd9a080fc
:: Running /tmp/workdir/go/bin/go with args ["/tmp/workdir/go/bin/go" "test" "-short" "golang.org/x/tools/..."] and env ["PATH=/tmp/workdir/go/bin:/bin:/sbin:/usr/bin:/usr/local/bin" "PWD=/" "HOME=/" "RC_PID=24" "WORKDIR=/tmp/workdir" "GOROOT_BOOTSTRAP=/tmp/workdir/go1.4" "GO_BUILDER_NAME=freebsd-386-12_0" "GOARCH=386" "GOHOSTARCH=386" "GOROOT=/tmp/workdir/go" "GOPATH=/tmp/workdir/gopath" "GOPROXY=http://10.240.0.50:30156" "TMPDIR=/tmp/workdir/tmp" "GOCACHE=/tmp/workdir/gocache"] in dir /tmp/workdir/gopath/src/golang.org/x/tools | 1.0 | x/build/app: build dashboard for subrepos displays wrong columns - For example, on x/tools, the Go revision 64b1889 line on freebsd-10_3/386 shows the following:
freebsd-386-12_0 at 64b1889e2d98e336160cad337a7781c720696290 building tools at c74ccfb859c02fdeda7e69caf9e7fdccd9a080fc
:: Running /tmp/workdir/go/bin/go with args ["/tmp/workdir/go/bin/go" "test" "-short" "golang.org/x/tools/..."] and env ["PATH=/tmp/workdir/go/bin:/bin:/sbin:/usr/bin:/usr/local/bin" "PWD=/" "HOME=/" "RC_PID=24" "WORKDIR=/tmp/workdir" "GOROOT_BOOTSTRAP=/tmp/workdir/go1.4" "GO_BUILDER_NAME=freebsd-386-12_0" "GOARCH=386" "GOHOSTARCH=386" "GOROOT=/tmp/workdir/go" "GOPATH=/tmp/workdir/gopath" "GOPROXY=http://10.240.0.50:30156" "TMPDIR=/tmp/workdir/tmp" "GOCACHE=/tmp/workdir/gocache"] in dir /tmp/workdir/gopath/src/golang.org/x/tools | build | x build app build dashboard for subrepos displays wrong columns for example on x tools the go revision line on freebsd shows the following freebsd at building tools at running tmp workdir go bin go with args and env in dir tmp workdir gopath src golang org x tools | 1 |
22,974 | 7,259,990,449 | IssuesEvent | 2018-02-18 02:55:18 | minecraft-dev/MinecraftDev | https://api.github.com/repos/minecraft-dev/MinecraftDev | closed | Add Gradle compileOnly support. | build: gradle status: stale | Due to some issues I have come across with another Gradle plugin, I now have to use the `compileOnly` scope for the Bukkit/Paper APIs (Haven't tested with any others as I do not use them right now) instead of the default `compile`.
As a result of this MinecraftDev no longer shows the icon pertaining to my API type in the top left of the project view, implying that MinecraftDev no longer acknowledges my plugin as a Minecraft plugin.
For those we have no idea what I mean:
Default:
```gradle
dependencies {
compile 'org.bukkit:bukkit:1.11.2-R0.1-SNAPSHOT'
}
```
CompileOnly:
```gradle
dependencies {
compileOnly 'org.bukkit:bukkit:1.11.2-R0.1-SNAPSHOT'
}
``` | 1.0 | Add Gradle compileOnly support. - Due to some issues I have come across with another Gradle plugin, I now have to use the `compileOnly` scope for the Bukkit/Paper APIs (Haven't tested with any others as I do not use them right now) instead of the default `compile`.
As a result of this MinecraftDev no longer shows the icon pertaining to my API type in the top left of the project view, implying that MinecraftDev no longer acknowledges my plugin as a Minecraft plugin.
For those we have no idea what I mean:
Default:
```gradle
dependencies {
compile 'org.bukkit:bukkit:1.11.2-R0.1-SNAPSHOT'
}
```
CompileOnly:
```gradle
dependencies {
compileOnly 'org.bukkit:bukkit:1.11.2-R0.1-SNAPSHOT'
}
``` | build | add gradle compileonly support due to some issues i have come across with another gradle plugin i now have to use the compileonly scope for the bukkit paper apis haven t tested with any others as i do not use them right now instead of the default compile as a result of this minecraftdev no longer shows the icon pertaining to my api type in the top left of the project view implying that minecraftdev no longer acknowledges my plugin as a minecraft plugin for those we have no idea what i mean default gradle dependencies compile org bukkit bukkit snapshot compileonly gradle dependencies compileonly org bukkit bukkit snapshot | 1 |
2,878 | 3,025,186,370 | IssuesEvent | 2015-08-03 06:19:09 | FRosner/spawncamping-dds | https://api.github.com/repos/FRosner/spawncamping-dds | closed | Build fails on systems with older git | Bug Build | ```git status -sb``` in build.sbt should be replaced with ```git status --procelain```, in order to retain a version-independent output -- the ```-b``` suboption to ```-s``` is not supported by git 1.7.
The S3-relevant parts in buildt.sbt in general should probably be moved to a separate location (and imported/referenced from build.sbt), in order to make debugging local building easier. | 1.0 | Build fails on systems with older git - ```git status -sb``` in build.sbt should be replaced with ```git status --procelain```, in order to retain a version-independent output -- the ```-b``` suboption to ```-s``` is not supported by git 1.7.
The S3-relevant parts in buildt.sbt in general should probably be moved to a separate location (and imported/referenced from build.sbt), in order to make debugging local building easier. | build | build fails on systems with older git git status sb in build sbt should be replaced with git status procelain in order to retain a version independent output the b suboption to s is not supported by git the relevant parts in buildt sbt in general should probably be moved to a separate location and imported referenced from build sbt in order to make debugging local building easier | 1 |
6,035 | 3,733,516,043 | IssuesEvent | 2016-03-08 00:43:08 | couchbase/couchbase-lite-android | https://api.github.com/repos/couchbase/couchbase-lite-android | closed | testInvokeProcessorAfterReachingCapacity failed on jenkins | Build | http://mobile.jenkins.couchbase.com/view/Couchbase_Lite/job/couchbase-lite-android-build/518/artifact/logcat.txt
```
01-10 13:18:18.574 I/TestRunner( 1088): failed: testInvokeProcessorAfterReachingCapacity(com.couchbase.lite.support.BatcherTest)
01-10 13:18:18.574 I/TestRunner( 1088): ----- begin exception -----
01-10 13:18:18.584 I/TestRunner( 1088):
01-10 13:18:18.584 I/TestRunner( 1088): junit.framework.AssertionFailedError
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.fail(Assert.java:48)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.assertTrue(Assert.java:20)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.assertTrue(Assert.java:27)
01-10 13:18:18.584 I/TestRunner( 1088): at com.couchbase.lite.support.BatcherTest.testInvokeProcessorAfterReachingCapacity(BatcherTest.java:402)
01-10 13:18:18.584 I/TestRunner( 1088): at java.lang.reflect.Method.invokeNative(Native Method)
01-10 13:18:18.584 I/TestRunner( 1088): at java.lang.reflect.Method.invoke(Method.java:515)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.runTest(TestCase.java:168)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.runBare(TestCase.java:134)
01-10 13:18:18.584 I/TestRunner( 1088): at com.couchbase.lite.LiteTestCase.runBare(LiteTestCase.java:20)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult$1.protect(TestResult.java:115)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult.runProtected(TestResult.java:133)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult.run(TestResult.java:118)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.run(TestCase.java:124)
01-10 13:18:18.584 I/TestRunner( 1088): at android.test.AndroidTestRunner.runTest(AndroidTestRunner.java:191)
01-10 13:18:18.584 I/TestRunner( 1088): at android.test.AndroidTestRunner.runTest(AndroidTestRunner.java:176)
01-10 13:18:18.584 I/TestRunner( 1088): at android.test.InstrumentationTestRunner.onStart(InstrumentationTestRunner.java:554)
01-10 13:18:18.584 I/TestRunner( 1088): at android.app.Instrumentation$InstrumentationThread.run(Instrumentation.java:1701)
01-10 13:18:18.584 I/TestRunner( 1088): ----- end exception -----
01-10 13:18:18.604 I/TestRunner( 1088): finished: testInvokeProcessorAfterReachingCapacity(com.couchbase.lite.support.BatcherTest)
01-10 13:18:18.644 I/TestRunner( 1088): started: testWaitForPendingFutures(com.couchbase.lite.support.BatcherTest)
``` | 1.0 | testInvokeProcessorAfterReachingCapacity failed on jenkins - http://mobile.jenkins.couchbase.com/view/Couchbase_Lite/job/couchbase-lite-android-build/518/artifact/logcat.txt
```
01-10 13:18:18.574 I/TestRunner( 1088): failed: testInvokeProcessorAfterReachingCapacity(com.couchbase.lite.support.BatcherTest)
01-10 13:18:18.574 I/TestRunner( 1088): ----- begin exception -----
01-10 13:18:18.584 I/TestRunner( 1088):
01-10 13:18:18.584 I/TestRunner( 1088): junit.framework.AssertionFailedError
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.fail(Assert.java:48)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.assertTrue(Assert.java:20)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.assertTrue(Assert.java:27)
01-10 13:18:18.584 I/TestRunner( 1088): at com.couchbase.lite.support.BatcherTest.testInvokeProcessorAfterReachingCapacity(BatcherTest.java:402)
01-10 13:18:18.584 I/TestRunner( 1088): at java.lang.reflect.Method.invokeNative(Native Method)
01-10 13:18:18.584 I/TestRunner( 1088): at java.lang.reflect.Method.invoke(Method.java:515)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.runTest(TestCase.java:168)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.runBare(TestCase.java:134)
01-10 13:18:18.584 I/TestRunner( 1088): at com.couchbase.lite.LiteTestCase.runBare(LiteTestCase.java:20)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult$1.protect(TestResult.java:115)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult.runProtected(TestResult.java:133)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult.run(TestResult.java:118)
01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.run(TestCase.java:124)
01-10 13:18:18.584 I/TestRunner( 1088): at android.test.AndroidTestRunner.runTest(AndroidTestRunner.java:191)
01-10 13:18:18.584 I/TestRunner( 1088): at android.test.AndroidTestRunner.runTest(AndroidTestRunner.java:176)
01-10 13:18:18.584 I/TestRunner( 1088): at android.test.InstrumentationTestRunner.onStart(InstrumentationTestRunner.java:554)
01-10 13:18:18.584 I/TestRunner( 1088): at android.app.Instrumentation$InstrumentationThread.run(Instrumentation.java:1701)
01-10 13:18:18.584 I/TestRunner( 1088): ----- end exception -----
01-10 13:18:18.604 I/TestRunner( 1088): finished: testInvokeProcessorAfterReachingCapacity(com.couchbase.lite.support.BatcherTest)
01-10 13:18:18.644 I/TestRunner( 1088): started: testWaitForPendingFutures(com.couchbase.lite.support.BatcherTest)
``` | build | testinvokeprocessorafterreachingcapacity failed on jenkins i testrunner failed testinvokeprocessorafterreachingcapacity com couchbase lite support batchertest i testrunner begin exception i testrunner i testrunner junit framework assertionfailederror i testrunner at junit framework assert fail assert java i testrunner at junit framework assert asserttrue assert java i testrunner at junit framework assert asserttrue assert java i testrunner at com couchbase lite support batchertest testinvokeprocessorafterreachingcapacity batchertest java i testrunner at java lang reflect method invokenative native method i testrunner at java lang reflect method invoke method java i testrunner at junit framework testcase runtest testcase java i testrunner at junit framework testcase runbare testcase java i testrunner at com couchbase lite litetestcase runbare litetestcase java i testrunner at junit framework testresult protect testresult java i testrunner at junit framework testresult runprotected testresult java i testrunner at junit framework testresult run testresult java i testrunner at junit framework testcase run testcase java i testrunner at android test androidtestrunner runtest androidtestrunner java i testrunner at android test androidtestrunner runtest androidtestrunner java i testrunner at android test instrumentationtestrunner onstart instrumentationtestrunner java i testrunner at android app instrumentation instrumentationthread run instrumentation java i testrunner end exception i testrunner finished testinvokeprocessorafterreachingcapacity com couchbase lite support batchertest i testrunner started testwaitforpendingfutures com couchbase lite support batchertest | 1 |
259,971 | 8,202,171,170 | IssuesEvent | 2018-09-02 05:11:47 | hack4impact-uiuc/h4i-recruitment | https://api.github.com/repos/hack4impact-uiuc/h4i-recruitment | closed | Add Categories given to a Candidate by an Interviewer | Priority: High | Look into the Software Dev Rubric. Each Interviewer should give a Candidate a `category`, which is an additional categorization besides the qualitative ranking. Make a new dropdown
Thus you need to:
- [x] add `category` input where the interviewer could add their choice.
- [x] add a text input where the interviewer would add their reasoning
- [ ] show the `category` given by the interviewers in the candidate's page | 1.0 | Add Categories given to a Candidate by an Interviewer - Look into the Software Dev Rubric. Each Interviewer should give a Candidate a `category`, which is an additional categorization besides the qualitative ranking. Make a new dropdown
Thus you need to:
- [x] add `category` input where the interviewer could add their choice.
- [x] add a text input where the interviewer would add their reasoning
- [ ] show the `category` given by the interviewers in the candidate's page | non_build | add categories given to a candidate by an interviewer look into the software dev rubric each interviewer should give a candidate a category which is an additional categorization besides the qualitative ranking make a new dropdown thus you need to add category input where the interviewer could add their choice add a text input where the interviewer would add their reasoning show the category given by the interviewers in the candidate s page | 0 |
23,288 | 10,867,656,765 | IssuesEvent | 2019-11-15 00:43:45 | omisego/plasma-contracts | https://api.github.com/repos/omisego/plasma-contracts | closed | DOS / griefing attacks on calls from processExits() | bug security | ## Issue Type
```
[ x ] bug report
[ ] feature request
```
## Current Behavior
`processExits` is susceptible to attacks. The first scenario is a DOS attack that would break processing of items in the `PriorityQueue`. Consider the following scenario that in a reduced version mimics how `PriorityQueue` is used for exit game.
1.) Add items to the queue
2.) Wait a defined period of time `minExitPeriod`
3.) Reduce the queue by N items (in order they were added) and while doing so make a call to another untrusted contract/EOA (all contracts that are not part of the deployment)
If during `removeItemsfromQueue` a contract is called that drains the entire gas stipend then even only one item from the queue could fail and permanently lock the queue. The scenario below fails when processing two items but this depends on the amount of gas used after the failed call and how much gas was provided. Likely exiting to `GasDos` will beak the exit game queue with only one item due to more instruction being executed after a call.
```
pragma solidity 0.5.11;
import "../../src/framework/utils/PriorityQueue.sol";
contract PriorityQueueDOS {
PriorityQueue pq;
address callee;
uint numberOfRuns;
constructor (address _callee) public {
pq = new PriorityQueue();
callee = _callee;
}
function addItemstoQueue(uint _numberOfRuns) public {
numberOfRuns = _numberOfRuns;
for (uint i=0; i<numberOfRuns; i++){
pq.insert(1);
}
assert(pq.currentSize() == numberOfRuns);
}
function removeItemsfromQueue() public{
for (uint i=0; i<numberOfRuns; i++){
callee.call.value(0)("");
pq.delMin();
}
assert(pq.currentSize() == 0);
}
}
contract GasDos {
function () external payable{
while (1==1){
}
}
}
```
The second attack scenario is a griefing attack. The paradigm of using a queue where items are strictly processed after their ordering and where users might call `processExits()` for exiting UTXOs that they have not initiated, is by design susceptible to this type of attack. The problem can only be mitigated by limiting the amount of gas that is provided to a call. There will be proportional griefing attack possible where a malicious user starts a large number of exits and pays X amount in fees and "good users" are interested in getting their exists processed and they have to pay the Y amount in processing fees to get to process their exits. Malicious users can use gas siphoning attacks to balance the spent fee ratio more in their favor.
## Suggested Fix
I propose to set the fees for all external calls to untrusted addresses as low as possible. We probably want to provide enough so that a fallback function can process it and emit an event.We also need to consider the call `SafeERC20`.
We should make the gas stipend update-able so that we can respond to changing gas prices in the future. Let's consider a similar mechanism than we have for `updateBondSize`. Up for discussion.
As for the griefing attack I don't see how we can fix that with the current design. I think a malicious user needs proportionally more gas than a 'good user' but I want to look more into the economics. | True | DOS / griefing attacks on calls from processExits() - ## Issue Type
```
[ x ] bug report
[ ] feature request
```
## Current Behavior
`processExits` is susceptible to attacks. The first scenario is a DOS attack that would break processing of items in the `PriorityQueue`. Consider the following scenario that in a reduced version mimics how `PriorityQueue` is used for exit game.
1.) Add items to the queue
2.) Wait a defined period of time `minExitPeriod`
3.) Reduce the queue by N items (in order they were added) and while doing so make a call to another untrusted contract/EOA (all contracts that are not part of the deployment)
If during `removeItemsfromQueue` a contract is called that drains the entire gas stipend then even only one item from the queue could fail and permanently lock the queue. The scenario below fails when processing two items but this depends on the amount of gas used after the failed call and how much gas was provided. Likely exiting to `GasDos` will beak the exit game queue with only one item due to more instruction being executed after a call.
```
pragma solidity 0.5.11;
import "../../src/framework/utils/PriorityQueue.sol";
contract PriorityQueueDOS {
PriorityQueue pq;
address callee;
uint numberOfRuns;
constructor (address _callee) public {
pq = new PriorityQueue();
callee = _callee;
}
function addItemstoQueue(uint _numberOfRuns) public {
numberOfRuns = _numberOfRuns;
for (uint i=0; i<numberOfRuns; i++){
pq.insert(1);
}
assert(pq.currentSize() == numberOfRuns);
}
function removeItemsfromQueue() public{
for (uint i=0; i<numberOfRuns; i++){
callee.call.value(0)("");
pq.delMin();
}
assert(pq.currentSize() == 0);
}
}
contract GasDos {
function () external payable{
while (1==1){
}
}
}
```
The second attack scenario is a griefing attack. The paradigm of using a queue where items are strictly processed after their ordering and where users might call `processExits()` for exiting UTXOs that they have not initiated, is by design susceptible to this type of attack. The problem can only be mitigated by limiting the amount of gas that is provided to a call. There will be proportional griefing attack possible where a malicious user starts a large number of exits and pays X amount in fees and "good users" are interested in getting their exists processed and they have to pay the Y amount in processing fees to get to process their exits. Malicious users can use gas siphoning attacks to balance the spent fee ratio more in their favor.
## Suggested Fix
I propose to set the fees for all external calls to untrusted addresses as low as possible. We probably want to provide enough so that a fallback function can process it and emit an event.We also need to consider the call `SafeERC20`.
We should make the gas stipend update-able so that we can respond to changing gas prices in the future. Let's consider a similar mechanism than we have for `updateBondSize`. Up for discussion.
As for the griefing attack I don't see how we can fix that with the current design. I think a malicious user needs proportionally more gas than a 'good user' but I want to look more into the economics. | non_build | dos griefing attacks on calls from processexits issue type bug report feature request current behavior processexits is susceptible to attacks the first scenario is a dos attack that would break processing of items in the priorityqueue consider the following scenario that in a reduced version mimics how priorityqueue is used for exit game add items to the queue wait a defined period of time minexitperiod reduce the queue by n items in order they were added and while doing so make a call to another untrusted contract eoa all contracts that are not part of the deployment if during removeitemsfromqueue a contract is called that drains the entire gas stipend then even only one item from the queue could fail and permanently lock the queue the scenario below fails when processing two items but this depends on the amount of gas used after the failed call and how much gas was provided likely exiting to gasdos will beak the exit game queue with only one item due to more instruction being executed after a call pragma solidity import src framework utils priorityqueue sol contract priorityqueuedos priorityqueue pq address callee uint numberofruns constructor address callee public pq new priorityqueue callee callee function additemstoqueue uint numberofruns public numberofruns numberofruns for uint i i numberofruns i pq insert assert pq currentsize numberofruns function removeitemsfromqueue public for uint i i numberofruns i callee call value pq delmin assert pq currentsize contract gasdos function external payable while the second attack scenario is a griefing attack the paradigm of using a queue where items are strictly processed after their ordering and where users might call processexits for exiting utxos that they have not initiated is by design susceptible to this type of attack the problem can only be mitigated by limiting the amount of gas that is provided to a call there will be proportional griefing attack possible where a malicious user starts a large number of exits and pays x amount in fees and good users are interested in getting their exists processed and they have to pay the y amount in processing fees to get to process their exits malicious users can use gas siphoning attacks to balance the spent fee ratio more in their favor suggested fix i propose to set the fees for all external calls to untrusted addresses as low as possible we probably want to provide enough so that a fallback function can process it and emit an event we also need to consider the call we should make the gas stipend update able so that we can respond to changing gas prices in the future let s consider a similar mechanism than we have for updatebondsize up for discussion as for the griefing attack i don t see how we can fix that with the current design i think a malicious user needs proportionally more gas than a good user but i want to look more into the economics | 0 |
546,545 | 16,014,403,551 | IssuesEvent | 2021-04-20 14:26:43 | enso-org/ide | https://api.github.com/repos/enso-org/ide | closed | Support new `executionContext/visualisationEvaluationFailed` notification | Category: Controllers Category: GUI Difficulty: Core Contributor Priority: High Type: Enhancement | <!--
Please ensure that you check the latest version of Enso IDE to see if your feature
has been implemented.
-->
### General Summary
<!--
- Describe the feature you are requesting.
-->
enso-org/enso#1671 adds new `executionContext/visualisationEvaluationFailed` notification about visualisation errors.
In general, there are two kinds of visualisation errors.
1. [`VisualisationExpressionError`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#visualisationexpressionerror), the error reply on [`executionContext/attachVisualisation`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#executioncontextattachvisualisation) or [`executionContext/modifyVisualisation`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#executioncontextmodifyvisualisation) request. This error is sent when the text visualisation expression is compiled into an Enso function.
2. (new) `executionContext/visualisationEvaluationFailed` notification sent during the program execution, when the computed value is applied to the visualisation function.
### Motivation
<!--
- A description of the motivation for adding this feature to Enso IDE.
- Ideally this would include use-cases that support the feature.
-->
Show visualisation errors in IDE. | 1.0 | Support new `executionContext/visualisationEvaluationFailed` notification - <!--
Please ensure that you check the latest version of Enso IDE to see if your feature
has been implemented.
-->
### General Summary
<!--
- Describe the feature you are requesting.
-->
enso-org/enso#1671 adds new `executionContext/visualisationEvaluationFailed` notification about visualisation errors.
In general, there are two kinds of visualisation errors.
1. [`VisualisationExpressionError`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#visualisationexpressionerror), the error reply on [`executionContext/attachVisualisation`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#executioncontextattachvisualisation) or [`executionContext/modifyVisualisation`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#executioncontextmodifyvisualisation) request. This error is sent when the text visualisation expression is compiled into an Enso function.
2. (new) `executionContext/visualisationEvaluationFailed` notification sent during the program execution, when the computed value is applied to the visualisation function.
### Motivation
<!--
- A description of the motivation for adding this feature to Enso IDE.
- Ideally this would include use-cases that support the feature.
-->
Show visualisation errors in IDE. | non_build | support new executioncontext visualisationevaluationfailed notification please ensure that you check the latest version of enso ide to see if your feature has been implemented general summary describe the feature you are requesting enso org enso adds new executioncontext visualisationevaluationfailed notification about visualisation errors in general there are two kinds of visualisation errors the error reply on or request this error is sent when the text visualisation expression is compiled into an enso function new executioncontext visualisationevaluationfailed notification sent during the program execution when the computed value is applied to the visualisation function motivation a description of the motivation for adding this feature to enso ide ideally this would include use cases that support the feature show visualisation errors in ide | 0 |
163,848 | 13,928,614,280 | IssuesEvent | 2020-10-21 21:48:04 | AssemblyScript/working-group | https://api.github.com/repos/AssemblyScript/working-group | closed | AssemblyScript Public Meeting #23 - October 21st, 2020 | documentation enhancement good first issue help wanted | # Date and Time
This public meeting will take place: October 21st, 2020, 18:00 UTC (11:00 AM US PDT, UTC -8)
# General Agenda
* Agenda Items from comments left on this Github issue
* Additional in-meeting comments / discussion
* [If time allows] Recap of [WebAssembly CG meeting](https://github.com/WebAssembly/meetings) if anyone attended
Feel free to comment on this issue if you have any agenda items you would like to bring up.
Meeting Notes will be placed on this issue for those who cannot make the meeting.
# Meeting Information
We will use the AssemblyScript discord for meetings: https://discord.gg/vSBbJ5T . We will meet in the "meeting" voice channel :+1:
Anyone who is contributing to the AssemblyScript project, building something with AssemblyScript, interested in the Assembly project or WebAssembly in general, is welcome to join! 😄
Notes will be taken by the host, and posted after the meeting. Notes are free to be edited through comments on the meeting notes at a later time. | 1.0 | AssemblyScript Public Meeting #23 - October 21st, 2020 - # Date and Time
This public meeting will take place: October 21st, 2020, 18:00 UTC (11:00 AM US PDT, UTC -8)
# General Agenda
* Agenda Items from comments left on this Github issue
* Additional in-meeting comments / discussion
* [If time allows] Recap of [WebAssembly CG meeting](https://github.com/WebAssembly/meetings) if anyone attended
Feel free to comment on this issue if you have any agenda items you would like to bring up.
Meeting Notes will be placed on this issue for those who cannot make the meeting.
# Meeting Information
We will use the AssemblyScript discord for meetings: https://discord.gg/vSBbJ5T . We will meet in the "meeting" voice channel :+1:
Anyone who is contributing to the AssemblyScript project, building something with AssemblyScript, interested in the Assembly project or WebAssembly in general, is welcome to join! 😄
Notes will be taken by the host, and posted after the meeting. Notes are free to be edited through comments on the meeting notes at a later time. | non_build | assemblyscript public meeting october date and time this public meeting will take place october utc am us pdt utc general agenda agenda items from comments left on this github issue additional in meeting comments discussion recap of if anyone attended feel free to comment on this issue if you have any agenda items you would like to bring up meeting notes will be placed on this issue for those who cannot make the meeting meeting information we will use the assemblyscript discord for meetings we will meet in the meeting voice channel anyone who is contributing to the assemblyscript project building something with assemblyscript interested in the assembly project or webassembly in general is welcome to join 😄 notes will be taken by the host and posted after the meeting notes are free to be edited through comments on the meeting notes at a later time | 0 |
59,854 | 14,664,237,575 | IssuesEvent | 2020-12-29 11:31:12 | sandboxie-plus/Sandboxie | https://api.github.com/repos/sandboxie-plus/Sandboxie | closed | Sandboxie-Plus-x64-v0.5.2.exe says it is going to install 0.5.1 | fixed in next build | When I run Sandboxie-Plus-x64-v0.5.2.exe I get this screen :

| 1.0 | Sandboxie-Plus-x64-v0.5.2.exe says it is going to install 0.5.1 - When I run Sandboxie-Plus-x64-v0.5.2.exe I get this screen :

| build | sandboxie plus exe says it is going to install when i run sandboxie plus exe i get this screen | 1 |
443,574 | 12,796,349,959 | IssuesEvent | 2020-07-02 10:17:04 | trufflesuite/truffle | https://api.github.com/repos/trufflesuite/truffle | closed | Latest Truffle breaks our box | bug needs reproduced priority2 ⚠️ | ## Issue
Our flashloan-box worked previously with Truffle v.5.10. It no longer works in latest Truffle.
## Steps to Reproduce
Our Flashloan-box: https://github.com/aave/flashloan-box
1. Using Truffle v.5.10, follow instructions of box. Everything should work.
2. Using latest Truffle version, follow instructions.
3. It fails on Step 8, when defining `const f = await Flashloan.deployed()`
## Expected Behavior
It should work as previous Truffle version.
## Actual Results
Errors with `ReferenceERrror: f is no defined`
## Environment
* Operating System: macOS
* Ethereum client:
* Truffle version (`truffle version`): 5.1.24
* node version (`node --version`): 11.15
* npm version (`npm --version`): 6.14.4
| 1.0 | Latest Truffle breaks our box - ## Issue
Our flashloan-box worked previously with Truffle v.5.10. It no longer works in latest Truffle.
## Steps to Reproduce
Our Flashloan-box: https://github.com/aave/flashloan-box
1. Using Truffle v.5.10, follow instructions of box. Everything should work.
2. Using latest Truffle version, follow instructions.
3. It fails on Step 8, when defining `const f = await Flashloan.deployed()`
## Expected Behavior
It should work as previous Truffle version.
## Actual Results
Errors with `ReferenceERrror: f is no defined`
## Environment
* Operating System: macOS
* Ethereum client:
* Truffle version (`truffle version`): 5.1.24
* node version (`node --version`): 11.15
* npm version (`npm --version`): 6.14.4
| non_build | latest truffle breaks our box issue our flashloan box worked previously with truffle v it no longer works in latest truffle steps to reproduce our flashloan box using truffle v follow instructions of box everything should work using latest truffle version follow instructions it fails on step when defining const f await flashloan deployed expected behavior it should work as previous truffle version actual results errors with referenceerrror f is no defined environment operating system macos ethereum client truffle version truffle version node version node version npm version npm version | 0 |
41,336 | 10,702,768,859 | IssuesEvent | 2019-10-24 08:12:50 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | What are the .keep files created when creating a ballerina projects? | Area/BuildTools BetaTesting Priority/Low Type/Docs Type/Question | "resources" folders in ballerina project contains a .keep file in each "resources" folder. What are they used for? Better if we can add a small description to docs explaining the use of them. | 1.0 | What are the .keep files created when creating a ballerina projects? - "resources" folders in ballerina project contains a .keep file in each "resources" folder. What are they used for? Better if we can add a small description to docs explaining the use of them. | build | what are the keep files created when creating a ballerina projects resources folders in ballerina project contains a keep file in each resources folder what are they used for better if we can add a small description to docs explaining the use of them | 1 |
147,675 | 5,643,789,862 | IssuesEvent | 2017-04-07 01:57:48 | leo-project/leofs | https://api.github.com/repos/leo-project/leofs | opened | [leofs_doctor] Add compaction related info | Improve Priority-MIDDLE | With #671, we can include the compaction related info in https://github.com/leo-project/leofs_doctor when the target node is leo_storage. | 1.0 | [leofs_doctor] Add compaction related info - With #671, we can include the compaction related info in https://github.com/leo-project/leofs_doctor when the target node is leo_storage. | non_build | add compaction related info with we can include the compaction related info in when the target node is leo storage | 0 |
213,652 | 16,529,004,201 | IssuesEvent | 2021-05-27 01:39:36 | ppb/pursuedpybear | https://api.github.com/repos/ppb/pursuedpybear | closed | Add a Failer to test_signal_once | Hacktoberfest new contributor tests | Relevant test: https://github.com/ppb/pursuedpybear/blob/canon/tests/test_engine.py#L60
We need to add a Failer with a reasonable timeout to this test to prevent infinite loops causing CI time outs.
Failer defaults to 1 second, which is reasonable for most tests that will fail with it.
Import `ppb.testutils.Failer` and add it to the basic_systems in the test in question. | 1.0 | Add a Failer to test_signal_once - Relevant test: https://github.com/ppb/pursuedpybear/blob/canon/tests/test_engine.py#L60
We need to add a Failer with a reasonable timeout to this test to prevent infinite loops causing CI time outs.
Failer defaults to 1 second, which is reasonable for most tests that will fail with it.
Import `ppb.testutils.Failer` and add it to the basic_systems in the test in question. | non_build | add a failer to test signal once relevant test we need to add a failer with a reasonable timeout to this test to prevent infinite loops causing ci time outs failer defaults to second which is reasonable for most tests that will fail with it import ppb testutils failer and add it to the basic systems in the test in question | 0 |
12,196 | 8,637,134,730 | IssuesEvent | 2018-11-23 10:13:59 | kyma-project/kyma | https://api.github.com/repos/kyma-project/kyma | closed | Missing authentication and authorization of Jaeger API clients | area/security area/tracing security/medium wontfix | Easiness of attack: Advanced
Impact: Moderate
## Description
Jaeger exposes 2 kinds of APIs to the local network:
- a rest query API, typically used by the Jaeger UI
- a trace ingestion API, using the Zipkin protocol (possibly there are other ingestion APIs, too) None of these APIs authenticate their clients. All clients are authorized to perform any operations that are exposed by the respective API.
## Risk:
An attacker who has gained access to the local network can read and write arbitrary traces. That compromises confidentiality and integrity.
## Suggested Mitigation:
Protect the rest query API by istio. If necessary, expose it to the Internet by an ingress controller. Make sure that Jaeger UI users are authenticated (via dex) and enforce RBAC rules for Jaeger.
Consider protecting trace ingestion with istio.
Disable unused ingestion APIs, if Jaeger still exposes any. | True | Missing authentication and authorization of Jaeger API clients - Easiness of attack: Advanced
Impact: Moderate
## Description
Jaeger exposes 2 kinds of APIs to the local network:
- a rest query API, typically used by the Jaeger UI
- a trace ingestion API, using the Zipkin protocol (possibly there are other ingestion APIs, too) None of these APIs authenticate their clients. All clients are authorized to perform any operations that are exposed by the respective API.
## Risk:
An attacker who has gained access to the local network can read and write arbitrary traces. That compromises confidentiality and integrity.
## Suggested Mitigation:
Protect the rest query API by istio. If necessary, expose it to the Internet by an ingress controller. Make sure that Jaeger UI users are authenticated (via dex) and enforce RBAC rules for Jaeger.
Consider protecting trace ingestion with istio.
Disable unused ingestion APIs, if Jaeger still exposes any. | non_build | missing authentication and authorization of jaeger api clients easiness of attack advanced impact moderate description jaeger exposes kinds of apis to the local network a rest query api typically used by the jaeger ui a trace ingestion api using the zipkin protocol possibly there are other ingestion apis too none of these apis authenticate their clients all clients are authorized to perform any operations that are exposed by the respective api risk an attacker who has gained access to the local network can read and write arbitrary traces that compromises confidentiality and integrity suggested mitigation protect the rest query api by istio if necessary expose it to the internet by an ingress controller make sure that jaeger ui users are authenticated via dex and enforce rbac rules for jaeger consider protecting trace ingestion with istio disable unused ingestion apis if jaeger still exposes any | 0 |
63,931 | 15,759,536,607 | IssuesEvent | 2021-03-31 08:02:08 | kubernetes-sigs/prometheus-adapter | https://api.github.com/repos/kubernetes-sigs/prometheus-adapter | closed | make docker-build does not work | build-deploy kind/bug lifecycle/rotten | 1. make docker-build is not working
2. Even if we make it work, does it work with Kubernetes 1.9 ? | 1.0 | make docker-build does not work - 1. make docker-build is not working
2. Even if we make it work, does it work with Kubernetes 1.9 ? | build | make docker build does not work make docker build is not working even if we make it work does it work with kubernetes | 1 |
93,678 | 27,013,798,636 | IssuesEvent | 2023-02-10 17:27:53 | microsoft/appcenter | https://api.github.com/repos/microsoft/appcenter | closed | org.jetbrains.kotlin.gradle.internal.KaptWithoutKotlincTask$KaptExecutionWorkAction | feature request build | I am trying to generate release build through appcenter but in local setup i am able to generate android apk.
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':app:kaptReleaseKotlin'.
> A failure occurred while executing org.jetbrains.kotlin.gradle.internal.KaptWithoutKotlincTask$KaptExecutionWorkAction
> java.lang.reflect.InvocationTargetException (no error message)
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org/
BUILD FAILED in 14m 52s | 1.0 | org.jetbrains.kotlin.gradle.internal.KaptWithoutKotlincTask$KaptExecutionWorkAction - I am trying to generate release build through appcenter but in local setup i am able to generate android apk.
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task ':app:kaptReleaseKotlin'.
> A failure occurred while executing org.jetbrains.kotlin.gradle.internal.KaptWithoutKotlincTask$KaptExecutionWorkAction
> java.lang.reflect.InvocationTargetException (no error message)
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org/
BUILD FAILED in 14m 52s | build | org jetbrains kotlin gradle internal kaptwithoutkotlinctask kaptexecutionworkaction i am trying to generate release build through appcenter but in local setup i am able to generate android apk failure build failed with an exception what went wrong execution failed for task app kaptreleasekotlin a failure occurred while executing org jetbrains kotlin gradle internal kaptwithoutkotlinctask kaptexecutionworkaction java lang reflect invocationtargetexception no error message try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights get more help at build failed in | 1 |
19,397 | 6,718,969,580 | IssuesEvent | 2017-10-15 18:54:04 | quasar-analytics/quasar | https://api.github.com/repos/quasar-analytics/quasar | closed | ServiceSpec fails on Windows | epic: Quasar Doesn't Build/Work on Windows status: In review topic: CI (Travis) type: Bug | ```
[error] x mount filesystem (2 seconds, 234 ms)
[error] '\/-(false)' is not \/- with value'true' (ServiceSpec.scala:116)
```
This error appears to be stable and reproducible. I'm pretty sure it's an actual failure. | 1.0 | ServiceSpec fails on Windows - ```
[error] x mount filesystem (2 seconds, 234 ms)
[error] '\/-(false)' is not \/- with value'true' (ServiceSpec.scala:116)
```
This error appears to be stable and reproducible. I'm pretty sure it's an actual failure. | build | servicespec fails on windows x mount filesystem seconds ms false is not with value true servicespec scala this error appears to be stable and reproducible i m pretty sure it s an actual failure | 1 |
46,946 | 24,794,790,556 | IssuesEvent | 2022-10-24 16:18:42 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | closed | Slow JSONExtract with LowCardinality(String) tuples | performance |
`JSONExtract` of Tuples with `LowCardinality(String)` is significantly slower than with regular strings: ~2.5 slower.
Tested on CH 21.8.8.29
## Unexpectedly slow example
**60MB/s**
```
cat test.ndjson | clickhouse local -q "
SELECT JSONExtract(event, 'Tuple(
a LowCardinality(String),
b LowCardinality(String),
c LowCardinality(String),
d LowCardinality(String)
)') AS json FROM table" --input-format LineAsString --output-format Native --structure "
event String" --verbose --logger.level trace >> /dev/null
```
## Expectedly fast example
**150 MB/s**
```
time cat test.ndjson | clickhouse local -q "
SELECT JSONExtract(event, 'Tuple(
a String,
b String,
c String,
d String
)') AS json FROM table" --input-format LineAsString --output-format Native --structure "
event String" --verbose --logger.level trace >> /dev/null
```
## Workaround
**140 MB/s**
```
cat test.ndjson | clickhouse local -q "
SELECT
toLowCardinality(json.1) AS a,
toLowCardinality(json.2) AS b,
toLowCardinality(json.3) AS c,
toLowCardinality(json.4) AS d
FROM (SELECT JSONExtract(event, 'Tuple(
a String,
b String,
c String,
d String
)') AS json FROM table)" --input-format LineAsString --output-format Native --structure "
event String" --verbose --logger.level trace >> /dev/null
```
cc: @Algunenano
[test.ndjson.zip](https://github.com/ClickHouse/ClickHouse/files/7347611/test.ndjson.zip)
| True | Slow JSONExtract with LowCardinality(String) tuples -
`JSONExtract` of Tuples with `LowCardinality(String)` is significantly slower than with regular strings: ~2.5 slower.
Tested on CH 21.8.8.29
## Unexpectedly slow example
**60MB/s**
```
cat test.ndjson | clickhouse local -q "
SELECT JSONExtract(event, 'Tuple(
a LowCardinality(String),
b LowCardinality(String),
c LowCardinality(String),
d LowCardinality(String)
)') AS json FROM table" --input-format LineAsString --output-format Native --structure "
event String" --verbose --logger.level trace >> /dev/null
```
## Expectedly fast example
**150 MB/s**
```
time cat test.ndjson | clickhouse local -q "
SELECT JSONExtract(event, 'Tuple(
a String,
b String,
c String,
d String
)') AS json FROM table" --input-format LineAsString --output-format Native --structure "
event String" --verbose --logger.level trace >> /dev/null
```
## Workaround
**140 MB/s**
```
cat test.ndjson | clickhouse local -q "
SELECT
toLowCardinality(json.1) AS a,
toLowCardinality(json.2) AS b,
toLowCardinality(json.3) AS c,
toLowCardinality(json.4) AS d
FROM (SELECT JSONExtract(event, 'Tuple(
a String,
b String,
c String,
d String
)') AS json FROM table)" --input-format LineAsString --output-format Native --structure "
event String" --verbose --logger.level trace >> /dev/null
```
cc: @Algunenano
[test.ndjson.zip](https://github.com/ClickHouse/ClickHouse/files/7347611/test.ndjson.zip)
| non_build | slow jsonextract with lowcardinality string tuples jsonextract of tuples with lowcardinality string is significantly slower than with regular strings slower tested on ch unexpectedly slow example s cat test ndjson clickhouse local q select jsonextract event tuple a lowcardinality string b lowcardinality string c lowcardinality string d lowcardinality string as json from table input format lineasstring output format native structure event string verbose logger level trace dev null expectedly fast example mb s time cat test ndjson clickhouse local q select jsonextract event tuple a string b string c string d string as json from table input format lineasstring output format native structure event string verbose logger level trace dev null workaround mb s cat test ndjson clickhouse local q select tolowcardinality json as a tolowcardinality json as b tolowcardinality json as c tolowcardinality json as d from select jsonextract event tuple a string b string c string d string as json from table input format lineasstring output format native structure event string verbose logger level trace dev null cc algunenano | 0 |
23,659 | 7,359,199,517 | IssuesEvent | 2018-03-10 03:14:23 | ild-games/Spool | https://api.github.com/repos/ild-games/Spool | opened | Configure code coverage to include all sources, not just files that are touched by a unit test | build | Angular code coverage testing is done via `ng test --code-coverage`. This uses istanbul under the hood by angular-cli to run code coverage checks. However, by default this will only cover source files that are covered by unit tests. This turns the code coverage check into a not very helpful metric, as we are only getting "What percentage of code paths weren't hit that could have been hit based on your existing unit tests" instead of "What percentage of your code based isn't unit tested". | 1.0 | Configure code coverage to include all sources, not just files that are touched by a unit test - Angular code coverage testing is done via `ng test --code-coverage`. This uses istanbul under the hood by angular-cli to run code coverage checks. However, by default this will only cover source files that are covered by unit tests. This turns the code coverage check into a not very helpful metric, as we are only getting "What percentage of code paths weren't hit that could have been hit based on your existing unit tests" instead of "What percentage of your code based isn't unit tested". | build | configure code coverage to include all sources not just files that are touched by a unit test angular code coverage testing is done via ng test code coverage this uses istanbul under the hood by angular cli to run code coverage checks however by default this will only cover source files that are covered by unit tests this turns the code coverage check into a not very helpful metric as we are only getting what percentage of code paths weren t hit that could have been hit based on your existing unit tests instead of what percentage of your code based isn t unit tested | 1 |
102,447 | 32,007,115,681 | IssuesEvent | 2023-09-21 15:28:01 | pwa-builder/PWABuilder | https://api.github.com/repos/pwa-builder/PWABuilder | closed | [PWA Builder]: Spin button form field does not have label. | bug :bug: needs triage :mag: needs attention :wave: a11y A11yWCAG Accessibility [E+D] A11ySev2 Product-PWABuilder A11yAuto HCL-E+D CT-July23 | ### What happened?
Spin button form field present on the URL does not have label, narrator does not announce the information of the 'Padding' label.
### How do we reproduce the behavior?
**Repro Steps:**
1. Open the PWA Builder URL in Anaheim dev browser.
2. PWA Builder page will appear.
3. Now Run the Fast Pass from AI4W.
4. Observe the issue.
### What do you expect to happen?
Spin button form field present on the URL must be associated with the label 'Padding'.
### What environment were you using?
**Test Environment:**
OS Build: Windows 11 Enterprise Insider Preview 22H2 (23493.1000)
Edge Browser: Version 115.0.1880.3 (Official build) dev (64-bit)
URL: https://www.pwabuilder.com/imageGenerator
Tool: AI4W
### Additional context
**Element path:**
app-index,image-generator,sl-input,#input
**Snippet:**
<input part="input" id="input" class="input__control" aria-describedby="help-text" type="number" title="" name="padding" required="" placeholder="" min="0" max="1" step="0.1" spellcheck="true">
**How to fix:**
Fix all of the following:
Form element has explicit <label> that is hidden
**User Experience:**
If the label of the form element will not be defined then the users depend on screen reader will not get the information of the purpose of the spin button, and they may get confused in accessing the UI properly.
**WCAG Reference:**
https://www.w3.org/WAI/WCAG21/Understanding/name-role-value.html
**Attachments:**


| 1.0 | [PWA Builder]: Spin button form field does not have label. - ### What happened?
Spin button form field present on the URL does not have label, narrator does not announce the information of the 'Padding' label.
### How do we reproduce the behavior?
**Repro Steps:**
1. Open the PWA Builder URL in Anaheim dev browser.
2. PWA Builder page will appear.
3. Now Run the Fast Pass from AI4W.
4. Observe the issue.
### What do you expect to happen?
Spin button form field present on the URL must be associated with the label 'Padding'.
### What environment were you using?
**Test Environment:**
OS Build: Windows 11 Enterprise Insider Preview 22H2 (23493.1000)
Edge Browser: Version 115.0.1880.3 (Official build) dev (64-bit)
URL: https://www.pwabuilder.com/imageGenerator
Tool: AI4W
### Additional context
**Element path:**
app-index,image-generator,sl-input,#input
**Snippet:**
<input part="input" id="input" class="input__control" aria-describedby="help-text" type="number" title="" name="padding" required="" placeholder="" min="0" max="1" step="0.1" spellcheck="true">
**How to fix:**
Fix all of the following:
Form element has explicit <label> that is hidden
**User Experience:**
If the label of the form element will not be defined then the users depend on screen reader will not get the information of the purpose of the spin button, and they may get confused in accessing the UI properly.
**WCAG Reference:**
https://www.w3.org/WAI/WCAG21/Understanding/name-role-value.html
**Attachments:**


| build | spin button form field does not have label what happened spin button form field present on the url does not have label narrator does not announce the information of the padding label how do we reproduce the behavior repro steps open the pwa builder url in anaheim dev browser pwa builder page will appear now run the fast pass from observe the issue what do you expect to happen spin button form field present on the url must be associated with the label padding what environment were you using test environment os build windows enterprise insider preview edge browser version official build dev bit url tool additional context element path app index image generator sl input input snippet how to fix fix all of the following form element has explicit that is hidden user experience if the label of the form element will not be defined then the users depend on screen reader will not get the information of the purpose of the spin button and they may get confused in accessing the ui properly wcag reference attachments | 1 |
26,685 | 7,858,659,330 | IssuesEvent | 2018-06-21 14:30:07 | ShaikASK/Testing | https://api.github.com/repos/ShaikASK/Testing | closed | Safari (10) /IE (11) : New Hires : Vertical scroll bar is missing in new hires screen as a result of which only 50% of screen is visible | Defect HR Admin Module HR User Module New Hire P2 Release #3 Build 2 | Steps :
1.Launch the URL
2.Sign in as HR user
3.Click on New Hires from side menu
4.Navigate to "View New Hires" page
5.Click on + icon beside "New Hires" tab
Experienced Behaviour : Observed that Vertical scroll bar is missing in new hires screen as a result of which only 50% of screen is visible when check with safari (10) & IE (11) (Refer screen shot)
Expected Behaviour : Ensure that application should display vertical scroll bar in new hires screen and it should should display proper UI

| 1.0 | Safari (10) /IE (11) : New Hires : Vertical scroll bar is missing in new hires screen as a result of which only 50% of screen is visible - Steps :
1.Launch the URL
2.Sign in as HR user
3.Click on New Hires from side menu
4.Navigate to "View New Hires" page
5.Click on + icon beside "New Hires" tab
Experienced Behaviour : Observed that Vertical scroll bar is missing in new hires screen as a result of which only 50% of screen is visible when check with safari (10) & IE (11) (Refer screen shot)
Expected Behaviour : Ensure that application should display vertical scroll bar in new hires screen and it should should display proper UI

| build | safari ie new hires vertical scroll bar is missing in new hires screen as a result of which only of screen is visible steps launch the url sign in as hr user click on new hires from side menu navigate to view new hires page click on icon beside new hires tab experienced behaviour observed that vertical scroll bar is missing in new hires screen as a result of which only of screen is visible when check with safari ie refer screen shot expected behaviour ensure that application should display vertical scroll bar in new hires screen and it should should display proper ui | 1 |
166,711 | 14,075,868,325 | IssuesEvent | 2020-11-04 09:40:04 | packethost/metabot | https://api.github.com/repos/packethost/metabot | opened | Uniform Standards Request: Maintained Repository | documentation enhancement good first issue help wanted | Hello!
We believe this repository is Maintained and therefore needs the following files updated:
* [ ] Flagged Public
* [ ] [The Maintained badge](https://github.com/packethost/standards/blob/master/glossary.md#maintained-badge)
* [ ] [The Maintained Statement in the README.md](https://github.com/packethost/standards/blob/master/glossary.md#maintained-statement)
* [ ] [Getting Started Tutorial](https://github.com/packethost/standards/blob/master/glossary.md#getting-started-tutorial)
* [ ] [README.md](https://github.com/packethost/standards/blob/master/glossary.md#readmemd)
* [ ] [LICENSE](https://github.com/packethost/standards/blob/master/glossary.md#license)
* [ ] [CI/CD](https://github.com/packethost/standards/blob/master/glossary.md#cicd)
* [ ] [OWNERS.md](https://github.com/packethost/standards/blob/master/glossary.md#ownersmd)
* [ ] [Developer Certificate of Origin](https://github.com/packethost/standards/blob/master/glossary.md#developer-certificate-of-origin)
* [ ] At least two [maintainers](https://github.com/packethost/standards/blob/master/glossary.md#maintainer)
* [ ] [A Manifest File](https://github.com/packethost/standards/blob/master/glossary.md#manifest-file), if relevant
* [ ] [Code of Conduct](https://github.com/packethost/standards/blob/master/glossary.md#code-of-conduct)
* [ ] [How to Contribute](https://github.com/packethost/standards/blob/master/glossary.md#contributors-guide)
* [ ] [SUPPORT.md](https://github.com/packethost/standards/blob/master/glossary.md#supportmd)
* [ ] [RELEASE.md](https://github.com/packethost/standards/blob/master/glossary.md#releasemd)
If you feel the repository should be experimental or end of life or that you'll need assistance to update these files, please let us know by filing an issue with https://github.com/packethost/standards.
## The Uniform Standards Project
Packet maintains a number of public repositories that help customers to run various workloads on Packet. These repositories are in various states of completeness and quality, and being public, developers often find them and start using them. This creates problems:
* Developers using low-quality repositories may infer that Packet generally provides a low quality experience.
* Many of our repositories are put online with no formal communication with, or training for, customer success. This leads to a below average support experience when things do go wrong.
* We spend a huge amount of time supporting users through various channels when with better upfront planning, documentation and testing much of this support work could be eliminated.
To that end, we propose three tiers of repositories: [Private](https://github.com/packethost/standards#private-tier-minimum-requirements), [Experimental](https://github.com/packethost/standards#experimental-tier-minimum-requirements), and [Maintained](https://github.com/packethost/standards#maintained-tier-minimum-requirements).
As a resource and example of a maintained repository, we've created https://github.com/packethost/standards. This is also where you can file any requests for assistance or modification of scope.
### The Goal
Our repositories should be the example from which adjacent, competing, projects look for inspiration.
Each repository should not look entirely different from other repositories in the ecosystem, having a different layout, a different testing model, or a different logging model, for example, without reason or recommendation from the subject matter experts from the community.
We should share our improvements with each ecosystem while seeking and respecting the feedback of these communities.
Whether or not strict guidelines have been provided for the project type, our repositories should ensure that the same components are offered across the board. How these components are provided may vary, based on the conventions of the project type. GitHub provides general guidance on this which they have integrated into their user experience.
| 1.0 | Uniform Standards Request: Maintained Repository - Hello!
We believe this repository is Maintained and therefore needs the following files updated:
* [ ] Flagged Public
* [ ] [The Maintained badge](https://github.com/packethost/standards/blob/master/glossary.md#maintained-badge)
* [ ] [The Maintained Statement in the README.md](https://github.com/packethost/standards/blob/master/glossary.md#maintained-statement)
* [ ] [Getting Started Tutorial](https://github.com/packethost/standards/blob/master/glossary.md#getting-started-tutorial)
* [ ] [README.md](https://github.com/packethost/standards/blob/master/glossary.md#readmemd)
* [ ] [LICENSE](https://github.com/packethost/standards/blob/master/glossary.md#license)
* [ ] [CI/CD](https://github.com/packethost/standards/blob/master/glossary.md#cicd)
* [ ] [OWNERS.md](https://github.com/packethost/standards/blob/master/glossary.md#ownersmd)
* [ ] [Developer Certificate of Origin](https://github.com/packethost/standards/blob/master/glossary.md#developer-certificate-of-origin)
* [ ] At least two [maintainers](https://github.com/packethost/standards/blob/master/glossary.md#maintainer)
* [ ] [A Manifest File](https://github.com/packethost/standards/blob/master/glossary.md#manifest-file), if relevant
* [ ] [Code of Conduct](https://github.com/packethost/standards/blob/master/glossary.md#code-of-conduct)
* [ ] [How to Contribute](https://github.com/packethost/standards/blob/master/glossary.md#contributors-guide)
* [ ] [SUPPORT.md](https://github.com/packethost/standards/blob/master/glossary.md#supportmd)
* [ ] [RELEASE.md](https://github.com/packethost/standards/blob/master/glossary.md#releasemd)
If you feel the repository should be experimental or end of life or that you'll need assistance to update these files, please let us know by filing an issue with https://github.com/packethost/standards.
## The Uniform Standards Project
Packet maintains a number of public repositories that help customers to run various workloads on Packet. These repositories are in various states of completeness and quality, and being public, developers often find them and start using them. This creates problems:
* Developers using low-quality repositories may infer that Packet generally provides a low quality experience.
* Many of our repositories are put online with no formal communication with, or training for, customer success. This leads to a below average support experience when things do go wrong.
* We spend a huge amount of time supporting users through various channels when with better upfront planning, documentation and testing much of this support work could be eliminated.
To that end, we propose three tiers of repositories: [Private](https://github.com/packethost/standards#private-tier-minimum-requirements), [Experimental](https://github.com/packethost/standards#experimental-tier-minimum-requirements), and [Maintained](https://github.com/packethost/standards#maintained-tier-minimum-requirements).
As a resource and example of a maintained repository, we've created https://github.com/packethost/standards. This is also where you can file any requests for assistance or modification of scope.
### The Goal
Our repositories should be the example from which adjacent, competing, projects look for inspiration.
Each repository should not look entirely different from other repositories in the ecosystem, having a different layout, a different testing model, or a different logging model, for example, without reason or recommendation from the subject matter experts from the community.
We should share our improvements with each ecosystem while seeking and respecting the feedback of these communities.
Whether or not strict guidelines have been provided for the project type, our repositories should ensure that the same components are offered across the board. How these components are provided may vary, based on the conventions of the project type. GitHub provides general guidance on this which they have integrated into their user experience.
| non_build | uniform standards request maintained repository hello we believe this repository is maintained and therefore needs the following files updated flagged public at least two if relevant if you feel the repository should be experimental or end of life or that you ll need assistance to update these files please let us know by filing an issue with the uniform standards project packet maintains a number of public repositories that help customers to run various workloads on packet these repositories are in various states of completeness and quality and being public developers often find them and start using them this creates problems developers using low quality repositories may infer that packet generally provides a low quality experience many of our repositories are put online with no formal communication with or training for customer success this leads to a below average support experience when things do go wrong we spend a huge amount of time supporting users through various channels when with better upfront planning documentation and testing much of this support work could be eliminated to that end we propose three tiers of repositories and as a resource and example of a maintained repository we ve created this is also where you can file any requests for assistance or modification of scope the goal our repositories should be the example from which adjacent competing projects look for inspiration each repository should not look entirely different from other repositories in the ecosystem having a different layout a different testing model or a different logging model for example without reason or recommendation from the subject matter experts from the community we should share our improvements with each ecosystem while seeking and respecting the feedback of these communities whether or not strict guidelines have been provided for the project type our repositories should ensure that the same components are offered across the board how these components are provided may vary based on the conventions of the project type github provides general guidance on this which they have integrated into their user experience | 0 |
689,938 | 23,640,915,825 | IssuesEvent | 2022-08-25 16:58:26 | deckhouse/deckhouse | https://api.github.com/repos/deckhouse/deckhouse | closed | [cloud-provider-aws] Add ability to configure additional policies for IAM roles | area/cloud-provider type/feature-request type/good-first-issue priority/backlog | ### Preflight Checklist
- [X] I agree to follow the [Code of Conduct](https://github.com/deckhouse/deckhouse/blob/main/CODE_OF_CONDUCT.md) that this project adheres to.
- [X] I have searched the [issue tracker](https://github.com/deckhouse/deckhouse/issues) for an issue that matches the one I want to file, without success.
### Use case. Why is this important?
There are several cases when you might need to give access to AWS API for applications running on nodes. We have a basic role that is attached to every node in the cluster. Additional policies can be attached to it.
### Proposed Solution
Add additionalRolePolicies parameter in AWSClusterConfiguration.
### Additional Information
_No response_ | 1.0 | [cloud-provider-aws] Add ability to configure additional policies for IAM roles - ### Preflight Checklist
- [X] I agree to follow the [Code of Conduct](https://github.com/deckhouse/deckhouse/blob/main/CODE_OF_CONDUCT.md) that this project adheres to.
- [X] I have searched the [issue tracker](https://github.com/deckhouse/deckhouse/issues) for an issue that matches the one I want to file, without success.
### Use case. Why is this important?
There are several cases when you might need to give access to AWS API for applications running on nodes. We have a basic role that is attached to every node in the cluster. Additional policies can be attached to it.
### Proposed Solution
Add additionalRolePolicies parameter in AWSClusterConfiguration.
### Additional Information
_No response_ | non_build | add ability to configure additional policies for iam roles preflight checklist i agree to follow the that this project adheres to i have searched the for an issue that matches the one i want to file without success use case why is this important there are several cases when you might need to give access to aws api for applications running on nodes we have a basic role that is attached to every node in the cluster additional policies can be attached to it proposed solution add additionalrolepolicies parameter in awsclusterconfiguration additional information no response | 0 |
2,961 | 5,832,709,801 | IssuesEvent | 2017-05-08 22:37:46 | Cxbx-Reloaded/Cxbx-Reloaded | https://api.github.com/repos/Cxbx-Reloaded/Cxbx-Reloaded | closed | [PAL] Lego Star Wars crashes at language selection screen | game-compatibility | Tested on Cxbx-Reloaded build 7c266ad (Apr 26 2017)
Pretty self explanatory. In Hardware mode, the loading screen will come up with working video and immediately crash. Software mode has an immediate crash with no audio or video.
Hardware mode crash: 
Hardware mode Kernel Debug file: [KrnlDebug.txt](https://github.com/Cxbx-Reloaded/Cxbx-Reloaded/files/963443/KrnlDebug.txt)
Settings used:
Display Adapter: NVIDIA GeForce GTX 1060 6GB
Direct3D Device: Direct3D HAL (Hardware)
Video Resolution: Automatic (Default) | True | [PAL] Lego Star Wars crashes at language selection screen - Tested on Cxbx-Reloaded build 7c266ad (Apr 26 2017)
Pretty self explanatory. In Hardware mode, the loading screen will come up with working video and immediately crash. Software mode has an immediate crash with no audio or video.
Hardware mode crash: 
Hardware mode Kernel Debug file: [KrnlDebug.txt](https://github.com/Cxbx-Reloaded/Cxbx-Reloaded/files/963443/KrnlDebug.txt)
Settings used:
Display Adapter: NVIDIA GeForce GTX 1060 6GB
Direct3D Device: Direct3D HAL (Hardware)
Video Resolution: Automatic (Default) | non_build | lego star wars crashes at language selection screen tested on cxbx reloaded build apr pretty self explanatory in hardware mode the loading screen will come up with working video and immediately crash software mode has an immediate crash with no audio or video hardware mode crash hardware mode kernel debug file settings used display adapter nvidia geforce gtx device hal hardware video resolution automatic default | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.