Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
853
labels
stringlengths
4
898
body
stringlengths
2
262k
index
stringclasses
13 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
250k
binary_label
int64
0
1
179,918
21,606,569,830
IssuesEvent
2022-05-04 04:25:43
testdemo1227/Demo-WebGoat1
https://api.github.com/repos/testdemo1227/Demo-WebGoat1
opened
jstl-1.2.jar: 1 vulnerabilities (highest severity is: 7.3)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jstl-1.2.jar</b></p></summary> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /ository/javax/servlet/jstl/1.2/jstl-1.2.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2015-0254](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | jstl-1.2.jar | Direct | org.apache.taglibs:taglibs-standard-impl:1.2.3 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-0254</summary> ### Vulnerable Library - <b>jstl-1.2.jar</b></p> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /ository/javax/servlet/jstl/1.2/jstl-1.2.jar</p> <p> Dependency Hierarchy: - :x: **jstl-1.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Apache Standard Taglibs before 1.2.3 allows remote attackers to execute arbitrary code or conduct external XML entity (XXE) attacks via a crafted XSLT extension in a (1) <x:parse> or (2) <x:transform> JSTL XML tag. <p>Publish Date: 2015-03-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254>CVE-2015-0254</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tomcat.apache.org/taglibs/standard/">https://tomcat.apache.org/taglibs/standard/</a></p> <p>Release Date: 2015-03-09</p> <p>Fix Resolution: org.apache.taglibs:taglibs-standard-impl:1.2.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"javax.servlet","packageName":"jstl","packageVersion":"1.2","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"javax.servlet:jstl:1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.taglibs:taglibs-standard-impl:1.2.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-0254","vulnerabilityDetails":"Apache Standard Taglibs before 1.2.3 allows remote attackers to execute arbitrary code or conduct external XML entity (XXE) attacks via a crafted XSLT extension in a (1) \u003cx:parse\u003e or (2) \u003cx:transform\u003e JSTL XML tag.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> -->
True
jstl-1.2.jar: 1 vulnerabilities (highest severity is: 7.3) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jstl-1.2.jar</b></p></summary> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /ository/javax/servlet/jstl/1.2/jstl-1.2.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2015-0254](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | jstl-1.2.jar | Direct | org.apache.taglibs:taglibs-standard-impl:1.2.3 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-0254</summary> ### Vulnerable Library - <b>jstl-1.2.jar</b></p> <p></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /ository/javax/servlet/jstl/1.2/jstl-1.2.jar</p> <p> Dependency Hierarchy: - :x: **jstl-1.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/testdemo1227/Demo-WebGoat1/commit/4bc811a006e2ea4234387fe41dc021407d92aae0">4bc811a006e2ea4234387fe41dc021407d92aae0</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Apache Standard Taglibs before 1.2.3 allows remote attackers to execute arbitrary code or conduct external XML entity (XXE) attacks via a crafted XSLT extension in a (1) <x:parse> or (2) <x:transform> JSTL XML tag. <p>Publish Date: 2015-03-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254>CVE-2015-0254</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tomcat.apache.org/taglibs/standard/">https://tomcat.apache.org/taglibs/standard/</a></p> <p>Release Date: 2015-03-09</p> <p>Fix Resolution: org.apache.taglibs:taglibs-standard-impl:1.2.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"javax.servlet","packageName":"jstl","packageVersion":"1.2","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"javax.servlet:jstl:1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.taglibs:taglibs-standard-impl:1.2.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-0254","vulnerabilityDetails":"Apache Standard Taglibs before 1.2.3 allows remote attackers to execute arbitrary code or conduct external XML entity (XXE) attacks via a crafted XSLT extension in a (1) \u003cx:parse\u003e or (2) \u003cx:transform\u003e JSTL XML tag.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-0254","cvss3Severity":"high","cvss3Score":"7.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> -->
non_build
jstl jar vulnerabilities highest severity is vulnerable library jstl jar path to dependency file pom xml path to vulnerable library ository javax servlet jstl jstl jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high jstl jar direct org apache taglibs taglibs standard impl details cve vulnerable library jstl jar path to dependency file pom xml path to vulnerable library ository javax servlet jstl jstl jar dependency hierarchy x jstl jar vulnerable library found in head commit a href found in base branch master vulnerability details apache standard taglibs before allows remote attackers to execute arbitrary code or conduct external xml entity xxe attacks via a crafted xslt extension in a or jstl xml tag publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache taglibs taglibs standard impl step up your open source security game with whitesource istransitivedependency false dependencytree javax servlet jstl isminimumfixversionavailable true minimumfixversion org apache taglibs taglibs standard impl isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails apache standard taglibs before allows remote attackers to execute arbitrary code or conduct external xml entity xxe attacks via a crafted xslt extension in a parse or transform jstl xml tag vulnerabilityurl
0
100,251
30,655,612,913
IssuesEvent
2023-07-25 11:56:29
dotnet/fsharp
https://api.github.com/repos/dotnet/fsharp
closed
Generate deterministic MVID for reference assemblies based on contents of the assembly
Feature Improvement Area-Compiler-CodeGen Area-Build
Continuation of reference assembly work, initially done in https://github.com/dotnet/fsharp/pull/12334 Generate deterministic MVID for reference assemblies based on contents of the assembly where the content also includes signature and optimization data; basically SHA1 of the bytes of the entire assembly. > **Note** > We need to keep in mind, that we store some extra information in sigdata, which can affect MVID generation: > - the metadata for "private" declarations in implementation files that don't have signature files > - the metadata for "internal" declarations when there are no InternalsVisibleTo in the assembly > > Related issue: https://github.com/dotnet/fsharp/issues/12746
1.0
Generate deterministic MVID for reference assemblies based on contents of the assembly - Continuation of reference assembly work, initially done in https://github.com/dotnet/fsharp/pull/12334 Generate deterministic MVID for reference assemblies based on contents of the assembly where the content also includes signature and optimization data; basically SHA1 of the bytes of the entire assembly. > **Note** > We need to keep in mind, that we store some extra information in sigdata, which can affect MVID generation: > - the metadata for "private" declarations in implementation files that don't have signature files > - the metadata for "internal" declarations when there are no InternalsVisibleTo in the assembly > > Related issue: https://github.com/dotnet/fsharp/issues/12746
build
generate deterministic mvid for reference assemblies based on contents of the assembly continuation of reference assembly work initially done in generate deterministic mvid for reference assemblies based on contents of the assembly where the content also includes signature and optimization data basically of the bytes of the entire assembly note we need to keep in mind that we store some extra information in sigdata which can affect mvid generation the metadata for private declarations in implementation files that don t have signature files the metadata for internal declarations when there are no internalsvisibleto in the assembly related issue
1
67,178
16,829,040,879
IssuesEvent
2021-06-17 23:45:26
golang/go
https://api.github.com/repos/golang/go
reopened
x/build: new builder for cmd/compile's "unified IR" mode
Builders NeedsFix new-builder
I'd like to add a builder for testing the compiler's new "unified IR" mode. This has landed on dev.typeparams, and it took all of 1 commits for us to accidentally break it again. Proposed build config (caveat: this is just copy/paste/tweaked based on some of the other compiler experiment configs; I don't feel strongly about any of these if other changes would be appropriate): ``` addBuilder(BuildConfig{ Name: "linux-amd64-unified-ir", HostType: "host-linux-buster", Notes: "builder with GO_GCFLAGS=-d=unified=1, see golang.org/issue/46786", buildsRepo: func(repo, branch, goBranch string) bool { return repo == "go" && branch == "dev.typeparams" }, env: []string{ "GO_DISABLE_OUTBOUND_NETWORK=1", "GO_GCFLAGS=-d=unified=1", }, GoDeps: []string{ "cf1ae5fc364eb7f2ee5203e4c5e30411c3cfe01f", // dev.typeparams commit that added -d=unified }, }) ``` Caveat: It's expected that the `go tool dist test test:0_1` step will fail with unified IR currently. (There's a list of tests that are known to fail with the new types2 type checker, but test/run.go doesn't know -d=unified enables this too. I will fix this shortly.)
2.0
x/build: new builder for cmd/compile's "unified IR" mode - I'd like to add a builder for testing the compiler's new "unified IR" mode. This has landed on dev.typeparams, and it took all of 1 commits for us to accidentally break it again. Proposed build config (caveat: this is just copy/paste/tweaked based on some of the other compiler experiment configs; I don't feel strongly about any of these if other changes would be appropriate): ``` addBuilder(BuildConfig{ Name: "linux-amd64-unified-ir", HostType: "host-linux-buster", Notes: "builder with GO_GCFLAGS=-d=unified=1, see golang.org/issue/46786", buildsRepo: func(repo, branch, goBranch string) bool { return repo == "go" && branch == "dev.typeparams" }, env: []string{ "GO_DISABLE_OUTBOUND_NETWORK=1", "GO_GCFLAGS=-d=unified=1", }, GoDeps: []string{ "cf1ae5fc364eb7f2ee5203e4c5e30411c3cfe01f", // dev.typeparams commit that added -d=unified }, }) ``` Caveat: It's expected that the `go tool dist test test:0_1` step will fail with unified IR currently. (There's a list of tests that are known to fail with the new types2 type checker, but test/run.go doesn't know -d=unified enables this too. I will fix this shortly.)
build
x build new builder for cmd compile s unified ir mode i d like to add a builder for testing the compiler s new unified ir mode this has landed on dev typeparams and it took all of commits for us to accidentally break it again proposed build config caveat this is just copy paste tweaked based on some of the other compiler experiment configs i don t feel strongly about any of these if other changes would be appropriate addbuilder buildconfig name linux unified ir hosttype host linux buster notes builder with go gcflags d unified see golang org issue buildsrepo func repo branch gobranch string bool return repo go branch dev typeparams env string go disable outbound network go gcflags d unified godeps string dev typeparams commit that added d unified caveat it s expected that the go tool dist test test step will fail with unified ir currently there s a list of tests that are known to fail with the new type checker but test run go doesn t know d unified enables this too i will fix this shortly
1
382,412
11,305,860,862
IssuesEvent
2020-01-18 09:27:26
acidanthera/bugtracker
https://api.github.com/repos/acidanthera/bugtracker
closed
AppleSupportPkg\Platform\ApfsDriverLoader: warning C4244: '=': conversion from '...' to '...', possible loss of data
priority:normal project:apfs
``` AppleSupportPkg\Platform\ApfsDriverLoader\ApfsDriverLoader.c(381): warning C4244: '=': conversion from 'EFI_LBA' to 'UINTN', possible loss of data AppleSupportPkg\Platform\ApfsDriverLoader\ApfsDriverLoader.c(979): warning C4244: '=': conversion from 'UINT64' to 'UINTN', possible loss of data ```
1.0
AppleSupportPkg\Platform\ApfsDriverLoader: warning C4244: '=': conversion from '...' to '...', possible loss of data - ``` AppleSupportPkg\Platform\ApfsDriverLoader\ApfsDriverLoader.c(381): warning C4244: '=': conversion from 'EFI_LBA' to 'UINTN', possible loss of data AppleSupportPkg\Platform\ApfsDriverLoader\ApfsDriverLoader.c(979): warning C4244: '=': conversion from 'UINT64' to 'UINTN', possible loss of data ```
non_build
applesupportpkg platform apfsdriverloader warning conversion from to possible loss of data applesupportpkg platform apfsdriverloader apfsdriverloader c warning conversion from efi lba to uintn possible loss of data applesupportpkg platform apfsdriverloader apfsdriverloader c warning conversion from to uintn possible loss of data
0
374,320
11,083,931,556
IssuesEvent
2019-12-13 15:29:06
googleapis/google-cloud-cpp-spanner
https://api.github.com/repos/googleapis/google-cloud-cpp-spanner
closed
Properly handle Session NOT_FOUND
priority: p2 type: bug
@coryan reported that one of his tests failed with the following error: `Exception caught in HTTP handler: Permanent failure in RunTransactionWithPolicies: Session not found: <XXX> [NOT FOUND]` We should handle this error internally and avoid exposing it to the user. From https://cloud.google.com/spanner/docs/sessions ``` Attempts to use a deleted session result in NOT_FOUND. If you encounter this error, create and use a new session, add the new session to the cache, and remove the deleted session from the cache. ``` However, I'm not entirely sure how to accomplish this given that a `Transaction` is bound to a `Session`. For RW transactions using `RunTransaction` we should be able to treat it similarly to an `ABORTED` and retry the whole transaction with a new session. For RO transactions or non-idempotent mutations, it's less clear what we should do. The page linked-to above gives some general advice, but nothing specifically about this case.
1.0
Properly handle Session NOT_FOUND - @coryan reported that one of his tests failed with the following error: `Exception caught in HTTP handler: Permanent failure in RunTransactionWithPolicies: Session not found: <XXX> [NOT FOUND]` We should handle this error internally and avoid exposing it to the user. From https://cloud.google.com/spanner/docs/sessions ``` Attempts to use a deleted session result in NOT_FOUND. If you encounter this error, create and use a new session, add the new session to the cache, and remove the deleted session from the cache. ``` However, I'm not entirely sure how to accomplish this given that a `Transaction` is bound to a `Session`. For RW transactions using `RunTransaction` we should be able to treat it similarly to an `ABORTED` and retry the whole transaction with a new session. For RO transactions or non-idempotent mutations, it's less clear what we should do. The page linked-to above gives some general advice, but nothing specifically about this case.
non_build
properly handle session not found coryan reported that one of his tests failed with the following error exception caught in http handler permanent failure in runtransactionwithpolicies session not found we should handle this error internally and avoid exposing it to the user from attempts to use a deleted session result in not found if you encounter this error create and use a new session add the new session to the cache and remove the deleted session from the cache however i m not entirely sure how to accomplish this given that a transaction is bound to a session for rw transactions using runtransaction we should be able to treat it similarly to an aborted and retry the whole transaction with a new session for ro transactions or non idempotent mutations it s less clear what we should do the page linked to above gives some general advice but nothing specifically about this case
0
27,496
7,970,704,837
IssuesEvent
2018-07-16 13:35:47
ninia/jep
https://api.github.com/repos/ninia/jep
closed
pip install : no such file or directory for classlist_7.txt
build
When installing jep with pip i get the error error: [Errno 2] No such file or directory: 'src/main/resources/jep/classlist_7.txt' - OS Platform, Distribution, and Version: Windows 7 - Python Distribution and Version: Python 3.6 - Java Distribution and Version: Java 1.8
1.0
pip install : no such file or directory for classlist_7.txt - When installing jep with pip i get the error error: [Errno 2] No such file or directory: 'src/main/resources/jep/classlist_7.txt' - OS Platform, Distribution, and Version: Windows 7 - Python Distribution and Version: Python 3.6 - Java Distribution and Version: Java 1.8
build
pip install no such file or directory for classlist txt when installing jep with pip i get the error error no such file or directory src main resources jep classlist txt os platform distribution and version windows python distribution and version python java distribution and version java
1
19,809
3,263,407,907
IssuesEvent
2015-10-22 04:26:56
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
closed
Is the Amber netcdf writer fully compliant with the official spec?
Component-Writers defect Format-Amber Priority-High
I have had a problem reading AMBER .ncdf files into the AMBER analysis tool cpptraj: ``` CPPTRAJ: Trajectory Analysis. V14.22 ___ ___ ___ ___ | \/ | \/ | \/ | _|_/\_|_/\_|_/\_|_ Reading 'penta.top' as Amber Topology > trajin mdanalout.ncdf Reading 'mdanalout.ncdf' as Amber NetCDF NETCDF Error: NetCDF: Attribute not found Error: Getting length for attribute title NETCDF Error: NetCDF: Variable not found Error: Getting spatial VID Error: Could not set up mdanalout.ncdf for reading. Error: Could not set up input trajectory 'mdanalout.ncdf'. ``` At first I thought this was an issue with cpptraj and raised it as an issue there (https://github.com/Amber-MD/cpptraj/issues/123) , but it seems it may be a MDAnalysis issue.
1.0
Is the Amber netcdf writer fully compliant with the official spec? - I have had a problem reading AMBER .ncdf files into the AMBER analysis tool cpptraj: ``` CPPTRAJ: Trajectory Analysis. V14.22 ___ ___ ___ ___ | \/ | \/ | \/ | _|_/\_|_/\_|_/\_|_ Reading 'penta.top' as Amber Topology > trajin mdanalout.ncdf Reading 'mdanalout.ncdf' as Amber NetCDF NETCDF Error: NetCDF: Attribute not found Error: Getting length for attribute title NETCDF Error: NetCDF: Variable not found Error: Getting spatial VID Error: Could not set up mdanalout.ncdf for reading. Error: Could not set up input trajectory 'mdanalout.ncdf'. ``` At first I thought this was an issue with cpptraj and raised it as an issue there (https://github.com/Amber-MD/cpptraj/issues/123) , but it seems it may be a MDAnalysis issue.
non_build
is the amber netcdf writer fully compliant with the official spec i have had a problem reading amber ncdf files into the amber analysis tool cpptraj cpptraj trajectory analysis reading penta top as amber topology trajin mdanalout ncdf reading mdanalout ncdf as amber netcdf netcdf error netcdf attribute not found error getting length for attribute title netcdf error netcdf variable not found error getting spatial vid error could not set up mdanalout ncdf for reading error could not set up input trajectory mdanalout ncdf at first i thought this was an issue with cpptraj and raised it as an issue there but it seems it may be a mdanalysis issue
0
283,930
21,337,808,079
IssuesEvent
2022-04-18 16:35:26
pyrsia/pyrsia
https://api.github.com/repos/pyrsia/pyrsia
opened
Build on "Good PR" for Reviewing Suggestions
documentation
There were a few items noted during the reviewers policy change, https://github.com/pyrsia/.github/pull/6#issuecomment-1098362917 - Try to avoid requesting changes - How to change review (in other words, dismiss your own approval) Other ideas are welcomed!
1.0
Build on "Good PR" for Reviewing Suggestions - There were a few items noted during the reviewers policy change, https://github.com/pyrsia/.github/pull/6#issuecomment-1098362917 - Try to avoid requesting changes - How to change review (in other words, dismiss your own approval) Other ideas are welcomed!
non_build
build on good pr for reviewing suggestions there were a few items noted during the reviewers policy change try to avoid requesting changes how to change review in other words dismiss your own approval other ideas are welcomed
0
669
3,964,688,323
IssuesEvent
2016-05-03 02:43:27
comozilla/Parapara-Canvas-Editor
https://api.github.com/repos/comozilla/Parapara-Canvas-Editor
closed
Viewクラスたちを継承できるようにしたい!
architecture code enhancement wontfix
`ColorPickerView`、`LineWidthPickerView`は、どちらも、 コンストラクタで引数として element、drawingConfig を受け取っていて、 それらをメンバにしているところが共通であるため、 継承したい!(jsで継承とかってできるのかなあ)
1.0
Viewクラスたちを継承できるようにしたい! - `ColorPickerView`、`LineWidthPickerView`は、どちらも、 コンストラクタで引数として element、drawingConfig を受け取っていて、 それらをメンバにしているところが共通であるため、 継承したい!(jsで継承とかってできるのかなあ)
non_build
viewクラスたちを継承できるようにしたい! colorpickerview 、 linewidthpickerview は、どちらも、 コンストラクタで引数として element、drawingconfig を受け取っていて、 それらをメンバにしているところが共通であるため、 継承したい!(jsで継承とかってできるのかなあ)
0
70,986
18,362,452,703
IssuesEvent
2021-10-09 13:07:31
beaverbuilder/feature-requests
https://api.github.com/repos/beaverbuilder/feature-requests
opened
Module slider
Beaver Builder
Imagine a slider in which you could put different kinds of modules. One 'pane' might be an image, the other a paragraph, and yet the other a callout. It would allow to build up a 'story' or 'action' set that gets the user to ultimately do something. There are plenty of other benefits, such as alternating between two modules. Imagine a photo, then details of that photo, then a photo again, and so on...
1.0
Module slider - Imagine a slider in which you could put different kinds of modules. One 'pane' might be an image, the other a paragraph, and yet the other a callout. It would allow to build up a 'story' or 'action' set that gets the user to ultimately do something. There are plenty of other benefits, such as alternating between two modules. Imagine a photo, then details of that photo, then a photo again, and so on...
build
module slider imagine a slider in which you could put different kinds of modules one pane might be an image the other a paragraph and yet the other a callout it would allow to build up a story or action set that gets the user to ultimately do something there are plenty of other benefits such as alternating between two modules imagine a photo then details of that photo then a photo again and so on
1
275,953
30,310,126,927
IssuesEvent
2023-07-10 12:17:17
openraven/magpie
https://api.github.com/repos/openraven/magpie
closed
Assets tracking rule with AWS Tags
aws security rules
People are wanting to ensure tags are on assets. In general security teams are focused on having asset owners and manual data / application tagging. It would be great to have a rule that returns all ARN's that don't have tags. Ideally the rule would allow the user to read in an array of expected tags as well so they can add missing tags as well as the case where there are no tags. This is likely related to #247 the policy rule actions ticket. ┆Issue is synchronized with this [Jira Task](https://openraven.atlassian.net/browse/RAD-248) by [Unito](https://www.unito.io)
True
Assets tracking rule with AWS Tags - People are wanting to ensure tags are on assets. In general security teams are focused on having asset owners and manual data / application tagging. It would be great to have a rule that returns all ARN's that don't have tags. Ideally the rule would allow the user to read in an array of expected tags as well so they can add missing tags as well as the case where there are no tags. This is likely related to #247 the policy rule actions ticket. ┆Issue is synchronized with this [Jira Task](https://openraven.atlassian.net/browse/RAD-248) by [Unito](https://www.unito.io)
non_build
assets tracking rule with aws tags people are wanting to ensure tags are on assets in general security teams are focused on having asset owners and manual data application tagging it would be great to have a rule that returns all arn s that don t have tags ideally the rule would allow the user to read in an array of expected tags as well so they can add missing tags as well as the case where there are no tags this is likely related to the policy rule actions ticket ┆issue is synchronized with this by
0
169,144
20,828,057,698
IssuesEvent
2022-03-19 01:27:20
Seagate/cortx-re
https://api.github.com/repos/Seagate/cortx-re
opened
CVE-2022-24302 (Medium) detected in paramiko-2.7.1-py2.py3-none-any.whl
security vulnerability
## CVE-2022-24302 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-2.7.1-py2.py3-none-any.whl</b></p></summary> <p>SSH2 protocol library</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /docker/cortx-deploy/python_requirements.txt</p> <p>Path to vulnerable library: /docker/cortx-deploy/python_requirements.txt,/scripts/third-party-rpm/python_requirements.txt,/docker/cortx-deploy/cortx-rgw/python_requirements.txt,/docker/cortx-deploy/cortx-control/python_requirements.txt,/docker/cortx-deploy/cortx-data/python_requirements.txt</p> <p> Dependency Hierarchy: - :x: **paramiko-2.7.1-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure. <p>Publish Date: 2022-03-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302>CVE-2022-24302</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.paramiko.org/changelog.html">https://www.paramiko.org/changelog.html</a></p> <p>Release Date: 2022-03-17</p> <p>Fix Resolution: paramiko - 2.10.1</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"paramiko","packageVersion":"2.7.1","packageFilePaths":["/docker/cortx-deploy/python_requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"paramiko:2.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"paramiko - 2.10.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-24302","vulnerabilityDetails":"In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2022-24302 (Medium) detected in paramiko-2.7.1-py2.py3-none-any.whl - ## CVE-2022-24302 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>paramiko-2.7.1-py2.py3-none-any.whl</b></p></summary> <p>SSH2 protocol library</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/06/1e/1e08baaaf6c3d3df1459fd85f0e7d2d6aa916f33958f151ee1ecc9800971/paramiko-2.7.1-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /docker/cortx-deploy/python_requirements.txt</p> <p>Path to vulnerable library: /docker/cortx-deploy/python_requirements.txt,/scripts/third-party-rpm/python_requirements.txt,/docker/cortx-deploy/cortx-rgw/python_requirements.txt,/docker/cortx-deploy/cortx-control/python_requirements.txt,/docker/cortx-deploy/cortx-data/python_requirements.txt</p> <p> Dependency Hierarchy: - :x: **paramiko-2.7.1-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure. <p>Publish Date: 2022-03-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302>CVE-2022-24302</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.paramiko.org/changelog.html">https://www.paramiko.org/changelog.html</a></p> <p>Release Date: 2022-03-17</p> <p>Fix Resolution: paramiko - 2.10.1</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"paramiko","packageVersion":"2.7.1","packageFilePaths":["/docker/cortx-deploy/python_requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"paramiko:2.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"paramiko - 2.10.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-24302","vulnerabilityDetails":"In Paramiko before 2.10.1, a race condition (between creation and chmod) in the write_private_key_file function could allow unauthorized information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24302","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
non_build
cve medium detected in paramiko none any whl cve medium severity vulnerability vulnerable library paramiko none any whl protocol library library home page a href path to dependency file docker cortx deploy python requirements txt path to vulnerable library docker cortx deploy python requirements txt scripts third party rpm python requirements txt docker cortx deploy cortx rgw python requirements txt docker cortx deploy cortx control python requirements txt docker cortx deploy cortx data python requirements txt dependency hierarchy x paramiko none any whl vulnerable library found in base branch main vulnerability details in paramiko before a race condition between creation and chmod in the write private key file function could allow unauthorized information disclosure publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution paramiko rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree paramiko isminimumfixversionavailable true minimumfixversion paramiko isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in paramiko before a race condition between creation and chmod in the write private key file function could allow unauthorized information disclosure vulnerabilityurl
0
42,374
10,964,126,942
IssuesEvent
2019-11-27 21:36:07
pytorch/vision
https://api.github.com/repos/pytorch/vision
closed
torchvision build from source failed
module: io needs reproduction topic: build topic: video
Environment: python3.7 cuda-10.0 pytorch successfully build from source on latest version ffmpeg ``` ffmpeg version 4.1.3-0york1~16.04 Copyright (c) 2000-2019 the FFmpeg developers built with gcc 5.4.0 (Ubuntu 5.4.0-6ubuntu1~16.04.11) 20160609 configuration: --prefix=/usr --extra-version='0york1~16.04' --toolchain=hardened --libdir=/usr/lib/x86_64-linux-gnu --incdir=/usr/include/x86_64-linux-gnu --arch=amd64 --enable-gpl --disable-stripping --enable-avresample --disable-filter=resample --enable-avisynth --enable-gnutls --enable-ladspa --enable-libaom --enable-libass --enable-libbluray --enable-libbs2b --enable-libcaca --enable-libcdio --enable-libcodec2 --enable-libflite --enable-libfontconfig --enable-libfreetype --enable-libfribidi --enable-libgme --enable-libgsm --enable-libjack --enable-libmp3lame --enable-libmysofa --enable-libopenjpeg --enable-libopenmpt --enable-libopus --enable-libpulse --enable-librsvg --enable-librubberband --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libspeex --enable-libssh --enable-libtheora --enable-libtwolame --enable-libvidstab --enable-libvorbis --enable-libvpx --enable-libwavpack --enable-libwebp --enable-libx265 --enable-libxml2 --enable-libxvid --enable-libzmq --enable-libzvbi --enable-lv2 --enable-omx --enable-openal --enable-opengl --enable-sdl2 --enable-nonfree --enable-libfdk-aac --enable-libdc1394 --enable-libdrm --enable-libiec61883 --enable-chromaprint --enable-frei0r --enable-libx264 --enable-shared libavutil 56. 22.100 / 56. 22.100 libavcodec 58. 35.100 / 58. 35.100 libavformat 58. 20.100 / 58. 20.100 libavdevice 58. 5.100 / 58. 5.100 libavfilter 7. 40.101 / 7. 40.101 libavresample 4. 0. 0 / 4. 0. 0 libswscale 5. 3.100 / 5. 3.100 libswresample 3. 3.100 / 3. 3.100 libpostproc 55. 3.100 / 55. 3.100 Hyper fast Audio and Video encoder ``` when run `python setup.py install` in `vision` directory: ``` cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++ /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::openCodecContext()’: /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:31:23: error: ‘AVStream {aka struct AVStream}’ has no member named ‘codecpar’ auto codec_id = st->codecpar->codec_id; ^ /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:48:59: error: ‘AVStream {aka struct AVStream}’ has no member named ‘codecpar’ if ((ret = avcodec_parameters_to_context(codecCtx_, st->codecpar)) < 0) { ^ /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:48:67: error: ‘avcodec_parameters_to_context’ was not declared in this scope if ((ret = avcodec_parameters_to_context(codecCtx_, st->codecpar)) < 0) { ^ /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::sendPacket(const AVPacket*)’: /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:146:47: error: ‘avcodec_send_packet’ was not declared in this scope return avcodec_send_packet(codecCtx_, packet); ^ /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::receiveFrame()’: /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:150:52: error: ‘avcodec_receive_frame’ was not declared in this scope int ret = avcodec_receive_frame(codecCtx_, frame_); ^ error: command 'gcc' failed with exit status 1 ```
1.0
torchvision build from source failed - Environment: python3.7 cuda-10.0 pytorch successfully build from source on latest version ffmpeg ``` ffmpeg version 4.1.3-0york1~16.04 Copyright (c) 2000-2019 the FFmpeg developers built with gcc 5.4.0 (Ubuntu 5.4.0-6ubuntu1~16.04.11) 20160609 configuration: --prefix=/usr --extra-version='0york1~16.04' --toolchain=hardened --libdir=/usr/lib/x86_64-linux-gnu --incdir=/usr/include/x86_64-linux-gnu --arch=amd64 --enable-gpl --disable-stripping --enable-avresample --disable-filter=resample --enable-avisynth --enable-gnutls --enable-ladspa --enable-libaom --enable-libass --enable-libbluray --enable-libbs2b --enable-libcaca --enable-libcdio --enable-libcodec2 --enable-libflite --enable-libfontconfig --enable-libfreetype --enable-libfribidi --enable-libgme --enable-libgsm --enable-libjack --enable-libmp3lame --enable-libmysofa --enable-libopenjpeg --enable-libopenmpt --enable-libopus --enable-libpulse --enable-librsvg --enable-librubberband --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libspeex --enable-libssh --enable-libtheora --enable-libtwolame --enable-libvidstab --enable-libvorbis --enable-libvpx --enable-libwavpack --enable-libwebp --enable-libx265 --enable-libxml2 --enable-libxvid --enable-libzmq --enable-libzvbi --enable-lv2 --enable-omx --enable-openal --enable-opengl --enable-sdl2 --enable-nonfree --enable-libfdk-aac --enable-libdc1394 --enable-libdrm --enable-libiec61883 --enable-chromaprint --enable-frei0r --enable-libx264 --enable-shared libavutil 56. 22.100 / 56. 22.100 libavcodec 58. 35.100 / 58. 35.100 libavformat 58. 20.100 / 58. 20.100 libavdevice 58. 5.100 / 58. 5.100 libavfilter 7. 40.101 / 7. 40.101 libavresample 4. 0. 0 / 4. 0. 0 libswscale 5. 3.100 / 5. 3.100 libswresample 3. 3.100 / 3. 3.100 libpostproc 55. 3.100 / 55. 3.100 Hyper fast Audio and Video encoder ``` when run `python setup.py install` in `vision` directory: ``` cc1plus: warning: command line option ‘-Wstrict-prototypes’ is valid for C/ObjC but not for C++ /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::openCodecContext()’: /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:31:23: error: ‘AVStream {aka struct AVStream}’ has no member named ‘codecpar’ auto codec_id = st->codecpar->codec_id; ^ /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:48:59: error: ‘AVStream {aka struct AVStream}’ has no member named ‘codecpar’ if ((ret = avcodec_parameters_to_context(codecCtx_, st->codecpar)) < 0) { ^ /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:48:67: error: ‘avcodec_parameters_to_context’ was not declared in this scope if ((ret = avcodec_parameters_to_context(codecCtx_, st->codecpar)) < 0) { ^ /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::sendPacket(const AVPacket*)’: /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:146:47: error: ‘avcodec_send_packet’ was not declared in this scope return avcodec_send_packet(codecCtx_, packet); ^ /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp: In member function ‘int FfmpegStream::receiveFrame()’: /home/yifan/temp/vision/torchvision/csrc/cpu/video_reader/FfmpegStream.cpp:150:52: error: ‘avcodec_receive_frame’ was not declared in this scope int ret = avcodec_receive_frame(codecCtx_, frame_); ^ error: command 'gcc' failed with exit status 1 ```
build
torchvision build from source failed environment cuda pytorch successfully build from source on latest version ffmpeg ffmpeg version copyright c the ffmpeg developers built with gcc ubuntu configuration prefix usr extra version toolchain hardened libdir usr lib linux gnu incdir usr include linux gnu arch enable gpl disable stripping enable avresample disable filter resample enable avisynth enable gnutls enable ladspa enable libaom enable libass enable libbluray enable enable libcaca enable libcdio enable enable libflite enable libfontconfig enable libfreetype enable libfribidi enable libgme enable libgsm enable libjack enable enable libmysofa enable libopenjpeg enable libopenmpt enable libopus enable libpulse enable librsvg enable librubberband enable libshine enable libsnappy enable libsoxr enable libspeex enable libssh enable libtheora enable libtwolame enable libvidstab enable libvorbis enable libvpx enable libwavpack enable libwebp enable enable enable libxvid enable libzmq enable libzvbi enable enable omx enable openal enable opengl enable enable nonfree enable libfdk aac enable enable libdrm enable enable chromaprint enable enable enable shared libavutil libavcodec libavformat libavdevice libavfilter libavresample libswscale libswresample libpostproc hyper fast audio and video encoder when run python setup py install in vision directory warning command line option ‘ wstrict prototypes’ is valid for c objc but not for c home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp in member function ‘int ffmpegstream opencodeccontext ’ home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp error ‘avstream aka struct avstream ’ has no member named ‘codecpar’ auto codec id st codecpar codec id home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp error ‘avstream aka struct avstream ’ has no member named ‘codecpar’ if ret avcodec parameters to context codecctx st codecpar home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp error ‘avcodec parameters to context’ was not declared in this scope if ret avcodec parameters to context codecctx st codecpar home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp in member function ‘int ffmpegstream sendpacket const avpacket ’ home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp error ‘avcodec send packet’ was not declared in this scope return avcodec send packet codecctx packet home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp in member function ‘int ffmpegstream receiveframe ’ home yifan temp vision torchvision csrc cpu video reader ffmpegstream cpp error ‘avcodec receive frame’ was not declared in this scope int ret avcodec receive frame codecctx frame error command gcc failed with exit status
1
107,765
23,479,858,358
IssuesEvent
2022-08-17 09:35:56
zond/dipact
https://api.github.com/repos/zond/dipact
closed
Remove Color.js component
code quality
This component is not necessary as there is a color input type built in to browsers.
1.0
Remove Color.js component - This component is not necessary as there is a color input type built in to browsers.
non_build
remove color js component this component is not necessary as there is a color input type built in to browsers
0
637,404
20,627,326,806
IssuesEvent
2022-03-08 00:35:21
NRCan/GSC-Field-Application
https://api.github.com/repos/NRCan/GSC-Field-Application
closed
Location - edit fails on saving
bug low priority
For testing purposes I took some stations with a really bad location, actually pretty much nothing. Then I went into the location form to edit it and clicked saved without modifying anything. Now I can't get out of the location form or even edit it to fix the problem. There is a warning message that pops on screen and prevents user from closing it or going elsewhere. ![image](https://user-images.githubusercontent.com/19910413/148241432-1e02d944-d8f5-4054-93e6-c4fbae6ae3c9.png)
1.0
Location - edit fails on saving - For testing purposes I took some stations with a really bad location, actually pretty much nothing. Then I went into the location form to edit it and clicked saved without modifying anything. Now I can't get out of the location form or even edit it to fix the problem. There is a warning message that pops on screen and prevents user from closing it or going elsewhere. ![image](https://user-images.githubusercontent.com/19910413/148241432-1e02d944-d8f5-4054-93e6-c4fbae6ae3c9.png)
non_build
location edit fails on saving for testing purposes i took some stations with a really bad location actually pretty much nothing then i went into the location form to edit it and clicked saved without modifying anything now i can t get out of the location form or even edit it to fix the problem there is a warning message that pops on screen and prevents user from closing it or going elsewhere
0
89,905
25,920,949,408
IssuesEvent
2022-12-15 21:53:30
brtnfld/cgnsjira
https://api.github.com/repos/brtnfld/cgnsjira
opened
[CGNS-164] Limit on size of writing a single array for parallel library: 2Gb limit and 4Gb windows (no limit in serial library)
bug To Do Major Build
> This issue has been migrated from the Forge. Read the [original ticket here](https://cgnsorg.atlassian.net/browse/CGNS-164). - _**Reporter:**_ None - _**Created at:**_ Thu, 6 Jun 2019 07:35:56 -0500 <p>(1) There are no issues writing large files on either Windows or Linux when serial CGNS used<br/> (2) When parallel CGNS is used, there appears to be a limit on the size of a single array that can be written. For Windows the limit appears to be 4Gb. For Linux the limit appears to be 2Gb. The overall file size can be larger than 4Gb, provided no single array written is >4Gb (windows) or 2Gb (Linux)</p> <p>So, in summary, the issue is with writing large single arrays on both platforms when using the parallel library. This is happening even when using cgp_open with a serial communicator. </p> <p>Since the CFD solution we are writing is written across a number of separate arrays (one for velocity, another the mesh etc etc) we have not hit this restriction yet, but it could happen with very large meshes, so this needs to be fixed. </p> <p>How did I come to this conclusion?</p> <p>I used the attached Large.cpp file to write >4Gb files on windows and Linux. This only uses the serial cg_open call and writes a single large array. This works fine on both platforms.<br/> Our existing Forte simulator is using parallel CGNS 3.3.0 and I used it to write a 20Gb file on windows when running a real engine simulation. However, although the file is 20Gb, no single array is >=4Gb.<br/> To determine the limit for writing single arrays in parallel, I modified my test code to include an option to write multiple arrays (attached). The first arg is the size of each array and the second the number of arrays. Running with 4 MPI processes and args 2 10 - i.e. write 10 arrays each of size 2Gb, leading to a 20Gb file, works on Windows. This code writes serially using cgp_open and a serial communicator and then repeats the same write in parallel. However, as soon as each single array is >=4Gb, the write fails. On Linux, the limit is 2Gb.</p> <p>Windows:<br/> mpiexec -localonly -n 2 CGNSLarge_sections.exe 4 2<br/> Large CGNS file test program, size=4Gb, written in 2 slices<br/> Open in serial<br/> Opening test.cgns...<br/> Now write the data......1......2......done<br/> cgio_write_all_data:H5Dwrite:write to node data failed</p> <p>Linux:<br/> mpirun -n 2 CGNSLarge_sections_new 2 1<br/> Large CGNS file test program, size=2Gb, written in 1 slices <br/> Open in serial <br/> Opening test.cgns...<br/> Now write the data......1......done<br/> File written successfully<br/> File closed successfully<br/> Re-open the file.....done<br/> ...1...268434944 2.68435e+08 <> 0</p>
1.0
[CGNS-164] Limit on size of writing a single array for parallel library: 2Gb limit and 4Gb windows (no limit in serial library) - > This issue has been migrated from the Forge. Read the [original ticket here](https://cgnsorg.atlassian.net/browse/CGNS-164). - _**Reporter:**_ None - _**Created at:**_ Thu, 6 Jun 2019 07:35:56 -0500 <p>(1) There are no issues writing large files on either Windows or Linux when serial CGNS used<br/> (2) When parallel CGNS is used, there appears to be a limit on the size of a single array that can be written. For Windows the limit appears to be 4Gb. For Linux the limit appears to be 2Gb. The overall file size can be larger than 4Gb, provided no single array written is >4Gb (windows) or 2Gb (Linux)</p> <p>So, in summary, the issue is with writing large single arrays on both platforms when using the parallel library. This is happening even when using cgp_open with a serial communicator. </p> <p>Since the CFD solution we are writing is written across a number of separate arrays (one for velocity, another the mesh etc etc) we have not hit this restriction yet, but it could happen with very large meshes, so this needs to be fixed. </p> <p>How did I come to this conclusion?</p> <p>I used the attached Large.cpp file to write >4Gb files on windows and Linux. This only uses the serial cg_open call and writes a single large array. This works fine on both platforms.<br/> Our existing Forte simulator is using parallel CGNS 3.3.0 and I used it to write a 20Gb file on windows when running a real engine simulation. However, although the file is 20Gb, no single array is >=4Gb.<br/> To determine the limit for writing single arrays in parallel, I modified my test code to include an option to write multiple arrays (attached). The first arg is the size of each array and the second the number of arrays. Running with 4 MPI processes and args 2 10 - i.e. write 10 arrays each of size 2Gb, leading to a 20Gb file, works on Windows. This code writes serially using cgp_open and a serial communicator and then repeats the same write in parallel. However, as soon as each single array is >=4Gb, the write fails. On Linux, the limit is 2Gb.</p> <p>Windows:<br/> mpiexec -localonly -n 2 CGNSLarge_sections.exe 4 2<br/> Large CGNS file test program, size=4Gb, written in 2 slices<br/> Open in serial<br/> Opening test.cgns...<br/> Now write the data......1......2......done<br/> cgio_write_all_data:H5Dwrite:write to node data failed</p> <p>Linux:<br/> mpirun -n 2 CGNSLarge_sections_new 2 1<br/> Large CGNS file test program, size=2Gb, written in 1 slices <br/> Open in serial <br/> Opening test.cgns...<br/> Now write the data......1......done<br/> File written successfully<br/> File closed successfully<br/> Re-open the file.....done<br/> ...1...268434944 2.68435e+08 <> 0</p>
build
limit on size of writing a single array for parallel library limit and windows no limit in serial library this issue has been migrated from the forge read the reporter none created at thu jun there are no issues writing large files on either windows or linux when serial cgns used when parallel cgns is used there appears to be a limit on the size of a single array that can be written for windows the limit appears to be for linux the limit appears to be the overall file size can be larger than provided no single array written is windows or linux so in summary the issue is with writing large single arrays on both platforms when using the parallel library this is happening even when using cgp open with a serial communicator since the cfd solution we are writing is written across a number of separate arrays one for velocity another the mesh etc etc we have not hit this restriction yet but it could happen with very large meshes so this needs to be fixed how did i come to this conclusion i used the attached large cpp file to write files on windows and linux this only uses the serial cg open call and writes a single large array this works fine on both platforms our existing forte simulator is using parallel cgns and i used it to write a file on windows when running a real engine simulation however although the file is no single array is to determine the limit for writing single arrays in parallel i modified my test code to include an option to write multiple arrays attached the first arg is the size of each array and the second the number of arrays running with mpi processes and args i e write arrays each of size leading to a file works on windows this code writes serially using cgp open and a serial communicator and then repeats the same write in parallel however as soon as each single array is the write fails on linux the limit is windows mpiexec localonly n cgnslarge sections exe large cgns file test program size written in slices open in serial opening test cgns now write the data done cgio write all data write to node data failed linux mpirun n cgnslarge sections new large cgns file test program size written in slices open in serial opening test cgns now write the data done file written successfully file closed successfully re open the file done
1
8,563
4,273,866,807
IssuesEvent
2016-07-13 18:39:27
haskell/cabal
https://api.github.com/repos/haskell/cabal
closed
new-repl with other ghci
component: nix-local-build
I'd like to use `cabal new-repl --with-ghc=intero`, this used to work with `cabal repl`. But now I get the following error: ``` jacco@xw4600:~/projects/leksah/leksah$ cabal new-repl --with-ghc intero cabal: '/home/jacco/.local/bin/intero' exited with an error: intero: UnhelpfulSpan "on the commandline": cannot use `--interactive' with `--supported-languages' Usage: For basic information, try the `--help' option. ``` This also happens with any other ghci (e.g. `cabal new-repl --with-ghc=ghci`)
1.0
new-repl with other ghci - I'd like to use `cabal new-repl --with-ghc=intero`, this used to work with `cabal repl`. But now I get the following error: ``` jacco@xw4600:~/projects/leksah/leksah$ cabal new-repl --with-ghc intero cabal: '/home/jacco/.local/bin/intero' exited with an error: intero: UnhelpfulSpan "on the commandline": cannot use `--interactive' with `--supported-languages' Usage: For basic information, try the `--help' option. ``` This also happens with any other ghci (e.g. `cabal new-repl --with-ghc=ghci`)
build
new repl with other ghci i d like to use cabal new repl with ghc intero this used to work with cabal repl but now i get the following error jacco projects leksah leksah cabal new repl with ghc intero cabal home jacco local bin intero exited with an error intero unhelpfulspan on the commandline cannot use interactive with supported languages usage for basic information try the help option this also happens with any other ghci e g cabal new repl with ghc ghci
1
76,452
21,427,646,769
IssuesEvent
2022-04-23 00:00:43
PowerShell/PowerShell
https://api.github.com/repos/PowerShell/PowerShell
closed
PowerShell v7.2.1 - Windows Update Loop
Issue-Question Area-Maintainers-Build Resolution-Answered
### Prerequisites - [X] Write a descriptive title. - [X] Make sure you are able to repro it on the [latest released version](https://github.com/PowerShell/PowerShell/releases) - [X] Search the existing issues. - [X] Refer to the [FAQ](https://github.com/PowerShell/PowerShell/blob/master/docs/FAQ.md). - [X] Refer to [Differences between Windows PowerShell 5.1 and PowerShell](https://docs.microsoft.com/powershell/scripting/whats-new/differences-from-windows-powershell). ### Steps to reproduce Run Windows Update with PowerShell installed and check for updates. ### Expected behavior ```console PowerShell v7.2.1 installs and completes. ``` ### Actual behavior ```console Same version of PowerShell is installed and again and again. ``` ### Environment data ```powershell Name Value ---- ----- PSVersion 7.2.1 PSEdition Core GitCommitId 7.2.1 OS Microsoft Windows 10.0.19044 Platform Win32NT PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…} PSRemotingProtocolVersion 2.3 SerializationVersion 1.1.0.1 WSManStackVersion 3.0 ``` ### OS Data ``` Version Caption ------- ------- 10.0.19044 Microsoft Windows 10 Pro ``` ### Windows update log Please see [this gist](https://gist.githubusercontent.com/hl2guide/dffed20e92f4346395eb2ae0cc7317cf/raw/3ff41f46d544fc72b4507bfe55819a546c726018/gistfile1.txt). ### Visuals _No response_
1.0
PowerShell v7.2.1 - Windows Update Loop - ### Prerequisites - [X] Write a descriptive title. - [X] Make sure you are able to repro it on the [latest released version](https://github.com/PowerShell/PowerShell/releases) - [X] Search the existing issues. - [X] Refer to the [FAQ](https://github.com/PowerShell/PowerShell/blob/master/docs/FAQ.md). - [X] Refer to [Differences between Windows PowerShell 5.1 and PowerShell](https://docs.microsoft.com/powershell/scripting/whats-new/differences-from-windows-powershell). ### Steps to reproduce Run Windows Update with PowerShell installed and check for updates. ### Expected behavior ```console PowerShell v7.2.1 installs and completes. ``` ### Actual behavior ```console Same version of PowerShell is installed and again and again. ``` ### Environment data ```powershell Name Value ---- ----- PSVersion 7.2.1 PSEdition Core GitCommitId 7.2.1 OS Microsoft Windows 10.0.19044 Platform Win32NT PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…} PSRemotingProtocolVersion 2.3 SerializationVersion 1.1.0.1 WSManStackVersion 3.0 ``` ### OS Data ``` Version Caption ------- ------- 10.0.19044 Microsoft Windows 10 Pro ``` ### Windows update log Please see [this gist](https://gist.githubusercontent.com/hl2guide/dffed20e92f4346395eb2ae0cc7317cf/raw/3ff41f46d544fc72b4507bfe55819a546c726018/gistfile1.txt). ### Visuals _No response_
build
powershell windows update loop prerequisites write a descriptive title make sure you are able to repro it on the search the existing issues refer to the refer to steps to reproduce run windows update with powershell installed and check for updates expected behavior console powershell installs and completes actual behavior console same version of powershell is installed and again and again environment data powershell name value psversion psedition core gitcommitid os microsoft windows platform pscompatibleversions … psremotingprotocolversion serializationversion wsmanstackversion os data version caption microsoft windows pro windows update log please see visuals no response
1
424,784
12,323,250,532
IssuesEvent
2020-05-13 11:50:24
Azure/ARO-RP
https://api.github.com/repos/Azure/ARO-RP
opened
APIServer status to metrics
priority-medium size-medium
APIserver status has very good insights on install failures and modes: ``` oc get kubeapiservers cluster -o yaml ... - lastTransitionTime: "2020-05-13T11:40:42Z" message: 3 nodes are active; 2 nodes are at revision 2; 1 nodes are at revision 3; 0 nodes have achieved new revision 4 status: "True" type: StaticPodsAvailable - lastTransitionTime: "2020-05-13T11:38:11Z" message: 2 nodes are at revision 2; 1 nodes are at revision 3; 0 nodes have achieved new revision 4 status: "True" type: NodeInstallerProgressing - lastTransitionTime: "2020-05-13T11:47:09Z" message: |- nodes/mjudeikis3-8bcbz-master-1 pods/kube-apiserver-mjudeikis3-8bcbz-master-1 container="kube-apiserver" is not ready nodes/mjudeikis3-8bcbz-master-1 pods/kube-apiserver-mjudeikis3-8bcbz-master-1 container="kube-apiserver" is waiting: "CrashLoopBackOff" - "back-off 20s restarting failed container=kube-apiserver pod=kube-apiserver-mjudeikis3-8bcbz-master-1_openshift-kube-apiserver(458b630d98792d8c21612b955384b7e7)" reason: Error status: "True" type: StaticPodsDegraded ``` We might want to consider those as metrics/logs for failed cluster install debugging when we don't have access to API server to use geneva.
1.0
APIServer status to metrics - APIserver status has very good insights on install failures and modes: ``` oc get kubeapiservers cluster -o yaml ... - lastTransitionTime: "2020-05-13T11:40:42Z" message: 3 nodes are active; 2 nodes are at revision 2; 1 nodes are at revision 3; 0 nodes have achieved new revision 4 status: "True" type: StaticPodsAvailable - lastTransitionTime: "2020-05-13T11:38:11Z" message: 2 nodes are at revision 2; 1 nodes are at revision 3; 0 nodes have achieved new revision 4 status: "True" type: NodeInstallerProgressing - lastTransitionTime: "2020-05-13T11:47:09Z" message: |- nodes/mjudeikis3-8bcbz-master-1 pods/kube-apiserver-mjudeikis3-8bcbz-master-1 container="kube-apiserver" is not ready nodes/mjudeikis3-8bcbz-master-1 pods/kube-apiserver-mjudeikis3-8bcbz-master-1 container="kube-apiserver" is waiting: "CrashLoopBackOff" - "back-off 20s restarting failed container=kube-apiserver pod=kube-apiserver-mjudeikis3-8bcbz-master-1_openshift-kube-apiserver(458b630d98792d8c21612b955384b7e7)" reason: Error status: "True" type: StaticPodsDegraded ``` We might want to consider those as metrics/logs for failed cluster install debugging when we don't have access to API server to use geneva.
non_build
apiserver status to metrics apiserver status has very good insights on install failures and modes oc get kubeapiservers cluster o yaml lasttransitiontime message nodes are active nodes are at revision nodes are at revision nodes have achieved new revision status true type staticpodsavailable lasttransitiontime message nodes are at revision nodes are at revision nodes have achieved new revision status true type nodeinstallerprogressing lasttransitiontime message nodes master pods kube apiserver master container kube apiserver is not ready nodes master pods kube apiserver master container kube apiserver is waiting crashloopbackoff back off restarting failed container kube apiserver pod kube apiserver master openshift kube apiserver reason error status true type staticpodsdegraded we might want to consider those as metrics logs for failed cluster install debugging when we don t have access to api server to use geneva
0
653,135
21,572,876,372
IssuesEvent
2022-05-02 10:23:22
clarin-eric/standards
https://api.github.com/repos/clarin-eric/standards
closed
add a schema for domain data
priority SIS:schemas
We need a schema to constrain the domain data (including domain groupings). It's not a critical addition, but ...
1.0
add a schema for domain data - We need a schema to constrain the domain data (including domain groupings). It's not a critical addition, but ...
non_build
add a schema for domain data we need a schema to constrain the domain data including domain groupings it s not a critical addition but
0
17,162
22,739,148,923
IssuesEvent
2022-07-07 00:44:42
MPMG-DCC-UFMG/C01
https://api.github.com/repos/MPMG-DCC-UFMG/C01
closed
Erro código 1006 ao coletar algumas páginas
[1] Bug [2] Baixa Prioridade [0] Desenvolvimento [3] Processamento Dinâmico
# Comportamento Esperado Realizar as coletas sem erro algum. # Comportamento Atual As coletas apresentam o erro code = 1006. Na maioria das vezes este erro não impede o funcionamento do coletor. Segue log exemplo para este problema: `2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - event = eof_received() 2021-08-03 19:55:04 [websockets.protocol] DEBUG: client ! failing CLOSING WebSocket connection with code 1006 2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - event = connection_lost(None) 2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - state = CLOSED 2021-08-03 19:55:04 [websockets.protocol] DEBUG: client x code = 1006, reason = [no reason] 2021-08-03 19:55:04 [scrapy.statscollectors] INFO: Dumping Scrapy stats: {'elapsed_time_seconds': 0.0047, 'finish_reason': 'finished', 'finish_time': datetime.datetime(2021, 8, 3, 19, 54, 57, 215496), 'log_count/DEBUG': 116, 'log_count/ERROR': 1, 'log_count/INFO': 10, 'log_count/WARNING': 3, 'memusage/max': 302215168, 'memusage/startup': 302215168, 'start_time': datetime.datetime(2021, 8, 3, 19, 54, 57, 210796)} 2021-08-03 19:55:04 [scrapy.core.engine] INFO: Spider closed (finished)` # Passos para reproduzir o erro Para recriar o problema, basta criar um coletor para a fonte "http://www.congonhas.mg.gov.br/index.php/licitacao-publica-prefeitura/", com um passo dinâmico de espera e um de salvar página. Pode ter alguma relação com o seguinte problema: https://github.com/miyakogi/pyppeteer/issues/62
1.0
Erro código 1006 ao coletar algumas páginas - # Comportamento Esperado Realizar as coletas sem erro algum. # Comportamento Atual As coletas apresentam o erro code = 1006. Na maioria das vezes este erro não impede o funcionamento do coletor. Segue log exemplo para este problema: `2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - event = eof_received() 2021-08-03 19:55:04 [websockets.protocol] DEBUG: client ! failing CLOSING WebSocket connection with code 1006 2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - event = connection_lost(None) 2021-08-03 19:55:04 [websockets.protocol] DEBUG: client - state = CLOSED 2021-08-03 19:55:04 [websockets.protocol] DEBUG: client x code = 1006, reason = [no reason] 2021-08-03 19:55:04 [scrapy.statscollectors] INFO: Dumping Scrapy stats: {'elapsed_time_seconds': 0.0047, 'finish_reason': 'finished', 'finish_time': datetime.datetime(2021, 8, 3, 19, 54, 57, 215496), 'log_count/DEBUG': 116, 'log_count/ERROR': 1, 'log_count/INFO': 10, 'log_count/WARNING': 3, 'memusage/max': 302215168, 'memusage/startup': 302215168, 'start_time': datetime.datetime(2021, 8, 3, 19, 54, 57, 210796)} 2021-08-03 19:55:04 [scrapy.core.engine] INFO: Spider closed (finished)` # Passos para reproduzir o erro Para recriar o problema, basta criar um coletor para a fonte "http://www.congonhas.mg.gov.br/index.php/licitacao-publica-prefeitura/", com um passo dinâmico de espera e um de salvar página. Pode ter alguma relação com o seguinte problema: https://github.com/miyakogi/pyppeteer/issues/62
non_build
erro código ao coletar algumas páginas comportamento esperado realizar as coletas sem erro algum comportamento atual as coletas apresentam o erro code na maioria das vezes este erro não impede o funcionamento do coletor segue log exemplo para este problema debug client event eof received debug client failing closing websocket connection with code debug client event connection lost none debug client state closed debug client x code reason info dumping scrapy stats elapsed time seconds finish reason finished finish time datetime datetime log count debug log count error log count info log count warning memusage max memusage startup start time datetime datetime info spider closed finished passos para reproduzir o erro para recriar o problema basta criar um coletor para a fonte com um passo dinâmico de espera e um de salvar página pode ter alguma relação com o seguinte problema
0
90,394
26,080,597,426
IssuesEvent
2022-12-25 09:23:59
OpenShot/openshot-qt
https://api.github.com/repos/OpenShot/openshot-qt
closed
Audio is not synced after exporting video
media-handling 📅 Daily Build
**Describe the bug:** I created a video with the latest daily build. The preview of the video is working just fine, that is, the video and the audio are in sync. Exporting the video with: * Target: MP4 (h.264) * Video Profile (HD 720p 30 fps (1280x720) * Quality (High) Result: the audio and video are out of sync Tried the same with 60 fps and then it is even worse. The audio starts later than the video. **Expected behavior:** Exported video and audio are in sync. **System Details:** - OpenShot Version [e.g. 2.4.3]: v2.6.1-dev-daily-9975-b890417a-46255e46, build date: 2022-10-20 20:23 - Operating System / Distro: Ubuntu 22.04.1 LTS **Log Files:** If you are experiencing a crash, please collect and attach logs of the problem. Log files can be found in the `.openshot_qt` folder in your user home directory. Log files over 2 MB in size will need to be compressed, please attach a ZIP file instead of the raw logs. **Exception / Stacktrace:** No crash or exception
1.0
Audio is not synced after exporting video - **Describe the bug:** I created a video with the latest daily build. The preview of the video is working just fine, that is, the video and the audio are in sync. Exporting the video with: * Target: MP4 (h.264) * Video Profile (HD 720p 30 fps (1280x720) * Quality (High) Result: the audio and video are out of sync Tried the same with 60 fps and then it is even worse. The audio starts later than the video. **Expected behavior:** Exported video and audio are in sync. **System Details:** - OpenShot Version [e.g. 2.4.3]: v2.6.1-dev-daily-9975-b890417a-46255e46, build date: 2022-10-20 20:23 - Operating System / Distro: Ubuntu 22.04.1 LTS **Log Files:** If you are experiencing a crash, please collect and attach logs of the problem. Log files can be found in the `.openshot_qt` folder in your user home directory. Log files over 2 MB in size will need to be compressed, please attach a ZIP file instead of the raw logs. **Exception / Stacktrace:** No crash or exception
build
audio is not synced after exporting video describe the bug i created a video with the latest daily build the preview of the video is working just fine that is the video and the audio are in sync exporting the video with target h video profile hd fps quality high result the audio and video are out of sync tried the same with fps and then it is even worse the audio starts later than the video expected behavior exported video and audio are in sync system details openshot version dev daily build date operating system distro ubuntu lts log files if you are experiencing a crash please collect and attach logs of the problem log files can be found in the openshot qt folder in your user home directory log files over mb in size will need to be compressed please attach a zip file instead of the raw logs exception stacktrace no crash or exception
1
692,461
23,735,440,669
IssuesEvent
2022-08-31 07:41:12
magento/magento2
https://api.github.com/repos/magento/magento2
closed
[Issue] Fixed the welcome message contains single quote character
Issue: Confirmed Component: Theme Reproduced on 2.4.x Progress: PR in progress Priority: P1 Area: Account
This issue is automatically created based on existing pull request: magento/magento2#34061: Fixed the welcome message contains single quote character --------- <!--- Thank you for contributing to Magento. To help us process this pull request we recommend that you add the following information: - Summary of the pull request, - Issue(s) related to the changes made, - Manual testing scenarios Fields marked with (*) are required. Please don't remove the template. --> <!--- Please provide a general summary of the Pull Request in the Title above --> ### Description (*) This PR is fixed for the issue, that if we put the single quote character to the header welcome message config, it will lead the frontend to have a bunch of console errors: > Uncaught SyntaxError: Unable to process binding "ifnot: function(){return customer().fullname }" Message: Unable to parse bindings. ### Manual testing scenarios (*) <!--- Please provide a set of unambiguous steps to test the proposed code change. Giving us manual testing scenarios will help with the processing and validation process. --> 1. From Admin, navigate to Content -> Design -> Configuration 2. Select a theme and Edit 3. In the edit page, expand the Header setting, put the `Welcome Text`, like "Welcome Khoi's store!" 4. Clear cache, then go to frontend, you'll see a JS error, and some JS functions not working because of this issue. ### Questions or comments Maybe we need to bind the message as a text, instead of HTML binding, like: > data-bind="text: '<?= $escaper->escapeHtmlAttr($welcomeMessage) ?>'" ### Contribution checklist (*) - [ ] Pull request has a meaningful description of its purpose - [ ] All commits are accompanied by meaningful commit messages - [ ] All new or changed code is covered with unit/integration tests (if applicable) - [ ] README.md files for modified modules are updated and included in the pull request if any [README.md predefined sections](https://github.com/magento/devdocs/wiki/Magento-module-README.md) require an update - [ ] All automated tests passed successfully (all builds are green)
1.0
[Issue] Fixed the welcome message contains single quote character - This issue is automatically created based on existing pull request: magento/magento2#34061: Fixed the welcome message contains single quote character --------- <!--- Thank you for contributing to Magento. To help us process this pull request we recommend that you add the following information: - Summary of the pull request, - Issue(s) related to the changes made, - Manual testing scenarios Fields marked with (*) are required. Please don't remove the template. --> <!--- Please provide a general summary of the Pull Request in the Title above --> ### Description (*) This PR is fixed for the issue, that if we put the single quote character to the header welcome message config, it will lead the frontend to have a bunch of console errors: > Uncaught SyntaxError: Unable to process binding "ifnot: function(){return customer().fullname }" Message: Unable to parse bindings. ### Manual testing scenarios (*) <!--- Please provide a set of unambiguous steps to test the proposed code change. Giving us manual testing scenarios will help with the processing and validation process. --> 1. From Admin, navigate to Content -> Design -> Configuration 2. Select a theme and Edit 3. In the edit page, expand the Header setting, put the `Welcome Text`, like "Welcome Khoi's store!" 4. Clear cache, then go to frontend, you'll see a JS error, and some JS functions not working because of this issue. ### Questions or comments Maybe we need to bind the message as a text, instead of HTML binding, like: > data-bind="text: '<?= $escaper->escapeHtmlAttr($welcomeMessage) ?>'" ### Contribution checklist (*) - [ ] Pull request has a meaningful description of its purpose - [ ] All commits are accompanied by meaningful commit messages - [ ] All new or changed code is covered with unit/integration tests (if applicable) - [ ] README.md files for modified modules are updated and included in the pull request if any [README.md predefined sections](https://github.com/magento/devdocs/wiki/Magento-module-README.md) require an update - [ ] All automated tests passed successfully (all builds are green)
non_build
fixed the welcome message contains single quote character this issue is automatically created based on existing pull request magento fixed the welcome message contains single quote character thank you for contributing to magento to help us process this pull request we recommend that you add the following information summary of the pull request issue s related to the changes made manual testing scenarios fields marked with are required please don t remove the template description this pr is fixed for the issue that if we put the single quote character to the header welcome message config it will lead the frontend to have a bunch of console errors uncaught syntaxerror unable to process binding ifnot function return customer fullname message unable to parse bindings manual testing scenarios please provide a set of unambiguous steps to test the proposed code change giving us manual testing scenarios will help with the processing and validation process from admin navigate to content design configuration select a theme and edit in the edit page expand the header setting put the welcome text like welcome khoi s store clear cache then go to frontend you ll see a js error and some js functions not working because of this issue questions or comments maybe we need to bind the message as a text instead of html binding like data bind text escapehtmlattr welcomemessage contribution checklist pull request has a meaningful description of its purpose all commits are accompanied by meaningful commit messages all new or changed code is covered with unit integration tests if applicable readme md files for modified modules are updated and included in the pull request if any require an update all automated tests passed successfully all builds are green
0
66,519
16,627,201,458
IssuesEvent
2021-06-03 11:08:22
zalando/restful-api-guidelines
https://api.github.com/repos/zalando/restful-api-guidelines
closed
Change build process to not need a personal access token
build
Currently, the build process to create the web side from the repository is [running on Travis CI](https://travis-ci.org/zalando/restful-api-guidelines), and publishes the result by pushing to the gh_pages branch. For that, it's using a `GH_TOKEN` variable configured in the [project settings](https://travis-ci.org/zalando/restful-api-guidelines/settings), which is a [personal access token](https://help.github.com/en/github/authenticating-to-github/creating-a-personal-access-token-for-the-command-line). This token was previously created by @maxim-tschumak, and after he left the organization, and thus lost the write rights, a new one was created today by me (@ePaul). This token gives read and write rights to all repositories I have access to, in my name. Unfortunately there seems to be no way of creating a token which has access just to a single repository. This induces those risks: * If I ever leave the organization, my token automatically loses write access to the repository, and thus breaks the publishing of new versions. * If something goes wrong and some attacker gets hold of the token, that person can impersonate me on all repositories – both my personal ones, and ones in organizations where I have access. We should think about setting up the build process in a way that a personal token is not needed. A possible solution might be using a repository-specific [deployment key](https://developer.github.com/v3/guides/managing-deploy-keys/#deploy-keys).
1.0
Change build process to not need a personal access token - Currently, the build process to create the web side from the repository is [running on Travis CI](https://travis-ci.org/zalando/restful-api-guidelines), and publishes the result by pushing to the gh_pages branch. For that, it's using a `GH_TOKEN` variable configured in the [project settings](https://travis-ci.org/zalando/restful-api-guidelines/settings), which is a [personal access token](https://help.github.com/en/github/authenticating-to-github/creating-a-personal-access-token-for-the-command-line). This token was previously created by @maxim-tschumak, and after he left the organization, and thus lost the write rights, a new one was created today by me (@ePaul). This token gives read and write rights to all repositories I have access to, in my name. Unfortunately there seems to be no way of creating a token which has access just to a single repository. This induces those risks: * If I ever leave the organization, my token automatically loses write access to the repository, and thus breaks the publishing of new versions. * If something goes wrong and some attacker gets hold of the token, that person can impersonate me on all repositories – both my personal ones, and ones in organizations where I have access. We should think about setting up the build process in a way that a personal token is not needed. A possible solution might be using a repository-specific [deployment key](https://developer.github.com/v3/guides/managing-deploy-keys/#deploy-keys).
build
change build process to not need a personal access token currently the build process to create the web side from the repository is and publishes the result by pushing to the gh pages branch for that it s using a gh token variable configured in the which is a this token was previously created by maxim tschumak and after he left the organization and thus lost the write rights a new one was created today by me epaul this token gives read and write rights to all repositories i have access to in my name unfortunately there seems to be no way of creating a token which has access just to a single repository this induces those risks if i ever leave the organization my token automatically loses write access to the repository and thus breaks the publishing of new versions if something goes wrong and some attacker gets hold of the token that person can impersonate me on all repositories – both my personal ones and ones in organizations where i have access we should think about setting up the build process in a way that a personal token is not needed a possible solution might be using a repository specific
1
89,741
25,894,424,002
IssuesEvent
2022-12-14 20:57:48
elastic/beats
https://api.github.com/repos/elastic/beats
closed
Build 470 for 7.17 with status FAILURE
automation ci-reported Team:Elastic-Agent-Data-Plane build-failures
## :broken_heart: Tests Failed <!-- BUILD BADGES--> > _the below badges are clickable and redirect to their specific view in the CI or DOCS_ [![Pipeline View](https://img.shields.io/badge/pipeline-pipeline%20-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//pipeline) [![Test View](https://img.shields.io/badge/test-test-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//tests) [![Changes](https://img.shields.io/badge/changes-changes-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//changes) [![Artifacts](https://img.shields.io/badge/artifacts-artifacts-yellow)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//artifacts) [![preview](https://img.shields.io/badge/docs-preview-yellowgreen)](http://beats_null.docs-preview.app.elstc.co/diff) [![preview](https://img.shields.io/badge/elastic-observability-blue)](https://ci-stats.elastic.co/app/apm/services/beats-ci/transactions/view?rangeFrom=2022-08-25T12:53:35.690Z&rangeTo=2022-08-25T13:13:35.690Z&transactionName=BUILD+Beats%2Fbeats%2F7.17&transactionType=job&latencyAggregationType=avg&traceId=5f1bada60f5f914aa4f76d955c272a38&transactionId=80fe1f1eb38b0665) <!-- BUILD SUMMARY--> <details><summary>Expand to view the summary</summary> <p> #### Build stats * Start Time: 2022-08-25T13:03:35.690+0000 * Duration: 113 min 7 sec #### Test stats :test_tube: | Test | Results | | ------------ | :-----------------------------: | | Failed | 2 | | Passed | 25709 | | Skipped | 1884 | | Total | 27595 | </p> </details> <!-- TEST RESULTS IF ANY--> ### Test errors [![2](https://img.shields.io/badge/2%20-red)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//tests) <details><summary>Expand to view the tests failures</summary><p> ##### `Build&Test / metricbeat-goIntegTest / TestFetch – github.com/elastic/beats/v7/metricbeat/module/jolokia/jmx` <ul> <details><summary>Expand to view the error details</summary><p> ``` Failed ``` </p></details> <details><summary>Expand to view the stacktrace</summary><p> ``` === RUN TestFetch Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:33: failed to start service "jolokia: exit status 1 Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:33: failed to start service "jolokia: exit status 1 Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:33: failed to start service "jolokia: exit status 1 jmx_integration_test.go:33: getting host for jolokia: no container running for service --- FAIL: TestFetch (6.64s) ``` </p></details> </ul> ##### `Build&Test / metricbeat-goIntegTest / TestData – github.com/elastic/beats/v7/metricbeat/module/jolokia/jmx` <ul> <details><summary>Expand to view the error details</summary><p> ``` Failed ``` </p></details> <details><summary>Expand to view the stacktrace</summary><p> ``` === RUN TestData Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:45: failed to start service "jolokia: exit status 1 Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:45: failed to start service "jolokia: exit status 1 Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:45: failed to start service "jolokia: exit status 1 jmx_integration_test.go:45: getting host for jolokia: no container running for service --- FAIL: TestData (6.54s) ``` </p></details> </ul> </p></details> <!-- STEPS ERRORS IF ANY --> ### Steps errors [![9](https://img.shields.io/badge/9%20-red)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//pipeline) <details><summary>Expand to view the steps failures</summary> <p> ##### `filebeat-unitTest - mage build unitTest` <ul> <li>Took 6 min 51 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14102/log/?start=0">here</a></li> <li>Description: <code>mage build unitTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 28 min 46 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15344/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 19 min 50 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/21860/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 19 min 55 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/22367/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-pythonIntegTest - mage pythonIntegTest` <ul> <li>Took 2 min 34 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14743/log/?start=0">here</a></li> <li>Description: <code>mage pythonIntegTest</code></l1> </ul> ##### `metricbeat-pythonIntegTest - mage pythonIntegTest` <ul> <li>Took 0 min 23 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15511/log/?start=0">here</a></li> <li>Description: <code>mage pythonIntegTest</code></l1> </ul> ##### `metricbeat-pythonIntegTest - mage pythonIntegTest` <ul> <li>Took 0 min 22 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15716/log/?start=0">here</a></li> <li>Description: <code>mage pythonIntegTest</code></l1> </ul> ##### `x-pack/filebeat-arm-arm - mage build unitTest` <ul> <li>Took 8 min 2 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14923/log/?start=0">here</a></li> <li>Description: <code>mage build unitTest</code></l1> </ul> ##### `Error signal` <ul> <li>Took 0 min 0 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/22383/log/?start=0">here</a></li> <li>Description: <code>Error "hudson.AbortException: script returned exit code 1"</code></l1> </ul> </p> </details>
1.0
Build 470 for 7.17 with status FAILURE - ## :broken_heart: Tests Failed <!-- BUILD BADGES--> > _the below badges are clickable and redirect to their specific view in the CI or DOCS_ [![Pipeline View](https://img.shields.io/badge/pipeline-pipeline%20-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//pipeline) [![Test View](https://img.shields.io/badge/test-test-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//tests) [![Changes](https://img.shields.io/badge/changes-changes-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//changes) [![Artifacts](https://img.shields.io/badge/artifacts-artifacts-yellow)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//artifacts) [![preview](https://img.shields.io/badge/docs-preview-yellowgreen)](http://beats_null.docs-preview.app.elstc.co/diff) [![preview](https://img.shields.io/badge/elastic-observability-blue)](https://ci-stats.elastic.co/app/apm/services/beats-ci/transactions/view?rangeFrom=2022-08-25T12:53:35.690Z&rangeTo=2022-08-25T13:13:35.690Z&transactionName=BUILD+Beats%2Fbeats%2F7.17&transactionType=job&latencyAggregationType=avg&traceId=5f1bada60f5f914aa4f76d955c272a38&transactionId=80fe1f1eb38b0665) <!-- BUILD SUMMARY--> <details><summary>Expand to view the summary</summary> <p> #### Build stats * Start Time: 2022-08-25T13:03:35.690+0000 * Duration: 113 min 7 sec #### Test stats :test_tube: | Test | Results | | ------------ | :-----------------------------: | | Failed | 2 | | Passed | 25709 | | Skipped | 1884 | | Total | 27595 | </p> </details> <!-- TEST RESULTS IF ANY--> ### Test errors [![2](https://img.shields.io/badge/2%20-red)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//tests) <details><summary>Expand to view the tests failures</summary><p> ##### `Build&Test / metricbeat-goIntegTest / TestFetch – github.com/elastic/beats/v7/metricbeat/module/jolokia/jmx` <ul> <details><summary>Expand to view the error details</summary><p> ``` Failed ``` </p></details> <details><summary>Expand to view the stacktrace</summary><p> ``` === RUN TestFetch Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:33: failed to start service "jolokia: exit status 1 Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:33: failed to start service "jolokia: exit status 1 Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:33: failed to start service "jolokia: exit status 1 jmx_integration_test.go:33: getting host for jolokia: no container running for service --- FAIL: TestFetch (6.64s) ``` </p></details> </ul> ##### `Build&Test / metricbeat-goIntegTest / TestData – github.com/elastic/beats/v7/metricbeat/module/jolokia/jmx` <ul> <details><summary>Expand to view the error details</summary><p> ``` Failed ``` </p></details> <details><summary>Expand to view the stacktrace</summary><p> ``` === RUN TestData Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:45: failed to start service "jolokia: exit status 1 Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:45: failed to start service "jolokia: exit status 1 Found orphan containers (metricbeat_7_17_7_ecdd50eff4-snapshot_http_1, metricbeat_7_17_7_ecdd50eff4-snapshot_haproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_golang_1, metricbeat_7_17_7_ecdd50eff4-snapshot_etcd_1, metricbeat_7_17_7_ecdd50eff4-snapshot_envoyproxy_1, metricbeat_7_17_7_ecdd50eff4-snapshot_dropwizard_1, metricbeat_7_17_7_ecdd50eff4-snapshot_couchdb_1, metricbeat_7_17_7_ecdd50eff4-snapshot_logstash_1, metricbeat_7_17_7_ecdd50eff4-snapshot_kafka_1) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up. Building jolokia Step 1/14 : FROM java:8-jdk-alpine Service "jolokia" failed to build: manifest for java:8-jdk-alpine not found: manifest unknown: manifest unknown jmx_integration_test.go:45: failed to start service "jolokia: exit status 1 jmx_integration_test.go:45: getting host for jolokia: no container running for service --- FAIL: TestData (6.54s) ``` </p></details> </ul> </p></details> <!-- STEPS ERRORS IF ANY --> ### Steps errors [![9](https://img.shields.io/badge/9%20-red)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F7.17/detail/7.17/470//pipeline) <details><summary>Expand to view the steps failures</summary> <p> ##### `filebeat-unitTest - mage build unitTest` <ul> <li>Took 6 min 51 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14102/log/?start=0">here</a></li> <li>Description: <code>mage build unitTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 28 min 46 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15344/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 19 min 50 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/21860/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 19 min 55 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/22367/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-pythonIntegTest - mage pythonIntegTest` <ul> <li>Took 2 min 34 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14743/log/?start=0">here</a></li> <li>Description: <code>mage pythonIntegTest</code></l1> </ul> ##### `metricbeat-pythonIntegTest - mage pythonIntegTest` <ul> <li>Took 0 min 23 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15511/log/?start=0">here</a></li> <li>Description: <code>mage pythonIntegTest</code></l1> </ul> ##### `metricbeat-pythonIntegTest - mage pythonIntegTest` <ul> <li>Took 0 min 22 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/15716/log/?start=0">here</a></li> <li>Description: <code>mage pythonIntegTest</code></l1> </ul> ##### `x-pack/filebeat-arm-arm - mage build unitTest` <ul> <li>Took 8 min 2 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/14923/log/?start=0">here</a></li> <li>Description: <code>mage build unitTest</code></l1> </ul> ##### `Error signal` <ul> <li>Took 0 min 0 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/7.17/runs/470/steps/22383/log/?start=0">here</a></li> <li>Description: <code>Error "hudson.AbortException: script returned exit code 1"</code></l1> </ul> </p> </details>
build
build for with status failure broken heart tests failed the below badges are clickable and redirect to their specific view in the ci or docs expand to view the summary build stats start time duration min sec test stats test tube test results failed passed skipped total test errors expand to view the tests failures build test metricbeat gointegtest testfetch – github com elastic beats metricbeat module jolokia jmx expand to view the error details failed expand to view the stacktrace run testfetch found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status jmx integration test go getting host for jolokia no container running for service fail testfetch build test metricbeat gointegtest testdata – github com elastic beats metricbeat module jolokia jmx expand to view the error details failed expand to view the stacktrace run testdata found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status found orphan containers metricbeat snapshot http metricbeat snapshot haproxy metricbeat snapshot golang metricbeat snapshot etcd metricbeat snapshot envoyproxy metricbeat snapshot dropwizard metricbeat snapshot couchdb metricbeat snapshot logstash metricbeat snapshot kafka for this project if you removed or renamed this service in your compose file you can run this command with the remove orphans flag to clean it up building jolokia step from java jdk alpine service jolokia failed to build manifest for java jdk alpine not found manifest unknown manifest unknown jmx integration test go failed to start service jolokia exit status jmx integration test go getting host for jolokia no container running for service fail testdata steps errors expand to view the steps failures filebeat unittest mage build unittest took min sec view more details a href description mage build unittest metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest metricbeat pythonintegtest mage pythonintegtest took min sec view more details a href description mage pythonintegtest metricbeat pythonintegtest mage pythonintegtest took min sec view more details a href description mage pythonintegtest metricbeat pythonintegtest mage pythonintegtest took min sec view more details a href description mage pythonintegtest x pack filebeat arm arm mage build unittest took min sec view more details a href description mage build unittest error signal took min sec view more details a href description error hudson abortexception script returned exit code
1
72,707
31,769,020,249
IssuesEvent
2023-09-12 10:32:16
gauravrs18/issue_onboarding
https://api.github.com/repos/gauravrs18/issue_onboarding
closed
dev-angular-code-account-services-new-connection-component-approve-component -consumer-details-component -connect-component -reject-button-component
CX-account-services
dev-angular-code-account-services-new-connection-component-approve-component -consumer-details-component -connect-component -reject-button-component
1.0
dev-angular-code-account-services-new-connection-component-approve-component -consumer-details-component -connect-component -reject-button-component - dev-angular-code-account-services-new-connection-component-approve-component -consumer-details-component -connect-component -reject-button-component
non_build
dev angular code account services new connection component approve component consumer details component connect component reject button component dev angular code account services new connection component approve component consumer details component connect component reject button component
0
62,332
15,219,448,375
IssuesEvent
2021-02-17 19:13:07
spring-projects/spring-security
https://api.github.com/repos/spring-projects/spring-security
closed
Update Git workflows to require JDK 11
in: build type: enhancement
As part of updating the build to require JDK 11, the Git workflows need to change.
1.0
Update Git workflows to require JDK 11 - As part of updating the build to require JDK 11, the Git workflows need to change.
build
update git workflows to require jdk as part of updating the build to require jdk the git workflows need to change
1
14,467
17,571,238,762
IssuesEvent
2021-08-14 18:43:02
elastic/beats
https://api.github.com/repos/elastic/beats
closed
Add glob or recursive option for processor `drop_fields`
enhancement libbeat :Processors Team:Integrations Stalled
**Describe the enhancement:** Add glob or recursive option for processor `drop_fields` which would be able to check the entire document for the field to remove. **Describe a specific use case for the enhancement or feature:** The processor works only with statically defined list of the fields and also only with fields at root level. If user needs to check the entire document for a field and does not know exact position, the only option is a script. Thank you
1.0
Add glob or recursive option for processor `drop_fields` - **Describe the enhancement:** Add glob or recursive option for processor `drop_fields` which would be able to check the entire document for the field to remove. **Describe a specific use case for the enhancement or feature:** The processor works only with statically defined list of the fields and also only with fields at root level. If user needs to check the entire document for a field and does not know exact position, the only option is a script. Thank you
non_build
add glob or recursive option for processor drop fields describe the enhancement add glob or recursive option for processor drop fields which would be able to check the entire document for the field to remove describe a specific use case for the enhancement or feature the processor works only with statically defined list of the fields and also only with fields at root level if user needs to check the entire document for a field and does not know exact position the only option is a script thank you
0
55,690
13,650,581,899
IssuesEvent
2020-09-26 19:52:51
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
Add `x.py setup` or similar
A-contributor-roadblock A-rustbuild C-feature-request
It has been [impressed](https://github.com/rust-lang/rust/issues/76371#issuecomment-687618541) [into](https://github.com/rust-lang/rust/issues/76165#issuecomment-687634489) [me](https://github.com/rust-lang/rust/pull/73964#issuecomment-653134345) [several](https://github.com/rust-lang/rust/pull/76446#issuecomment-688417692) [times](https://github.com/rust-lang/rust/pull/76417#issuecomment-687881451) that there are no good defaults for everyone. Instead, what if people could choose their own defaults, but in a guided way? I'm imagining something like this: ``` $ x.py setup Welcome to the Rust project! What do you want to do with x.py? a) Contribute to the standard library b) Contribute to the compiler c) Install Rust from source Please choose one (a/b/c): b Changed the following settings: - `incremental = true` - `debug = true` Do you plan to modify codegen/LLVM? (y/[n]): n # maybe some magic with $(which llvm-config) here; but won't be necessary after https://github.com/rust-lang/rust/pull/76349 To get started, try one of the following commands: - `x.py check` - `x.py build library/std` - `x.py test src/test/ui` ``` This would be entirely opt-in, if you don't like it, just don't run the command and modify `config.toml.example` as before. This would work best in conjunction with #76165, but does not require it. cc @Lokathor, @thomcc
1.0
Add `x.py setup` or similar - It has been [impressed](https://github.com/rust-lang/rust/issues/76371#issuecomment-687618541) [into](https://github.com/rust-lang/rust/issues/76165#issuecomment-687634489) [me](https://github.com/rust-lang/rust/pull/73964#issuecomment-653134345) [several](https://github.com/rust-lang/rust/pull/76446#issuecomment-688417692) [times](https://github.com/rust-lang/rust/pull/76417#issuecomment-687881451) that there are no good defaults for everyone. Instead, what if people could choose their own defaults, but in a guided way? I'm imagining something like this: ``` $ x.py setup Welcome to the Rust project! What do you want to do with x.py? a) Contribute to the standard library b) Contribute to the compiler c) Install Rust from source Please choose one (a/b/c): b Changed the following settings: - `incremental = true` - `debug = true` Do you plan to modify codegen/LLVM? (y/[n]): n # maybe some magic with $(which llvm-config) here; but won't be necessary after https://github.com/rust-lang/rust/pull/76349 To get started, try one of the following commands: - `x.py check` - `x.py build library/std` - `x.py test src/test/ui` ``` This would be entirely opt-in, if you don't like it, just don't run the command and modify `config.toml.example` as before. This would work best in conjunction with #76165, but does not require it. cc @Lokathor, @thomcc
build
add x py setup or similar it has been that there are no good defaults for everyone instead what if people could choose their own defaults but in a guided way i m imagining something like this x py setup welcome to the rust project what do you want to do with x py a contribute to the standard library b contribute to the compiler c install rust from source please choose one a b c b changed the following settings incremental true debug true do you plan to modify codegen llvm y n maybe some magic with which llvm config here but won t be necessary after to get started try one of the following commands x py check x py build library std x py test src test ui this would be entirely opt in if you don t like it just don t run the command and modify config toml example as before this would work best in conjunction with but does not require it cc lokathor thomcc
1
86,871
3,734,396,324
IssuesEvent
2016-03-08 06:36:53
notsecure/uTox
https://api.github.com/repos/notsecure/uTox
closed
Avatars should look better
enhancement Friends mid_priority user_interface
Two things: - They should be resized to 40x40 using cubic filtering(maybe this can be stolen from GIMP's source code?) so they look better(and look the same for every OS, currently Windows' built-in filtering looks worse) - They should use rounded corners as in the mockup: https://github.com/ItsDuke/Tox-UI
1.0
Avatars should look better - Two things: - They should be resized to 40x40 using cubic filtering(maybe this can be stolen from GIMP's source code?) so they look better(and look the same for every OS, currently Windows' built-in filtering looks worse) - They should use rounded corners as in the mockup: https://github.com/ItsDuke/Tox-UI
non_build
avatars should look better two things they should be resized to using cubic filtering maybe this can be stolen from gimp s source code so they look better and look the same for every os currently windows built in filtering looks worse they should use rounded corners as in the mockup
0
127,451
18,010,479,620
IssuesEvent
2021-09-16 08:01:32
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
opened
CVE-2019-19927 (Medium) detected in linux-yocto-4.1v4.1.17
security vulnerability
## CVE-2019-19927 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.1v4.1.17</b></p></summary> <p> <p>[no description]</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.1>https://git.yoctoproject.org/git/linux-yocto-4.1</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/gpu/drm/ttm/ttm_page_alloc.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/gpu/drm/ttm/ttm_page_alloc.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel 5.0.0-rc7 (as distributed in ubuntu/linux.git on kernel.ubuntu.com), mounting a crafted f2fs filesystem image and performing some operations can lead to slab-out-of-bounds read access in ttm_put_pages in drivers/gpu/drm/ttm/ttm_page_alloc.c. This is related to the vmwgfx or ttm module. <p>Publish Date: 2019-12-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19927>CVE-2019-19927</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/torvalds/linux/tree/v5.1-rc6">https://github.com/torvalds/linux/tree/v5.1-rc6</a></p> <p>Release Date: 2019-12-31</p> <p>Fix Resolution: v5.1-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-19927 (Medium) detected in linux-yocto-4.1v4.1.17 - ## CVE-2019-19927 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.1v4.1.17</b></p></summary> <p> <p>[no description]</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.1>https://git.yoctoproject.org/git/linux-yocto-4.1</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/gpu/drm/ttm/ttm_page_alloc.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/gpu/drm/ttm/ttm_page_alloc.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel 5.0.0-rc7 (as distributed in ubuntu/linux.git on kernel.ubuntu.com), mounting a crafted f2fs filesystem image and performing some operations can lead to slab-out-of-bounds read access in ttm_put_pages in drivers/gpu/drm/ttm/ttm_page_alloc.c. This is related to the vmwgfx or ttm module. <p>Publish Date: 2019-12-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19927>CVE-2019-19927</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/torvalds/linux/tree/v5.1-rc6">https://github.com/torvalds/linux/tree/v5.1-rc6</a></p> <p>Release Date: 2019-12-31</p> <p>Fix Resolution: v5.1-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_build
cve medium detected in linux yocto cve medium severity vulnerability vulnerable library linux yocto library home page a href found in head commit a href vulnerable source files drivers gpu drm ttm ttm page alloc c drivers gpu drm ttm ttm page alloc c vulnerability details in the linux kernel as distributed in ubuntu linux git on kernel ubuntu com mounting a crafted filesystem image and performing some operations can lead to slab out of bounds read access in ttm put pages in drivers gpu drm ttm ttm page alloc c this is related to the vmwgfx or ttm module publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
137,179
30,643,463,371
IssuesEvent
2023-07-25 01:15:26
libai8723/front-end-dev-notes-bignerdbook
https://api.github.com/repos/libai8723/front-end-dev-notes-bignerdbook
closed
LeetCode. 1218. Longest Arithmetic Subsequence of Given Difference
DP LeetCode
拥有给定公差的最长算术子序列。 https://leetcode.com/problems/longest-arithmetic-subsequence-of-given-difference/description/ 这个题算是做完了,耗费了我前前后后一周的时间,主要还是想法不对。一开始没有想到DP的做法,而是直接想到了排序,二分,外加缓存的优化,所以虽然还是accept了,但是整体的性能差距很大。 ### 1. 朴素的想法 1. 首先把数组排序了,排序按照数字的大小排序,其次按照数字本身在arr的序号排序,这样相同值的数字,最小序号的都排在前面了。 2. 对于同样取值的数字,如果是以这类数字作为起点的话,那么肯定是找序号最好的那个开始就可以了,因为序号大的同数值类型的作为起点,肯定长度不会超过序号小的。 3. 如果一个数字在前序的搜索中被用过了,那么就要考虑排除了。因为要么之前已经搜索过了,要么就是搜索到这个数字的时候,对应的长度肯定没办法超过之前的了。这就是我的缓存的原理。 ### 2. DP的搞法 1. DP的核心都是空间换时间,那么这道题的想法就是,对于已经看过的arr[0]~arr[i],定义一个dp函数:dp(number) = max_length 2. 这里的这个公式的含义比较特殊,这里的number是arr[0~i]这个区间内的一个数字,dp(number)的含义是,以number作为结束的,以difference为公差的,最长子数组的长度。 3. 这里我想过一些问题,那么如果number在arr[0~i]区间有重复怎么办?好像也不影响,这个函数只是表达,存在一个number结尾的,公差是difference的数组,长度是max_length 4. 解决了这个公式的定义之后,现在看一下怎么存储,因为入参只有一个number,取值也是一个整数max_length,这样搞一个map就可以了,比较高效。 5. 然后再看当i扩展到i+1的时候怎么办呢?number = arr[i+1],这时候要回看,已有的map(其实就是dp函数的定义)中是否有一个dp(number - difference)的定义,如果有的话,那么dp(number) = dp(number -difference) + 1, 同时判断一下,dp(number)是否大于accumulated_max_length,如果大于的话,就更新accumulated_max_length,如果dp(number - difference)不存在,就默认dp(number)为1. 6. 这样一路推导下去,就搞定了,map的最大的大小就是arr中distinct value count的大小,在空间上也是非常高效的。
1.0
LeetCode. 1218. Longest Arithmetic Subsequence of Given Difference - 拥有给定公差的最长算术子序列。 https://leetcode.com/problems/longest-arithmetic-subsequence-of-given-difference/description/ 这个题算是做完了,耗费了我前前后后一周的时间,主要还是想法不对。一开始没有想到DP的做法,而是直接想到了排序,二分,外加缓存的优化,所以虽然还是accept了,但是整体的性能差距很大。 ### 1. 朴素的想法 1. 首先把数组排序了,排序按照数字的大小排序,其次按照数字本身在arr的序号排序,这样相同值的数字,最小序号的都排在前面了。 2. 对于同样取值的数字,如果是以这类数字作为起点的话,那么肯定是找序号最好的那个开始就可以了,因为序号大的同数值类型的作为起点,肯定长度不会超过序号小的。 3. 如果一个数字在前序的搜索中被用过了,那么就要考虑排除了。因为要么之前已经搜索过了,要么就是搜索到这个数字的时候,对应的长度肯定没办法超过之前的了。这就是我的缓存的原理。 ### 2. DP的搞法 1. DP的核心都是空间换时间,那么这道题的想法就是,对于已经看过的arr[0]~arr[i],定义一个dp函数:dp(number) = max_length 2. 这里的这个公式的含义比较特殊,这里的number是arr[0~i]这个区间内的一个数字,dp(number)的含义是,以number作为结束的,以difference为公差的,最长子数组的长度。 3. 这里我想过一些问题,那么如果number在arr[0~i]区间有重复怎么办?好像也不影响,这个函数只是表达,存在一个number结尾的,公差是difference的数组,长度是max_length 4. 解决了这个公式的定义之后,现在看一下怎么存储,因为入参只有一个number,取值也是一个整数max_length,这样搞一个map就可以了,比较高效。 5. 然后再看当i扩展到i+1的时候怎么办呢?number = arr[i+1],这时候要回看,已有的map(其实就是dp函数的定义)中是否有一个dp(number - difference)的定义,如果有的话,那么dp(number) = dp(number -difference) + 1, 同时判断一下,dp(number)是否大于accumulated_max_length,如果大于的话,就更新accumulated_max_length,如果dp(number - difference)不存在,就默认dp(number)为1. 6. 这样一路推导下去,就搞定了,map的最大的大小就是arr中distinct value count的大小,在空间上也是非常高效的。
non_build
leetcode longest arithmetic subsequence of given difference 拥有给定公差的最长算术子序列。 这个题算是做完了,耗费了我前前后后一周的时间,主要还是想法不对。一开始没有想到dp的做法,而是直接想到了排序,二分,外加缓存的优化,所以虽然还是accept了,但是整体的性能差距很大。 朴素的想法 首先把数组排序了,排序按照数字的大小排序,其次按照数字本身在arr的序号排序,这样相同值的数字,最小序号的都排在前面了。 对于同样取值的数字,如果是以这类数字作为起点的话,那么肯定是找序号最好的那个开始就可以了,因为序号大的同数值类型的作为起点,肯定长度不会超过序号小的。 如果一个数字在前序的搜索中被用过了,那么就要考虑排除了。因为要么之前已经搜索过了,要么就是搜索到这个数字的时候,对应的长度肯定没办法超过之前的了。这就是我的缓存的原理。 dp的搞法 dp的核心都是空间换时间,那么这道题的想法就是,对于已经看过的arr arr ,定义一个dp函数:dp number max length 这里的这个公式的含义比较特殊,这里的number是arr 这个区间内的一个数字,dp number 的含义是,以number作为结束的,以difference为公差的,最长子数组的长度。 这里我想过一些问题,那么如果number在arr 区间有重复怎么办?好像也不影响,这个函数只是表达,存在一个number结尾的,公差是difference的数组,长度是max length 解决了这个公式的定义之后,现在看一下怎么存储,因为入参只有一个number,取值也是一个整数max length,这样搞一个map就可以了,比较高效。 然后再看当i扩展到i ?number arr ,这时候要回看,已有的map(其实就是dp函数的定义)中是否有一个dp number difference 的定义,如果有的话,那么dp number dp number difference , 同时判断一下,dp number 是否大于accumulated max length,如果大于的话,就更新accumulated max length,如果dp number difference 不存在,就默认dp number 这样一路推导下去,就搞定了,map的最大的大小就是arr中distinct value count的大小,在空间上也是非常高效的。
0
36,442
17,692,740,580
IssuesEvent
2021-08-24 12:06:52
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Global sound issues
Performance Sound Bug
So for a while now there's been a number off issues around admin loaded sounds. The first is that some sounds, such as emitters firing, will interrupt the sound being played, rather like there aren't enough channels to play both sounds at the same time. Its particularly noticeable when multiple emitters are set up and it will make it stutter, due to their short sound length and regular triggering of said effect. There also seems to be some major issue of lag being created when they are uploaded. It will cause the server to hang for at least 1 second during a full 1mb song with a population of around 60 or higher (havent tried it on low pop). Its hard to tell if its just a population increase thats effecting it, but Im fairly certain it didnt used to be this bad in terms of lag. I realise this issue is relatively vague but its more a call to arms to do some further testing and seeing if anyone might know of what might have changed to the code of late (never noticed these issues before number value sound levels became a thing (i.e. when it was just on or off as a pref)
True
Global sound issues - So for a while now there's been a number off issues around admin loaded sounds. The first is that some sounds, such as emitters firing, will interrupt the sound being played, rather like there aren't enough channels to play both sounds at the same time. Its particularly noticeable when multiple emitters are set up and it will make it stutter, due to their short sound length and regular triggering of said effect. There also seems to be some major issue of lag being created when they are uploaded. It will cause the server to hang for at least 1 second during a full 1mb song with a population of around 60 or higher (havent tried it on low pop). Its hard to tell if its just a population increase thats effecting it, but Im fairly certain it didnt used to be this bad in terms of lag. I realise this issue is relatively vague but its more a call to arms to do some further testing and seeing if anyone might know of what might have changed to the code of late (never noticed these issues before number value sound levels became a thing (i.e. when it was just on or off as a pref)
non_build
global sound issues so for a while now there s been a number off issues around admin loaded sounds the first is that some sounds such as emitters firing will interrupt the sound being played rather like there aren t enough channels to play both sounds at the same time its particularly noticeable when multiple emitters are set up and it will make it stutter due to their short sound length and regular triggering of said effect there also seems to be some major issue of lag being created when they are uploaded it will cause the server to hang for at least second during a full song with a population of around or higher havent tried it on low pop its hard to tell if its just a population increase thats effecting it but im fairly certain it didnt used to be this bad in terms of lag i realise this issue is relatively vague but its more a call to arms to do some further testing and seeing if anyone might know of what might have changed to the code of late never noticed these issues before number value sound levels became a thing i e when it was just on or off as a pref
0
35,206
16,985,382,652
IssuesEvent
2021-06-30 13:52:30
supaero-aircraft-design/FAST-GA
https://api.github.com/repos/supaero-aircraft-design/FAST-GA
closed
Calculer et utiliser au bon endroit la trainée des trains d'atterrissage
aerodynamics methodology performances
Principalement au décollage. On part du principe que pour un train d'atterrissage non rétractable on paye la pénalité de traînée à tout moment. Pour un train rétractable, le cd0 du aux trains n'est additionné que sur le phases basse vitesse. Sachant que la montée et la descente sont des phases en haute-vitesse, ça reste juste.
True
Calculer et utiliser au bon endroit la trainée des trains d'atterrissage - Principalement au décollage. On part du principe que pour un train d'atterrissage non rétractable on paye la pénalité de traînée à tout moment. Pour un train rétractable, le cd0 du aux trains n'est additionné que sur le phases basse vitesse. Sachant que la montée et la descente sont des phases en haute-vitesse, ça reste juste.
non_build
calculer et utiliser au bon endroit la trainée des trains d atterrissage principalement au décollage on part du principe que pour un train d atterrissage non rétractable on paye la pénalité de traînée à tout moment pour un train rétractable le du aux trains n est additionné que sur le phases basse vitesse sachant que la montée et la descente sont des phases en haute vitesse ça reste juste
0
653,597
21,608,147,096
IssuesEvent
2022-05-04 07:09:46
cyrusae/highlighter-public
https://api.github.com/repos/cyrusae/highlighter-public
opened
Clean paragraphs on the go
bug database frontend Priority: ++
Doesn't appear to be reliably saving, so may want to revisit how that's working
1.0
Clean paragraphs on the go - Doesn't appear to be reliably saving, so may want to revisit how that's working
non_build
clean paragraphs on the go doesn t appear to be reliably saving so may want to revisit how that s working
0
334,046
29,820,341,007
IssuesEvent
2023-06-17 01:30:45
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix device.test_as_ivy_dev
Sub Task Failing Test
| | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix device.test_as_ivy_dev - | | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5229705239/jobs/9442842920"><img src=https://img.shields.io/badge/-success-success></a>
non_build
fix device test as ivy dev jax a href src torch a href src tensorflow a href src numpy a href src paddle a href src
0
238,387
18,240,125,839
IssuesEvent
2021-10-01 11:55:28
hashicorp/terraform-ls
https://api.github.com/repos/hashicorp/terraform-ls
opened
docs: Reflect tfvars support in USAGE (client setup guides)
documentation
Currently we do not actively encourage clients to detect and send tfvars files to the language server here: https://github.com/hashicorp/terraform-ls/blob/main/docs/USAGE.md The only mention so far is in the dev-oriented docs: https://github.com/hashicorp/terraform-ls/blob/main/docs/language-clients.md
1.0
docs: Reflect tfvars support in USAGE (client setup guides) - Currently we do not actively encourage clients to detect and send tfvars files to the language server here: https://github.com/hashicorp/terraform-ls/blob/main/docs/USAGE.md The only mention so far is in the dev-oriented docs: https://github.com/hashicorp/terraform-ls/blob/main/docs/language-clients.md
non_build
docs reflect tfvars support in usage client setup guides currently we do not actively encourage clients to detect and send tfvars files to the language server here the only mention so far is in the dev oriented docs
0
32,849
8,960,682,081
IssuesEvent
2019-01-28 07:04:54
akka/akka
https://api.github.com/repos/akka/akka
closed
sbt 1.2.7 broke unidoc/javadoc with jdk9
t:build
The update to sbt 1.2.7 in https://github.com/akka/akka/commit/6ba4c341f5d2a090c28e9ec92d0ea05f7188de61 broke unidoc/javadoc Reproduce with sbt 1.2.7: ``` sbt -Dakka.genjavadoc.enabled=true > clean > unidoc ... javadoc exited with exit code 1 ``` I have verified that it was not the `classpath 4.4.12` update that caused it.
1.0
sbt 1.2.7 broke unidoc/javadoc with jdk9 - The update to sbt 1.2.7 in https://github.com/akka/akka/commit/6ba4c341f5d2a090c28e9ec92d0ea05f7188de61 broke unidoc/javadoc Reproduce with sbt 1.2.7: ``` sbt -Dakka.genjavadoc.enabled=true > clean > unidoc ... javadoc exited with exit code 1 ``` I have verified that it was not the `classpath 4.4.12` update that caused it.
build
sbt broke unidoc javadoc with the update to sbt in broke unidoc javadoc reproduce with sbt sbt dakka genjavadoc enabled true clean unidoc javadoc exited with exit code i have verified that it was not the classpath update that caused it
1
286,793
21,609,480,041
IssuesEvent
2022-05-04 08:36:49
WordPress/Documentation-Issue-Tracker
https://api.github.com/repos/WordPress/Documentation-Issue-Tracker
closed
[HelpHub] Paragraph Block
user documentation 5.7 5.9 block editor
Article: https://wordpress.org/support/article/paragraph-block/ ## Updating for 5.7 - [ ] Add Keyboard Input inline format - [26801](https://github.com/WordPress/gutenberg/pull/26801) ## General - [ ] make sure all screenshots are relevant for current version (5.9) - [ ] Video updates - [ ] update changelog at the end of the article Issue migrated from Trello: https://trello.com/c/mWOy1uu2/1-paragraph-block
1.0
[HelpHub] Paragraph Block - Article: https://wordpress.org/support/article/paragraph-block/ ## Updating for 5.7 - [ ] Add Keyboard Input inline format - [26801](https://github.com/WordPress/gutenberg/pull/26801) ## General - [ ] make sure all screenshots are relevant for current version (5.9) - [ ] Video updates - [ ] update changelog at the end of the article Issue migrated from Trello: https://trello.com/c/mWOy1uu2/1-paragraph-block
non_build
paragraph block article updating for add keyboard input inline format general make sure all screenshots are relevant for current version video updates update changelog at the end of the article issue migrated from trello
0
24,178
2,666,730,519
IssuesEvent
2015-03-21 21:25:16
Templarian/MaterialDesign
https://api.github.com/repos/Templarian/MaterialDesign
closed
H(eading)1 ... H(eading)6 and P(aragraph) icons for editor
High Priority Icon Request
These are often used in embedded editors and would be nice to have - [x] format-header-pound - [x] format-header-1 - [x] format-header-2 - [x] format-header-3 - [x] format-header-4 - [x] format-header-5 - [x] format-header-6 - [x] format-paragraph
1.0
H(eading)1 ... H(eading)6 and P(aragraph) icons for editor - These are often used in embedded editors and would be nice to have - [x] format-header-pound - [x] format-header-1 - [x] format-header-2 - [x] format-header-3 - [x] format-header-4 - [x] format-header-5 - [x] format-header-6 - [x] format-paragraph
non_build
h eading h eading and p aragraph icons for editor these are often used in embedded editors and would be nice to have format header pound format header format header format header format header format header format header format paragraph
0
56,283
13,784,651,070
IssuesEvent
2020-10-08 21:14:47
microsoft/vscode-cpptools
https://api.github.com/repos/microsoft/vscode-cpptools
closed
${vcpkgRoot} variable is not expanded in tasks.json
Visual Studio Code build/tasks duplicate question
${vcpkgRoot} variable is expanded in c_cpp_properties.json (for example in includePath) but not in tasks.json. I need this expansion to write a cmake task command. Something like this: ``` { "label": "cmake", "type": "shell", "command": "cmake", "args": [ "${workspaceFolder}", "-B", "${workspaceFolder}/build", "-DCMAKE_TOOLCHAIN_FILE=${vcpkgRoot}/scripts/buildsystems/vcpkg.cmake" ], "group": "build", "problemMatcher": [] } ``` Thanks
1.0
${vcpkgRoot} variable is not expanded in tasks.json - ${vcpkgRoot} variable is expanded in c_cpp_properties.json (for example in includePath) but not in tasks.json. I need this expansion to write a cmake task command. Something like this: ``` { "label": "cmake", "type": "shell", "command": "cmake", "args": [ "${workspaceFolder}", "-B", "${workspaceFolder}/build", "-DCMAKE_TOOLCHAIN_FILE=${vcpkgRoot}/scripts/buildsystems/vcpkg.cmake" ], "group": "build", "problemMatcher": [] } ``` Thanks
build
vcpkgroot variable is not expanded in tasks json vcpkgroot variable is expanded in c cpp properties json for example in includepath but not in tasks json i need this expansion to write a cmake task command something like this label cmake type shell command cmake args workspacefolder b workspacefolder build dcmake toolchain file vcpkgroot scripts buildsystems vcpkg cmake group build problemmatcher thanks
1
143,371
19,177,931,077
IssuesEvent
2021-12-04 00:07:11
samq-ghdemo/js-monorepo
https://api.github.com/repos/samq-ghdemo/js-monorepo
opened
CVE-2020-7608 (Medium) detected in multiple libraries
security vulnerability
## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-2.4.1.tgz</b>, <b>yargs-parser-8.1.0.tgz</b>, <b>yargs-parser-9.0.2.tgz</b></p></summary> <p> <details><summary><b>yargs-parser-2.4.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz</a></p> <p>Path to dependency file: js-monorepo/NodeGoat/package.json</p> <p>Path to vulnerable library: js-monorepo/NodeGoat/node_modules/nyc/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - realize-package-specifier-3.0.3.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - :x: **yargs-parser-2.4.1.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-8.1.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz</a></p> <p>Path to dependency file: js-monorepo/nodejs-goof/package.json</p> <p>Path to vulnerable library: js-monorepo/nodejs-goof/node_modules/nyc/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - tap-11.1.5.tgz (Root Library) - nyc-11.9.0.tgz - :x: **yargs-parser-8.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-9.0.2.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz</a></p> <p>Path to dependency file: js-monorepo/nodejs-goof/package.json</p> <p>Path to vulnerable library: js-monorepo/nodejs-goof/node_modules/nyc/node_modules/yargs/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - tap-11.1.5.tgz (Root Library) - nyc-11.9.0.tgz - yargs-11.1.0.tgz - :x: **yargs-parser-9.0.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/js-monorepo/commit/f3701923c18333c1e4e49bf595dd36b3f186812f">f3701923c18333c1e4e49bf595dd36b3f186812f</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: 5.0.1;13.1.2;15.0.1;18.1.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"2.4.1","packageFilePaths":["/NodeGoat/package.json"],"isTransitiveDependency":true,"dependencyTree":"realize-package-specifier:3.0.3;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;yargs-parser:2.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"8.1.0","packageFilePaths":["/nodejs-goof/package.json"],"isTransitiveDependency":true,"dependencyTree":"tap:11.1.5;nyc:11.9.0;yargs-parser:8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"9.0.2","packageFilePaths":["/nodejs-goof/package.json"],"isTransitiveDependency":true,"dependencyTree":"tap:11.1.5;nyc:11.9.0;yargs:11.1.0;yargs-parser:9.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7608 (Medium) detected in multiple libraries - ## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-2.4.1.tgz</b>, <b>yargs-parser-8.1.0.tgz</b>, <b>yargs-parser-9.0.2.tgz</b></p></summary> <p> <details><summary><b>yargs-parser-2.4.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz</a></p> <p>Path to dependency file: js-monorepo/NodeGoat/package.json</p> <p>Path to vulnerable library: js-monorepo/NodeGoat/node_modules/nyc/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - realize-package-specifier-3.0.3.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - :x: **yargs-parser-2.4.1.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-8.1.0.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-8.1.0.tgz</a></p> <p>Path to dependency file: js-monorepo/nodejs-goof/package.json</p> <p>Path to vulnerable library: js-monorepo/nodejs-goof/node_modules/nyc/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - tap-11.1.5.tgz (Root Library) - nyc-11.9.0.tgz - :x: **yargs-parser-8.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>yargs-parser-9.0.2.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-9.0.2.tgz</a></p> <p>Path to dependency file: js-monorepo/nodejs-goof/package.json</p> <p>Path to vulnerable library: js-monorepo/nodejs-goof/node_modules/nyc/node_modules/yargs/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - tap-11.1.5.tgz (Root Library) - nyc-11.9.0.tgz - yargs-11.1.0.tgz - :x: **yargs-parser-9.0.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/js-monorepo/commit/f3701923c18333c1e4e49bf595dd36b3f186812f">f3701923c18333c1e4e49bf595dd36b3f186812f</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2">https://github.com/yargs/yargs-parser/commit/63810ca1ae1a24b08293a4d971e70e058c7a41e2</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: 5.0.1;13.1.2;15.0.1;18.1.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"2.4.1","packageFilePaths":["/NodeGoat/package.json"],"isTransitiveDependency":true,"dependencyTree":"realize-package-specifier:3.0.3;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;yargs-parser:2.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"8.1.0","packageFilePaths":["/nodejs-goof/package.json"],"isTransitiveDependency":true,"dependencyTree":"tap:11.1.5;nyc:11.9.0;yargs-parser:8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false},{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"9.0.2","packageFilePaths":["/nodejs-goof/package.json"],"isTransitiveDependency":true,"dependencyTree":"tap:11.1.5;nyc:11.9.0;yargs:11.1.0;yargs-parser:9.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.1;13.1.2;15.0.1;18.1.1","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_build
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries yargs parser tgz yargs parser tgz yargs parser tgz yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file js monorepo nodegoat package json path to vulnerable library js monorepo nodegoat node modules nyc node modules yargs parser package json dependency hierarchy realize package specifier tgz root library grunt contrib nodeunit tgz nodeunit tgz tap tgz nyc tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file js monorepo nodejs goof package json path to vulnerable library js monorepo nodejs goof node modules nyc node modules yargs parser package json dependency hierarchy tap tgz root library nyc tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file js monorepo nodejs goof package json path to vulnerable library js monorepo nodejs goof node modules nyc node modules yargs node modules yargs parser package json dependency hierarchy tap tgz root library nyc tgz yargs tgz x yargs parser tgz vulnerable library found in head commit a href found in base branch main vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree realize package specifier grunt contrib nodeunit nodeunit tap nyc yargs parser isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename yargs parser packageversion packagefilepaths istransitivedependency true dependencytree tap nyc yargs parser isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename yargs parser packageversion packagefilepaths istransitivedependency true dependencytree tap nyc yargs yargs parser isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload vulnerabilityurl
0
349,292
31,791,005,868
IssuesEvent
2023-09-13 03:18:36
TencentBlueKing/bk-cmdb
https://api.github.com/repos/TencentBlueKing/bk-cmdb
closed
【CMDB+3.11.2-alpha1】业务拓扑下导出的主机数据前提提示成功,实际上Excel表格里面没有数据
for test priority: Urgent
问题描述: 业务拓扑下导出的主机数据前提提示成功,实际上Excel表格里面没有塑化剂 一,前置条件: 业务拓扑下存在主机,并在业务拓扑页面 二,操作步骤: 1.勾选多条主机 2.点击更多--导出选中 ![image](https://github.com/TencentBlueKing/bk-cmdb/assets/132563061/77ec7a01-a5bc-45b9-b6f2-049c82ba9771) 3.点击下一步-->开始导出 ![image](https://github.com/TencentBlueKing/bk-cmdb/assets/132563061/aa8aba0d-2986-4449-8f51-213b52f9897d) 三,预期结果: 数据导出成功,Excel表格正常查看到导出的数据 四,实际结果: 前端提示导出成功,Excel表格里面没有数据 ![image](https://github.com/TencentBlueKing/bk-cmdb/assets/132563061/3cb2af0e-acc3-4866-a76c-d2f0bb5f9172)
1.0
【CMDB+3.11.2-alpha1】业务拓扑下导出的主机数据前提提示成功,实际上Excel表格里面没有数据 - 问题描述: 业务拓扑下导出的主机数据前提提示成功,实际上Excel表格里面没有塑化剂 一,前置条件: 业务拓扑下存在主机,并在业务拓扑页面 二,操作步骤: 1.勾选多条主机 2.点击更多--导出选中 ![image](https://github.com/TencentBlueKing/bk-cmdb/assets/132563061/77ec7a01-a5bc-45b9-b6f2-049c82ba9771) 3.点击下一步-->开始导出 ![image](https://github.com/TencentBlueKing/bk-cmdb/assets/132563061/aa8aba0d-2986-4449-8f51-213b52f9897d) 三,预期结果: 数据导出成功,Excel表格正常查看到导出的数据 四,实际结果: 前端提示导出成功,Excel表格里面没有数据 ![image](https://github.com/TencentBlueKing/bk-cmdb/assets/132563061/3cb2af0e-acc3-4866-a76c-d2f0bb5f9172)
non_build
【cmdb 】业务拓扑下导出的主机数据前提提示成功,实际上excel表格里面没有数据 问题描述: 业务拓扑下导出的主机数据前提提示成功,实际上excel表格里面没有塑化剂 一,前置条件: 业务拓扑下存在主机,并在业务拓扑页面 二,操作步骤: 勾选多条主机 点击更多 导出选中 点击下一步 开始导出 三,预期结果: 数据导出成功,excel表格正常查看到导出的数据 四,实际结果: 前端提示导出成功,excel表格里面没有数据
0
568,873
16,989,623,224
IssuesEvent
2021-06-30 18:36:41
crispy-computing-machine/Winbinder
https://api.github.com/repos/crispy-computing-machine/Winbinder
closed
Allocate memory dynamically in various places
Low priority enhancement help wanted
Allocate memory dynamically instead of limited static memory - In wbtemp_create_menu(): allocate accel[] dynamically like pitem - Allocate tab controls and pages dynamically: TABDATA should allocate items dynamically; now it is carrying a somewhat large data structure
1.0
Allocate memory dynamically in various places - Allocate memory dynamically instead of limited static memory - In wbtemp_create_menu(): allocate accel[] dynamically like pitem - Allocate tab controls and pages dynamically: TABDATA should allocate items dynamically; now it is carrying a somewhat large data structure
non_build
allocate memory dynamically in various places allocate memory dynamically instead of limited static memory in wbtemp create menu allocate accel dynamically like pitem allocate tab controls and pages dynamically tabdata should allocate items dynamically now it is carrying a somewhat large data structure
0
105,395
13,183,326,087
IssuesEvent
2020-08-12 17:18:41
COVID19Tracking/website
https://api.github.com/repos/COVID19Tracking/website
closed
Remove Florida exception made in #1250
DESIGN DEV
See #1250 -- we should no longer need this exception since FL has been reporting hosp. data for over 30 days now.
1.0
Remove Florida exception made in #1250 - See #1250 -- we should no longer need this exception since FL has been reporting hosp. data for over 30 days now.
non_build
remove florida exception made in see we should no longer need this exception since fl has been reporting hosp data for over days now
0
180,534
6,650,521,737
IssuesEvent
2017-09-28 16:36:51
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.yam.com - see bug description
browser-firefox priority-normal status-invalid
<!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: web --> **URL**: https://www.yam.com/ **Browser / Version**: Firefox 58.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Something else **Description**: The connection is unsafe. **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2017/9/f13f8200-7ae0-4f1b-96dc-018e374e39ed-thumb.jpeg)](https://webcompat.com/uploads/2017/9/f13f8200-7ae0-4f1b-96dc-018e374e39ed.jpeg) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.yam.com - see bug description - <!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: web --> **URL**: https://www.yam.com/ **Browser / Version**: Firefox 58.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Something else **Description**: The connection is unsafe. **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2017/9/f13f8200-7ae0-4f1b-96dc-018e374e39ed-thumb.jpeg)](https://webcompat.com/uploads/2017/9/f13f8200-7ae0-4f1b-96dc-018e374e39ed.jpeg) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_build
see bug description url browser version firefox operating system windows tested another browser no problem type something else description the connection is unsafe steps to reproduce from with ❤️
0
4,906
3,477,900,140
IssuesEvent
2015-12-28 07:16:21
rubinius/rubinius
https://api.github.com/repos/rubinius/rubinius
closed
Rubinius fails to compile: '_setjmp' was not declared in this scope
build Solaris
Hi I'm trying to compile Rubinius pulled from the git repo as of tonight within a Solaris zone and I keep getting the same error. I've tried the current git repo and the last two releases. I first attempted this with Ruby 1.9.3-xxx installed, but after reading on the Rubinius website, I restarted the template with a fresh Ruby 2.0.0-xxx install. They all error out in the same place: ``` Running 191 tasks using 1 parallel threads 2: CXX vm/accessor_primitives.cpp 2: CXX vm/agent.cpp 2: CXX vm/agent_components.cpp 2: CXX vm/arguments.cpp 2: CXX vm/auxiliary_threads.cpp 2: CXX vm/builtin/access_variable.cpp 2: CXX vm/builtin/alias.cpp 2: CXX vm/builtin/array.cpp 2: CXX vm/builtin/atomic.cpp 2: CXX vm/builtin/autoload.cpp 2: CXX vm/builtin/bignum.cpp 2: CXX vm/builtin/block_as_method.cpp 2: CXX vm/builtin/block_environment.cpp 2: CXX vm/builtin/byte_array.cpp 2: CXX vm/builtin/call_custom_cache.cpp 2: CXX vm/builtin/call_site.cpp 2: CXX vm/builtin/call_unit.cpp 2: CXX vm/builtin/call_unit_adapter.cpp 2: CXX vm/builtin/channel.cpp 2: CXX vm/builtin/character.cpp 2: CXX vm/builtin/class.cpp 2: CXX vm/builtin/compact_lookup_table.cpp 2: CXX vm/builtin/compiled_code.cpp 2: CXX vm/builtin/constant_cache.cpp 2: CXX vm/builtin/constant_scope.cpp 2: CXX vm/builtin/constant_table.cpp 2: CXX vm/builtin/data.cpp 2: CXX vm/builtin/dir.cpp 2: CXX vm/builtin/encoding.cpp 2: CXX vm/builtin/exception.cpp 2: CXX vm/builtin/executable.cpp 2: CXX vm/builtin/ffi_pointer.cpp 2: CXX vm/builtin/fiber.cpp 2: CXX vm/builtin/find_object.cpp 2: CXX vm/builtin/fixnum.cpp 2: CXX vm/builtin/float.cpp 2: CXX vm/builtin/heap_dump.cpp 2: CXX vm/builtin/immediates.cpp 2: CXX vm/builtin/integer.cpp 2: CXX vm/builtin/io.cpp 2: CXX vm/builtin/iseq.cpp 2: CXX vm/builtin/list.cpp 2: CXX vm/builtin/location.cpp 2: CXX vm/builtin/lookup_table.cpp 2: CXX vm/builtin/method_table.cpp 2: CXX vm/builtin/module.cpp 2: CXX vm/builtin/mono_inline_cache.cpp 2: CXX vm/builtin/native_function.cpp 2: CXX vm/builtin/native_method.cpp vm/builtin/native_method.cpp: In static member function 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&)': vm/builtin/native_method.cpp:691:9: error: there are no arguments to '_setjmp' that depend on a template parameter, so a declaration of '_setjmp' must be available [-fpermissive] vm/builtin/native_method.cpp:691:9: note: (if you use '-fpermissive', G++ will accept your code, but allowing the use of an undeclared name is deprecated) vm/builtin/native_method.cpp:702:9: error: there are no arguments to '_setjmp' that depend on a template parameter, so a declaration of '_setjmp' must be available [-fpermissive] vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::ZeroArguments; rubinius::Executable = rubinius::Executable]': vm/builtin/native_method.cpp:776:82: required from here vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:691:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:702:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::OneArgument; rubinius::Executable = rubinius::Executable]': vm/builtin/native_method.cpp:780:80: required from here vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:691:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:702:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::TwoArguments; rubinius::Executable = rubinius::Executable]': vm/builtin/native_method.cpp:784:81: required from here vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:691:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:702:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::ThreeArguments; rubinius::Executable = rubinius::Executable]': vm/builtin/native_method.cpp:788:83: required from here vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:691:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:702:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::GenericArguments; rubinius::Executable = rubinius::Executable]': vm/builtin/native_method.cpp:792:85: required from here vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:691:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:702:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' Error: g++ -I/usr/local/src/rubinius/vm -I/usr/local/src/rubinius/vm/include -I/usr/local/src/rubinius/vm/builtin -I. -Ivm/test/cxxtest -I/usr/local/src/rubinius/vendor/udis86 -I/usr/local/src/rubinius/vendor/libffi/include -Ivendor/double-conversion/src -DHAVE_CONFIG_H -I/usr/local/src/rubinius/vm/include/capi -I/usr/local/src/rubinius/vendor/oniguruma -I/usr/local/src/rubinius/vendor/libtommath -pipe -Wall -fno-omit-frame-pointer -g -I/usr/local/include -I/opt/local/include -fPIC -Wno-strict-aliasing -O2 -DHAS_EXECINFO -DHAVE_CLOCK_GETTIME -DHAVE_NL_LANGINFO -DHAVE_POSIX_FADVISE -DHAVE_STRNLEN -DHAVE_TIMEZONE -DHAVE_TZNAME -DHAVE_DAYLIGHT -DHAVE_ALLOCA_H -DHAVE_STRING_H -DHAVE_SYS_TIME_H -DHAVE_SYS_TIMES_H -DHAVE_SYS_TYPES_H -DHAVE_UNISTD_H -DHAVE_STDARG_H -I/usr/local/include -D_DEBUG -include llvm/Support/Solaris.h -D_GNU_SOURCE -fPIC -DENABLE_LLVM -Wno-unused-function -Werror -DRBX_PROFILER -D__STDC_LIMIT_MACROS -D__STDC_CONSTANT_MACROS -D_LARGEFILE_SOURCE -D_FILE_OFFSET_BITS=64 -fno-rtti -fvisibility-inlines-hidden -c -o vm/builtin/artifacts/native_method.cpp.o vm/builtin/native_method.cpp rake aborted! Error compiling Tasks: TOP => default => spec => build => build:build => vm/vm (See full trace by running task with --trace) ``` Installed packages: I built LLVM from source as it wasn't in the repos, using LLVM-3.3 ``` binutils-2.23.2nb1 GNU binary utilities bison-3.0 GNU yacc(1) replacement bmake-20110606nb1 Portable (autoconf) version of NetBSD 'make' utility bootstrap-mk-files-20130912 *.mk files for the bootstrap bmake utility bsdinstall-20130905 Portable version of the BSD install(1) program bzip2-1.0.6nb1 Block-sorting file compressor changepass-1.3.3 Simple tool to set user passwords non-interactively cloog-0.18.0nb1 Code generator for loop optimization (used by gcc) coreutils-8.13nb6 GNU basic file, shell and text manipulation utilities curl-7.32.0 Client that groks URLs cyrus-sasl-2.1.26nb3 Simple Authentication and Security Layer db4-4.8.30 Berkeley DB version 4 from Oracle diffutils-3.3 GNU diff utilities - find the differences between files duo-unix-1.9.5 2 Factor Authentication provided by Duo Security expat-2.1.0 XML parser library written in C findutils-4.2.33nb3 The GNU find, locate, updatedb, and xargs utilities fontconfig-2.10.95 Library for configuring and customizing font access freetype2-2.4.12 Font rendering engine and library API gawk-4.1.0 GNU awk gcc47-4.7.3nb1 The GNU Compiler Collection (GCC) - 4.7 Release Series gcc47-libs-4.7.3nb3 The GNU Compiler Collection (GCC) support shared libraries genbfpw-0 Simple tool to generate Blowfish crypts of passwords gettext-lib-0.18.3.1 Internationalized Message Handling Library (libintl) ghostscript-9.05nb6 Meta-package for installing the preferred ghostscript version ghostscript-fonts-8.11nb3 Postscript fonts for Aladdin Ghostscript ghostscript-gpl-9.05nb6 Postscript interpreter git-base-1.8.4nb1 GIT Tree History Storage Tool (base package) gmake-3.82nb7 GNU version of 'make' utility gmp-5.1.2 Library for arbitrary precision arithmetic grep-2.14nb2 GNU grep groff-1.20.1nb12 GNU roff text processing suite gsed-4.2.2nb4 GNU implementation of sed, the POSIX stream editor gtar-base-1.26nb2 The GNU tape archiver with remote magnetic tape support guardtime-1.0.12 Command line client for GuardTime keyless signature service isl-0.11.2 Integer set library required by gcc graphite jasper-1.900.1nb7 Software-based reference implementation of the JPEG-2000 codec jbigkit-2.0 JBIG-KIT lossless image compression library jpeg-9 IJG's jpeg compression utilities json-c-0.10 JSON library in C less-458 Pager similar to more and pg libICE-1.0.8 Inter Client Exchange (ICE) library for X libSM-1.2.2 X Session Management Library libX11-1.6.2 Base X libraries from modular Xorg X11 libXau-1.0.8 Authorization Protocol for X from X.org libXaw-1.0.12 X Athena Widgets Library from modular Xorg X11 libXdmcp-1.1.1 X Display Manager Control Protocol library from X.org libXext-1.3.2 X Extension library libXmu-1.1.2 X Miscellaneous Utilities library libXpm-3.5.11 X PixMap Library from modular Xorg X11 libXt-1.1.4 X Toolkit Intrinsics library libarchive-2.8.4nb2 Library to read/create different archive formats libee-0.4.1 Event Expression Library inspired by CEE libestr-0.1.5 Library for some string essentials libffi-3.0.13 Foreign function interface libgcrypt-1.5.3 GNU cryptographic library libgetopt-1.4.4 Library for handling --long options libgpg-error-1.12 Definitions of common error values for all GnuPG components libguardtime-0.3.11 GuardTime Client C SDK libiconv-1.14nb2 Character set conversion library libidn-1.28 Internationalized Domain Names command line tool liblognorm-0.3.7 Tool to normalize log data libpaper-1.1.24 Paper size handling library libssh2-1.2.2nb2 SSH2 protocol library libuuid-2.19.1 Generate unique identifiers for objects libxcb-1.9.1 X protocol C-language Binding libxml2-2.9.1 XML parser library from the GNOME project libyaml-0.1.4 YAML 1.1 parser and emitter written in C m4-1.4.16nb3 GNU version of UNIX m4 macro language processor mit-krb5-1.10.6 MIT Kerberos 5 authentication system mozilla-rootcerts-1.0.20121229nb1 Root CA certificates from the Mozilla Project mpfr-3.1.2 GMP-based library for multiple-precision floating-point computations nawk-20050424nb3 Brian Kernighan's pattern-directed scanning and processing language nbsed-20120308 NetBSD-current's sed(1) ncurses-5.9nb1 CRT screen handling and optimization package netpbm-10.35.80nb7 Toolkit for conversion of images between different formats nodejs-0.10.22 V8 JavaScript for clients and servers openldap-client-2.4.36 Lightweight Directory Access Protocol libraries and client programs openssl-1.0.1enb2 Secure Socket Layer and cryptographic library p5-Authen-SASL-2.16nb1 Perl module to handle SASL authentication p5-Digest-HMAC-1.03nb2 Perl5 module for HMAC p5-Digest-MD5-2.53 Perl5 extension interface for MD5 algorithm p5-Digest-SHA-5.85 Perl module for SHA-1/224/256/384/512 algorithms p5-Email-Valid-1.192 Perl5 module for testing validity of an email address p5-Error-0.17021 Perl extension module for try/throw/catch exception handling p5-GSSAPI-0.28nb4 Perl extension providing access to the GSSAPIv2 library p5-IO-Socket-INET6-2.71 Perl object interface for AF_INET|AF_INET6 domain sockets p5-IO-Socket-SSL-1.953 Perl5 SSL socket interface class p5-MIME-Base64-3.14 Perl5 module for Base64 and Quoted-Printable encodings p5-MailTools-2.12nb1 Perl5 modules related to mail applications p5-Net-DNS-0.72 Perl5 module for DNS resolution p5-Net-IP-1.26nb1 Perl extension for manipulating IPv4/IPv6 addresses p5-Net-LibIDN-0.12nb5 Perl bindings for GNU Libidn p5-Net-SMTP-SSL-1.01nb4 Perl5 module providing SSL support for Net::SMTP p5-Net-SSLeay-1.55 Perl5 module for using OpenSSL p5-Socket6-0.23nb4 Perl5 module to support getaddrinfo() and getnameinfo() p5-TimeDate-2.30 Perl5 TimeDate distribution patch-2.5.9nb2 Patch files using diff output pcre-8.33 Perl Compatible Regular Expressions library perl-5.18.1nb1 Practical Extraction and Report Language pkg_install-20130902 Package management and administration tools for pkgsrc pkg_install-info-4.5nb3 Standalone GNU info file installation utility pkgin-0.6.4nb1 Apt / yum like tool for managing pkgsrc binary packages png-1.6.6 Library for manipulating PNG images postfix-2.9.8 Postfix SMTP server and tools ppl-0.11.2nb2 The Parma Polyhedra Library. Used by gcc for loop optimization psutils-1.17nb4 Utilities for manipulating PostScript documents python27-2.7.5nb3 Interpreted, interactive, object-oriented programming language readline-6.2 GNU library that can recall and edit previous input rsyslog-7.4.4nb2 The enhanced syslogd for Unix ruby-2.0.0p353 Wrapper package for Ruby programming language ruby200-base-2.0.0p353 Ruby 2.0.0 release minimum base package ruby200-bundler-1.3.5 Manage your application's dependencies ruby200-rake-10.0.4 Ruby Make ruby200-rubygems-2.0.10 Ruby standard for publishing and managing third party libraries sdc-manta-1.2.2 Node.js SDK for Manta sdc-node-0.10.22 V8 JavaScript for clients and servers sdc-smartdc-7.1.1 Client SDK and CLI for the Joyent SmartDataCenter API smtools-20131031 Joyent tools relevant to SmartOS and SmartMachines sqlite3-3.8.0.2 SQL Database Engine in a C Library sudo-1.7.10p7 Allow others to run commands as root tcp_wrappers-7.6.4 Monitor and filter incoming requests for network services tiff-4.0.3nb6 Library and tools for reading and writing TIFF data files xmlcatmgr-2.2nb1 XML and SGML catalog manager xz-5.0.5 XZ utilities zlib-1.2.8 General purpose data compression library zoneinit-1.6.5 Joyent Instance initialization toolchain ```
1.0
Rubinius fails to compile: '_setjmp' was not declared in this scope - Hi I'm trying to compile Rubinius pulled from the git repo as of tonight within a Solaris zone and I keep getting the same error. I've tried the current git repo and the last two releases. I first attempted this with Ruby 1.9.3-xxx installed, but after reading on the Rubinius website, I restarted the template with a fresh Ruby 2.0.0-xxx install. They all error out in the same place: ``` Running 191 tasks using 1 parallel threads 2: CXX vm/accessor_primitives.cpp 2: CXX vm/agent.cpp 2: CXX vm/agent_components.cpp 2: CXX vm/arguments.cpp 2: CXX vm/auxiliary_threads.cpp 2: CXX vm/builtin/access_variable.cpp 2: CXX vm/builtin/alias.cpp 2: CXX vm/builtin/array.cpp 2: CXX vm/builtin/atomic.cpp 2: CXX vm/builtin/autoload.cpp 2: CXX vm/builtin/bignum.cpp 2: CXX vm/builtin/block_as_method.cpp 2: CXX vm/builtin/block_environment.cpp 2: CXX vm/builtin/byte_array.cpp 2: CXX vm/builtin/call_custom_cache.cpp 2: CXX vm/builtin/call_site.cpp 2: CXX vm/builtin/call_unit.cpp 2: CXX vm/builtin/call_unit_adapter.cpp 2: CXX vm/builtin/channel.cpp 2: CXX vm/builtin/character.cpp 2: CXX vm/builtin/class.cpp 2: CXX vm/builtin/compact_lookup_table.cpp 2: CXX vm/builtin/compiled_code.cpp 2: CXX vm/builtin/constant_cache.cpp 2: CXX vm/builtin/constant_scope.cpp 2: CXX vm/builtin/constant_table.cpp 2: CXX vm/builtin/data.cpp 2: CXX vm/builtin/dir.cpp 2: CXX vm/builtin/encoding.cpp 2: CXX vm/builtin/exception.cpp 2: CXX vm/builtin/executable.cpp 2: CXX vm/builtin/ffi_pointer.cpp 2: CXX vm/builtin/fiber.cpp 2: CXX vm/builtin/find_object.cpp 2: CXX vm/builtin/fixnum.cpp 2: CXX vm/builtin/float.cpp 2: CXX vm/builtin/heap_dump.cpp 2: CXX vm/builtin/immediates.cpp 2: CXX vm/builtin/integer.cpp 2: CXX vm/builtin/io.cpp 2: CXX vm/builtin/iseq.cpp 2: CXX vm/builtin/list.cpp 2: CXX vm/builtin/location.cpp 2: CXX vm/builtin/lookup_table.cpp 2: CXX vm/builtin/method_table.cpp 2: CXX vm/builtin/module.cpp 2: CXX vm/builtin/mono_inline_cache.cpp 2: CXX vm/builtin/native_function.cpp 2: CXX vm/builtin/native_method.cpp vm/builtin/native_method.cpp: In static member function 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&)': vm/builtin/native_method.cpp:691:9: error: there are no arguments to '_setjmp' that depend on a template parameter, so a declaration of '_setjmp' must be available [-fpermissive] vm/builtin/native_method.cpp:691:9: note: (if you use '-fpermissive', G++ will accept your code, but allowing the use of an undeclared name is deprecated) vm/builtin/native_method.cpp:702:9: error: there are no arguments to '_setjmp' that depend on a template parameter, so a declaration of '_setjmp' must be available [-fpermissive] vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::ZeroArguments; rubinius::Executable = rubinius::Executable]': vm/builtin/native_method.cpp:776:82: required from here vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:691:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:702:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::OneArgument; rubinius::Executable = rubinius::Executable]': vm/builtin/native_method.cpp:780:80: required from here vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:691:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:702:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::TwoArguments; rubinius::Executable = rubinius::Executable]': vm/builtin/native_method.cpp:784:81: required from here vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:691:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:702:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::ThreeArguments; rubinius::Executable = rubinius::Executable]': vm/builtin/native_method.cpp:788:83: required from here vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:691:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:702:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp: In instantiation of 'static rubinius::Object* rubinius::NativeMethod::executor_implementation(rubinius::State*, rubinius::CallFrame*, rubinius::Executable*, rubinius::Module*, rubinius::Arguments&) [with ArgumentHandler = rubinius::GenericArguments; rubinius::Executable = rubinius::Executable]': vm/builtin/native_method.cpp:792:85: required from here vm/builtin/native_method.cpp:691:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:691:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' vm/builtin/native_method.cpp:702:9: error: '_setjmp' was not declared in this scope vm/builtin/native_method.cpp:702:9: note: suggested alternative: In file included from /usr/include/setjmp.h:34:0, from /usr/local/src/rubinius/vm/vm.hpp:25, from /usr/local/src/rubinius/vm/builtin/object.hpp:6, from /usr/local/src/rubinius/vm/builtin/array.hpp:4, from /usr/local/src/rubinius/vm/arguments.hpp:5, from vm/builtin/native_method.cpp:1: /usr/include/iso/setjmp_iso.h:94:12: note: 'std::_setjmp' Error: g++ -I/usr/local/src/rubinius/vm -I/usr/local/src/rubinius/vm/include -I/usr/local/src/rubinius/vm/builtin -I. -Ivm/test/cxxtest -I/usr/local/src/rubinius/vendor/udis86 -I/usr/local/src/rubinius/vendor/libffi/include -Ivendor/double-conversion/src -DHAVE_CONFIG_H -I/usr/local/src/rubinius/vm/include/capi -I/usr/local/src/rubinius/vendor/oniguruma -I/usr/local/src/rubinius/vendor/libtommath -pipe -Wall -fno-omit-frame-pointer -g -I/usr/local/include -I/opt/local/include -fPIC -Wno-strict-aliasing -O2 -DHAS_EXECINFO -DHAVE_CLOCK_GETTIME -DHAVE_NL_LANGINFO -DHAVE_POSIX_FADVISE -DHAVE_STRNLEN -DHAVE_TIMEZONE -DHAVE_TZNAME -DHAVE_DAYLIGHT -DHAVE_ALLOCA_H -DHAVE_STRING_H -DHAVE_SYS_TIME_H -DHAVE_SYS_TIMES_H -DHAVE_SYS_TYPES_H -DHAVE_UNISTD_H -DHAVE_STDARG_H -I/usr/local/include -D_DEBUG -include llvm/Support/Solaris.h -D_GNU_SOURCE -fPIC -DENABLE_LLVM -Wno-unused-function -Werror -DRBX_PROFILER -D__STDC_LIMIT_MACROS -D__STDC_CONSTANT_MACROS -D_LARGEFILE_SOURCE -D_FILE_OFFSET_BITS=64 -fno-rtti -fvisibility-inlines-hidden -c -o vm/builtin/artifacts/native_method.cpp.o vm/builtin/native_method.cpp rake aborted! Error compiling Tasks: TOP => default => spec => build => build:build => vm/vm (See full trace by running task with --trace) ``` Installed packages: I built LLVM from source as it wasn't in the repos, using LLVM-3.3 ``` binutils-2.23.2nb1 GNU binary utilities bison-3.0 GNU yacc(1) replacement bmake-20110606nb1 Portable (autoconf) version of NetBSD 'make' utility bootstrap-mk-files-20130912 *.mk files for the bootstrap bmake utility bsdinstall-20130905 Portable version of the BSD install(1) program bzip2-1.0.6nb1 Block-sorting file compressor changepass-1.3.3 Simple tool to set user passwords non-interactively cloog-0.18.0nb1 Code generator for loop optimization (used by gcc) coreutils-8.13nb6 GNU basic file, shell and text manipulation utilities curl-7.32.0 Client that groks URLs cyrus-sasl-2.1.26nb3 Simple Authentication and Security Layer db4-4.8.30 Berkeley DB version 4 from Oracle diffutils-3.3 GNU diff utilities - find the differences between files duo-unix-1.9.5 2 Factor Authentication provided by Duo Security expat-2.1.0 XML parser library written in C findutils-4.2.33nb3 The GNU find, locate, updatedb, and xargs utilities fontconfig-2.10.95 Library for configuring and customizing font access freetype2-2.4.12 Font rendering engine and library API gawk-4.1.0 GNU awk gcc47-4.7.3nb1 The GNU Compiler Collection (GCC) - 4.7 Release Series gcc47-libs-4.7.3nb3 The GNU Compiler Collection (GCC) support shared libraries genbfpw-0 Simple tool to generate Blowfish crypts of passwords gettext-lib-0.18.3.1 Internationalized Message Handling Library (libintl) ghostscript-9.05nb6 Meta-package for installing the preferred ghostscript version ghostscript-fonts-8.11nb3 Postscript fonts for Aladdin Ghostscript ghostscript-gpl-9.05nb6 Postscript interpreter git-base-1.8.4nb1 GIT Tree History Storage Tool (base package) gmake-3.82nb7 GNU version of 'make' utility gmp-5.1.2 Library for arbitrary precision arithmetic grep-2.14nb2 GNU grep groff-1.20.1nb12 GNU roff text processing suite gsed-4.2.2nb4 GNU implementation of sed, the POSIX stream editor gtar-base-1.26nb2 The GNU tape archiver with remote magnetic tape support guardtime-1.0.12 Command line client for GuardTime keyless signature service isl-0.11.2 Integer set library required by gcc graphite jasper-1.900.1nb7 Software-based reference implementation of the JPEG-2000 codec jbigkit-2.0 JBIG-KIT lossless image compression library jpeg-9 IJG's jpeg compression utilities json-c-0.10 JSON library in C less-458 Pager similar to more and pg libICE-1.0.8 Inter Client Exchange (ICE) library for X libSM-1.2.2 X Session Management Library libX11-1.6.2 Base X libraries from modular Xorg X11 libXau-1.0.8 Authorization Protocol for X from X.org libXaw-1.0.12 X Athena Widgets Library from modular Xorg X11 libXdmcp-1.1.1 X Display Manager Control Protocol library from X.org libXext-1.3.2 X Extension library libXmu-1.1.2 X Miscellaneous Utilities library libXpm-3.5.11 X PixMap Library from modular Xorg X11 libXt-1.1.4 X Toolkit Intrinsics library libarchive-2.8.4nb2 Library to read/create different archive formats libee-0.4.1 Event Expression Library inspired by CEE libestr-0.1.5 Library for some string essentials libffi-3.0.13 Foreign function interface libgcrypt-1.5.3 GNU cryptographic library libgetopt-1.4.4 Library for handling --long options libgpg-error-1.12 Definitions of common error values for all GnuPG components libguardtime-0.3.11 GuardTime Client C SDK libiconv-1.14nb2 Character set conversion library libidn-1.28 Internationalized Domain Names command line tool liblognorm-0.3.7 Tool to normalize log data libpaper-1.1.24 Paper size handling library libssh2-1.2.2nb2 SSH2 protocol library libuuid-2.19.1 Generate unique identifiers for objects libxcb-1.9.1 X protocol C-language Binding libxml2-2.9.1 XML parser library from the GNOME project libyaml-0.1.4 YAML 1.1 parser and emitter written in C m4-1.4.16nb3 GNU version of UNIX m4 macro language processor mit-krb5-1.10.6 MIT Kerberos 5 authentication system mozilla-rootcerts-1.0.20121229nb1 Root CA certificates from the Mozilla Project mpfr-3.1.2 GMP-based library for multiple-precision floating-point computations nawk-20050424nb3 Brian Kernighan's pattern-directed scanning and processing language nbsed-20120308 NetBSD-current's sed(1) ncurses-5.9nb1 CRT screen handling and optimization package netpbm-10.35.80nb7 Toolkit for conversion of images between different formats nodejs-0.10.22 V8 JavaScript for clients and servers openldap-client-2.4.36 Lightweight Directory Access Protocol libraries and client programs openssl-1.0.1enb2 Secure Socket Layer and cryptographic library p5-Authen-SASL-2.16nb1 Perl module to handle SASL authentication p5-Digest-HMAC-1.03nb2 Perl5 module for HMAC p5-Digest-MD5-2.53 Perl5 extension interface for MD5 algorithm p5-Digest-SHA-5.85 Perl module for SHA-1/224/256/384/512 algorithms p5-Email-Valid-1.192 Perl5 module for testing validity of an email address p5-Error-0.17021 Perl extension module for try/throw/catch exception handling p5-GSSAPI-0.28nb4 Perl extension providing access to the GSSAPIv2 library p5-IO-Socket-INET6-2.71 Perl object interface for AF_INET|AF_INET6 domain sockets p5-IO-Socket-SSL-1.953 Perl5 SSL socket interface class p5-MIME-Base64-3.14 Perl5 module for Base64 and Quoted-Printable encodings p5-MailTools-2.12nb1 Perl5 modules related to mail applications p5-Net-DNS-0.72 Perl5 module for DNS resolution p5-Net-IP-1.26nb1 Perl extension for manipulating IPv4/IPv6 addresses p5-Net-LibIDN-0.12nb5 Perl bindings for GNU Libidn p5-Net-SMTP-SSL-1.01nb4 Perl5 module providing SSL support for Net::SMTP p5-Net-SSLeay-1.55 Perl5 module for using OpenSSL p5-Socket6-0.23nb4 Perl5 module to support getaddrinfo() and getnameinfo() p5-TimeDate-2.30 Perl5 TimeDate distribution patch-2.5.9nb2 Patch files using diff output pcre-8.33 Perl Compatible Regular Expressions library perl-5.18.1nb1 Practical Extraction and Report Language pkg_install-20130902 Package management and administration tools for pkgsrc pkg_install-info-4.5nb3 Standalone GNU info file installation utility pkgin-0.6.4nb1 Apt / yum like tool for managing pkgsrc binary packages png-1.6.6 Library for manipulating PNG images postfix-2.9.8 Postfix SMTP server and tools ppl-0.11.2nb2 The Parma Polyhedra Library. Used by gcc for loop optimization psutils-1.17nb4 Utilities for manipulating PostScript documents python27-2.7.5nb3 Interpreted, interactive, object-oriented programming language readline-6.2 GNU library that can recall and edit previous input rsyslog-7.4.4nb2 The enhanced syslogd for Unix ruby-2.0.0p353 Wrapper package for Ruby programming language ruby200-base-2.0.0p353 Ruby 2.0.0 release minimum base package ruby200-bundler-1.3.5 Manage your application's dependencies ruby200-rake-10.0.4 Ruby Make ruby200-rubygems-2.0.10 Ruby standard for publishing and managing third party libraries sdc-manta-1.2.2 Node.js SDK for Manta sdc-node-0.10.22 V8 JavaScript for clients and servers sdc-smartdc-7.1.1 Client SDK and CLI for the Joyent SmartDataCenter API smtools-20131031 Joyent tools relevant to SmartOS and SmartMachines sqlite3-3.8.0.2 SQL Database Engine in a C Library sudo-1.7.10p7 Allow others to run commands as root tcp_wrappers-7.6.4 Monitor and filter incoming requests for network services tiff-4.0.3nb6 Library and tools for reading and writing TIFF data files xmlcatmgr-2.2nb1 XML and SGML catalog manager xz-5.0.5 XZ utilities zlib-1.2.8 General purpose data compression library zoneinit-1.6.5 Joyent Instance initialization toolchain ```
build
rubinius fails to compile setjmp was not declared in this scope hi i m trying to compile rubinius pulled from the git repo as of tonight within a solaris zone and i keep getting the same error i ve tried the current git repo and the last two releases i first attempted this with ruby xxx installed but after reading on the rubinius website i restarted the template with a fresh ruby xxx install they all error out in the same place running tasks using parallel threads cxx vm accessor primitives cpp cxx vm agent cpp cxx vm agent components cpp cxx vm arguments cpp cxx vm auxiliary threads cpp cxx vm builtin access variable cpp cxx vm builtin alias cpp cxx vm builtin array cpp cxx vm builtin atomic cpp cxx vm builtin autoload cpp cxx vm builtin bignum cpp cxx vm builtin block as method cpp cxx vm builtin block environment cpp cxx vm builtin byte array cpp cxx vm builtin call custom cache cpp cxx vm builtin call site cpp cxx vm builtin call unit cpp cxx vm builtin call unit adapter cpp cxx vm builtin channel cpp cxx vm builtin character cpp cxx vm builtin class cpp cxx vm builtin compact lookup table cpp cxx vm builtin compiled code cpp cxx vm builtin constant cache cpp cxx vm builtin constant scope cpp cxx vm builtin constant table cpp cxx vm builtin data cpp cxx vm builtin dir cpp cxx vm builtin encoding cpp cxx vm builtin exception cpp cxx vm builtin executable cpp cxx vm builtin ffi pointer cpp cxx vm builtin fiber cpp cxx vm builtin find object cpp cxx vm builtin fixnum cpp cxx vm builtin float cpp cxx vm builtin heap dump cpp cxx vm builtin immediates cpp cxx vm builtin integer cpp cxx vm builtin io cpp cxx vm builtin iseq cpp cxx vm builtin list cpp cxx vm builtin location cpp cxx vm builtin lookup table cpp cxx vm builtin method table cpp cxx vm builtin module cpp cxx vm builtin mono inline cache cpp cxx vm builtin native function cpp cxx vm builtin native method cpp vm builtin native method cpp in static member function static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp error there are no arguments to setjmp that depend on a template parameter so a declaration of setjmp must be available vm builtin native method cpp note if you use fpermissive g will accept your code but allowing the use of an undeclared name is deprecated vm builtin native method cpp error there are no arguments to setjmp that depend on a template parameter so a declaration of setjmp must be available vm builtin native method cpp in instantiation of static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp required from here vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp in instantiation of static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp required from here vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp in instantiation of static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp required from here vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp in instantiation of static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp required from here vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp in instantiation of static rubinius object rubinius nativemethod executor implementation rubinius state rubinius callframe rubinius executable rubinius module rubinius arguments vm builtin native method cpp required from here vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp vm builtin native method cpp error setjmp was not declared in this scope vm builtin native method cpp note suggested alternative in file included from usr include setjmp h from usr local src rubinius vm vm hpp from usr local src rubinius vm builtin object hpp from usr local src rubinius vm builtin array hpp from usr local src rubinius vm arguments hpp from vm builtin native method cpp usr include iso setjmp iso h note std setjmp error g i usr local src rubinius vm i usr local src rubinius vm include i usr local src rubinius vm builtin i ivm test cxxtest i usr local src rubinius vendor i usr local src rubinius vendor libffi include ivendor double conversion src dhave config h i usr local src rubinius vm include capi i usr local src rubinius vendor oniguruma i usr local src rubinius vendor libtommath pipe wall fno omit frame pointer g i usr local include i opt local include fpic wno strict aliasing dhas execinfo dhave clock gettime dhave nl langinfo dhave posix fadvise dhave strnlen dhave timezone dhave tzname dhave daylight dhave alloca h dhave string h dhave sys time h dhave sys times h dhave sys types h dhave unistd h dhave stdarg h i usr local include d debug include llvm support solaris h d gnu source fpic denable llvm wno unused function werror drbx profiler d stdc limit macros d stdc constant macros d largefile source d file offset bits fno rtti fvisibility inlines hidden c o vm builtin artifacts native method cpp o vm builtin native method cpp rake aborted error compiling tasks top default spec build build build vm vm see full trace by running task with trace installed packages i built llvm from source as it wasn t in the repos using llvm binutils gnu binary utilities bison gnu yacc replacement bmake portable autoconf version of netbsd make utility bootstrap mk files mk files for the bootstrap bmake utility bsdinstall portable version of the bsd install program block sorting file compressor changepass simple tool to set user passwords non interactively cloog code generator for loop optimization used by gcc coreutils gnu basic file shell and text manipulation utilities curl client that groks urls cyrus sasl simple authentication and security layer berkeley db version from oracle diffutils gnu diff utilities find the differences between files duo unix factor authentication provided by duo security expat xml parser library written in c findutils the gnu find locate updatedb and xargs utilities fontconfig library for configuring and customizing font access font rendering engine and library api gawk gnu awk the gnu compiler collection gcc release series libs the gnu compiler collection gcc support shared libraries genbfpw simple tool to generate blowfish crypts of passwords gettext lib internationalized message handling library libintl ghostscript meta package for installing the preferred ghostscript version ghostscript fonts postscript fonts for aladdin ghostscript ghostscript gpl postscript interpreter git base git tree history storage tool base package gmake gnu version of make utility gmp library for arbitrary precision arithmetic grep gnu grep groff gnu roff text processing suite gsed gnu implementation of sed the posix stream editor gtar base the gnu tape archiver with remote magnetic tape support guardtime command line client for guardtime keyless signature service isl integer set library required by gcc graphite jasper software based reference implementation of the jpeg codec jbigkit jbig kit lossless image compression library jpeg ijg s jpeg compression utilities json c json library in c less pager similar to more and pg libice inter client exchange ice library for x libsm x session management library base x libraries from modular xorg libxau authorization protocol for x from x org libxaw x athena widgets library from modular xorg libxdmcp x display manager control protocol library from x org libxext x extension library libxmu x miscellaneous utilities library libxpm x pixmap library from modular xorg libxt x toolkit intrinsics library libarchive library to read create different archive formats libee event expression library inspired by cee libestr library for some string essentials libffi foreign function interface libgcrypt gnu cryptographic library libgetopt library for handling long options libgpg error definitions of common error values for all gnupg components libguardtime guardtime client c sdk libiconv character set conversion library libidn internationalized domain names command line tool liblognorm tool to normalize log data libpaper paper size handling library protocol library libuuid generate unique identifiers for objects libxcb x protocol c language binding xml parser library from the gnome project libyaml yaml parser and emitter written in c gnu version of unix macro language processor mit mit kerberos authentication system mozilla rootcerts root ca certificates from the mozilla project mpfr gmp based library for multiple precision floating point computations nawk brian kernighan s pattern directed scanning and processing language nbsed netbsd current s sed ncurses crt screen handling and optimization package netpbm toolkit for conversion of images between different formats nodejs javascript for clients and servers openldap client lightweight directory access protocol libraries and client programs openssl secure socket layer and cryptographic library authen sasl perl module to handle sasl authentication digest hmac module for hmac digest extension interface for algorithm digest sha perl module for sha algorithms email valid module for testing validity of an email address error perl extension module for try throw catch exception handling gssapi perl extension providing access to the library io socket perl object interface for af inet af domain sockets io socket ssl ssl socket interface class mime module for and quoted printable encodings mailtools modules related to mail applications net dns module for dns resolution net ip perl extension for manipulating addresses net libidn perl bindings for gnu libidn net smtp ssl module providing ssl support for net smtp net ssleay module for using openssl module to support getaddrinfo and getnameinfo timedate timedate distribution patch patch files using diff output pcre perl compatible regular expressions library perl practical extraction and report language pkg install package management and administration tools for pkgsrc pkg install info standalone gnu info file installation utility pkgin apt yum like tool for managing pkgsrc binary packages png library for manipulating png images postfix postfix smtp server and tools ppl the parma polyhedra library used by gcc for loop optimization psutils utilities for manipulating postscript documents interpreted interactive object oriented programming language readline gnu library that can recall and edit previous input rsyslog the enhanced syslogd for unix ruby wrapper package for ruby programming language base ruby release minimum base package bundler manage your application s dependencies rake ruby make rubygems ruby standard for publishing and managing third party libraries sdc manta node js sdk for manta sdc node javascript for clients and servers sdc smartdc client sdk and cli for the joyent smartdatacenter api smtools joyent tools relevant to smartos and smartmachines sql database engine in a c library sudo allow others to run commands as root tcp wrappers monitor and filter incoming requests for network services tiff library and tools for reading and writing tiff data files xmlcatmgr xml and sgml catalog manager xz xz utilities zlib general purpose data compression library zoneinit joyent instance initialization toolchain
1
167,764
13,041,626,133
IssuesEvent
2020-07-28 20:45:09
nextras/orm
https://api.github.com/repos/nextras/orm
closed
Removing entities in foreach
bug failing-testcase-needed
I get error `Property Smartlook\Domain\Project\ProjectPermission::$project is not nullable.`, when in 2nd foreach run call remove method. Code: ```php // un-assign user from projects foreach ($user->projectPermissions as $projectPermission) { $this->projectPermissionRepository->remove($projectPermission); } $this->userRepository->flush(); ``` Entity: ```php /** * @property-read int $id {primary-proxy} * @property Project $project {m:1 Project::$userPermissions} {primary} * @property User $user {m:1 User::$projectPermissions} {primary} * @property bool $isOwner * @property-read int $permission */ class ProjectPermission extends Entity ``` **Versions::** - Database: MySQL 5.5 - Orm: master - Dbal: master
1.0
Removing entities in foreach - I get error `Property Smartlook\Domain\Project\ProjectPermission::$project is not nullable.`, when in 2nd foreach run call remove method. Code: ```php // un-assign user from projects foreach ($user->projectPermissions as $projectPermission) { $this->projectPermissionRepository->remove($projectPermission); } $this->userRepository->flush(); ``` Entity: ```php /** * @property-read int $id {primary-proxy} * @property Project $project {m:1 Project::$userPermissions} {primary} * @property User $user {m:1 User::$projectPermissions} {primary} * @property bool $isOwner * @property-read int $permission */ class ProjectPermission extends Entity ``` **Versions::** - Database: MySQL 5.5 - Orm: master - Dbal: master
non_build
removing entities in foreach i get error property smartlook domain project projectpermission project is not nullable when in foreach run call remove method code php un assign user from projects foreach user projectpermissions as projectpermission this projectpermissionrepository remove projectpermission this userrepository flush entity php property read int id primary proxy property project project m project userpermissions primary property user user m user projectpermissions primary property bool isowner property read int permission class projectpermission extends entity versions database mysql orm master dbal master
0
127,235
12,310,004,304
IssuesEvent
2020-05-12 09:53:01
jstockwin/py-pdf-parser
https://api.github.com/repos/jstockwin/py-pdf-parser
closed
Add some examples to the documentation
Component - Documentation Difficulty - Easy Priority - Low Type - Enhancement
It would be good to have some example use cases in the documentation somewhere.
1.0
Add some examples to the documentation - It would be good to have some example use cases in the documentation somewhere.
non_build
add some examples to the documentation it would be good to have some example use cases in the documentation somewhere
0
48,541
12,214,707,488
IssuesEvent
2020-05-01 10:45:33
woocommerce/woocommerce-admin
https://api.github.com/repos/woocommerce/woocommerce-admin
closed
Webpack builds files in project root
Build [Type] Bug
**Describe the bug** Occassionally, the new optimized files introduced in https://github.com/woocommerce/woocommerce-admin/pull/4094 get built directly in the wc-admin root instead of the `dist` folder. This seems to happen on branch change. This may just be a point of frustration temporarily and not long-term since it's unlikely we'll be switching back and forth between branches with this change for too long. We may decide that this is an edge case and does not need a fix. If we do decide to fix this, my best guess is that this is stemming from the introduction of the `path` property in webpack `output`. **To Reproduce** Steps to reproduce the behavior: 1. Go to an older branch or commit. 1. Run `npm start`. 1. Switch to the current master. 1. Note the files are built in the root directory. **Expected behavior** Expected to only have those files built into `dist` folder. **Screenshots** <img width="518" alt="Screen Shot 2020-04-30 at 3 01 56 PM" src="https://user-images.githubusercontent.com/10561050/80713600-6ccd3180-8afc-11ea-9b79-f78b11521090.png">
1.0
Webpack builds files in project root - **Describe the bug** Occassionally, the new optimized files introduced in https://github.com/woocommerce/woocommerce-admin/pull/4094 get built directly in the wc-admin root instead of the `dist` folder. This seems to happen on branch change. This may just be a point of frustration temporarily and not long-term since it's unlikely we'll be switching back and forth between branches with this change for too long. We may decide that this is an edge case and does not need a fix. If we do decide to fix this, my best guess is that this is stemming from the introduction of the `path` property in webpack `output`. **To Reproduce** Steps to reproduce the behavior: 1. Go to an older branch or commit. 1. Run `npm start`. 1. Switch to the current master. 1. Note the files are built in the root directory. **Expected behavior** Expected to only have those files built into `dist` folder. **Screenshots** <img width="518" alt="Screen Shot 2020-04-30 at 3 01 56 PM" src="https://user-images.githubusercontent.com/10561050/80713600-6ccd3180-8afc-11ea-9b79-f78b11521090.png">
build
webpack builds files in project root describe the bug occassionally the new optimized files introduced in get built directly in the wc admin root instead of the dist folder this seems to happen on branch change this may just be a point of frustration temporarily and not long term since it s unlikely we ll be switching back and forth between branches with this change for too long we may decide that this is an edge case and does not need a fix if we do decide to fix this my best guess is that this is stemming from the introduction of the path property in webpack output to reproduce steps to reproduce the behavior go to an older branch or commit run npm start switch to the current master note the files are built in the root directory expected behavior expected to only have those files built into dist folder screenshots img width alt screen shot at pm src
1
411,081
27,812,515,618
IssuesEvent
2023-03-18 09:43:22
caodaion/caodaion.github.io
https://api.github.com/repos/caodaion/caodaion.github.io
closed
DOCUMENTATION | PROJECT DOCUMENTATION | Requirement specification
documentation
If you are leading a software development or system implementation project – such as an ERP rollout, then this is a must have document that is required to be prepared. A requirement specification is a detailed document that serves as a guide to the development team on the various features that need to be present in the system. Usually, a requirement specification follows a modular approach which means that you identify various modules (or processes) and then drill down/detail out the requirements at a detailed level. To describe each feature, a diagrammatic approach along with a detailed logic of the feature is usually the best way to go about. Examples of diagrams that may be included are: Context diagram: A bird’s eye level view of the module/feature showing the various users that interact with the system Process flow diagram: usually a flow chart or swim lane diagram defining the proces When you have had a business analyst prepare the above, what you have is the functional requirement – meaning the various features/functionality to be present in the system. However, it has been my experience that many times, people forget to write out non-functional requirements. A non functional requirement is, in short, a system requirement but it cannot be directly connected with any module/feature/functionality. Example of non-functional requirements are: - Number of concurrent users, of the system - Any particular performance criteria that is present e.g. a transaction should be approved/rejected in 20 seconds The next phase of the project after analysis is the design (or implementation) phase and in order to do a good design, non-functional requirements come in handy. The point that you have to remember is that if you have not catered to the non-functional requirements, then even if during the final user acceptance testing, it is seen that you have catered to all the functional requirements, even then your project will not get passed. In short, non-functional requirements are usually a very small part of the requirement, but they are perhaps the most neglected/least documented, and as a good project manager it is your job to document the non-functional requirements and see to it that the system design caters to these. (Check out this page from California Polytechnic State University for an [overview of non-functional requirements](http://users.csc.calpoly.edu/~jdalbey/SWE/QA/nonfunctional.html)). If you are looking for a Requirement Specification template, check out [my requirement specification template](https://www.tacticalprojectmanager.com/requirement-specification-template-requirement-numbering/) which supports automatic numbering of requirements and matching of customer & product requirements (to ensure traceability).
1.0
DOCUMENTATION | PROJECT DOCUMENTATION | Requirement specification - If you are leading a software development or system implementation project – such as an ERP rollout, then this is a must have document that is required to be prepared. A requirement specification is a detailed document that serves as a guide to the development team on the various features that need to be present in the system. Usually, a requirement specification follows a modular approach which means that you identify various modules (or processes) and then drill down/detail out the requirements at a detailed level. To describe each feature, a diagrammatic approach along with a detailed logic of the feature is usually the best way to go about. Examples of diagrams that may be included are: Context diagram: A bird’s eye level view of the module/feature showing the various users that interact with the system Process flow diagram: usually a flow chart or swim lane diagram defining the proces When you have had a business analyst prepare the above, what you have is the functional requirement – meaning the various features/functionality to be present in the system. However, it has been my experience that many times, people forget to write out non-functional requirements. A non functional requirement is, in short, a system requirement but it cannot be directly connected with any module/feature/functionality. Example of non-functional requirements are: - Number of concurrent users, of the system - Any particular performance criteria that is present e.g. a transaction should be approved/rejected in 20 seconds The next phase of the project after analysis is the design (or implementation) phase and in order to do a good design, non-functional requirements come in handy. The point that you have to remember is that if you have not catered to the non-functional requirements, then even if during the final user acceptance testing, it is seen that you have catered to all the functional requirements, even then your project will not get passed. In short, non-functional requirements are usually a very small part of the requirement, but they are perhaps the most neglected/least documented, and as a good project manager it is your job to document the non-functional requirements and see to it that the system design caters to these. (Check out this page from California Polytechnic State University for an [overview of non-functional requirements](http://users.csc.calpoly.edu/~jdalbey/SWE/QA/nonfunctional.html)). If you are looking for a Requirement Specification template, check out [my requirement specification template](https://www.tacticalprojectmanager.com/requirement-specification-template-requirement-numbering/) which supports automatic numbering of requirements and matching of customer & product requirements (to ensure traceability).
non_build
documentation project documentation requirement specification if you are leading a software development or system implementation project – such as an erp rollout then this is a must have document that is required to be prepared a requirement specification is a detailed document that serves as a guide to the development team on the various features that need to be present in the system usually a requirement specification follows a modular approach which means that you identify various modules or processes and then drill down detail out the requirements at a detailed level to describe each feature a diagrammatic approach along with a detailed logic of the feature is usually the best way to go about examples of diagrams that may be included are context diagram a bird’s eye level view of the module feature showing the various users that interact with the system process flow diagram usually a flow chart or swim lane diagram defining the proces when you have had a business analyst prepare the above what you have is the functional requirement – meaning the various features functionality to be present in the system however it has been my experience that many times people forget to write out non functional requirements a non functional requirement is in short a system requirement but it cannot be directly connected with any module feature functionality example of non functional requirements are number of concurrent users of the system any particular performance criteria that is present e g a transaction should be approved rejected in seconds the next phase of the project after analysis is the design or implementation phase and in order to do a good design non functional requirements come in handy the point that you have to remember is that if you have not catered to the non functional requirements then even if during the final user acceptance testing it is seen that you have catered to all the functional requirements even then your project will not get passed in short non functional requirements are usually a very small part of the requirement but they are perhaps the most neglected least documented and as a good project manager it is your job to document the non functional requirements and see to it that the system design caters to these check out this page from california polytechnic state university for an if you are looking for a requirement specification template check out which supports automatic numbering of requirements and matching of customer product requirements to ensure traceability
0
23,531
22,146,650,047
IssuesEvent
2022-06-03 12:46:28
WDscholia/scholia
https://api.github.com/repos/WDscholia/scholia
closed
Link co-occurring topics to topic comparison aspect
enhancement usability
Example screenshot from https://tools.wmflabs.org/scholia/topic/Q5227350 : <img width="1156" alt="Screen Shot 2020-05-29 at 22 58 18" src="https://user-images.githubusercontent.com/465923/83318149-e6d9fe80-a1ff-11ea-84c0-5909982827b8.png"> Probably best to link it via the count value, i.e. the "383" would become [383](https://tools.wmflabs.org/scholia/topics/Q5227350,Q45933174).
True
Link co-occurring topics to topic comparison aspect - Example screenshot from https://tools.wmflabs.org/scholia/topic/Q5227350 : <img width="1156" alt="Screen Shot 2020-05-29 at 22 58 18" src="https://user-images.githubusercontent.com/465923/83318149-e6d9fe80-a1ff-11ea-84c0-5909982827b8.png"> Probably best to link it via the count value, i.e. the "383" would become [383](https://tools.wmflabs.org/scholia/topics/Q5227350,Q45933174).
non_build
link co occurring topics to topic comparison aspect example screenshot from img width alt screen shot at src probably best to link it via the count value i e the would become
0
21,923
30,446,460,024
IssuesEvent
2023-07-15 18:31:24
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
pyutils 0.0.1b18 has 2 GuardDog issues
guarddog typosquatting silent-process-execution
https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b18", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:214", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpooezcz_t/pyutils" } }```
1.0
pyutils 0.0.1b18 has 2 GuardDog issues - https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b18", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:214", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpooezcz_t/pyutils" } }```
non_build
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt pytils python utils silent process execution location pyutils exec utils py pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmpooezcz t pyutils
0
32,866
8,966,831,388
IssuesEvent
2019-01-29 00:34:57
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
Issue with training of object detection api (ssd_mobilenet_v2_coco.config)
comp:apis stat:awaiting response type:build/install
Command Used: python train.py --logtostderr --train_dir=training/ --pipeline_config_path=training/ssd_mobilenet_v2_coco.config Please guide me if i am doing something wrong. Error Trace: Traceback (most recent call last): File "train.py", line 163, in <module> tf.app.run() File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/tensorflow/python/platform/app.py", line 125, in run _sys.exit(main(argv)) File "train.py", line 91, in main FLAGS.pipeline_config_path) File "/home/tayyab/Desktop/models/object_detection/utils/config_util.py", line 43, in get_configs_from_pipeline_file text_format.Merge(proto_str, pipeline_config) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 536, in Merge descriptor_pool=descriptor_pool) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 590, in MergeLines return parser.MergeLines(lines, message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 623, in MergeLines self._ParseOrMerge(lines, message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 638, in _ParseOrMerge self._MergeField(tokenizer, message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField merger(tokenizer, message, field) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField self._MergeField(tokenizer, sub_message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField merger(tokenizer, message, field) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField self._MergeField(tokenizer, sub_message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField merger(tokenizer, message, field) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField self._MergeField(tokenizer, sub_message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 730, in _MergeField (message_descriptor.full_name, name)) google.protobuf.text_format.ParseError: 86:7 : Message type "object_detection.protos.SsdFeatureExtractor" has no field named "use_depthwise".
1.0
Issue with training of object detection api (ssd_mobilenet_v2_coco.config) - Command Used: python train.py --logtostderr --train_dir=training/ --pipeline_config_path=training/ssd_mobilenet_v2_coco.config Please guide me if i am doing something wrong. Error Trace: Traceback (most recent call last): File "train.py", line 163, in <module> tf.app.run() File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/tensorflow/python/platform/app.py", line 125, in run _sys.exit(main(argv)) File "train.py", line 91, in main FLAGS.pipeline_config_path) File "/home/tayyab/Desktop/models/object_detection/utils/config_util.py", line 43, in get_configs_from_pipeline_file text_format.Merge(proto_str, pipeline_config) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 536, in Merge descriptor_pool=descriptor_pool) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 590, in MergeLines return parser.MergeLines(lines, message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 623, in MergeLines self._ParseOrMerge(lines, message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 638, in _ParseOrMerge self._MergeField(tokenizer, message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField merger(tokenizer, message, field) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField self._MergeField(tokenizer, sub_message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField merger(tokenizer, message, field) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField self._MergeField(tokenizer, sub_message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 763, in _MergeField merger(tokenizer, message, field) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 837, in _MergeMessageField self._MergeField(tokenizer, sub_message) File "/home/tayyab/Desktop/tensor_android/local/lib/python2.7/site-packages/google/protobuf/text_format.py", line 730, in _MergeField (message_descriptor.full_name, name)) google.protobuf.text_format.ParseError: 86:7 : Message type "object_detection.protos.SsdFeatureExtractor" has no field named "use_depthwise".
build
issue with training of object detection api ssd mobilenet coco config command used python train py logtostderr train dir training pipeline config path training ssd mobilenet coco config please guide me if i am doing something wrong error trace traceback most recent call last file train py line in tf app run file home tayyab desktop tensor android local lib site packages tensorflow python platform app py line in run sys exit main argv file train py line in main flags pipeline config path file home tayyab desktop models object detection utils config util py line in get configs from pipeline file text format merge proto str pipeline config file home tayyab desktop tensor android local lib site packages google protobuf text format py line in merge descriptor pool descriptor pool file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergelines return parser mergelines lines message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergelines self parseormerge lines message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in parseormerge self mergefield tokenizer message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergefield merger tokenizer message field file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergemessagefield self mergefield tokenizer sub message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergefield merger tokenizer message field file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergemessagefield self mergefield tokenizer sub message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergefield merger tokenizer message field file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergemessagefield self mergefield tokenizer sub message file home tayyab desktop tensor android local lib site packages google protobuf text format py line in mergefield message descriptor full name name google protobuf text format parseerror message type object detection protos ssdfeatureextractor has no field named use depthwise
1
456,821
13,151,008,267
IssuesEvent
2020-08-09 14:38:24
chrisjsewell/docutils
https://api.github.com/repos/chrisjsewell/docutils
opened
support unicode box characters for table markup [SF:feature-requests:6]
feature-requests open priority-5
author: trentbuck created: 2006-03-05 23:41:02 assigned: None SF_url: https://sourceforge.net/p/docutils/feature-requests/6 See first attachment. --- commenter: trentbuck posted: 2006-03-05 23:41:02 title: #6 support unicode box characters for table markup attachments: - https://sourceforge.net/p/docutils/feature-requests/_discuss/thread/03996de8/4305/attachment/tables.txt issue description --- commenter: goodger posted: 2006-03-06 20:39:33 title: #6 support unicode box characters for table markup Logged In: YES user\_id=7733 Unicode contains many characters that would be suitable for reST markup. Few are currently supported. Adding this feature would not be exceedingly difficult, but I'd classify it as low-priority. We depend on developers to implement features that interest them; patches are welcome\! Not a bug; changed to "feature request". --- commenter: fwiemann posted: 2006-05-03 21:42:09 title: #6 support unicode box characters for table markup Logged In: YES user\_id=1374215 This is not \*quite\* easy to implement because we'd need to distinguish different types of intersection characters: not only "+" for everything, but "┌", "┬", "┐", etc. --- commenter: milde posted: 2015-02-19 13:01:05.637000 title: #6 support unicode box characters for table markup It should suffice to map all intersection characters to "+".
1.0
support unicode box characters for table markup [SF:feature-requests:6] - author: trentbuck created: 2006-03-05 23:41:02 assigned: None SF_url: https://sourceforge.net/p/docutils/feature-requests/6 See first attachment. --- commenter: trentbuck posted: 2006-03-05 23:41:02 title: #6 support unicode box characters for table markup attachments: - https://sourceforge.net/p/docutils/feature-requests/_discuss/thread/03996de8/4305/attachment/tables.txt issue description --- commenter: goodger posted: 2006-03-06 20:39:33 title: #6 support unicode box characters for table markup Logged In: YES user\_id=7733 Unicode contains many characters that would be suitable for reST markup. Few are currently supported. Adding this feature would not be exceedingly difficult, but I'd classify it as low-priority. We depend on developers to implement features that interest them; patches are welcome\! Not a bug; changed to "feature request". --- commenter: fwiemann posted: 2006-05-03 21:42:09 title: #6 support unicode box characters for table markup Logged In: YES user\_id=1374215 This is not \*quite\* easy to implement because we'd need to distinguish different types of intersection characters: not only "+" for everything, but "┌", "┬", "┐", etc. --- commenter: milde posted: 2015-02-19 13:01:05.637000 title: #6 support unicode box characters for table markup It should suffice to map all intersection characters to "+".
non_build
support unicode box characters for table markup author trentbuck created assigned none sf url see first attachment commenter trentbuck posted title support unicode box characters for table markup attachments issue description commenter goodger posted title support unicode box characters for table markup logged in yes user id unicode contains many characters that would be suitable for rest markup few are currently supported adding this feature would not be exceedingly difficult but i d classify it as low priority we depend on developers to implement features that interest them patches are welcome not a bug changed to feature request commenter fwiemann posted title support unicode box characters for table markup logged in yes user id this is not quite easy to implement because we d need to distinguish different types of intersection characters not only for everything but ┌ ┬ ┐ etc commenter milde posted title support unicode box characters for table markup it should suffice to map all intersection characters to
0
21,738
7,065,112,867
IssuesEvent
2018-01-06 16:07:39
jupyterlab/jupyterlab
https://api.github.com/repos/jupyterlab/jupyterlab
closed
Rogue intransigent extension
status:Needs Discussion tag:Build System
#I was working with @ian-r-rose on the jupyterlab-latex extension, and in the course of attempting to debug it, we attempted to link to the extension in a number of ways. Most notably, we used a number of methods including - `jlpm install && jlpm run build && jupyterlab labextension install` - `jupyterlab labextension link` - `jlpm run add:sibling ../jupyterlab-latex` (from inside the jupyterlab repo) We were running `jupyterlab` both with the public packages (`jupyter lab --watch` with `jlpm run watch` in `jupyterlab-latex`) and (`jupyter lab --dev-mode --watch` after running `jlpm run add:sibling ../jupyterlab-latex`). Unfortunately, I've not been able to make the extension go away after many attempts in a variety of methods: - `jlpm run clean` (in both jupyterlab-latex and jupyterlab) - `jlpm run clean:slate` (in jupyterlab) - `git clean -xfd` (in jupyterlab-latex) - `git clean -xfd && jlpm install && jlpm run build` (in jupyterlab) - `jupyterlab labextension uninstall @jupyterlab/latex` - `jupyterlab labextension unlink @jupyterlab/latex` - `jupyter_conf_search latex` → leading to me removing the `"jupyterlab_latex": true,` line Even after all this I'm still getting the claim `@jupyterlab/latex` exists when I run `jupyter labextension list` ``` Build recommended: @jupyterlab/latex needs to be removed @jupyterlab/latex needs to be removed ``` and with `import ipdb; ipdb.set_trace()` inside the `jupyterlab/commands.py` `build_check` I'm still seeing both ``` 'extensions': {'@jupyterlab/application-extension': '', ⋮ '@jupyterlab/latex': 'lib/latex', ⋮ ``` ``` 'linkedPackages': {'@jupyterlab/latex': '/Users/mpacer/jupyter/jupyterlab-latex'}, ``` and ``` 'mimeExtensions': {'@jupyterlab/json-extension': '', '@jupyterlab/latex': 'lib/pdf', '@jupyterlab/pdf-extension': '', ⋮ ``` I'm going to keep trying to debug this but this is still way too complicated in order to figure this out. Happy to help figure out how to make this nicer.
1.0
Rogue intransigent extension - #I was working with @ian-r-rose on the jupyterlab-latex extension, and in the course of attempting to debug it, we attempted to link to the extension in a number of ways. Most notably, we used a number of methods including - `jlpm install && jlpm run build && jupyterlab labextension install` - `jupyterlab labextension link` - `jlpm run add:sibling ../jupyterlab-latex` (from inside the jupyterlab repo) We were running `jupyterlab` both with the public packages (`jupyter lab --watch` with `jlpm run watch` in `jupyterlab-latex`) and (`jupyter lab --dev-mode --watch` after running `jlpm run add:sibling ../jupyterlab-latex`). Unfortunately, I've not been able to make the extension go away after many attempts in a variety of methods: - `jlpm run clean` (in both jupyterlab-latex and jupyterlab) - `jlpm run clean:slate` (in jupyterlab) - `git clean -xfd` (in jupyterlab-latex) - `git clean -xfd && jlpm install && jlpm run build` (in jupyterlab) - `jupyterlab labextension uninstall @jupyterlab/latex` - `jupyterlab labextension unlink @jupyterlab/latex` - `jupyter_conf_search latex` → leading to me removing the `"jupyterlab_latex": true,` line Even after all this I'm still getting the claim `@jupyterlab/latex` exists when I run `jupyter labextension list` ``` Build recommended: @jupyterlab/latex needs to be removed @jupyterlab/latex needs to be removed ``` and with `import ipdb; ipdb.set_trace()` inside the `jupyterlab/commands.py` `build_check` I'm still seeing both ``` 'extensions': {'@jupyterlab/application-extension': '', ⋮ '@jupyterlab/latex': 'lib/latex', ⋮ ``` ``` 'linkedPackages': {'@jupyterlab/latex': '/Users/mpacer/jupyter/jupyterlab-latex'}, ``` and ``` 'mimeExtensions': {'@jupyterlab/json-extension': '', '@jupyterlab/latex': 'lib/pdf', '@jupyterlab/pdf-extension': '', ⋮ ``` I'm going to keep trying to debug this but this is still way too complicated in order to figure this out. Happy to help figure out how to make this nicer.
build
rogue intransigent extension i was working with ian r rose on the jupyterlab latex extension and in the course of attempting to debug it we attempted to link to the extension in a number of ways most notably we used a number of methods including jlpm install jlpm run build jupyterlab labextension install jupyterlab labextension link jlpm run add sibling jupyterlab latex from inside the jupyterlab repo we were running jupyterlab both with the public packages jupyter lab watch with jlpm run watch in jupyterlab latex and jupyter lab dev mode watch after running jlpm run add sibling jupyterlab latex unfortunately i ve not been able to make the extension go away after many attempts in a variety of methods jlpm run clean in both jupyterlab latex and jupyterlab jlpm run clean slate in jupyterlab git clean xfd in jupyterlab latex git clean xfd jlpm install jlpm run build in jupyterlab jupyterlab labextension uninstall jupyterlab latex jupyterlab labextension unlink jupyterlab latex jupyter conf search latex → leading to me removing the jupyterlab latex true line even after all this i m still getting the claim jupyterlab latex exists when i run jupyter labextension list build recommended jupyterlab latex needs to be removed jupyterlab latex needs to be removed and with import ipdb ipdb set trace inside the jupyterlab commands py build check i m still seeing both extensions jupyterlab application extension ⋮ jupyterlab latex lib latex ⋮ linkedpackages jupyterlab latex users mpacer jupyter jupyterlab latex and mimeextensions jupyterlab json extension jupyterlab latex lib pdf jupyterlab pdf extension ⋮ i m going to keep trying to debug this but this is still way too complicated in order to figure this out happy to help figure out how to make this nicer
1
289,741
21,790,654,232
IssuesEvent
2022-05-14 21:10:23
sirfuzzalot/textual-inputs
https://api.github.com/repos/sirfuzzalot/textual-inputs
closed
Add Contributing Guide
documentation
**USER STORY** As a developer I want clear guidance on how I may contribute my work to this project **REQUIREMENTS** - add a contributing guide
1.0
Add Contributing Guide - **USER STORY** As a developer I want clear guidance on how I may contribute my work to this project **REQUIREMENTS** - add a contributing guide
non_build
add contributing guide user story as a developer i want clear guidance on how i may contribute my work to this project requirements add a contributing guide
0
84,844
10,568,085,964
IssuesEvent
2019-10-06 10:21:55
elementary/files
https://api.github.com/repos/elementary/files
opened
Give aggregate info in overlay in ListView
Needs Design Priority: Wishlist
## Prerequisites - [* ] I have searched open and closed issues for duplicates. ## Feature In ListView the information overlay is suppressed. This makes sense for single files as the same information is available in the view. However, it would be useful, and more consistent to show aggregate information when multiple files are selected, as happens in the other views. **Describe the solution you'd like** Do not suppress the information overlay when more than one file is selected when in ListView. **Existing work** All the required code exists in Files - it should just require tweaking the suppression logic.
1.0
Give aggregate info in overlay in ListView - ## Prerequisites - [* ] I have searched open and closed issues for duplicates. ## Feature In ListView the information overlay is suppressed. This makes sense for single files as the same information is available in the view. However, it would be useful, and more consistent to show aggregate information when multiple files are selected, as happens in the other views. **Describe the solution you'd like** Do not suppress the information overlay when more than one file is selected when in ListView. **Existing work** All the required code exists in Files - it should just require tweaking the suppression logic.
non_build
give aggregate info in overlay in listview prerequisites i have searched open and closed issues for duplicates feature in listview the information overlay is suppressed this makes sense for single files as the same information is available in the view however it would be useful and more consistent to show aggregate information when multiple files are selected as happens in the other views describe the solution you d like do not suppress the information overlay when more than one file is selected when in listview existing work all the required code exists in files it should just require tweaking the suppression logic
0
37,403
15,285,064,554
IssuesEvent
2021-02-23 13:05:04
alexwonet/challenge-new
https://api.github.com/repos/alexwonet/challenge-new
closed
Your Own IOT Service
Your Own IOT Service
# Cloud deployment Your challenge is to develop and deploy your own IoT service to the SAP Cloud Platform CloudFoundry environment. ## Step 1 - Prepare service First we must make sure that your new service implements the Open Service Broker API. ### The Open Service Broker API spec and what it's good for https://ict.swisscom.ch/2019/12/open-service-broker-api/ https://github.com/openservicebrokerapi/servicebroker/blob/master/spec.md ### Endpoints to implement: - Catalog (GET) - Provision (PUT) - Bind (GET, PUT) --- ### Requirements 1.1. Your service should allow to create a service instance that match an active device on http://localhost:9090/device. 1.2. The device binding should return the full device credentials as ``` { "credentials": { "connection_string": "device://438bcdca-c714-4c4a-8ba0-da975dec49f1/fNhptfPn", "device_type": "ROUTER", "id": "438bcdca-c714-4c4a-8ba0-da975dec49f1", "state": "Online", } } ``` 1.3. Your service should support basic authentication and allow access for admin:admin. https://docs.iris-go.com/iris/security/basicauth 1.4. The create instance endpoint should support the device_type service instance param (ROUTER,MODEM,SWITCH) See provisioning parameters: https://github.com/openservicebrokerapi/servicebroker/blob/master/spec.md#provisioning ### Recommended Golang libs https://github.com/pmorie/osb-broker-lib This library is most useful if you want to build your own broker from scratch Golang server library: https://www.iris-go.com/docs/# Connecting to the /device endpoint https://github.com/go-resty/resty osb-broker-lib with Golang server example: ```go package main import ( "github.com/fatih/structs" "github.com/gavv/httpexpect" "github.com/kataras/iris/v12" "github.com/onsi/ginkgo" "github.wdf.sap.corp/cloudenablers/challenger/pkg/types" "github.wdf.sap.corp/cloudenablers/challenger/test/utils" "gopkg.in/yaml.v3" "net" "net/http" "time" osb "sigs.k8s.io/go-open-service-broker-client/v2" ) func main() { app := iris.New() brokerAPIs := app.Party("/v2") { brokerAPIs.Get("/catalog", GetCatalog) brokerAPIs.Get("/service_instances/{instance_id:string}/service_bindings/{binding_id:string}", GetServiceBinding) brokerAPIs.Put("/service_instances/{instance_id:string}", CreateServiceInstance) brokerAPIs.Put("/service_instances/{instance_id:string}/service_bindings/{binding_id:string}", CreateServiceBinding) } serverPort := ":" + os.Getenv("PORT") app.Listen(serverPort) } func CreateServiceInstance(ctx iris.Context) { provisionDetails := &osb.ProvisionRequest{} instanceID = ctx.Params().Get("instance_id") deviceType := provisionDetails.Parameters["device_type"] iotBackendURL := "http://localhost:9090" apiCookieKey := "YOUR-API_COOKIE" reqBody, _ := ctx.GetBody() json.Unmarshal(reqBody, provisionDetails) //find device that matches thew provided deviceType in the provisionDetails deviceType := provisionDetails.Parameters["device_type"] client := getClient(iotBackendURL, apiCookieKey) pages := utils.GetDevicesPages(client) // Replace with your client impl (resty) for _, item := range pages {} // go over the device service pages result ctx.JSON(&osb.ProvisionResponse{}) } func GetServiceBinding(ctx iris.Context) { ctx.JSON(&osb.GetBindingResponse{ Credentials: structs.Map(foundItem), }) } func CreateServiceBinding(ctx iris.Context) { ctx.JSON(&osb.BindResponse{ Credentials: structs.Map(foundItem), }) } func GetCatalog(ctx iris.Context) { response := &osb.CatalogResponse{} data := ` --- services: - name: your-iot-service id: 4f6e6cf6-ffdd-425f-a2c7-3c9258ad246e-alexwonet description: The example service! bindable: true metadata: displayName: "Example service" imageUrl: https://avatars2.githubusercontent.com/u/19862012?s=200&v=4 plans: - name: iot-service id: iot-service-alexwonet description: The default plan for the service free: true ` yaml.Unmarshal([]byte(data), &response) ctx.JSON(response) } ``` ### Running Golang apps on cloudfoundry: Define a correct port for the server to listen on: ``` serverPort := ":" + os.Getenv("PORT") app.Listen(serverPort) ``` ### define your cloudfoundry manifest.yml ``` applications: - name: alexwonet-challenge-app instances: 1 buildpacks: - go_buildpack ``` ## Step 2 - Deploy Your Service Create a new [trial account](https://www.sap.com/products/cloud-platform.html) on SAP CF - (Hint) [tutorial](https://developers.sap.com/tutorials/hcp-create-trial-account.html). [Deploy](https://docs.cloudfoundry.org/devguide/deploy-apps/manifest.html) the service and create a [service broker](https://docs.cloudfoundry.org/services/managing-service-brokers.html). Lastly, create a [service instance](https://docs.cloudfoundry.org/devguide/services/managing-services.html) and [bind](http://cli.cloudfoundry.org/en-US/v6/bind-service.html) the app! If all works well go back to http://localhost:9090 and try out your service. ### Try out will try to confirm that 1. We can create a service instance -> select a real device from the "/device" list api 2. Bind to it -> get the credentials to use the device 3. Confirm the device actually matches an active & cool device (not above 29°C) ## Step 3 - Validate & pass CI In your selected "challenge-new" repository create: `.github/workflows/golangci-lint.yml` ``` GET /device/{id}/status name: golangci-lint on: push: tags: - v* branches: - master - main pull_request: release: types: # This configuration does not affect the page_build event above - created jobs: golangci: name: lint runs-on: ubuntu-latest steps: - uses: actions/checkout@v2 - name: golangci-lint uses: golangci/golangci-lint-action@v2 with: # Required: the version of golangci-lint is required and must be specified without patch version: we always use the latest patch version. version: v1.33 # Optional: working directory, useful for monorepos # working-directory: somedir # Optional: golangci-lint command line arguments. # args: --issues-exit-code=0 # Optional: show only new issues if it's a pull request. The default value is `false`. # only-new-issues: true ``` 3.1 Make sure that your go code for this challenge is in the repository master branch. 3.2 Fix all lint issues that golangci will report (we want our code to be clean) 3.3 Check that the CI pass (In github actions) & Create a new release If all works well go back to http://localhost:9090 and validate your release pass CI! ### Use full Links https://github.com/golangci/golangci-lint https://github.com/features/actions Good luck!!!
1.0
Your Own IOT Service - # Cloud deployment Your challenge is to develop and deploy your own IoT service to the SAP Cloud Platform CloudFoundry environment. ## Step 1 - Prepare service First we must make sure that your new service implements the Open Service Broker API. ### The Open Service Broker API spec and what it's good for https://ict.swisscom.ch/2019/12/open-service-broker-api/ https://github.com/openservicebrokerapi/servicebroker/blob/master/spec.md ### Endpoints to implement: - Catalog (GET) - Provision (PUT) - Bind (GET, PUT) --- ### Requirements 1.1. Your service should allow to create a service instance that match an active device on http://localhost:9090/device. 1.2. The device binding should return the full device credentials as ``` { "credentials": { "connection_string": "device://438bcdca-c714-4c4a-8ba0-da975dec49f1/fNhptfPn", "device_type": "ROUTER", "id": "438bcdca-c714-4c4a-8ba0-da975dec49f1", "state": "Online", } } ``` 1.3. Your service should support basic authentication and allow access for admin:admin. https://docs.iris-go.com/iris/security/basicauth 1.4. The create instance endpoint should support the device_type service instance param (ROUTER,MODEM,SWITCH) See provisioning parameters: https://github.com/openservicebrokerapi/servicebroker/blob/master/spec.md#provisioning ### Recommended Golang libs https://github.com/pmorie/osb-broker-lib This library is most useful if you want to build your own broker from scratch Golang server library: https://www.iris-go.com/docs/# Connecting to the /device endpoint https://github.com/go-resty/resty osb-broker-lib with Golang server example: ```go package main import ( "github.com/fatih/structs" "github.com/gavv/httpexpect" "github.com/kataras/iris/v12" "github.com/onsi/ginkgo" "github.wdf.sap.corp/cloudenablers/challenger/pkg/types" "github.wdf.sap.corp/cloudenablers/challenger/test/utils" "gopkg.in/yaml.v3" "net" "net/http" "time" osb "sigs.k8s.io/go-open-service-broker-client/v2" ) func main() { app := iris.New() brokerAPIs := app.Party("/v2") { brokerAPIs.Get("/catalog", GetCatalog) brokerAPIs.Get("/service_instances/{instance_id:string}/service_bindings/{binding_id:string}", GetServiceBinding) brokerAPIs.Put("/service_instances/{instance_id:string}", CreateServiceInstance) brokerAPIs.Put("/service_instances/{instance_id:string}/service_bindings/{binding_id:string}", CreateServiceBinding) } serverPort := ":" + os.Getenv("PORT") app.Listen(serverPort) } func CreateServiceInstance(ctx iris.Context) { provisionDetails := &osb.ProvisionRequest{} instanceID = ctx.Params().Get("instance_id") deviceType := provisionDetails.Parameters["device_type"] iotBackendURL := "http://localhost:9090" apiCookieKey := "YOUR-API_COOKIE" reqBody, _ := ctx.GetBody() json.Unmarshal(reqBody, provisionDetails) //find device that matches thew provided deviceType in the provisionDetails deviceType := provisionDetails.Parameters["device_type"] client := getClient(iotBackendURL, apiCookieKey) pages := utils.GetDevicesPages(client) // Replace with your client impl (resty) for _, item := range pages {} // go over the device service pages result ctx.JSON(&osb.ProvisionResponse{}) } func GetServiceBinding(ctx iris.Context) { ctx.JSON(&osb.GetBindingResponse{ Credentials: structs.Map(foundItem), }) } func CreateServiceBinding(ctx iris.Context) { ctx.JSON(&osb.BindResponse{ Credentials: structs.Map(foundItem), }) } func GetCatalog(ctx iris.Context) { response := &osb.CatalogResponse{} data := ` --- services: - name: your-iot-service id: 4f6e6cf6-ffdd-425f-a2c7-3c9258ad246e-alexwonet description: The example service! bindable: true metadata: displayName: "Example service" imageUrl: https://avatars2.githubusercontent.com/u/19862012?s=200&v=4 plans: - name: iot-service id: iot-service-alexwonet description: The default plan for the service free: true ` yaml.Unmarshal([]byte(data), &response) ctx.JSON(response) } ``` ### Running Golang apps on cloudfoundry: Define a correct port for the server to listen on: ``` serverPort := ":" + os.Getenv("PORT") app.Listen(serverPort) ``` ### define your cloudfoundry manifest.yml ``` applications: - name: alexwonet-challenge-app instances: 1 buildpacks: - go_buildpack ``` ## Step 2 - Deploy Your Service Create a new [trial account](https://www.sap.com/products/cloud-platform.html) on SAP CF - (Hint) [tutorial](https://developers.sap.com/tutorials/hcp-create-trial-account.html). [Deploy](https://docs.cloudfoundry.org/devguide/deploy-apps/manifest.html) the service and create a [service broker](https://docs.cloudfoundry.org/services/managing-service-brokers.html). Lastly, create a [service instance](https://docs.cloudfoundry.org/devguide/services/managing-services.html) and [bind](http://cli.cloudfoundry.org/en-US/v6/bind-service.html) the app! If all works well go back to http://localhost:9090 and try out your service. ### Try out will try to confirm that 1. We can create a service instance -> select a real device from the "/device" list api 2. Bind to it -> get the credentials to use the device 3. Confirm the device actually matches an active & cool device (not above 29°C) ## Step 3 - Validate & pass CI In your selected "challenge-new" repository create: `.github/workflows/golangci-lint.yml` ``` GET /device/{id}/status name: golangci-lint on: push: tags: - v* branches: - master - main pull_request: release: types: # This configuration does not affect the page_build event above - created jobs: golangci: name: lint runs-on: ubuntu-latest steps: - uses: actions/checkout@v2 - name: golangci-lint uses: golangci/golangci-lint-action@v2 with: # Required: the version of golangci-lint is required and must be specified without patch version: we always use the latest patch version. version: v1.33 # Optional: working directory, useful for monorepos # working-directory: somedir # Optional: golangci-lint command line arguments. # args: --issues-exit-code=0 # Optional: show only new issues if it's a pull request. The default value is `false`. # only-new-issues: true ``` 3.1 Make sure that your go code for this challenge is in the repository master branch. 3.2 Fix all lint issues that golangci will report (we want our code to be clean) 3.3 Check that the CI pass (In github actions) & Create a new release If all works well go back to http://localhost:9090 and validate your release pass CI! ### Use full Links https://github.com/golangci/golangci-lint https://github.com/features/actions Good luck!!!
non_build
your own iot service cloud deployment your challenge is to develop and deploy your own iot service to the sap cloud platform cloudfoundry environment step prepare service first we must make sure that your new service implements the open service broker api the open service broker api spec and what it s good for endpoints to implement catalog get provision put bind get put requirements your service should allow to create a service instance that match an active device on the device binding should return the full device credentials as credentials connection string device fnhptfpn device type router id state online your service should support basic authentication and allow access for admin admin the create instance endpoint should support the device type service instance param router modem switch see provisioning parameters recommended golang libs this library is most useful if you want to build your own broker from scratch golang server library connecting to the device endpoint osb broker lib with golang server example go package main import github com fatih structs github com gavv httpexpect github com kataras iris github com onsi ginkgo github wdf sap corp cloudenablers challenger pkg types github wdf sap corp cloudenablers challenger test utils gopkg in yaml net net http time osb sigs io go open service broker client func main app iris new brokerapis app party brokerapis get catalog getcatalog brokerapis get service instances instance id string service bindings binding id string getservicebinding brokerapis put service instances instance id string createserviceinstance brokerapis put service instances instance id string service bindings binding id string createservicebinding serverport os getenv port app listen serverport func createserviceinstance ctx iris context provisiondetails osb provisionrequest instanceid ctx params get instance id devicetype provisiondetails parameters iotbackendurl apicookiekey your api cookie reqbody ctx getbody json unmarshal reqbody provisiondetails find device that matches thew provided devicetype in the provisiondetails devicetype provisiondetails parameters client getclient iotbackendurl apicookiekey pages utils getdevicespages client replace with your client impl resty for item range pages go over the device service pages result ctx json osb provisionresponse func getservicebinding ctx iris context ctx json osb getbindingresponse credentials structs map founditem func createservicebinding ctx iris context ctx json osb bindresponse credentials structs map founditem func getcatalog ctx iris context response osb catalogresponse data services name your iot service id ffdd alexwonet description the example service bindable true metadata displayname example service imageurl plans name iot service id iot service alexwonet description the default plan for the service free true yaml unmarshal byte data response ctx json response running golang apps on cloudfoundry define a correct port for the server to listen on serverport os getenv port app listen serverport define your cloudfoundry manifest yml applications name alexwonet challenge app instances buildpacks go buildpack step deploy your service create a new on sap cf hint the service and create a lastly create a and the app if all works well go back to and try out your service try out will try to confirm that we can create a service instance select a real device from the device list api bind to it get the credentials to use the device confirm the device actually matches an active cool device not above °c step validate pass ci in your selected challenge new repository create github workflows golangci lint yml get device id status name golangci lint on push tags v branches master main pull request release types this configuration does not affect the page build event above created jobs golangci name lint runs on ubuntu latest steps uses actions checkout name golangci lint uses golangci golangci lint action with required the version of golangci lint is required and must be specified without patch version we always use the latest patch version version optional working directory useful for monorepos working directory somedir optional golangci lint command line arguments args issues exit code optional show only new issues if it s a pull request the default value is false only new issues true make sure that your go code for this challenge is in the repository master branch fix all lint issues that golangci will report we want our code to be clean check that the ci pass in github actions create a new release if all works well go back to and validate your release pass ci use full links good luck
0
93,383
26,935,230,333
IssuesEvent
2023-02-07 20:03:40
openhwgroup/cva6
https://api.github.com/repos/openhwgroup/cva6
closed
Unable to view waveform of variable in Questasim
Component:Tool-and-build Type:Question Status:Invalid
In corev_apu/fpga/src, the axi2apb directory contains the axi2apb_64_32.sv. axi2apb_64_32.sv has two variables CS, NS of type enum logic. I am trying to simulate a testbench for this module using Questasim. As shown in the attached screenshot, these two variables CS, NS don't appear in the simulation window. I wish to view the waveforms for these two variables. Can you please advise how to view them in Questasim? ![Screenshot from 2022-09-21 19-46-27](https://user-images.githubusercontent.com/114074746/191496169-8c9c3557-9c0b-4607-bdfb-b00424d16f0b.png)
1.0
Unable to view waveform of variable in Questasim - In corev_apu/fpga/src, the axi2apb directory contains the axi2apb_64_32.sv. axi2apb_64_32.sv has two variables CS, NS of type enum logic. I am trying to simulate a testbench for this module using Questasim. As shown in the attached screenshot, these two variables CS, NS don't appear in the simulation window. I wish to view the waveforms for these two variables. Can you please advise how to view them in Questasim? ![Screenshot from 2022-09-21 19-46-27](https://user-images.githubusercontent.com/114074746/191496169-8c9c3557-9c0b-4607-bdfb-b00424d16f0b.png)
build
unable to view waveform of variable in questasim in corev apu fpga src the directory contains the sv sv has two variables cs ns of type enum logic i am trying to simulate a testbench for this module using questasim as shown in the attached screenshot these two variables cs ns don t appear in the simulation window i wish to view the waveforms for these two variables can you please advise how to view them in questasim
1
538,880
15,780,282,852
IssuesEvent
2021-04-01 09:45:44
sopra-fs21-group-03/Server
https://api.github.com/repos/sopra-fs21-group-03/Server
closed
Input in fields is handled correctly and wehen registering/loging in the correct api call is made
medium priority task
Time estimate: 2h This task is part of user story #1
1.0
Input in fields is handled correctly and wehen registering/loging in the correct api call is made - Time estimate: 2h This task is part of user story #1
non_build
input in fields is handled correctly and wehen registering loging in the correct api call is made time estimate this task is part of user story
0
821,440
30,823,411,723
IssuesEvent
2023-08-01 18:03:55
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
profiler: TestAgentIntegration failed
type: bug priority: p1 api: cloudprofiler flakybot: issue
Note: #7152 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: b2964a92d093a3d751e8b43aa70471bd989b1244 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f532917b-d222-41c3-89ee-274229ab20e1), [Sponge](http://sponge2/f532917b-d222-41c3-89ee-274229ab20e1) status: failed <details><summary>Test output</summary><br><pre> proftest.go:481: Serial port output for profiler-test-gomaster-2023-07-20-15-47-00-115169-pdt: \033[2J\033[01;01H\033[=3h\033[2J\033[01;01H\033[2J\033[01;01H\033[=3h\033[2J\033[01;01HBdsDxe: loading Boot0001 "UEFI Google PersistentDisk " from PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0) BdsDxe: starting Boot0001 "UEFI Google PersistentDisk " from PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0) UEFI: Attempting to start image. Description: UEFI Google PersistentDisk FilePath: PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0) OptionNumber: 1. \033[0m\033[30m\033[47mWelcome to GRUB! \033[0m\033[37m\033[40m\033[0m\033[30m\033[40m\033[2J\033[01;01H\033[0m\033[37m\033[40m\033[0m\033[30m\033[40m\033[2J\033[01;01H\033[0m\033[37m\033[40m Booting `Debian GNU/Linux' Loading Linux 5.10.0-23-cloud-amd64 ... Loading initial ramdisk ... error: no suitable video mode found. Booting in blind mode Thu Jul 20 22:47:26 UTC 2023: + retry apt-get update Thu Jul 20 22:47:26 UTC 2023: + for i in {1..3} Thu Jul 20 22:47:26 UTC 2023: + '[' 1 == 1 ']' Thu Jul 20 22:47:26 UTC 2023: + apt-get update Thu Jul 20 22:47:28 UTC 2023: + return 0 Thu Jul 20 22:47:28 UTC 2023: + retry apt-get -y -q install git Thu Jul 20 22:47:28 UTC 2023: + for i in {1..3} Thu Jul 20 22:47:28 UTC 2023: + '[' 1 == 1 ']' Thu Jul 20 22:47:28 UTC 2023: + apt-get -y -q install git Thu Jul 20 22:47:29 UTC 2023: debconf: unable to initialize frontend: Dialog Thu Jul 20 22:47:29 UTC 2023: debconf: (TERM is not set, so the dialog frontend is not usable.) Thu Jul 20 22:47:29 UTC 2023: debconf: falling back to frontend: Readline Thu Jul 20 22:47:29 UTC 2023: debconf: unable to initialize frontend: Readline Thu Jul 20 22:47:29 UTC 2023: debconf: (Can't locate Term/ReadLine.pm in @INC (you may need to install the Term::ReadLine module) (@INC contains: /etc/perl /usr/local/lib/x86_64-linux-gnu/perl/5.32.1 /usr/local/share/perl/5.32.1 /usr/lib/x86_64-linux-gnu/perl5/5.32 /usr/share/perl5 /usr/lib/x86_64-linux-gnu/perl-base /usr/lib/x86_64-linux-gnu/perl/5.32 /usr/share/perl/5.32 /usr/local/lib/site_perl) at /usr/share/perl5/Debconf/FrontEnd/Readline.pm line 7, <> line 8.) Thu Jul 20 22:47:29 UTC 2023: debconf: falling back to frontend: Teletype Thu Jul 20 22:47:29 UTC 2023: dpkg-preconfigure: unable to re-open stdin: Thu Jul 20 22:47:39 UTC 2023: + return 0 Thu Jul 20 22:47:39 UTC 2023: + mkdir -p /tmp/gocache Thu Jul 20 22:47:39 UTC 2023: + export GOCACHE=/tmp/gocache Thu Jul 20 22:47:39 UTC 2023: + GOCACHE=/tmp/gocache Thu Jul 20 22:47:39 UTC 2023: + '[' master = master ']' Thu Jul 20 22:47:39 UTC 2023: + retry apt-get -y -q install gcc Thu Jul 20 22:47:39 UTC 2023: + for i in {1..3} Thu Jul 20 22:47:39 UTC 2023: + '[' 1 == 1 ']' Thu Jul 20 22:47:39 UTC 2023: + apt-get -y -q install gcc Thu Jul 20 22:47:40 UTC 2023: debconf: unable to initialize frontend: Dialog Thu Jul 20 22:47:40 UTC 2023: debconf: (TERM is not set, so the dialog frontend is not usable.) Thu Jul 20 22:47:40 UTC 2023: debconf: falling back to frontend: Readline Thu Jul 20 22:47:40 UTC 2023: debconf: unable to initialize frontend: Readline Thu Jul 20 22:47:40 UTC 2023: debconf: (This frontend requires a controlling tty.) Thu Jul 20 22:47:40 UTC 2023: debconf: falling back to frontend: Teletype Thu Jul 20 22:47:40 UTC 2023: dpkg-preconfigure: unable to re-open stdin: Thu Jul 20 22:48:13 UTC 2023: + return 0 Thu Jul 20 22:48:13 UTC 2023: + mkdir -p /tmp/bin Thu Jul 20 22:48:13 UTC 2023: + retry curl -sL -o /tmp/bin/gimme https://raw.githubusercontent.com/travis-ci/gimme/master/gimme Thu Jul 20 22:48:13 UTC 2023: + for i in {1..3} Thu Jul 20 22:48:13 UTC 2023: + '[' 1 == 1 ']' Thu Jul 20 22:48:13 UTC 2023: + curl -sL -o /tmp/bin/gimme https://raw.githubusercontent.com/travis-ci/gimme/master/gimme Thu Jul 20 22:48:13 UTC 2023: + return 0 Thu Jul 20 22:48:13 UTC 2023: + chmod +x /tmp/bin/gimme Thu Jul 20 22:48:14 UTC 2023: + export PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/tmp/bin Thu Jul 20 22:48:14 UTC 2023: + PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/tmp/bin Thu Jul 20 22:48:14 UTC 2023: + retry gimme_retrier Thu Jul 20 22:48:14 UTC 2023: + for i in {1..3} Thu Jul 20 22:48:14 UTC 2023: + '[' 1 == 1 ']' Thu Jul 20 22:48:14 UTC 2023: + gimme_retrier Thu Jul 20 22:48:14 UTC 2023: ++ gimme master Thu Jul 20 22:49:18 UTC 2023: I don't have any idea what to do with 'master'. Thu Jul 20 22:49:18 UTC 2023: (using download type 'auto') Thu Jul 20 22:49:18 UTC 2023: + eval '' Thu Jul 20 22:49:18 UTC 2023: + which go Thu Jul 20 22:49:18 UTC 2023: + for i in {1..3} Thu Jul 20 22:49:18 UTC 2023: + '[' 2 == 1 ']' Thu Jul 20 22:49:18 UTC 2023: + sleep 10 Thu Jul 20 22:49:28 UTC 2023: + gimme_retrier Thu Jul 20 22:49:28 UTC 2023: ++ gimme master Thu Jul 20 22:49:29 UTC 2023: I don't have any idea what to do with 'master'. Thu Jul 20 22:49:29 UTC 2023: (using download type 'auto') Thu Jul 20 22:49:29 UTC 2023: + eval '' Thu Jul 20 22:49:29 UTC 2023: + which go Thu Jul 20 22:49:29 UTC 2023: + for i in {1..3} Thu Jul 20 22:49:29 UTC 2023: + '[' 3 == 1 ']' Thu Jul 20 22:49:29 UTC 2023: + sleep 10 Thu Jul 20 22:49:39 UTC 2023: + gimme_retrier Thu Jul 20 22:49:39 UTC 2023: ++ gimme master Thu Jul 20 22:49:40 UTC 2023: I don't have any idea what to do with 'master'. Thu Jul 20 22:49:40 UTC 2023: (using download type 'auto') Thu Jul 20 22:49:40 UTC 2023: + eval '' Thu Jul 20 22:49:40 UTC 2023: + which go Thu Jul 20 22:49:40 UTC 2023: + return 1 Thu Jul 20 22:49:40 UTC 2023: ++ echo 'failed to set up or run the benchmark' Thu Jul 20 22:49:40 UTC 2023: failed to set up or run the benchmark Thu Jul 20 22:49:40 UTC 2023: ++ sleep 300 integration_test.go:364: PollAndLogSerialPort() got error: failed to execute the prober benchmark script</pre></details>
1.0
profiler: TestAgentIntegration failed - Note: #7152 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: b2964a92d093a3d751e8b43aa70471bd989b1244 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f532917b-d222-41c3-89ee-274229ab20e1), [Sponge](http://sponge2/f532917b-d222-41c3-89ee-274229ab20e1) status: failed <details><summary>Test output</summary><br><pre> proftest.go:481: Serial port output for profiler-test-gomaster-2023-07-20-15-47-00-115169-pdt: \033[2J\033[01;01H\033[=3h\033[2J\033[01;01H\033[2J\033[01;01H\033[=3h\033[2J\033[01;01HBdsDxe: loading Boot0001 "UEFI Google PersistentDisk " from PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0) BdsDxe: starting Boot0001 "UEFI Google PersistentDisk " from PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0) UEFI: Attempting to start image. Description: UEFI Google PersistentDisk FilePath: PciRoot(0x0)/Pci(0x3,0x0)/Scsi(0x1,0x0) OptionNumber: 1. \033[0m\033[30m\033[47mWelcome to GRUB! \033[0m\033[37m\033[40m\033[0m\033[30m\033[40m\033[2J\033[01;01H\033[0m\033[37m\033[40m\033[0m\033[30m\033[40m\033[2J\033[01;01H\033[0m\033[37m\033[40m Booting `Debian GNU/Linux' Loading Linux 5.10.0-23-cloud-amd64 ... Loading initial ramdisk ... error: no suitable video mode found. Booting in blind mode Thu Jul 20 22:47:26 UTC 2023: + retry apt-get update Thu Jul 20 22:47:26 UTC 2023: + for i in {1..3} Thu Jul 20 22:47:26 UTC 2023: + '[' 1 == 1 ']' Thu Jul 20 22:47:26 UTC 2023: + apt-get update Thu Jul 20 22:47:28 UTC 2023: + return 0 Thu Jul 20 22:47:28 UTC 2023: + retry apt-get -y -q install git Thu Jul 20 22:47:28 UTC 2023: + for i in {1..3} Thu Jul 20 22:47:28 UTC 2023: + '[' 1 == 1 ']' Thu Jul 20 22:47:28 UTC 2023: + apt-get -y -q install git Thu Jul 20 22:47:29 UTC 2023: debconf: unable to initialize frontend: Dialog Thu Jul 20 22:47:29 UTC 2023: debconf: (TERM is not set, so the dialog frontend is not usable.) Thu Jul 20 22:47:29 UTC 2023: debconf: falling back to frontend: Readline Thu Jul 20 22:47:29 UTC 2023: debconf: unable to initialize frontend: Readline Thu Jul 20 22:47:29 UTC 2023: debconf: (Can't locate Term/ReadLine.pm in @INC (you may need to install the Term::ReadLine module) (@INC contains: /etc/perl /usr/local/lib/x86_64-linux-gnu/perl/5.32.1 /usr/local/share/perl/5.32.1 /usr/lib/x86_64-linux-gnu/perl5/5.32 /usr/share/perl5 /usr/lib/x86_64-linux-gnu/perl-base /usr/lib/x86_64-linux-gnu/perl/5.32 /usr/share/perl/5.32 /usr/local/lib/site_perl) at /usr/share/perl5/Debconf/FrontEnd/Readline.pm line 7, <> line 8.) Thu Jul 20 22:47:29 UTC 2023: debconf: falling back to frontend: Teletype Thu Jul 20 22:47:29 UTC 2023: dpkg-preconfigure: unable to re-open stdin: Thu Jul 20 22:47:39 UTC 2023: + return 0 Thu Jul 20 22:47:39 UTC 2023: + mkdir -p /tmp/gocache Thu Jul 20 22:47:39 UTC 2023: + export GOCACHE=/tmp/gocache Thu Jul 20 22:47:39 UTC 2023: + GOCACHE=/tmp/gocache Thu Jul 20 22:47:39 UTC 2023: + '[' master = master ']' Thu Jul 20 22:47:39 UTC 2023: + retry apt-get -y -q install gcc Thu Jul 20 22:47:39 UTC 2023: + for i in {1..3} Thu Jul 20 22:47:39 UTC 2023: + '[' 1 == 1 ']' Thu Jul 20 22:47:39 UTC 2023: + apt-get -y -q install gcc Thu Jul 20 22:47:40 UTC 2023: debconf: unable to initialize frontend: Dialog Thu Jul 20 22:47:40 UTC 2023: debconf: (TERM is not set, so the dialog frontend is not usable.) Thu Jul 20 22:47:40 UTC 2023: debconf: falling back to frontend: Readline Thu Jul 20 22:47:40 UTC 2023: debconf: unable to initialize frontend: Readline Thu Jul 20 22:47:40 UTC 2023: debconf: (This frontend requires a controlling tty.) Thu Jul 20 22:47:40 UTC 2023: debconf: falling back to frontend: Teletype Thu Jul 20 22:47:40 UTC 2023: dpkg-preconfigure: unable to re-open stdin: Thu Jul 20 22:48:13 UTC 2023: + return 0 Thu Jul 20 22:48:13 UTC 2023: + mkdir -p /tmp/bin Thu Jul 20 22:48:13 UTC 2023: + retry curl -sL -o /tmp/bin/gimme https://raw.githubusercontent.com/travis-ci/gimme/master/gimme Thu Jul 20 22:48:13 UTC 2023: + for i in {1..3} Thu Jul 20 22:48:13 UTC 2023: + '[' 1 == 1 ']' Thu Jul 20 22:48:13 UTC 2023: + curl -sL -o /tmp/bin/gimme https://raw.githubusercontent.com/travis-ci/gimme/master/gimme Thu Jul 20 22:48:13 UTC 2023: + return 0 Thu Jul 20 22:48:13 UTC 2023: + chmod +x /tmp/bin/gimme Thu Jul 20 22:48:14 UTC 2023: + export PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/tmp/bin Thu Jul 20 22:48:14 UTC 2023: + PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/tmp/bin Thu Jul 20 22:48:14 UTC 2023: + retry gimme_retrier Thu Jul 20 22:48:14 UTC 2023: + for i in {1..3} Thu Jul 20 22:48:14 UTC 2023: + '[' 1 == 1 ']' Thu Jul 20 22:48:14 UTC 2023: + gimme_retrier Thu Jul 20 22:48:14 UTC 2023: ++ gimme master Thu Jul 20 22:49:18 UTC 2023: I don't have any idea what to do with 'master'. Thu Jul 20 22:49:18 UTC 2023: (using download type 'auto') Thu Jul 20 22:49:18 UTC 2023: + eval '' Thu Jul 20 22:49:18 UTC 2023: + which go Thu Jul 20 22:49:18 UTC 2023: + for i in {1..3} Thu Jul 20 22:49:18 UTC 2023: + '[' 2 == 1 ']' Thu Jul 20 22:49:18 UTC 2023: + sleep 10 Thu Jul 20 22:49:28 UTC 2023: + gimme_retrier Thu Jul 20 22:49:28 UTC 2023: ++ gimme master Thu Jul 20 22:49:29 UTC 2023: I don't have any idea what to do with 'master'. Thu Jul 20 22:49:29 UTC 2023: (using download type 'auto') Thu Jul 20 22:49:29 UTC 2023: + eval '' Thu Jul 20 22:49:29 UTC 2023: + which go Thu Jul 20 22:49:29 UTC 2023: + for i in {1..3} Thu Jul 20 22:49:29 UTC 2023: + '[' 3 == 1 ']' Thu Jul 20 22:49:29 UTC 2023: + sleep 10 Thu Jul 20 22:49:39 UTC 2023: + gimme_retrier Thu Jul 20 22:49:39 UTC 2023: ++ gimme master Thu Jul 20 22:49:40 UTC 2023: I don't have any idea what to do with 'master'. Thu Jul 20 22:49:40 UTC 2023: (using download type 'auto') Thu Jul 20 22:49:40 UTC 2023: + eval '' Thu Jul 20 22:49:40 UTC 2023: + which go Thu Jul 20 22:49:40 UTC 2023: + return 1 Thu Jul 20 22:49:40 UTC 2023: ++ echo 'failed to set up or run the benchmark' Thu Jul 20 22:49:40 UTC 2023: failed to set up or run the benchmark Thu Jul 20 22:49:40 UTC 2023: ++ sleep 300 integration_test.go:364: PollAndLogSerialPort() got error: failed to execute the prober benchmark script</pre></details>
non_build
profiler testagentintegration failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output proftest go serial port output for profiler test gomaster pdt loading uefi google persistentdisk from pciroot pci scsi bdsdxe starting uefi google persistentdisk from pciroot pci scsi uefi attempting to start image description uefi google persistentdisk filepath pciroot pci scsi optionnumber to grub booting debian gnu linux loading linux cloud loading initial ramdisk error no suitable video mode found booting in blind mode thu jul utc retry apt get update thu jul utc for i in thu jul utc thu jul utc apt get update thu jul utc return thu jul utc retry apt get y q install git thu jul utc for i in thu jul utc thu jul utc apt get y q install git thu jul utc debconf unable to initialize frontend dialog thu jul utc debconf term is not set so the dialog frontend is not usable thu jul utc debconf falling back to frontend readline thu jul utc debconf unable to initialize frontend readline thu jul utc debconf can t locate term readline pm in inc you may need to install the term readline module inc contains etc perl usr local lib linux gnu perl usr local share perl usr lib linux gnu usr share usr lib linux gnu perl base usr lib linux gnu perl usr share perl usr local lib site perl at usr share debconf frontend readline pm line line thu jul utc debconf falling back to frontend teletype thu jul utc dpkg preconfigure unable to re open stdin thu jul utc return thu jul utc mkdir p tmp gocache thu jul utc export gocache tmp gocache thu jul utc gocache tmp gocache thu jul utc thu jul utc retry apt get y q install gcc thu jul utc for i in thu jul utc thu jul utc apt get y q install gcc thu jul utc debconf unable to initialize frontend dialog thu jul utc debconf term is not set so the dialog frontend is not usable thu jul utc debconf falling back to frontend readline thu jul utc debconf unable to initialize frontend readline thu jul utc debconf this frontend requires a controlling tty thu jul utc debconf falling back to frontend teletype thu jul utc dpkg preconfigure unable to re open stdin thu jul utc return thu jul utc mkdir p tmp bin thu jul utc retry curl sl o tmp bin gimme thu jul utc for i in thu jul utc thu jul utc curl sl o tmp bin gimme thu jul utc return thu jul utc chmod x tmp bin gimme thu jul utc export path usr local sbin usr local bin usr sbin usr bin sbin bin tmp bin thu jul utc path usr local sbin usr local bin usr sbin usr bin sbin bin tmp bin thu jul utc retry gimme retrier thu jul utc for i in thu jul utc thu jul utc gimme retrier thu jul utc gimme master thu jul utc i don t have any idea what to do with master thu jul utc using download type auto thu jul utc eval thu jul utc which go thu jul utc for i in thu jul utc thu jul utc sleep thu jul utc gimme retrier thu jul utc gimme master thu jul utc i don t have any idea what to do with master thu jul utc using download type auto thu jul utc eval thu jul utc which go thu jul utc for i in thu jul utc thu jul utc sleep thu jul utc gimme retrier thu jul utc gimme master thu jul utc i don t have any idea what to do with master thu jul utc using download type auto thu jul utc eval thu jul utc which go thu jul utc return thu jul utc echo failed to set up or run the benchmark thu jul utc failed to set up or run the benchmark thu jul utc sleep integration test go pollandlogserialport got error failed to execute the prober benchmark script
0
186,548
14,397,483,867
IssuesEvent
2020-12-03 08:14:30
zeek/spicy
https://api.github.com/repos/zeek/spicy
closed
Add unit tests for IntrusivePtr
Runtime Library Testing / CI
With #491 came a new IntrusivePtr class (adapted from Zeek). We need more test coverage for that.
1.0
Add unit tests for IntrusivePtr - With #491 came a new IntrusivePtr class (adapted from Zeek). We need more test coverage for that.
non_build
add unit tests for intrusiveptr with came a new intrusiveptr class adapted from zeek we need more test coverage for that
0
68,213
14,914,662,904
IssuesEvent
2021-01-22 15:42:37
finos/secref-data
https://api.github.com/repos/finos/secref-data
closed
Nov 10th 2020 - SecRef Meeting Minutes
Securities Reference Data meeting
# Security Reference Data Project Meeting ## Date Nov 10th 2020 - 11am EST / 4pm BST ## Untracked attendees - Lodovico Gavotti / Refinitiv - ... ## Agenda - [x] Convene & roll call (5mins) - [x] Introduce new joiners & volunteers - [x] Approve [Oct 27th](https://github.com/finos/secref-data/issues/40) meeting minutes - _meeting minutes were approved_ - [ ] Review submitted use cases [#33](https://github.com/finos/secref-data/issues/33) - (use cases around data ingestion from new providers and duplicates had been scheduled to be submitted by Oct 27th, 2020) - [ ] Review workflow [#26](https://github.com/finos/secref-data/issues/26) - [x] AOB, Q&A & Adjourn (5mins) ## Decisions Made - [ ] Confirmed that access to the mapping tool/data will be via API calls and that a GUI is a secondary priority. - [ ] Decision 2 - [ ] ... ## Action Items - [ ] @toshaellison to add the use cases discussed during the call to [https://github.com/finos/secref-data/issues/33](https://github.com/finos/secref-data/issues/33) - [ ] Discuss how new data sources will be mapped into the tool. - [ ] Discuss the process by which licenses can be validated when API calls are made. ### WebEx info - [WebEx Meeting URL](https://finos.webex.com/finos/j.php?MTID=mf43ec72e7ce8e771da12093c940c7ab9) - Meeting Number: 668 805 739 - Join by phone: +1.415.655.0003 USA Toll | +44-20319-88141 UK Toll | Access code: 668 805 739 - [Global call-in numbers](https://finos.webex.com/finos/globalcallin.php?serviceType=MC&ED=582460637&tollFree=0)
True
Nov 10th 2020 - SecRef Meeting Minutes - # Security Reference Data Project Meeting ## Date Nov 10th 2020 - 11am EST / 4pm BST ## Untracked attendees - Lodovico Gavotti / Refinitiv - ... ## Agenda - [x] Convene & roll call (5mins) - [x] Introduce new joiners & volunteers - [x] Approve [Oct 27th](https://github.com/finos/secref-data/issues/40) meeting minutes - _meeting minutes were approved_ - [ ] Review submitted use cases [#33](https://github.com/finos/secref-data/issues/33) - (use cases around data ingestion from new providers and duplicates had been scheduled to be submitted by Oct 27th, 2020) - [ ] Review workflow [#26](https://github.com/finos/secref-data/issues/26) - [x] AOB, Q&A & Adjourn (5mins) ## Decisions Made - [ ] Confirmed that access to the mapping tool/data will be via API calls and that a GUI is a secondary priority. - [ ] Decision 2 - [ ] ... ## Action Items - [ ] @toshaellison to add the use cases discussed during the call to [https://github.com/finos/secref-data/issues/33](https://github.com/finos/secref-data/issues/33) - [ ] Discuss how new data sources will be mapped into the tool. - [ ] Discuss the process by which licenses can be validated when API calls are made. ### WebEx info - [WebEx Meeting URL](https://finos.webex.com/finos/j.php?MTID=mf43ec72e7ce8e771da12093c940c7ab9) - Meeting Number: 668 805 739 - Join by phone: +1.415.655.0003 USA Toll | +44-20319-88141 UK Toll | Access code: 668 805 739 - [Global call-in numbers](https://finos.webex.com/finos/globalcallin.php?serviceType=MC&ED=582460637&tollFree=0)
non_build
nov secref meeting minutes security reference data project meeting date nov est bst untracked attendees lodovico gavotti refinitiv agenda convene roll call introduce new joiners volunteers approve meeting minutes meeting minutes were approved review submitted use cases use cases around data ingestion from new providers and duplicates had been scheduled to be submitted by oct review workflow aob q a adjourn decisions made confirmed that access to the mapping tool data will be via api calls and that a gui is a secondary priority decision action items toshaellison to add the use cases discussed during the call to discuss how new data sources will be mapped into the tool discuss the process by which licenses can be validated when api calls are made webex info meeting number join by phone usa toll uk toll access code
0
320,370
23,808,797,170
IssuesEvent
2022-09-04 12:56:11
WeCanWebserv/Webserv
https://api.github.com/repos/WeCanWebserv/Webserv
closed
[FEAT] HTTP Request message의 Format(token) 파악하기
documentation
## Description - HTTP Request message의 token들 파악하기 ## TO DO - [x] request-line 파악하기 - [x] header 파악하기 - [x] body 파악하기
1.0
[FEAT] HTTP Request message의 Format(token) 파악하기 - ## Description - HTTP Request message의 token들 파악하기 ## TO DO - [x] request-line 파악하기 - [x] header 파악하기 - [x] body 파악하기
non_build
http request message의 format token 파악하기 description http request message의 token들 파악하기 to do request line 파악하기 header 파악하기 body 파악하기
0
30,916
8,621,951,994
IssuesEvent
2018-11-20 18:50:58
hashicorp/packer
https://api.github.com/repos/hashicorp/packer
closed
Vsphere-iso validation is failing for RHEL7.5 VM deployment
builder/vmware-esxi question
I am trying to build a vm with vsphere-iso Plugin .. Below is my rhel75.Json file ... ``` { "builders":[ { "type": "vsphere-iso", "vcenter_server": "nahq-rvapl00072", "insecure_connection": "true", "username": "srv_vra_lab", "password": "XXXXXXXXXXX", "datacenter": "LAB-NUTANIX", "vm_name": "RHEL75_SB", "folder": "_Packer", "host": "HQLABINF002/nahq-rpehv00644", "cluster": "HQLABINF002", "resource_pool": "vRA-LAB-Site2-APP", "datastore": "lab-nutanix02-ds01", "CPUs": 8, "RAM": 10240, "disk_size": 20480, "guest_os_type": "rhel7_64Guest", "cdrom_adapter_type": "ide", "disk_adapter_type": "lsiLogic", "disk_thin_provisioned": "true", "network": "Site1-LAB_930", "network_adapter": "vmxnet3", "iso_datastore": "lab-nutanix01-ds01", "iso": "[LABX_TEMPLATES_NFS] ISOs/rhel75.iso", "communicator": "ssh", "ssh_username": "root", "ssh_password": "packer", "convert_to_template": "true" } ], "provisioners": [ { "type": "shell", "inline": [ "echo hello" ] } ] } ``` Validation is failing with below errors. [bodduls@packer]$ ./packer validate rhel75.json Template validation failed. Errors are shown below. Errors validating build 'vsphere-iso'. 3 errors occurred: * unknown configuration key: "cdrom_adapter_type" * unknown configuration key: "disk_adapter_type" * unknown configuration key: "disk_thin_provisioned" +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ When I remove these optional parameters from the file .. build is failing with the below message and from V-center logs I can see this is related to VirtualCdrom which is not mounting the specified ISO. [bodduls@ packer]$ ./packer build rhel75.json vsphere-iso output will be in this color. ==> vsphere-iso: Creating VM... Build 'vsphere-iso' errored: Invalid configuration for device '2'. ==> Some builds didn't complete successfully and had errors: --> vsphere-iso: Invalid configuration for device '2'. ==> Builds finished but no artifacts were created. [bodduls@packer]$ +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ I am using Packer v1.3.2 [bodduls@ packer]$ ./packer version Packer v1.3.2 [bodduls@ packer]$ [bodduls@labd-rclap00142 packer]$ ls -ltr total 151572 -rwxr-xr-x 1 bodduls bodduls 86309745 Oct 29 07:40 packer -rw-rw-r-- 1 bodduls bodduls 27459221 Oct 29 07:47 packer_1.3.2_linux_amd64.zip -rwxr-xr-x 1 bodduls bodduls 41392224 Nov 7 11:19 packer-builder-vsphere-iso.linux -rw-rw-r-- 1 bodduls bodduls 19466 Nov 7 14:52 crash.log -rw-rw-r-- 1 bodduls bodduls 1033 Nov 13 09:40 rhel75.json [bodduls@labd-rclap00142 packer]$ +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Any suggestion to overcome this ? +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
1.0
Vsphere-iso validation is failing for RHEL7.5 VM deployment - I am trying to build a vm with vsphere-iso Plugin .. Below is my rhel75.Json file ... ``` { "builders":[ { "type": "vsphere-iso", "vcenter_server": "nahq-rvapl00072", "insecure_connection": "true", "username": "srv_vra_lab", "password": "XXXXXXXXXXX", "datacenter": "LAB-NUTANIX", "vm_name": "RHEL75_SB", "folder": "_Packer", "host": "HQLABINF002/nahq-rpehv00644", "cluster": "HQLABINF002", "resource_pool": "vRA-LAB-Site2-APP", "datastore": "lab-nutanix02-ds01", "CPUs": 8, "RAM": 10240, "disk_size": 20480, "guest_os_type": "rhel7_64Guest", "cdrom_adapter_type": "ide", "disk_adapter_type": "lsiLogic", "disk_thin_provisioned": "true", "network": "Site1-LAB_930", "network_adapter": "vmxnet3", "iso_datastore": "lab-nutanix01-ds01", "iso": "[LABX_TEMPLATES_NFS] ISOs/rhel75.iso", "communicator": "ssh", "ssh_username": "root", "ssh_password": "packer", "convert_to_template": "true" } ], "provisioners": [ { "type": "shell", "inline": [ "echo hello" ] } ] } ``` Validation is failing with below errors. [bodduls@packer]$ ./packer validate rhel75.json Template validation failed. Errors are shown below. Errors validating build 'vsphere-iso'. 3 errors occurred: * unknown configuration key: "cdrom_adapter_type" * unknown configuration key: "disk_adapter_type" * unknown configuration key: "disk_thin_provisioned" +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ When I remove these optional parameters from the file .. build is failing with the below message and from V-center logs I can see this is related to VirtualCdrom which is not mounting the specified ISO. [bodduls@ packer]$ ./packer build rhel75.json vsphere-iso output will be in this color. ==> vsphere-iso: Creating VM... Build 'vsphere-iso' errored: Invalid configuration for device '2'. ==> Some builds didn't complete successfully and had errors: --> vsphere-iso: Invalid configuration for device '2'. ==> Builds finished but no artifacts were created. [bodduls@packer]$ +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ I am using Packer v1.3.2 [bodduls@ packer]$ ./packer version Packer v1.3.2 [bodduls@ packer]$ [bodduls@labd-rclap00142 packer]$ ls -ltr total 151572 -rwxr-xr-x 1 bodduls bodduls 86309745 Oct 29 07:40 packer -rw-rw-r-- 1 bodduls bodduls 27459221 Oct 29 07:47 packer_1.3.2_linux_amd64.zip -rwxr-xr-x 1 bodduls bodduls 41392224 Nov 7 11:19 packer-builder-vsphere-iso.linux -rw-rw-r-- 1 bodduls bodduls 19466 Nov 7 14:52 crash.log -rw-rw-r-- 1 bodduls bodduls 1033 Nov 13 09:40 rhel75.json [bodduls@labd-rclap00142 packer]$ +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Any suggestion to overcome this ? +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
build
vsphere iso validation is failing for vm deployment i am trying to build a vm with vsphere iso plugin below is my json file builders type vsphere iso vcenter server nahq insecure connection true username srv vra lab password xxxxxxxxxxx datacenter lab nutanix vm name sb folder packer host nahq cluster resource pool vra lab app datastore lab cpus ram disk size guest os type cdrom adapter type ide disk adapter type lsilogic disk thin provisioned true network lab network adapter iso datastore lab iso isos iso communicator ssh ssh username root ssh password packer convert to template true provisioners type shell inline validation is failing with below errors packer validate json template validation failed errors are shown below errors validating build vsphere iso errors occurred unknown configuration key cdrom adapter type unknown configuration key disk adapter type unknown configuration key disk thin provisioned when i remove these optional parameters from the file build is failing with the below message and from v center logs i can see this is related to virtualcdrom which is not mounting the specified iso packer build json vsphere iso output will be in this color vsphere iso creating vm build vsphere iso errored invalid configuration for device some builds didn t complete successfully and had errors vsphere iso invalid configuration for device builds finished but no artifacts were created i am using packer packer version packer ls ltr total rwxr xr x bodduls bodduls oct packer rw rw r bodduls bodduls oct packer linux zip rwxr xr x bodduls bodduls nov packer builder vsphere iso linux rw rw r bodduls bodduls nov crash log rw rw r bodduls bodduls nov json any suggestion to overcome this
1
13,514
5,392,625,845
IssuesEvent
2017-02-26 13:03:13
junit-team/junit5
https://api.github.com/repos/junit-team/junit5
closed
Upgrade to Gradle 3.4
build enhancement up-for-grabs
## Overview The build fails after upgrading to Gradle 3.4 with following stacktrace: ``` ... :junit-platform-gradle-plugin:spotlessCheck :junit-platform-gradle-plugin:compileTestJava NO-SOURCE :junit-platform-gradle-plugin:compileTestGroovy :junit-platform-gradle-plugin:processTestResources NO-SOURCE :junit-platform-gradle-plugin:testClasses :junit-platform-gradle-plugin:test org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec > adds dependencies to configuration FAILED java.lang.UnsupportedOperationException at org.gradle.api.internal.CompositeDomainObjectSet$DomainObjectCompositeCollection.containsAll(CompositeDomainObjectSet.java:228) at org.gradle.api.internal.DefaultDomainObjectCollection.containsAll(DefaultDomainObjectCollection.java:241) at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98) at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98) at org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec.adds dependencies to configuration(JUnitPlatformPluginSpec.groovy:251) 10 tests completed, 1 failed :junit-platform-gradle-plugin:test FAILED ``` ## Deliverables - [ ] Solve `UnsupportedOperationException` - ~Make use of https://docs.gradle.org/current/userguide/java_library_plugin.html~ Moved to #691
1.0
Upgrade to Gradle 3.4 - ## Overview The build fails after upgrading to Gradle 3.4 with following stacktrace: ``` ... :junit-platform-gradle-plugin:spotlessCheck :junit-platform-gradle-plugin:compileTestJava NO-SOURCE :junit-platform-gradle-plugin:compileTestGroovy :junit-platform-gradle-plugin:processTestResources NO-SOURCE :junit-platform-gradle-plugin:testClasses :junit-platform-gradle-plugin:test org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec > adds dependencies to configuration FAILED java.lang.UnsupportedOperationException at org.gradle.api.internal.CompositeDomainObjectSet$DomainObjectCompositeCollection.containsAll(CompositeDomainObjectSet.java:228) at org.gradle.api.internal.DefaultDomainObjectCollection.containsAll(DefaultDomainObjectCollection.java:241) at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98) at org.gradle.api.internal.DelegatingDomainObjectSet.containsAll(DelegatingDomainObjectSet.java:98) at org.junit.platform.gradle.plugin.JUnitPlatformPluginSpec.adds dependencies to configuration(JUnitPlatformPluginSpec.groovy:251) 10 tests completed, 1 failed :junit-platform-gradle-plugin:test FAILED ``` ## Deliverables - [ ] Solve `UnsupportedOperationException` - ~Make use of https://docs.gradle.org/current/userguide/java_library_plugin.html~ Moved to #691
build
upgrade to gradle overview the build fails after upgrading to gradle with following stacktrace junit platform gradle plugin spotlesscheck junit platform gradle plugin compiletestjava no source junit platform gradle plugin compiletestgroovy junit platform gradle plugin processtestresources no source junit platform gradle plugin testclasses junit platform gradle plugin test org junit platform gradle plugin junitplatformpluginspec adds dependencies to configuration failed java lang unsupportedoperationexception at org gradle api internal compositedomainobjectset domainobjectcompositecollection containsall compositedomainobjectset java at org gradle api internal defaultdomainobjectcollection containsall defaultdomainobjectcollection java at org gradle api internal delegatingdomainobjectset containsall delegatingdomainobjectset java at org gradle api internal delegatingdomainobjectset containsall delegatingdomainobjectset java at org junit platform gradle plugin junitplatformpluginspec adds dependencies to configuration junitplatformpluginspec groovy tests completed failed junit platform gradle plugin test failed deliverables solve unsupportedoperationexception make use of moved to
1
24,432
12,103,737,745
IssuesEvent
2020-04-20 18:56:28
Azure/azure-sdk-for-net
https://api.github.com/repos/Azure/azure-sdk-for-net
closed
Add ReceiptTypeConfidence field to USReceipt
Client Cognitive Services FormRecognizer
`ReceiptType` comes back as a field in `DocumentResults`, without a bounding box, but with a confidence value. Add a `float? ReceiptTypeConfidence` to `USReceipt` to capture the confidence.
1.0
Add ReceiptTypeConfidence field to USReceipt - `ReceiptType` comes back as a field in `DocumentResults`, without a bounding box, but with a confidence value. Add a `float? ReceiptTypeConfidence` to `USReceipt` to capture the confidence.
non_build
add receipttypeconfidence field to usreceipt receipttype comes back as a field in documentresults without a bounding box but with a confidence value add a float receipttypeconfidence to usreceipt to capture the confidence
0
563,731
16,704,702,966
IssuesEvent
2021-06-09 08:36:40
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.ebay.com - design is broken
browser-firefox engine-gecko priority-critical
<!-- @browser: Firefox 89.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:89.0) Gecko/20100101 Firefox/89.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/76109 --> **URL**: https://www.ebay.com/ **Browser / Version**: Firefox 89.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Design is broken **Description**: Images not loaded **Steps to Reproduce**: The Page not visible correctly, no image, no order <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/6/9f2b29a9-412b-4998-9d31-8c548ac8aaa9.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.ebay.com - design is broken - <!-- @browser: Firefox 89.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:89.0) Gecko/20100101 Firefox/89.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/76109 --> **URL**: https://www.ebay.com/ **Browser / Version**: Firefox 89.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Design is broken **Description**: Images not loaded **Steps to Reproduce**: The Page not visible correctly, no image, no order <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/6/9f2b29a9-412b-4998-9d31-8c548ac8aaa9.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_build
design is broken url browser version firefox operating system windows tested another browser yes chrome problem type design is broken description images not loaded steps to reproduce the page not visible correctly no image no order view the screenshot img alt screenshot src browser configuration none from with ❤️
0
196,568
6,935,334,656
IssuesEvent
2017-12-03 07:24:04
manrajsingh960/GetTogether
https://api.github.com/repos/manrajsingh960/GetTogether
closed
Joined Events (When data is deleted)
functional problem priority
The joined events list is cleared when you clear the data
1.0
Joined Events (When data is deleted) - The joined events list is cleared when you clear the data
non_build
joined events when data is deleted the joined events list is cleared when you clear the data
0
265,283
23,158,358,956
IssuesEvent
2022-07-29 15:03:18
redpanda-data/redpanda
https://api.github.com/repos/redpanda-data/redpanda
closed
Failure in `PartitionMovementTest`.`test_availability_when_one_node_down` (ECONNREFUSED)
kind/bug area/redpanda area/tests ci-failure
PartitionMovementTest.test_availability_when_one_node_down fails with ECONNREFUSED trying to query the `partitions/kafka/<topic>/0/replicas` endpoint, which is very similar to #5189. Same error, same query, same test class, but different test case--thus the new bug instead of reopening #5189. ``` PartitionMovementTest.test_availability_when_one_node_down (3/9 runs) failure at 2022-07-27T21:40:32.761Z: ConnectionError(MaxRetryError("HTTPConnectionPool(host='docker-rp-7', port=9644): Max retries exceeded with url: /v1/partitions/kafka/topic-ezddcdbxtc/0/replicas (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f6d6d258280>: Failed to establish a new connection: [Errno 111] Connection refused'))")) ``` from [build 13204](https://buildkite.com/redpanda/redpanda/builds/13204#01824125-c0c1-4155-a6c5-914fc29a724f)
1.0
Failure in `PartitionMovementTest`.`test_availability_when_one_node_down` (ECONNREFUSED) - PartitionMovementTest.test_availability_when_one_node_down fails with ECONNREFUSED trying to query the `partitions/kafka/<topic>/0/replicas` endpoint, which is very similar to #5189. Same error, same query, same test class, but different test case--thus the new bug instead of reopening #5189. ``` PartitionMovementTest.test_availability_when_one_node_down (3/9 runs) failure at 2022-07-27T21:40:32.761Z: ConnectionError(MaxRetryError("HTTPConnectionPool(host='docker-rp-7', port=9644): Max retries exceeded with url: /v1/partitions/kafka/topic-ezddcdbxtc/0/replicas (Caused by NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7f6d6d258280>: Failed to establish a new connection: [Errno 111] Connection refused'))")) ``` from [build 13204](https://buildkite.com/redpanda/redpanda/builds/13204#01824125-c0c1-4155-a6c5-914fc29a724f)
non_build
failure in partitionmovementtest test availability when one node down econnrefused partitionmovementtest test availability when one node down fails with econnrefused trying to query the partitions kafka replicas endpoint which is very similar to same error same query same test class but different test case thus the new bug instead of reopening partitionmovementtest test availability when one node down runs failure at connectionerror maxretryerror httpconnectionpool host docker rp port max retries exceeded with url partitions kafka topic ezddcdbxtc replicas caused by newconnectionerror failed to establish a new connection connection refused from
0
14,107
5,556,494,506
IssuesEvent
2017-03-24 09:20:31
ng-bootstrap/ng-bootstrap
https://api.github.com/repos/ng-bootstrap/ng-bootstrap
closed
angular 4 compatibility
type: build
### Bug description: See *how to reproduce* below. The issue is that currently the generated package.json has: "peerDependencies": { "@angular/common": "^2.0.0", "@angular/compiler": "^2.0.0", "@angular/core": "^2.0.0", "@angular/forms": "^2.0.0" }, which is too restrictive - according to the [official docs](https://nodejs.org/en/blog/npm/peer-dependencies/) : > One piece of advice: peer dependency requirements, unlike those for regular dependencies, should be lenient *Note*: this is a follow-up to https://github.com/ng-bootstrap/ng-bootstrap/pull/1390 which is indeed wrong - I didn't know the package.json was autogenerated and changed the actual dependencies (because I was looking at using the fork directly as a dependency of my project), but the issue remains. *Note2*: by using `"^2.0.0 || >=4.0.0-beta <5.0.0",` you would still be supporting 2.x. ### How to reproduce the issue: ```bash npm install -g @angular/cli ng new ngcli-ng4 --ng4 cd ngcli-ng4 npm install --save ng-bootstrap ``` output ``` ngcli-ng4@0.0.0 /ything/projects/tmk/tmp/ngcli-ng4 ├── UNMET PEER DEPENDENCY @angular/common@4.0.0-rc.5 ├── UNMET PEER DEPENDENCY @angular/compiler@4.0.0-rc.5 ├── UNMET PEER DEPENDENCY @angular/core@4.0.0-rc.5 ├── UNMET PEER DEPENDENCY @angular/forms@4.0.0-rc.5 ├─┬ ng-bootstrap@1.1.16-1 │ └── moment@2.15.1 └── UNMET PEER DEPENDENCY zone.js@0.7.8 npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@^1.0.0 (node_modules/chokidar/node_modules/fsevents): npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.1.1: wanted {"os":"darwin","arch":"any"} (current: {"os":"linux","arch":"x64"}) npm WARN @angular/core@4.0.0-rc.5 requires a peer of zone.js@^0.8.4 but none was installed. npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/common@^2.0.0 but none was installed. npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/compiler@^2.0.0 but none was installed. npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/core@^2.0.0 but none was installed. npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/forms@^2.0.0 but none was installed. ``` ### Versions * npm 3.10.10 * Angular: 4.0.0-rc.5 * ng-bootstrap: 1.1.16
1.0
angular 4 compatibility - ### Bug description: See *how to reproduce* below. The issue is that currently the generated package.json has: "peerDependencies": { "@angular/common": "^2.0.0", "@angular/compiler": "^2.0.0", "@angular/core": "^2.0.0", "@angular/forms": "^2.0.0" }, which is too restrictive - according to the [official docs](https://nodejs.org/en/blog/npm/peer-dependencies/) : > One piece of advice: peer dependency requirements, unlike those for regular dependencies, should be lenient *Note*: this is a follow-up to https://github.com/ng-bootstrap/ng-bootstrap/pull/1390 which is indeed wrong - I didn't know the package.json was autogenerated and changed the actual dependencies (because I was looking at using the fork directly as a dependency of my project), but the issue remains. *Note2*: by using `"^2.0.0 || >=4.0.0-beta <5.0.0",` you would still be supporting 2.x. ### How to reproduce the issue: ```bash npm install -g @angular/cli ng new ngcli-ng4 --ng4 cd ngcli-ng4 npm install --save ng-bootstrap ``` output ``` ngcli-ng4@0.0.0 /ything/projects/tmk/tmp/ngcli-ng4 ├── UNMET PEER DEPENDENCY @angular/common@4.0.0-rc.5 ├── UNMET PEER DEPENDENCY @angular/compiler@4.0.0-rc.5 ├── UNMET PEER DEPENDENCY @angular/core@4.0.0-rc.5 ├── UNMET PEER DEPENDENCY @angular/forms@4.0.0-rc.5 ├─┬ ng-bootstrap@1.1.16-1 │ └── moment@2.15.1 └── UNMET PEER DEPENDENCY zone.js@0.7.8 npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@^1.0.0 (node_modules/chokidar/node_modules/fsevents): npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.1.1: wanted {"os":"darwin","arch":"any"} (current: {"os":"linux","arch":"x64"}) npm WARN @angular/core@4.0.0-rc.5 requires a peer of zone.js@^0.8.4 but none was installed. npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/common@^2.0.0 but none was installed. npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/compiler@^2.0.0 but none was installed. npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/core@^2.0.0 but none was installed. npm WARN ng-bootstrap@1.1.16-1 requires a peer of @angular/forms@^2.0.0 but none was installed. ``` ### Versions * npm 3.10.10 * Angular: 4.0.0-rc.5 * ng-bootstrap: 1.1.16
build
angular compatibility bug description see how to reproduce below the issue is that currently the generated package json has peerdependencies angular common angular compiler angular core angular forms which is too restrictive according to the one piece of advice peer dependency requirements unlike those for regular dependencies should be lenient note this is a follow up to which is indeed wrong i didn t know the package json was autogenerated and changed the actual dependencies because i was looking at using the fork directly as a dependency of my project but the issue remains by using beta you would still be supporting x how to reproduce the issue bash npm install g angular cli ng new ngcli cd ngcli npm install save ng bootstrap output ngcli ything projects tmk tmp ngcli ├── unmet peer dependency angular common rc ├── unmet peer dependency angular compiler rc ├── unmet peer dependency angular core rc ├── unmet peer dependency angular forms rc ├─┬ ng bootstrap │ └── moment └── unmet peer dependency zone js npm warn optional skipping optional dependency fsevents node modules chokidar node modules fsevents npm warn notsup skipping optional dependency unsupported platform for fsevents wanted os darwin arch any current os linux arch npm warn angular core rc requires a peer of zone js but none was installed npm warn ng bootstrap requires a peer of angular common but none was installed npm warn ng bootstrap requires a peer of angular compiler but none was installed npm warn ng bootstrap requires a peer of angular core but none was installed npm warn ng bootstrap requires a peer of angular forms but none was installed versions npm angular rc ng bootstrap
1
147,408
19,522,528,587
IssuesEvent
2021-12-29 21:42:44
barreljan/netlog
https://api.github.com/repos/barreljan/netlog
closed
Sanitizing $_POST inputs to SQL
bug security
As there is a way to for instance cURL your way into a http POST, any key or value could be a risk. A potential issue is there with the new host in the settings.php page. If you could craft a few headers, it will hit line 128 (foreach loop) and then line 167 to start the seq. of adding a new host. The hostip or hosttype fields are not prepared, escaped of by whatever means so with knowlegde of the code, you could drop a table there or worse. Must investigate the index.php also however the search input is already in order with a prepared statement. Other lines must be investigated as well.
True
Sanitizing $_POST inputs to SQL - As there is a way to for instance cURL your way into a http POST, any key or value could be a risk. A potential issue is there with the new host in the settings.php page. If you could craft a few headers, it will hit line 128 (foreach loop) and then line 167 to start the seq. of adding a new host. The hostip or hosttype fields are not prepared, escaped of by whatever means so with knowlegde of the code, you could drop a table there or worse. Must investigate the index.php also however the search input is already in order with a prepared statement. Other lines must be investigated as well.
non_build
sanitizing post inputs to sql as there is a way to for instance curl your way into a http post any key or value could be a risk a potential issue is there with the new host in the settings php page if you could craft a few headers it will hit line foreach loop and then line to start the seq of adding a new host the hostip or hosttype fields are not prepared escaped of by whatever means so with knowlegde of the code you could drop a table there or worse must investigate the index php also however the search input is already in order with a prepared statement other lines must be investigated as well
0
5,056
3,497,535,023
IssuesEvent
2016-01-06 01:50:17
mxe/mxe
https://api.github.com/repos/mxe/mxe
closed
Create Travis CI example of using MXE binary distribution
build-pkg cmake
Current repo for this task: https://github.com/starius/mxe-travis-example (empty) Can we move it to github.com/mxe/mxe-travis-example when it's finished?
1.0
Create Travis CI example of using MXE binary distribution - Current repo for this task: https://github.com/starius/mxe-travis-example (empty) Can we move it to github.com/mxe/mxe-travis-example when it's finished?
build
create travis ci example of using mxe binary distribution current repo for this task empty can we move it to github com mxe mxe travis example when it s finished
1
74,295
20,109,360,022
IssuesEvent
2022-02-07 13:45:07
gakonst/foundry
https://api.github.com/repos/gakonst/foundry
closed
Seeing bad cache invalidation in makerdao/spells-mainnet
T-bug C-forge P-normal Cmd-forge-build
Steps to reproduce: 1. Fork https://github.com/makerdao/spells-mainnet and checkout commit be8c6e946a5fa7768d8af0f6e200712f397ad587 . 2. Remove the `--force` arg from `test-dssspell-forge.sh`. 3. Run `make test-forge`. 4. Tests should correctly mostly error (4 success, 9 failing) and report that the spell has already been cast. Cache has been built. 5. Edit `src/DssSpell.t.base.sol` line 294 and set the `deployed_spell` to `address(0)`. 6. Run `make test-forge`. Expected output: Most tests should start passing (11 succeed, 2 failing) with no mention of spell already cast. Actual output: Identical from the first test which is incorrect.
1.0
Seeing bad cache invalidation in makerdao/spells-mainnet - Steps to reproduce: 1. Fork https://github.com/makerdao/spells-mainnet and checkout commit be8c6e946a5fa7768d8af0f6e200712f397ad587 . 2. Remove the `--force` arg from `test-dssspell-forge.sh`. 3. Run `make test-forge`. 4. Tests should correctly mostly error (4 success, 9 failing) and report that the spell has already been cast. Cache has been built. 5. Edit `src/DssSpell.t.base.sol` line 294 and set the `deployed_spell` to `address(0)`. 6. Run `make test-forge`. Expected output: Most tests should start passing (11 succeed, 2 failing) with no mention of spell already cast. Actual output: Identical from the first test which is incorrect.
build
seeing bad cache invalidation in makerdao spells mainnet steps to reproduce fork and checkout commit remove the force arg from test dssspell forge sh run make test forge tests should correctly mostly error success failing and report that the spell has already been cast cache has been built edit src dssspell t base sol line and set the deployed spell to address run make test forge expected output most tests should start passing succeed failing with no mention of spell already cast actual output identical from the first test which is incorrect
1
74,683
9,104,687,630
IssuesEvent
2019-02-20 18:49:50
envoyproxy/envoy
https://api.github.com/repos/envoyproxy/envoy
closed
healthcheck: Allow non-200 status codes as healthy HTTP healthchecks
design proposal help wanted
*Title*: *healthcheck: Allow non-200 status codes as HTTP healthchecks* *Description*: We have certain services that do not return a 200 on a successful healthcheck. The most common one I've seen is returning an HTTP 204. Envoy marks these as unhealthy because it is checking [explicitly](https://github.com/envoyproxy/envoy/blob/master/source/common/upstream/health_checker_impl.cc#L180) for a 200 I think a good default would be 2xx or 3xx is considered healthy. Whether a given status code should be considered healthy is up for debate, but i think this is pretty standard for proxies right now, and so i think it makes sense to do this. For extra credit, we can have configuration for what constitutes healthy. e.g. regex check on the body of the response.
1.0
healthcheck: Allow non-200 status codes as healthy HTTP healthchecks - *Title*: *healthcheck: Allow non-200 status codes as HTTP healthchecks* *Description*: We have certain services that do not return a 200 on a successful healthcheck. The most common one I've seen is returning an HTTP 204. Envoy marks these as unhealthy because it is checking [explicitly](https://github.com/envoyproxy/envoy/blob/master/source/common/upstream/health_checker_impl.cc#L180) for a 200 I think a good default would be 2xx or 3xx is considered healthy. Whether a given status code should be considered healthy is up for debate, but i think this is pretty standard for proxies right now, and so i think it makes sense to do this. For extra credit, we can have configuration for what constitutes healthy. e.g. regex check on the body of the response.
non_build
healthcheck allow non status codes as healthy http healthchecks title healthcheck allow non status codes as http healthchecks description we have certain services that do not return a on a successful healthcheck the most common one i ve seen is returning an http envoy marks these as unhealthy because it is checking for a i think a good default would be or is considered healthy whether a given status code should be considered healthy is up for debate but i think this is pretty standard for proxies right now and so i think it makes sense to do this for extra credit we can have configuration for what constitutes healthy e g regex check on the body of the response
0
34,040
9,254,581,667
IssuesEvent
2019-03-16 00:41:24
golang/go
https://api.github.com/repos/golang/go
opened
x/build/app: build dashboard for subrepos displays wrong columns
Builders
For example, on x/tools, the Go revision 64b1889 line on freebsd-10_3/386 shows the following: freebsd-386-12_0 at 64b1889e2d98e336160cad337a7781c720696290 building tools at c74ccfb859c02fdeda7e69caf9e7fdccd9a080fc :: Running /tmp/workdir/go/bin/go with args ["/tmp/workdir/go/bin/go" "test" "-short" "golang.org/x/tools/..."] and env ["PATH=/tmp/workdir/go/bin:/bin:/sbin:/usr/bin:/usr/local/bin" "PWD=/" "HOME=/" "RC_PID=24" "WORKDIR=/tmp/workdir" "GOROOT_BOOTSTRAP=/tmp/workdir/go1.4" "GO_BUILDER_NAME=freebsd-386-12_0" "GOARCH=386" "GOHOSTARCH=386" "GOROOT=/tmp/workdir/go" "GOPATH=/tmp/workdir/gopath" "GOPROXY=http://10.240.0.50:30156" "TMPDIR=/tmp/workdir/tmp" "GOCACHE=/tmp/workdir/gocache"] in dir /tmp/workdir/gopath/src/golang.org/x/tools
1.0
x/build/app: build dashboard for subrepos displays wrong columns - For example, on x/tools, the Go revision 64b1889 line on freebsd-10_3/386 shows the following: freebsd-386-12_0 at 64b1889e2d98e336160cad337a7781c720696290 building tools at c74ccfb859c02fdeda7e69caf9e7fdccd9a080fc :: Running /tmp/workdir/go/bin/go with args ["/tmp/workdir/go/bin/go" "test" "-short" "golang.org/x/tools/..."] and env ["PATH=/tmp/workdir/go/bin:/bin:/sbin:/usr/bin:/usr/local/bin" "PWD=/" "HOME=/" "RC_PID=24" "WORKDIR=/tmp/workdir" "GOROOT_BOOTSTRAP=/tmp/workdir/go1.4" "GO_BUILDER_NAME=freebsd-386-12_0" "GOARCH=386" "GOHOSTARCH=386" "GOROOT=/tmp/workdir/go" "GOPATH=/tmp/workdir/gopath" "GOPROXY=http://10.240.0.50:30156" "TMPDIR=/tmp/workdir/tmp" "GOCACHE=/tmp/workdir/gocache"] in dir /tmp/workdir/gopath/src/golang.org/x/tools
build
x build app build dashboard for subrepos displays wrong columns for example on x tools the go revision line on freebsd shows the following freebsd at building tools at running tmp workdir go bin go with args and env in dir tmp workdir gopath src golang org x tools
1
22,974
7,259,990,449
IssuesEvent
2018-02-18 02:55:18
minecraft-dev/MinecraftDev
https://api.github.com/repos/minecraft-dev/MinecraftDev
closed
Add Gradle compileOnly support.
build: gradle status: stale
Due to some issues I have come across with another Gradle plugin, I now have to use the `compileOnly` scope for the Bukkit/Paper APIs (Haven't tested with any others as I do not use them right now) instead of the default `compile`. As a result of this MinecraftDev no longer shows the icon pertaining to my API type in the top left of the project view, implying that MinecraftDev no longer acknowledges my plugin as a Minecraft plugin. For those we have no idea what I mean: Default: ```gradle dependencies { compile 'org.bukkit:bukkit:1.11.2-R0.1-SNAPSHOT' } ``` CompileOnly: ```gradle dependencies { compileOnly 'org.bukkit:bukkit:1.11.2-R0.1-SNAPSHOT' } ```
1.0
Add Gradle compileOnly support. - Due to some issues I have come across with another Gradle plugin, I now have to use the `compileOnly` scope for the Bukkit/Paper APIs (Haven't tested with any others as I do not use them right now) instead of the default `compile`. As a result of this MinecraftDev no longer shows the icon pertaining to my API type in the top left of the project view, implying that MinecraftDev no longer acknowledges my plugin as a Minecraft plugin. For those we have no idea what I mean: Default: ```gradle dependencies { compile 'org.bukkit:bukkit:1.11.2-R0.1-SNAPSHOT' } ``` CompileOnly: ```gradle dependencies { compileOnly 'org.bukkit:bukkit:1.11.2-R0.1-SNAPSHOT' } ```
build
add gradle compileonly support due to some issues i have come across with another gradle plugin i now have to use the compileonly scope for the bukkit paper apis haven t tested with any others as i do not use them right now instead of the default compile as a result of this minecraftdev no longer shows the icon pertaining to my api type in the top left of the project view implying that minecraftdev no longer acknowledges my plugin as a minecraft plugin for those we have no idea what i mean default gradle dependencies compile org bukkit bukkit snapshot compileonly gradle dependencies compileonly org bukkit bukkit snapshot
1
2,878
3,025,186,370
IssuesEvent
2015-08-03 06:19:09
FRosner/spawncamping-dds
https://api.github.com/repos/FRosner/spawncamping-dds
closed
Build fails on systems with older git
Bug Build
```git status -sb``` in build.sbt should be replaced with ```git status --procelain```, in order to retain a version-independent output -- the ```-b``` suboption to ```-s``` is not supported by git 1.7. The S3-relevant parts in buildt.sbt in general should probably be moved to a separate location (and imported/referenced from build.sbt), in order to make debugging local building easier.
1.0
Build fails on systems with older git - ```git status -sb``` in build.sbt should be replaced with ```git status --procelain```, in order to retain a version-independent output -- the ```-b``` suboption to ```-s``` is not supported by git 1.7. The S3-relevant parts in buildt.sbt in general should probably be moved to a separate location (and imported/referenced from build.sbt), in order to make debugging local building easier.
build
build fails on systems with older git git status sb in build sbt should be replaced with git status procelain in order to retain a version independent output the b suboption to s is not supported by git the relevant parts in buildt sbt in general should probably be moved to a separate location and imported referenced from build sbt in order to make debugging local building easier
1
6,035
3,733,516,043
IssuesEvent
2016-03-08 00:43:08
couchbase/couchbase-lite-android
https://api.github.com/repos/couchbase/couchbase-lite-android
closed
testInvokeProcessorAfterReachingCapacity failed on jenkins
Build
http://mobile.jenkins.couchbase.com/view/Couchbase_Lite/job/couchbase-lite-android-build/518/artifact/logcat.txt ``` 01-10 13:18:18.574 I/TestRunner( 1088): failed: testInvokeProcessorAfterReachingCapacity(com.couchbase.lite.support.BatcherTest) 01-10 13:18:18.574 I/TestRunner( 1088): ----- begin exception ----- 01-10 13:18:18.584 I/TestRunner( 1088): 01-10 13:18:18.584 I/TestRunner( 1088): junit.framework.AssertionFailedError 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.fail(Assert.java:48) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.assertTrue(Assert.java:20) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.assertTrue(Assert.java:27) 01-10 13:18:18.584 I/TestRunner( 1088): at com.couchbase.lite.support.BatcherTest.testInvokeProcessorAfterReachingCapacity(BatcherTest.java:402) 01-10 13:18:18.584 I/TestRunner( 1088): at java.lang.reflect.Method.invokeNative(Native Method) 01-10 13:18:18.584 I/TestRunner( 1088): at java.lang.reflect.Method.invoke(Method.java:515) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.runTest(TestCase.java:168) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.runBare(TestCase.java:134) 01-10 13:18:18.584 I/TestRunner( 1088): at com.couchbase.lite.LiteTestCase.runBare(LiteTestCase.java:20) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult$1.protect(TestResult.java:115) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult.runProtected(TestResult.java:133) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult.run(TestResult.java:118) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.run(TestCase.java:124) 01-10 13:18:18.584 I/TestRunner( 1088): at android.test.AndroidTestRunner.runTest(AndroidTestRunner.java:191) 01-10 13:18:18.584 I/TestRunner( 1088): at android.test.AndroidTestRunner.runTest(AndroidTestRunner.java:176) 01-10 13:18:18.584 I/TestRunner( 1088): at android.test.InstrumentationTestRunner.onStart(InstrumentationTestRunner.java:554) 01-10 13:18:18.584 I/TestRunner( 1088): at android.app.Instrumentation$InstrumentationThread.run(Instrumentation.java:1701) 01-10 13:18:18.584 I/TestRunner( 1088): ----- end exception ----- 01-10 13:18:18.604 I/TestRunner( 1088): finished: testInvokeProcessorAfterReachingCapacity(com.couchbase.lite.support.BatcherTest) 01-10 13:18:18.644 I/TestRunner( 1088): started: testWaitForPendingFutures(com.couchbase.lite.support.BatcherTest) ```
1.0
testInvokeProcessorAfterReachingCapacity failed on jenkins - http://mobile.jenkins.couchbase.com/view/Couchbase_Lite/job/couchbase-lite-android-build/518/artifact/logcat.txt ``` 01-10 13:18:18.574 I/TestRunner( 1088): failed: testInvokeProcessorAfterReachingCapacity(com.couchbase.lite.support.BatcherTest) 01-10 13:18:18.574 I/TestRunner( 1088): ----- begin exception ----- 01-10 13:18:18.584 I/TestRunner( 1088): 01-10 13:18:18.584 I/TestRunner( 1088): junit.framework.AssertionFailedError 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.fail(Assert.java:48) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.assertTrue(Assert.java:20) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.Assert.assertTrue(Assert.java:27) 01-10 13:18:18.584 I/TestRunner( 1088): at com.couchbase.lite.support.BatcherTest.testInvokeProcessorAfterReachingCapacity(BatcherTest.java:402) 01-10 13:18:18.584 I/TestRunner( 1088): at java.lang.reflect.Method.invokeNative(Native Method) 01-10 13:18:18.584 I/TestRunner( 1088): at java.lang.reflect.Method.invoke(Method.java:515) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.runTest(TestCase.java:168) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.runBare(TestCase.java:134) 01-10 13:18:18.584 I/TestRunner( 1088): at com.couchbase.lite.LiteTestCase.runBare(LiteTestCase.java:20) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult$1.protect(TestResult.java:115) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult.runProtected(TestResult.java:133) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestResult.run(TestResult.java:118) 01-10 13:18:18.584 I/TestRunner( 1088): at junit.framework.TestCase.run(TestCase.java:124) 01-10 13:18:18.584 I/TestRunner( 1088): at android.test.AndroidTestRunner.runTest(AndroidTestRunner.java:191) 01-10 13:18:18.584 I/TestRunner( 1088): at android.test.AndroidTestRunner.runTest(AndroidTestRunner.java:176) 01-10 13:18:18.584 I/TestRunner( 1088): at android.test.InstrumentationTestRunner.onStart(InstrumentationTestRunner.java:554) 01-10 13:18:18.584 I/TestRunner( 1088): at android.app.Instrumentation$InstrumentationThread.run(Instrumentation.java:1701) 01-10 13:18:18.584 I/TestRunner( 1088): ----- end exception ----- 01-10 13:18:18.604 I/TestRunner( 1088): finished: testInvokeProcessorAfterReachingCapacity(com.couchbase.lite.support.BatcherTest) 01-10 13:18:18.644 I/TestRunner( 1088): started: testWaitForPendingFutures(com.couchbase.lite.support.BatcherTest) ```
build
testinvokeprocessorafterreachingcapacity failed on jenkins i testrunner failed testinvokeprocessorafterreachingcapacity com couchbase lite support batchertest i testrunner begin exception i testrunner i testrunner junit framework assertionfailederror i testrunner at junit framework assert fail assert java i testrunner at junit framework assert asserttrue assert java i testrunner at junit framework assert asserttrue assert java i testrunner at com couchbase lite support batchertest testinvokeprocessorafterreachingcapacity batchertest java i testrunner at java lang reflect method invokenative native method i testrunner at java lang reflect method invoke method java i testrunner at junit framework testcase runtest testcase java i testrunner at junit framework testcase runbare testcase java i testrunner at com couchbase lite litetestcase runbare litetestcase java i testrunner at junit framework testresult protect testresult java i testrunner at junit framework testresult runprotected testresult java i testrunner at junit framework testresult run testresult java i testrunner at junit framework testcase run testcase java i testrunner at android test androidtestrunner runtest androidtestrunner java i testrunner at android test androidtestrunner runtest androidtestrunner java i testrunner at android test instrumentationtestrunner onstart instrumentationtestrunner java i testrunner at android app instrumentation instrumentationthread run instrumentation java i testrunner end exception i testrunner finished testinvokeprocessorafterreachingcapacity com couchbase lite support batchertest i testrunner started testwaitforpendingfutures com couchbase lite support batchertest
1
259,971
8,202,171,170
IssuesEvent
2018-09-02 05:11:47
hack4impact-uiuc/h4i-recruitment
https://api.github.com/repos/hack4impact-uiuc/h4i-recruitment
closed
Add Categories given to a Candidate by an Interviewer
Priority: High
Look into the Software Dev Rubric. Each Interviewer should give a Candidate a `category`, which is an additional categorization besides the qualitative ranking. Make a new dropdown Thus you need to: - [x] add `category` input where the interviewer could add their choice. - [x] add a text input where the interviewer would add their reasoning - [ ] show the `category` given by the interviewers in the candidate's page
1.0
Add Categories given to a Candidate by an Interviewer - Look into the Software Dev Rubric. Each Interviewer should give a Candidate a `category`, which is an additional categorization besides the qualitative ranking. Make a new dropdown Thus you need to: - [x] add `category` input where the interviewer could add their choice. - [x] add a text input where the interviewer would add their reasoning - [ ] show the `category` given by the interviewers in the candidate's page
non_build
add categories given to a candidate by an interviewer look into the software dev rubric each interviewer should give a candidate a category which is an additional categorization besides the qualitative ranking make a new dropdown thus you need to add category input where the interviewer could add their choice add a text input where the interviewer would add their reasoning show the category given by the interviewers in the candidate s page
0
23,288
10,867,656,765
IssuesEvent
2019-11-15 00:43:45
omisego/plasma-contracts
https://api.github.com/repos/omisego/plasma-contracts
closed
DOS / griefing attacks on calls from processExits()
bug security
## Issue Type ``` [ x ] bug report [ ] feature request ``` ## Current Behavior `processExits` is susceptible to attacks. The first scenario is a DOS attack that would break processing of items in the `PriorityQueue`. Consider the following scenario that in a reduced version mimics how `PriorityQueue` is used for exit game. 1.) Add items to the queue 2.) Wait a defined period of time `minExitPeriod` 3.) Reduce the queue by N items (in order they were added) and while doing so make a call to another untrusted contract/EOA (all contracts that are not part of the deployment) If during `removeItemsfromQueue` a contract is called that drains the entire gas stipend then even only one item from the queue could fail and permanently lock the queue. The scenario below fails when processing two items but this depends on the amount of gas used after the failed call and how much gas was provided. Likely exiting to `GasDos` will beak the exit game queue with only one item due to more instruction being executed after a call. ``` pragma solidity 0.5.11; import "../../src/framework/utils/PriorityQueue.sol"; contract PriorityQueueDOS { PriorityQueue pq; address callee; uint numberOfRuns; constructor (address _callee) public { pq = new PriorityQueue(); callee = _callee; } function addItemstoQueue(uint _numberOfRuns) public { numberOfRuns = _numberOfRuns; for (uint i=0; i<numberOfRuns; i++){ pq.insert(1); } assert(pq.currentSize() == numberOfRuns); } function removeItemsfromQueue() public{ for (uint i=0; i<numberOfRuns; i++){ callee.call.value(0)(""); pq.delMin(); } assert(pq.currentSize() == 0); } } contract GasDos { function () external payable{ while (1==1){ } } } ``` The second attack scenario is a griefing attack. The paradigm of using a queue where items are strictly processed after their ordering and where users might call `processExits()` for exiting UTXOs that they have not initiated, is by design susceptible to this type of attack. The problem can only be mitigated by limiting the amount of gas that is provided to a call. There will be proportional griefing attack possible where a malicious user starts a large number of exits and pays X amount in fees and "good users" are interested in getting their exists processed and they have to pay the Y amount in processing fees to get to process their exits. Malicious users can use gas siphoning attacks to balance the spent fee ratio more in their favor. ## Suggested Fix I propose to set the fees for all external calls to untrusted addresses as low as possible. We probably want to provide enough so that a fallback function can process it and emit an event.We also need to consider the call `SafeERC20`. We should make the gas stipend update-able so that we can respond to changing gas prices in the future. Let's consider a similar mechanism than we have for `updateBondSize`. Up for discussion. As for the griefing attack I don't see how we can fix that with the current design. I think a malicious user needs proportionally more gas than a 'good user' but I want to look more into the economics.
True
DOS / griefing attacks on calls from processExits() - ## Issue Type ``` [ x ] bug report [ ] feature request ``` ## Current Behavior `processExits` is susceptible to attacks. The first scenario is a DOS attack that would break processing of items in the `PriorityQueue`. Consider the following scenario that in a reduced version mimics how `PriorityQueue` is used for exit game. 1.) Add items to the queue 2.) Wait a defined period of time `minExitPeriod` 3.) Reduce the queue by N items (in order they were added) and while doing so make a call to another untrusted contract/EOA (all contracts that are not part of the deployment) If during `removeItemsfromQueue` a contract is called that drains the entire gas stipend then even only one item from the queue could fail and permanently lock the queue. The scenario below fails when processing two items but this depends on the amount of gas used after the failed call and how much gas was provided. Likely exiting to `GasDos` will beak the exit game queue with only one item due to more instruction being executed after a call. ``` pragma solidity 0.5.11; import "../../src/framework/utils/PriorityQueue.sol"; contract PriorityQueueDOS { PriorityQueue pq; address callee; uint numberOfRuns; constructor (address _callee) public { pq = new PriorityQueue(); callee = _callee; } function addItemstoQueue(uint _numberOfRuns) public { numberOfRuns = _numberOfRuns; for (uint i=0; i<numberOfRuns; i++){ pq.insert(1); } assert(pq.currentSize() == numberOfRuns); } function removeItemsfromQueue() public{ for (uint i=0; i<numberOfRuns; i++){ callee.call.value(0)(""); pq.delMin(); } assert(pq.currentSize() == 0); } } contract GasDos { function () external payable{ while (1==1){ } } } ``` The second attack scenario is a griefing attack. The paradigm of using a queue where items are strictly processed after their ordering and where users might call `processExits()` for exiting UTXOs that they have not initiated, is by design susceptible to this type of attack. The problem can only be mitigated by limiting the amount of gas that is provided to a call. There will be proportional griefing attack possible where a malicious user starts a large number of exits and pays X amount in fees and "good users" are interested in getting their exists processed and they have to pay the Y amount in processing fees to get to process their exits. Malicious users can use gas siphoning attacks to balance the spent fee ratio more in their favor. ## Suggested Fix I propose to set the fees for all external calls to untrusted addresses as low as possible. We probably want to provide enough so that a fallback function can process it and emit an event.We also need to consider the call `SafeERC20`. We should make the gas stipend update-able so that we can respond to changing gas prices in the future. Let's consider a similar mechanism than we have for `updateBondSize`. Up for discussion. As for the griefing attack I don't see how we can fix that with the current design. I think a malicious user needs proportionally more gas than a 'good user' but I want to look more into the economics.
non_build
dos griefing attacks on calls from processexits issue type bug report feature request current behavior processexits is susceptible to attacks the first scenario is a dos attack that would break processing of items in the priorityqueue consider the following scenario that in a reduced version mimics how priorityqueue is used for exit game add items to the queue wait a defined period of time minexitperiod reduce the queue by n items in order they were added and while doing so make a call to another untrusted contract eoa all contracts that are not part of the deployment if during removeitemsfromqueue a contract is called that drains the entire gas stipend then even only one item from the queue could fail and permanently lock the queue the scenario below fails when processing two items but this depends on the amount of gas used after the failed call and how much gas was provided likely exiting to gasdos will beak the exit game queue with only one item due to more instruction being executed after a call pragma solidity import src framework utils priorityqueue sol contract priorityqueuedos priorityqueue pq address callee uint numberofruns constructor address callee public pq new priorityqueue callee callee function additemstoqueue uint numberofruns public numberofruns numberofruns for uint i i numberofruns i pq insert assert pq currentsize numberofruns function removeitemsfromqueue public for uint i i numberofruns i callee call value pq delmin assert pq currentsize contract gasdos function external payable while the second attack scenario is a griefing attack the paradigm of using a queue where items are strictly processed after their ordering and where users might call processexits for exiting utxos that they have not initiated is by design susceptible to this type of attack the problem can only be mitigated by limiting the amount of gas that is provided to a call there will be proportional griefing attack possible where a malicious user starts a large number of exits and pays x amount in fees and good users are interested in getting their exists processed and they have to pay the y amount in processing fees to get to process their exits malicious users can use gas siphoning attacks to balance the spent fee ratio more in their favor suggested fix i propose to set the fees for all external calls to untrusted addresses as low as possible we probably want to provide enough so that a fallback function can process it and emit an event we also need to consider the call we should make the gas stipend update able so that we can respond to changing gas prices in the future let s consider a similar mechanism than we have for updatebondsize up for discussion as for the griefing attack i don t see how we can fix that with the current design i think a malicious user needs proportionally more gas than a good user but i want to look more into the economics
0
546,545
16,014,403,551
IssuesEvent
2021-04-20 14:26:43
enso-org/ide
https://api.github.com/repos/enso-org/ide
closed
Support new `executionContext/visualisationEvaluationFailed` notification
Category: Controllers Category: GUI Difficulty: Core Contributor Priority: High Type: Enhancement
<!-- Please ensure that you check the latest version of Enso IDE to see if your feature has been implemented. --> ### General Summary <!-- - Describe the feature you are requesting. --> enso-org/enso#1671 adds new `executionContext/visualisationEvaluationFailed` notification about visualisation errors. In general, there are two kinds of visualisation errors. 1. [`VisualisationExpressionError`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#visualisationexpressionerror), the error reply on [`executionContext/attachVisualisation`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#executioncontextattachvisualisation) or [`executionContext/modifyVisualisation`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#executioncontextmodifyvisualisation) request. This error is sent when the text visualisation expression is compiled into an Enso function. 2. (new) `executionContext/visualisationEvaluationFailed` notification sent during the program execution, when the computed value is applied to the visualisation function. ### Motivation <!-- - A description of the motivation for adding this feature to Enso IDE. - Ideally this would include use-cases that support the feature. --> Show visualisation errors in IDE.
1.0
Support new `executionContext/visualisationEvaluationFailed` notification - <!-- Please ensure that you check the latest version of Enso IDE to see if your feature has been implemented. --> ### General Summary <!-- - Describe the feature you are requesting. --> enso-org/enso#1671 adds new `executionContext/visualisationEvaluationFailed` notification about visualisation errors. In general, there are two kinds of visualisation errors. 1. [`VisualisationExpressionError`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#visualisationexpressionerror), the error reply on [`executionContext/attachVisualisation`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#executioncontextattachvisualisation) or [`executionContext/modifyVisualisation`](https://enso.org/docs/developer/enso/language-server/protocol-language-server.html#executioncontextmodifyvisualisation) request. This error is sent when the text visualisation expression is compiled into an Enso function. 2. (new) `executionContext/visualisationEvaluationFailed` notification sent during the program execution, when the computed value is applied to the visualisation function. ### Motivation <!-- - A description of the motivation for adding this feature to Enso IDE. - Ideally this would include use-cases that support the feature. --> Show visualisation errors in IDE.
non_build
support new executioncontext visualisationevaluationfailed notification please ensure that you check the latest version of enso ide to see if your feature has been implemented general summary describe the feature you are requesting enso org enso adds new executioncontext visualisationevaluationfailed notification about visualisation errors in general there are two kinds of visualisation errors the error reply on or request this error is sent when the text visualisation expression is compiled into an enso function new executioncontext visualisationevaluationfailed notification sent during the program execution when the computed value is applied to the visualisation function motivation a description of the motivation for adding this feature to enso ide ideally this would include use cases that support the feature show visualisation errors in ide
0
163,848
13,928,614,280
IssuesEvent
2020-10-21 21:48:04
AssemblyScript/working-group
https://api.github.com/repos/AssemblyScript/working-group
closed
AssemblyScript Public Meeting #23 - October 21st, 2020
documentation enhancement good first issue help wanted
# Date and Time This public meeting will take place: October 21st, 2020, 18:00 UTC (11:00 AM US PDT, UTC -8) # General Agenda * Agenda Items from comments left on this Github issue * Additional in-meeting comments / discussion * [If time allows] Recap of [WebAssembly CG meeting](https://github.com/WebAssembly/meetings) if anyone attended Feel free to comment on this issue if you have any agenda items you would like to bring up. Meeting Notes will be placed on this issue for those who cannot make the meeting. # Meeting Information We will use the AssemblyScript discord for meetings: https://discord.gg/vSBbJ5T . We will meet in the "meeting" voice channel :+1: Anyone who is contributing to the AssemblyScript project, building something with AssemblyScript, interested in the Assembly project or WebAssembly in general, is welcome to join! 😄 Notes will be taken by the host, and posted after the meeting. Notes are free to be edited through comments on the meeting notes at a later time.
1.0
AssemblyScript Public Meeting #23 - October 21st, 2020 - # Date and Time This public meeting will take place: October 21st, 2020, 18:00 UTC (11:00 AM US PDT, UTC -8) # General Agenda * Agenda Items from comments left on this Github issue * Additional in-meeting comments / discussion * [If time allows] Recap of [WebAssembly CG meeting](https://github.com/WebAssembly/meetings) if anyone attended Feel free to comment on this issue if you have any agenda items you would like to bring up. Meeting Notes will be placed on this issue for those who cannot make the meeting. # Meeting Information We will use the AssemblyScript discord for meetings: https://discord.gg/vSBbJ5T . We will meet in the "meeting" voice channel :+1: Anyone who is contributing to the AssemblyScript project, building something with AssemblyScript, interested in the Assembly project or WebAssembly in general, is welcome to join! 😄 Notes will be taken by the host, and posted after the meeting. Notes are free to be edited through comments on the meeting notes at a later time.
non_build
assemblyscript public meeting october date and time this public meeting will take place october utc am us pdt utc general agenda agenda items from comments left on this github issue additional in meeting comments discussion recap of if anyone attended feel free to comment on this issue if you have any agenda items you would like to bring up meeting notes will be placed on this issue for those who cannot make the meeting meeting information we will use the assemblyscript discord for meetings we will meet in the meeting voice channel anyone who is contributing to the assemblyscript project building something with assemblyscript interested in the assembly project or webassembly in general is welcome to join 😄 notes will be taken by the host and posted after the meeting notes are free to be edited through comments on the meeting notes at a later time
0
59,854
14,664,237,575
IssuesEvent
2020-12-29 11:31:12
sandboxie-plus/Sandboxie
https://api.github.com/repos/sandboxie-plus/Sandboxie
closed
Sandboxie-Plus-x64-v0.5.2.exe says it is going to install 0.5.1
fixed in next build
When I run Sandboxie-Plus-x64-v0.5.2.exe I get this screen : ![image](https://user-images.githubusercontent.com/3707138/102995929-c15d6b00-452a-11eb-8228-4d435b1e9b19.png)
1.0
Sandboxie-Plus-x64-v0.5.2.exe says it is going to install 0.5.1 - When I run Sandboxie-Plus-x64-v0.5.2.exe I get this screen : ![image](https://user-images.githubusercontent.com/3707138/102995929-c15d6b00-452a-11eb-8228-4d435b1e9b19.png)
build
sandboxie plus exe says it is going to install when i run sandboxie plus exe i get this screen
1
443,574
12,796,349,959
IssuesEvent
2020-07-02 10:17:04
trufflesuite/truffle
https://api.github.com/repos/trufflesuite/truffle
closed
Latest Truffle breaks our box
bug needs reproduced priority2 ⚠️
## Issue Our flashloan-box worked previously with Truffle v.5.10. It no longer works in latest Truffle. ## Steps to Reproduce Our Flashloan-box: https://github.com/aave/flashloan-box 1. Using Truffle v.5.10, follow instructions of box. Everything should work. 2. Using latest Truffle version, follow instructions. 3. It fails on Step 8, when defining `const f = await Flashloan.deployed()` ## Expected Behavior It should work as previous Truffle version. ## Actual Results Errors with `ReferenceERrror: f is no defined` ## Environment * Operating System: macOS * Ethereum client: * Truffle version (`truffle version`): 5.1.24 * node version (`node --version`): 11.15 * npm version (`npm --version`): 6.14.4
1.0
Latest Truffle breaks our box - ## Issue Our flashloan-box worked previously with Truffle v.5.10. It no longer works in latest Truffle. ## Steps to Reproduce Our Flashloan-box: https://github.com/aave/flashloan-box 1. Using Truffle v.5.10, follow instructions of box. Everything should work. 2. Using latest Truffle version, follow instructions. 3. It fails on Step 8, when defining `const f = await Flashloan.deployed()` ## Expected Behavior It should work as previous Truffle version. ## Actual Results Errors with `ReferenceERrror: f is no defined` ## Environment * Operating System: macOS * Ethereum client: * Truffle version (`truffle version`): 5.1.24 * node version (`node --version`): 11.15 * npm version (`npm --version`): 6.14.4
non_build
latest truffle breaks our box issue our flashloan box worked previously with truffle v it no longer works in latest truffle steps to reproduce our flashloan box using truffle v follow instructions of box everything should work using latest truffle version follow instructions it fails on step when defining const f await flashloan deployed expected behavior it should work as previous truffle version actual results errors with referenceerrror f is no defined environment operating system macos ethereum client truffle version truffle version node version node version npm version npm version
0
41,336
10,702,768,859
IssuesEvent
2019-10-24 08:12:50
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
What are the .keep files created when creating a ballerina projects?
Area/BuildTools BetaTesting Priority/Low Type/Docs Type/Question
"resources" folders in ballerina project contains a .keep file in each "resources" folder. What are they used for? Better if we can add a small description to docs explaining the use of them.
1.0
What are the .keep files created when creating a ballerina projects? - "resources" folders in ballerina project contains a .keep file in each "resources" folder. What are they used for? Better if we can add a small description to docs explaining the use of them.
build
what are the keep files created when creating a ballerina projects resources folders in ballerina project contains a keep file in each resources folder what are they used for better if we can add a small description to docs explaining the use of them
1
147,675
5,643,789,862
IssuesEvent
2017-04-07 01:57:48
leo-project/leofs
https://api.github.com/repos/leo-project/leofs
opened
[leofs_doctor] Add compaction related info
Improve Priority-MIDDLE
With #671, we can include the compaction related info in https://github.com/leo-project/leofs_doctor when the target node is leo_storage.
1.0
[leofs_doctor] Add compaction related info - With #671, we can include the compaction related info in https://github.com/leo-project/leofs_doctor when the target node is leo_storage.
non_build
add compaction related info with we can include the compaction related info in when the target node is leo storage
0
213,652
16,529,004,201
IssuesEvent
2021-05-27 01:39:36
ppb/pursuedpybear
https://api.github.com/repos/ppb/pursuedpybear
closed
Add a Failer to test_signal_once
Hacktoberfest new contributor tests
Relevant test: https://github.com/ppb/pursuedpybear/blob/canon/tests/test_engine.py#L60 We need to add a Failer with a reasonable timeout to this test to prevent infinite loops causing CI time outs. Failer defaults to 1 second, which is reasonable for most tests that will fail with it. Import `ppb.testutils.Failer` and add it to the basic_systems in the test in question.
1.0
Add a Failer to test_signal_once - Relevant test: https://github.com/ppb/pursuedpybear/blob/canon/tests/test_engine.py#L60 We need to add a Failer with a reasonable timeout to this test to prevent infinite loops causing CI time outs. Failer defaults to 1 second, which is reasonable for most tests that will fail with it. Import `ppb.testutils.Failer` and add it to the basic_systems in the test in question.
non_build
add a failer to test signal once relevant test we need to add a failer with a reasonable timeout to this test to prevent infinite loops causing ci time outs failer defaults to second which is reasonable for most tests that will fail with it import ppb testutils failer and add it to the basic systems in the test in question
0
12,196
8,637,134,730
IssuesEvent
2018-11-23 10:13:59
kyma-project/kyma
https://api.github.com/repos/kyma-project/kyma
closed
Missing authentication and authorization of Jaeger API clients
area/security area/tracing security/medium wontfix
Easiness of attack: Advanced Impact: Moderate ## Description Jaeger exposes 2 kinds of APIs to the local network: - a rest query API, typically used by the Jaeger UI - a trace ingestion API, using the Zipkin protocol (possibly there are other ingestion APIs, too) None of these APIs authenticate their clients. All clients are authorized to perform any operations that are exposed by the respective API. ## Risk: An attacker who has gained access to the local network can read and write arbitrary traces. That compromises confidentiality and integrity. ## Suggested Mitigation: Protect the rest query API by istio. If necessary, expose it to the Internet by an ingress controller. Make sure that Jaeger UI users are authenticated (via dex) and enforce RBAC rules for Jaeger. Consider protecting trace ingestion with istio. Disable unused ingestion APIs, if Jaeger still exposes any.
True
Missing authentication and authorization of Jaeger API clients - Easiness of attack: Advanced Impact: Moderate ## Description Jaeger exposes 2 kinds of APIs to the local network: - a rest query API, typically used by the Jaeger UI - a trace ingestion API, using the Zipkin protocol (possibly there are other ingestion APIs, too) None of these APIs authenticate their clients. All clients are authorized to perform any operations that are exposed by the respective API. ## Risk: An attacker who has gained access to the local network can read and write arbitrary traces. That compromises confidentiality and integrity. ## Suggested Mitigation: Protect the rest query API by istio. If necessary, expose it to the Internet by an ingress controller. Make sure that Jaeger UI users are authenticated (via dex) and enforce RBAC rules for Jaeger. Consider protecting trace ingestion with istio. Disable unused ingestion APIs, if Jaeger still exposes any.
non_build
missing authentication and authorization of jaeger api clients easiness of attack advanced impact moderate description jaeger exposes kinds of apis to the local network a rest query api typically used by the jaeger ui a trace ingestion api using the zipkin protocol possibly there are other ingestion apis too none of these apis authenticate their clients all clients are authorized to perform any operations that are exposed by the respective api risk an attacker who has gained access to the local network can read and write arbitrary traces that compromises confidentiality and integrity suggested mitigation protect the rest query api by istio if necessary expose it to the internet by an ingress controller make sure that jaeger ui users are authenticated via dex and enforce rbac rules for jaeger consider protecting trace ingestion with istio disable unused ingestion apis if jaeger still exposes any
0
63,931
15,759,536,607
IssuesEvent
2021-03-31 08:02:08
kubernetes-sigs/prometheus-adapter
https://api.github.com/repos/kubernetes-sigs/prometheus-adapter
closed
make docker-build does not work
build-deploy kind/bug lifecycle/rotten
1. make docker-build is not working 2. Even if we make it work, does it work with Kubernetes 1.9 ?
1.0
make docker-build does not work - 1. make docker-build is not working 2. Even if we make it work, does it work with Kubernetes 1.9 ?
build
make docker build does not work make docker build is not working even if we make it work does it work with kubernetes
1
93,678
27,013,798,636
IssuesEvent
2023-02-10 17:27:53
microsoft/appcenter
https://api.github.com/repos/microsoft/appcenter
closed
org.jetbrains.kotlin.gradle.internal.KaptWithoutKotlincTask$KaptExecutionWorkAction
feature request build
I am trying to generate release build through appcenter but in local setup i am able to generate android apk. FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':app:kaptReleaseKotlin'. > A failure occurred while executing org.jetbrains.kotlin.gradle.internal.KaptWithoutKotlincTask$KaptExecutionWorkAction > java.lang.reflect.InvocationTargetException (no error message) * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org/ BUILD FAILED in 14m 52s
1.0
org.jetbrains.kotlin.gradle.internal.KaptWithoutKotlincTask$KaptExecutionWorkAction - I am trying to generate release build through appcenter but in local setup i am able to generate android apk. FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':app:kaptReleaseKotlin'. > A failure occurred while executing org.jetbrains.kotlin.gradle.internal.KaptWithoutKotlincTask$KaptExecutionWorkAction > java.lang.reflect.InvocationTargetException (no error message) * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org/ BUILD FAILED in 14m 52s
build
org jetbrains kotlin gradle internal kaptwithoutkotlinctask kaptexecutionworkaction i am trying to generate release build through appcenter but in local setup i am able to generate android apk failure build failed with an exception what went wrong execution failed for task app kaptreleasekotlin a failure occurred while executing org jetbrains kotlin gradle internal kaptwithoutkotlinctask kaptexecutionworkaction java lang reflect invocationtargetexception no error message try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights get more help at build failed in
1
19,397
6,718,969,580
IssuesEvent
2017-10-15 18:54:04
quasar-analytics/quasar
https://api.github.com/repos/quasar-analytics/quasar
closed
ServiceSpec fails on Windows
epic: Quasar Doesn't Build/Work on Windows status: In review topic: CI (Travis) type: Bug
``` [error] x mount filesystem (2 seconds, 234 ms) [error] '\/-(false)' is not \/- with value'true' (ServiceSpec.scala:116) ``` This error appears to be stable and reproducible. I'm pretty sure it's an actual failure.
1.0
ServiceSpec fails on Windows - ``` [error] x mount filesystem (2 seconds, 234 ms) [error] '\/-(false)' is not \/- with value'true' (ServiceSpec.scala:116) ``` This error appears to be stable and reproducible. I'm pretty sure it's an actual failure.
build
servicespec fails on windows x mount filesystem seconds ms false is not with value true servicespec scala this error appears to be stable and reproducible i m pretty sure it s an actual failure
1
46,946
24,794,790,556
IssuesEvent
2022-10-24 16:18:42
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
closed
Slow JSONExtract with LowCardinality(String) tuples
performance
`JSONExtract` of Tuples with `LowCardinality(String)` is significantly slower than with regular strings: ~2.5 slower. Tested on CH 21.8.8.29 ## Unexpectedly slow example **60MB/s** ``` cat test.ndjson | clickhouse local -q " SELECT JSONExtract(event, 'Tuple( a LowCardinality(String), b LowCardinality(String), c LowCardinality(String), d LowCardinality(String) )') AS json FROM table" --input-format LineAsString --output-format Native --structure " event String" --verbose --logger.level trace >> /dev/null ``` ## Expectedly fast example **150 MB/s** ``` time cat test.ndjson | clickhouse local -q " SELECT JSONExtract(event, 'Tuple( a String, b String, c String, d String )') AS json FROM table" --input-format LineAsString --output-format Native --structure " event String" --verbose --logger.level trace >> /dev/null ``` ## Workaround **140 MB/s** ``` cat test.ndjson | clickhouse local -q " SELECT toLowCardinality(json.1) AS a, toLowCardinality(json.2) AS b, toLowCardinality(json.3) AS c, toLowCardinality(json.4) AS d FROM (SELECT JSONExtract(event, 'Tuple( a String, b String, c String, d String )') AS json FROM table)" --input-format LineAsString --output-format Native --structure " event String" --verbose --logger.level trace >> /dev/null ``` cc: @Algunenano [test.ndjson.zip](https://github.com/ClickHouse/ClickHouse/files/7347611/test.ndjson.zip)
True
Slow JSONExtract with LowCardinality(String) tuples - `JSONExtract` of Tuples with `LowCardinality(String)` is significantly slower than with regular strings: ~2.5 slower. Tested on CH 21.8.8.29 ## Unexpectedly slow example **60MB/s** ``` cat test.ndjson | clickhouse local -q " SELECT JSONExtract(event, 'Tuple( a LowCardinality(String), b LowCardinality(String), c LowCardinality(String), d LowCardinality(String) )') AS json FROM table" --input-format LineAsString --output-format Native --structure " event String" --verbose --logger.level trace >> /dev/null ``` ## Expectedly fast example **150 MB/s** ``` time cat test.ndjson | clickhouse local -q " SELECT JSONExtract(event, 'Tuple( a String, b String, c String, d String )') AS json FROM table" --input-format LineAsString --output-format Native --structure " event String" --verbose --logger.level trace >> /dev/null ``` ## Workaround **140 MB/s** ``` cat test.ndjson | clickhouse local -q " SELECT toLowCardinality(json.1) AS a, toLowCardinality(json.2) AS b, toLowCardinality(json.3) AS c, toLowCardinality(json.4) AS d FROM (SELECT JSONExtract(event, 'Tuple( a String, b String, c String, d String )') AS json FROM table)" --input-format LineAsString --output-format Native --structure " event String" --verbose --logger.level trace >> /dev/null ``` cc: @Algunenano [test.ndjson.zip](https://github.com/ClickHouse/ClickHouse/files/7347611/test.ndjson.zip)
non_build
slow jsonextract with lowcardinality string tuples jsonextract of tuples with lowcardinality string is significantly slower than with regular strings slower tested on ch unexpectedly slow example s cat test ndjson clickhouse local q select jsonextract event tuple a lowcardinality string b lowcardinality string c lowcardinality string d lowcardinality string as json from table input format lineasstring output format native structure event string verbose logger level trace dev null expectedly fast example mb s time cat test ndjson clickhouse local q select jsonextract event tuple a string b string c string d string as json from table input format lineasstring output format native structure event string verbose logger level trace dev null workaround mb s cat test ndjson clickhouse local q select tolowcardinality json as a tolowcardinality json as b tolowcardinality json as c tolowcardinality json as d from select jsonextract event tuple a string b string c string d string as json from table input format lineasstring output format native structure event string verbose logger level trace dev null cc algunenano
0
23,659
7,359,199,517
IssuesEvent
2018-03-10 03:14:23
ild-games/Spool
https://api.github.com/repos/ild-games/Spool
opened
Configure code coverage to include all sources, not just files that are touched by a unit test
build
Angular code coverage testing is done via `ng test --code-coverage`. This uses istanbul under the hood by angular-cli to run code coverage checks. However, by default this will only cover source files that are covered by unit tests. This turns the code coverage check into a not very helpful metric, as we are only getting "What percentage of code paths weren't hit that could have been hit based on your existing unit tests" instead of "What percentage of your code based isn't unit tested".
1.0
Configure code coverage to include all sources, not just files that are touched by a unit test - Angular code coverage testing is done via `ng test --code-coverage`. This uses istanbul under the hood by angular-cli to run code coverage checks. However, by default this will only cover source files that are covered by unit tests. This turns the code coverage check into a not very helpful metric, as we are only getting "What percentage of code paths weren't hit that could have been hit based on your existing unit tests" instead of "What percentage of your code based isn't unit tested".
build
configure code coverage to include all sources not just files that are touched by a unit test angular code coverage testing is done via ng test code coverage this uses istanbul under the hood by angular cli to run code coverage checks however by default this will only cover source files that are covered by unit tests this turns the code coverage check into a not very helpful metric as we are only getting what percentage of code paths weren t hit that could have been hit based on your existing unit tests instead of what percentage of your code based isn t unit tested
1
102,447
32,007,115,681
IssuesEvent
2023-09-21 15:28:01
pwa-builder/PWABuilder
https://api.github.com/repos/pwa-builder/PWABuilder
closed
[PWA Builder]: Spin button form field does not have label.
bug :bug: needs triage :mag: needs attention :wave: a11y A11yWCAG Accessibility [E+D] A11ySev2 Product-PWABuilder A11yAuto HCL-E+D CT-July23
### What happened? Spin button form field present on the URL does not have label, narrator does not announce the information of the 'Padding' label. ### How do we reproduce the behavior? **Repro Steps:** 1. Open the PWA Builder URL in Anaheim dev browser. 2. PWA Builder page will appear. 3. Now Run the Fast Pass from AI4W. 4. Observe the issue. ### What do you expect to happen? Spin button form field present on the URL must be associated with the label 'Padding'. ### What environment were you using? **Test Environment:** OS Build: Windows 11 Enterprise Insider Preview 22H2 (23493.1000) Edge Browser: Version 115.0.1880.3 (Official build) dev (64-bit) URL: https://www.pwabuilder.com/imageGenerator Tool: AI4W ### Additional context **Element path:** app-index,image-generator,sl-input,#input **Snippet:** <input part="input" id="input" class="input__control" aria-describedby="help-text" type="number" title="" name="padding" required="" placeholder="" min="0" max="1" step="0.1" spellcheck="true"> **How to fix:** Fix all of the following: Form element has explicit <label> that is hidden **User Experience:** If the label of the form element will not be defined then the users depend on screen reader will not get the information of the purpose of the spin button, and they may get confused in accessing the UI properly. **WCAG Reference:** https://www.w3.org/WAI/WCAG21/Understanding/name-role-value.html **Attachments:** ![MAS4 1 2_Spin button form field does not have label](https://github.com/pwa-builder/PWABuilder/assets/117908911/1427e1bc-000b-4581-84f7-55f313510915) ![Code_Snip](https://github.com/pwa-builder/PWABuilder/assets/117908911/369d31f5-9c72-4fb1-ac5c-9fd005a5788a)
1.0
[PWA Builder]: Spin button form field does not have label. - ### What happened? Spin button form field present on the URL does not have label, narrator does not announce the information of the 'Padding' label. ### How do we reproduce the behavior? **Repro Steps:** 1. Open the PWA Builder URL in Anaheim dev browser. 2. PWA Builder page will appear. 3. Now Run the Fast Pass from AI4W. 4. Observe the issue. ### What do you expect to happen? Spin button form field present on the URL must be associated with the label 'Padding'. ### What environment were you using? **Test Environment:** OS Build: Windows 11 Enterprise Insider Preview 22H2 (23493.1000) Edge Browser: Version 115.0.1880.3 (Official build) dev (64-bit) URL: https://www.pwabuilder.com/imageGenerator Tool: AI4W ### Additional context **Element path:** app-index,image-generator,sl-input,#input **Snippet:** <input part="input" id="input" class="input__control" aria-describedby="help-text" type="number" title="" name="padding" required="" placeholder="" min="0" max="1" step="0.1" spellcheck="true"> **How to fix:** Fix all of the following: Form element has explicit <label> that is hidden **User Experience:** If the label of the form element will not be defined then the users depend on screen reader will not get the information of the purpose of the spin button, and they may get confused in accessing the UI properly. **WCAG Reference:** https://www.w3.org/WAI/WCAG21/Understanding/name-role-value.html **Attachments:** ![MAS4 1 2_Spin button form field does not have label](https://github.com/pwa-builder/PWABuilder/assets/117908911/1427e1bc-000b-4581-84f7-55f313510915) ![Code_Snip](https://github.com/pwa-builder/PWABuilder/assets/117908911/369d31f5-9c72-4fb1-ac5c-9fd005a5788a)
build
spin button form field does not have label what happened spin button form field present on the url does not have label narrator does not announce the information of the padding label how do we reproduce the behavior repro steps open the pwa builder url in anaheim dev browser pwa builder page will appear now run the fast pass from observe the issue what do you expect to happen spin button form field present on the url must be associated with the label padding what environment were you using test environment os build windows enterprise insider preview edge browser version official build dev bit url tool additional context element path app index image generator sl input input snippet how to fix fix all of the following form element has explicit that is hidden user experience if the label of the form element will not be defined then the users depend on screen reader will not get the information of the purpose of the spin button and they may get confused in accessing the ui properly wcag reference attachments
1
26,685
7,858,659,330
IssuesEvent
2018-06-21 14:30:07
ShaikASK/Testing
https://api.github.com/repos/ShaikASK/Testing
closed
Safari (10) /IE (11) : New Hires : Vertical scroll bar is missing in new hires screen as a result of which only 50% of screen is visible
Defect HR Admin Module HR User Module New Hire P2 Release #3 Build 2
Steps : 1.Launch the URL 2.Sign in as HR user 3.Click on New Hires from side menu 4.Navigate to "View New Hires" page 5.Click on + icon beside "New Hires" tab Experienced Behaviour : Observed that Vertical scroll bar is missing in new hires screen as a result of which only 50% of screen is visible when check with safari (10) & IE (11) (Refer screen shot) Expected Behaviour : Ensure that application should display vertical scroll bar in new hires screen and it should should display proper UI ![safari 10](https://user-images.githubusercontent.com/31955617/41638513-5f1aad44-7477-11e8-8702-87a862b7d49b.png)
1.0
Safari (10) /IE (11) : New Hires : Vertical scroll bar is missing in new hires screen as a result of which only 50% of screen is visible - Steps : 1.Launch the URL 2.Sign in as HR user 3.Click on New Hires from side menu 4.Navigate to "View New Hires" page 5.Click on + icon beside "New Hires" tab Experienced Behaviour : Observed that Vertical scroll bar is missing in new hires screen as a result of which only 50% of screen is visible when check with safari (10) & IE (11) (Refer screen shot) Expected Behaviour : Ensure that application should display vertical scroll bar in new hires screen and it should should display proper UI ![safari 10](https://user-images.githubusercontent.com/31955617/41638513-5f1aad44-7477-11e8-8702-87a862b7d49b.png)
build
safari ie new hires vertical scroll bar is missing in new hires screen as a result of which only of screen is visible steps launch the url sign in as hr user click on new hires from side menu navigate to view new hires page click on icon beside new hires tab experienced behaviour observed that vertical scroll bar is missing in new hires screen as a result of which only of screen is visible when check with safari ie refer screen shot expected behaviour ensure that application should display vertical scroll bar in new hires screen and it should should display proper ui
1
166,711
14,075,868,325
IssuesEvent
2020-11-04 09:40:04
packethost/metabot
https://api.github.com/repos/packethost/metabot
opened
Uniform Standards Request: Maintained Repository
documentation enhancement good first issue help wanted
Hello! We believe this repository is Maintained and therefore needs the following files updated: * [ ] Flagged Public * [ ] [The Maintained badge](https://github.com/packethost/standards/blob/master/glossary.md#maintained-badge) * [ ] [The Maintained Statement in the README.md](https://github.com/packethost/standards/blob/master/glossary.md#maintained-statement) * [ ] [Getting Started Tutorial](https://github.com/packethost/standards/blob/master/glossary.md#getting-started-tutorial) * [ ] [README.md](https://github.com/packethost/standards/blob/master/glossary.md#readmemd) * [ ] [LICENSE](https://github.com/packethost/standards/blob/master/glossary.md#license) * [ ] [CI/CD](https://github.com/packethost/standards/blob/master/glossary.md#cicd) * [ ] [OWNERS.md](https://github.com/packethost/standards/blob/master/glossary.md#ownersmd) * [ ] [Developer Certificate of Origin](https://github.com/packethost/standards/blob/master/glossary.md#developer-certificate-of-origin) * [ ] At least two [maintainers](https://github.com/packethost/standards/blob/master/glossary.md#maintainer) * [ ] [A Manifest File](https://github.com/packethost/standards/blob/master/glossary.md#manifest-file), if relevant * [ ] [Code of Conduct](https://github.com/packethost/standards/blob/master/glossary.md#code-of-conduct) * [ ] [How to Contribute](https://github.com/packethost/standards/blob/master/glossary.md#contributors-guide) * [ ] [SUPPORT.md](https://github.com/packethost/standards/blob/master/glossary.md#supportmd) * [ ] [RELEASE.md](https://github.com/packethost/standards/blob/master/glossary.md#releasemd) If you feel the repository should be experimental or end of life or that you'll need assistance to update these files, please let us know by filing an issue with https://github.com/packethost/standards. ## The Uniform Standards Project Packet maintains a number of public repositories that help customers to run various workloads on Packet. These repositories are in various states of completeness and quality, and being public, developers often find them and start using them. This creates problems: * Developers using low-quality repositories may infer that Packet generally provides a low quality experience. * Many of our repositories are put online with no formal communication with, or training for, customer success. This leads to a below average support experience when things do go wrong. * We spend a huge amount of time supporting users through various channels when with better upfront planning, documentation and testing much of this support work could be eliminated. To that end, we propose three tiers of repositories: [Private](https://github.com/packethost/standards#private-tier-minimum-requirements), [Experimental](https://github.com/packethost/standards#experimental-tier-minimum-requirements), and [Maintained](https://github.com/packethost/standards#maintained-tier-minimum-requirements). As a resource and example of a maintained repository, we've created https://github.com/packethost/standards. This is also where you can file any requests for assistance or modification of scope. ### The Goal Our repositories should be the example from which adjacent, competing, projects look for inspiration. Each repository should not look entirely different from other repositories in the ecosystem, having a different layout, a different testing model, or a different logging model, for example, without reason or recommendation from the subject matter experts from the community. We should share our improvements with each ecosystem while seeking and respecting the feedback of these communities. Whether or not strict guidelines have been provided for the project type, our repositories should ensure that the same components are offered across the board. How these components are provided may vary, based on the conventions of the project type. GitHub provides general guidance on this which they have integrated into their user experience.
1.0
Uniform Standards Request: Maintained Repository - Hello! We believe this repository is Maintained and therefore needs the following files updated: * [ ] Flagged Public * [ ] [The Maintained badge](https://github.com/packethost/standards/blob/master/glossary.md#maintained-badge) * [ ] [The Maintained Statement in the README.md](https://github.com/packethost/standards/blob/master/glossary.md#maintained-statement) * [ ] [Getting Started Tutorial](https://github.com/packethost/standards/blob/master/glossary.md#getting-started-tutorial) * [ ] [README.md](https://github.com/packethost/standards/blob/master/glossary.md#readmemd) * [ ] [LICENSE](https://github.com/packethost/standards/blob/master/glossary.md#license) * [ ] [CI/CD](https://github.com/packethost/standards/blob/master/glossary.md#cicd) * [ ] [OWNERS.md](https://github.com/packethost/standards/blob/master/glossary.md#ownersmd) * [ ] [Developer Certificate of Origin](https://github.com/packethost/standards/blob/master/glossary.md#developer-certificate-of-origin) * [ ] At least two [maintainers](https://github.com/packethost/standards/blob/master/glossary.md#maintainer) * [ ] [A Manifest File](https://github.com/packethost/standards/blob/master/glossary.md#manifest-file), if relevant * [ ] [Code of Conduct](https://github.com/packethost/standards/blob/master/glossary.md#code-of-conduct) * [ ] [How to Contribute](https://github.com/packethost/standards/blob/master/glossary.md#contributors-guide) * [ ] [SUPPORT.md](https://github.com/packethost/standards/blob/master/glossary.md#supportmd) * [ ] [RELEASE.md](https://github.com/packethost/standards/blob/master/glossary.md#releasemd) If you feel the repository should be experimental or end of life or that you'll need assistance to update these files, please let us know by filing an issue with https://github.com/packethost/standards. ## The Uniform Standards Project Packet maintains a number of public repositories that help customers to run various workloads on Packet. These repositories are in various states of completeness and quality, and being public, developers often find them and start using them. This creates problems: * Developers using low-quality repositories may infer that Packet generally provides a low quality experience. * Many of our repositories are put online with no formal communication with, or training for, customer success. This leads to a below average support experience when things do go wrong. * We spend a huge amount of time supporting users through various channels when with better upfront planning, documentation and testing much of this support work could be eliminated. To that end, we propose three tiers of repositories: [Private](https://github.com/packethost/standards#private-tier-minimum-requirements), [Experimental](https://github.com/packethost/standards#experimental-tier-minimum-requirements), and [Maintained](https://github.com/packethost/standards#maintained-tier-minimum-requirements). As a resource and example of a maintained repository, we've created https://github.com/packethost/standards. This is also where you can file any requests for assistance or modification of scope. ### The Goal Our repositories should be the example from which adjacent, competing, projects look for inspiration. Each repository should not look entirely different from other repositories in the ecosystem, having a different layout, a different testing model, or a different logging model, for example, without reason or recommendation from the subject matter experts from the community. We should share our improvements with each ecosystem while seeking and respecting the feedback of these communities. Whether or not strict guidelines have been provided for the project type, our repositories should ensure that the same components are offered across the board. How these components are provided may vary, based on the conventions of the project type. GitHub provides general guidance on this which they have integrated into their user experience.
non_build
uniform standards request maintained repository hello we believe this repository is maintained and therefore needs the following files updated flagged public at least two if relevant if you feel the repository should be experimental or end of life or that you ll need assistance to update these files please let us know by filing an issue with the uniform standards project packet maintains a number of public repositories that help customers to run various workloads on packet these repositories are in various states of completeness and quality and being public developers often find them and start using them this creates problems developers using low quality repositories may infer that packet generally provides a low quality experience many of our repositories are put online with no formal communication with or training for customer success this leads to a below average support experience when things do go wrong we spend a huge amount of time supporting users through various channels when with better upfront planning documentation and testing much of this support work could be eliminated to that end we propose three tiers of repositories and as a resource and example of a maintained repository we ve created this is also where you can file any requests for assistance or modification of scope the goal our repositories should be the example from which adjacent competing projects look for inspiration each repository should not look entirely different from other repositories in the ecosystem having a different layout a different testing model or a different logging model for example without reason or recommendation from the subject matter experts from the community we should share our improvements with each ecosystem while seeking and respecting the feedback of these communities whether or not strict guidelines have been provided for the project type our repositories should ensure that the same components are offered across the board how these components are provided may vary based on the conventions of the project type github provides general guidance on this which they have integrated into their user experience
0
689,938
23,640,915,825
IssuesEvent
2022-08-25 16:58:26
deckhouse/deckhouse
https://api.github.com/repos/deckhouse/deckhouse
closed
[cloud-provider-aws] Add ability to configure additional policies for IAM roles
area/cloud-provider type/feature-request type/good-first-issue priority/backlog
### Preflight Checklist - [X] I agree to follow the [Code of Conduct](https://github.com/deckhouse/deckhouse/blob/main/CODE_OF_CONDUCT.md) that this project adheres to. - [X] I have searched the [issue tracker](https://github.com/deckhouse/deckhouse/issues) for an issue that matches the one I want to file, without success. ### Use case. Why is this important? There are several cases when you might need to give access to AWS API for applications running on nodes. We have a basic role that is attached to every node in the cluster. Additional policies can be attached to it. ### Proposed Solution Add additionalRolePolicies parameter in AWSClusterConfiguration. ### Additional Information _No response_
1.0
[cloud-provider-aws] Add ability to configure additional policies for IAM roles - ### Preflight Checklist - [X] I agree to follow the [Code of Conduct](https://github.com/deckhouse/deckhouse/blob/main/CODE_OF_CONDUCT.md) that this project adheres to. - [X] I have searched the [issue tracker](https://github.com/deckhouse/deckhouse/issues) for an issue that matches the one I want to file, without success. ### Use case. Why is this important? There are several cases when you might need to give access to AWS API for applications running on nodes. We have a basic role that is attached to every node in the cluster. Additional policies can be attached to it. ### Proposed Solution Add additionalRolePolicies parameter in AWSClusterConfiguration. ### Additional Information _No response_
non_build
add ability to configure additional policies for iam roles preflight checklist i agree to follow the that this project adheres to i have searched the for an issue that matches the one i want to file without success use case why is this important there are several cases when you might need to give access to aws api for applications running on nodes we have a basic role that is attached to every node in the cluster additional policies can be attached to it proposed solution add additionalrolepolicies parameter in awsclusterconfiguration additional information no response
0
2,961
5,832,709,801
IssuesEvent
2017-05-08 22:37:46
Cxbx-Reloaded/Cxbx-Reloaded
https://api.github.com/repos/Cxbx-Reloaded/Cxbx-Reloaded
closed
[PAL] Lego Star Wars crashes at language selection screen
game-compatibility
Tested on Cxbx-Reloaded build 7c266ad (Apr 26 2017) Pretty self explanatory. In Hardware mode, the loading screen will come up with working video and immediately crash. Software mode has an immediate crash with no audio or video. Hardware mode crash: ![Photo 1](https://cloud.githubusercontent.com/assets/10300935/25513957/280fb6de-2c2d-11e7-8d3d-b958a6f90ab7.PNG) Hardware mode Kernel Debug file: [KrnlDebug.txt](https://github.com/Cxbx-Reloaded/Cxbx-Reloaded/files/963443/KrnlDebug.txt) Settings used: Display Adapter: NVIDIA GeForce GTX 1060 6GB Direct3D Device: Direct3D HAL (Hardware) Video Resolution: Automatic (Default)
True
[PAL] Lego Star Wars crashes at language selection screen - Tested on Cxbx-Reloaded build 7c266ad (Apr 26 2017) Pretty self explanatory. In Hardware mode, the loading screen will come up with working video and immediately crash. Software mode has an immediate crash with no audio or video. Hardware mode crash: ![Photo 1](https://cloud.githubusercontent.com/assets/10300935/25513957/280fb6de-2c2d-11e7-8d3d-b958a6f90ab7.PNG) Hardware mode Kernel Debug file: [KrnlDebug.txt](https://github.com/Cxbx-Reloaded/Cxbx-Reloaded/files/963443/KrnlDebug.txt) Settings used: Display Adapter: NVIDIA GeForce GTX 1060 6GB Direct3D Device: Direct3D HAL (Hardware) Video Resolution: Automatic (Default)
non_build
lego star wars crashes at language selection screen tested on cxbx reloaded build apr pretty self explanatory in hardware mode the loading screen will come up with working video and immediately crash software mode has an immediate crash with no audio or video hardware mode crash hardware mode kernel debug file settings used display adapter nvidia geforce gtx device hal hardware video resolution automatic default
0