Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
853
labels
stringlengths
4
898
body
stringlengths
2
262k
index
stringclasses
13 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
250k
binary_label
int64
0
1
20,356
6,849,919,325
IssuesEvent
2017-11-14 00:15:37
Microsoft/vscode
https://api.github.com/repos/Microsoft/vscode
closed
semver is defined twice
bug engineering vscode-build
Run the following in the root of our repo `tsc -p src/tsconfig.json --noEmit` and see errors about semver being defined twice. That's true, once via typings/semver.d.ts and once via package.json
1.0
semver is defined twice - Run the following in the root of our repo `tsc -p src/tsconfig.json --noEmit` and see errors about semver being defined twice. That's true, once via typings/semver.d.ts and once via package.json
build
semver is defined twice run the following in the root of our repo tsc p src tsconfig json noemit and see errors about semver being defined twice that s true once via typings semver d ts and once via package json
1
75,297
20,754,547,841
IssuesEvent
2022-03-15 10:54:41
LandSandBoat/server
https://api.github.com/repos/LandSandBoat/server
closed
[dbtool] Small deprecation warning on dbtool startup
build + ci
<!-- remove space and place 'x' mark between square [] brackets or click the checkbox after saving to affirm: --> - [x] I have paid attention to this example and will edit again if need be to not break the formatting, or I will be ignored - [x] I have searched existing [issues](https://github.com/LandSandBoat/server/issues) to see if the issue has already been opened - [x] I have checked the commit log to see if the issue has been resolved since my server was last updated - [x] I have read the [Contributing Guide](https://github.com/LandSandBoat/server/blob/base/CONTRIBUTING.md) - [x] I have specified what branch this happens on branch: EDITME **_Additional Information_** (Steps to reproduce/Expected behavior) **:** Everything still works, this just flashes for a second or two and then regular operation continues: ``` dbtool.py:8: DeprecationWarning: The distutils package is deprecated and slated for removal in Python 3.12. Use setuptools or check PEP 632 for potential alternatives import distutils.spawn ERROR: dbtool.py is designed to be run from the <root>/tools folder, not <root>. Please run from the tools folder. Press ENTER to continue... ```
1.0
[dbtool] Small deprecation warning on dbtool startup - <!-- remove space and place 'x' mark between square [] brackets or click the checkbox after saving to affirm: --> - [x] I have paid attention to this example and will edit again if need be to not break the formatting, or I will be ignored - [x] I have searched existing [issues](https://github.com/LandSandBoat/server/issues) to see if the issue has already been opened - [x] I have checked the commit log to see if the issue has been resolved since my server was last updated - [x] I have read the [Contributing Guide](https://github.com/LandSandBoat/server/blob/base/CONTRIBUTING.md) - [x] I have specified what branch this happens on branch: EDITME **_Additional Information_** (Steps to reproduce/Expected behavior) **:** Everything still works, this just flashes for a second or two and then regular operation continues: ``` dbtool.py:8: DeprecationWarning: The distutils package is deprecated and slated for removal in Python 3.12. Use setuptools or check PEP 632 for potential alternatives import distutils.spawn ERROR: dbtool.py is designed to be run from the <root>/tools folder, not <root>. Please run from the tools folder. Press ENTER to continue... ```
build
small deprecation warning on dbtool startup i have paid attention to this example and will edit again if need be to not break the formatting or i will be ignored i have searched existing to see if the issue has already been opened i have checked the commit log to see if the issue has been resolved since my server was last updated i have read the i have specified what branch this happens on branch editme additional information steps to reproduce expected behavior everything still works this just flashes for a second or two and then regular operation continues dbtool py deprecationwarning the distutils package is deprecated and slated for removal in python use setuptools or check pep for potential alternatives import distutils spawn error dbtool py is designed to be run from the tools folder not please run from the tools folder press enter to continue
1
82,424
15,646,559,083
IssuesEvent
2021-03-23 01:12:26
jgeraigery/linux
https://api.github.com/repos/jgeraigery/linux
opened
CVE-2020-12655 (Medium) detected in linuxv5.2
security vulnerability
## CVE-2020-12655 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux/fs/xfs/libxfs/xfs_alloc.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux/fs/xfs/libxfs/xfs_alloc.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in xfs_agf_verify in fs/xfs/libxfs/xfs_alloc.c in the Linux kernel through 5.6.10. Attackers may trigger a sync of excessive duration via an XFS v5 image with crafted metadata, aka CID-d0c7feaf8767. <p>Publish Date: 2020-05-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-12655>CVE-2020-12655</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-12655">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-12655</a></p> <p>Release Date: 2020-05-05</p> <p>Fix Resolution: v5.7-rc1</p> </p> </details> <p></p>
True
CVE-2020-12655 (Medium) detected in linuxv5.2 - ## CVE-2020-12655 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux/fs/xfs/libxfs/xfs_alloc.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux/fs/xfs/libxfs/xfs_alloc.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in xfs_agf_verify in fs/xfs/libxfs/xfs_alloc.c in the Linux kernel through 5.6.10. Attackers may trigger a sync of excessive duration via an XFS v5 image with crafted metadata, aka CID-d0c7feaf8767. <p>Publish Date: 2020-05-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-12655>CVE-2020-12655</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-12655">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-12655</a></p> <p>Release Date: 2020-05-05</p> <p>Fix Resolution: v5.7-rc1</p> </p> </details> <p></p>
non_build
cve medium detected in cve medium severity vulnerability vulnerable library linux kernel source tree library home page a href vulnerable source files linux fs xfs libxfs xfs alloc c linux fs xfs libxfs xfs alloc c vulnerability details an issue was discovered in xfs agf verify in fs xfs libxfs xfs alloc c in the linux kernel through attackers may trigger a sync of excessive duration via an xfs image with crafted metadata aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
19,793
14,565,437,677
IssuesEvent
2020-12-17 07:15:24
home-climate-control/esphome2influxdb
https://api.github.com/repos/home-climate-control/esphome2influxdb
closed
Provide autoconfiguration ability
usability
### Expected Behavior ESPHome devices appear automagically ### Actual Behavior Everything needs to be configured by hand ### Acceptance Criteria * ESPHome devices configured with [MQTT Client Component](https://esphome.io/components/mqtt.html) are recognized as such; * Configuration entry templates are emitted so that they can be incorporated into persistent configuration; * A flag must be provided to allow automatically discovered devices to start emitting InfluxDB feeds immediately, regardless of whether a persistent configuration entry exists.
True
Provide autoconfiguration ability - ### Expected Behavior ESPHome devices appear automagically ### Actual Behavior Everything needs to be configured by hand ### Acceptance Criteria * ESPHome devices configured with [MQTT Client Component](https://esphome.io/components/mqtt.html) are recognized as such; * Configuration entry templates are emitted so that they can be incorporated into persistent configuration; * A flag must be provided to allow automatically discovered devices to start emitting InfluxDB feeds immediately, regardless of whether a persistent configuration entry exists.
non_build
provide autoconfiguration ability expected behavior esphome devices appear automagically actual behavior everything needs to be configured by hand acceptance criteria esphome devices configured with are recognized as such configuration entry templates are emitted so that they can be incorporated into persistent configuration a flag must be provided to allow automatically discovered devices to start emitting influxdb feeds immediately regardless of whether a persistent configuration entry exists
0
84,272
24,265,956,909
IssuesEvent
2022-09-28 06:14:13
o3de/o3de
https://api.github.com/repos/o3de/o3de
opened
Building O3DE on VS2022 slower than VS2019 depending on configuration
kind/bug needs-triage sig/build sig/core
**Describe the bug** If a build that previously was running with code and artifacts from VC142 (Visual Studio 2019's C++ compiler) is switched to VC143 (Visual Studio 2022's compiler) on the same machine, there may be unexpectedly longer build times. This is inconsistent from build to build, however. **Assets required** O3DE code base from the Development branch **Steps to reproduce** Steps to reproduce the behavior: 1. Configure and build O3DE using VC142 on VS2019 2. Finish the build 3. Configure and build the same source using VC143 on VS2022, using default configurations 4. Finish the build 5. Note a slower overall build time **Expected behavior** The same or faster build time on VS2022 as VS2019 **Actual behavior** A slower overall build time (observed 15% slower in some cases) **Screenshots/Video** VS2019 Incremental Profile Build: https://jenkins.build.o3de.org/blue/organizations/jenkins/O3DE_periodic-incremental-daily/detail/development/113/pipeline/1054 (1h 16m 16s) VS2022 Incremental Profile Build: https://jenkins.build.o3de.org/blue/organizations/jenkins/O3DE/detail/development/3164/pipeline/378 (1h 40m 32s) **Found in Branch** Development **Desktop/Device (please complete the following information):** - Device: PC - OS: Windows/Windows Server - Version 10/2019 - CPU Intel Xeon Platinum 8000 and AMD Ryzen 5900x - GPU N/A - Memory 32 GB (Intel) and 64 GB (AMD) **Additional context** Core/Memory usage ratio from cl.exe/link.exe is also under investigation. Some reports on MS's side here: https://developercommunity.visualstudio.com/t/c-compile-in-vs-2022-is-much-slower-than-in-vs2019/1680228
1.0
Building O3DE on VS2022 slower than VS2019 depending on configuration - **Describe the bug** If a build that previously was running with code and artifacts from VC142 (Visual Studio 2019's C++ compiler) is switched to VC143 (Visual Studio 2022's compiler) on the same machine, there may be unexpectedly longer build times. This is inconsistent from build to build, however. **Assets required** O3DE code base from the Development branch **Steps to reproduce** Steps to reproduce the behavior: 1. Configure and build O3DE using VC142 on VS2019 2. Finish the build 3. Configure and build the same source using VC143 on VS2022, using default configurations 4. Finish the build 5. Note a slower overall build time **Expected behavior** The same or faster build time on VS2022 as VS2019 **Actual behavior** A slower overall build time (observed 15% slower in some cases) **Screenshots/Video** VS2019 Incremental Profile Build: https://jenkins.build.o3de.org/blue/organizations/jenkins/O3DE_periodic-incremental-daily/detail/development/113/pipeline/1054 (1h 16m 16s) VS2022 Incremental Profile Build: https://jenkins.build.o3de.org/blue/organizations/jenkins/O3DE/detail/development/3164/pipeline/378 (1h 40m 32s) **Found in Branch** Development **Desktop/Device (please complete the following information):** - Device: PC - OS: Windows/Windows Server - Version 10/2019 - CPU Intel Xeon Platinum 8000 and AMD Ryzen 5900x - GPU N/A - Memory 32 GB (Intel) and 64 GB (AMD) **Additional context** Core/Memory usage ratio from cl.exe/link.exe is also under investigation. Some reports on MS's side here: https://developercommunity.visualstudio.com/t/c-compile-in-vs-2022-is-much-slower-than-in-vs2019/1680228
build
building on slower than depending on configuration describe the bug if a build that previously was running with code and artifacts from visual studio s c compiler is switched to visual studio s compiler on the same machine there may be unexpectedly longer build times this is inconsistent from build to build however assets required code base from the development branch steps to reproduce steps to reproduce the behavior configure and build using on finish the build configure and build the same source using on using default configurations finish the build note a slower overall build time expected behavior the same or faster build time on as actual behavior a slower overall build time observed slower in some cases screenshots video incremental profile build incremental profile build found in branch development desktop device please complete the following information device pc os windows windows server version cpu intel xeon platinum and amd ryzen gpu n a memory gb intel and gb amd additional context core memory usage ratio from cl exe link exe is also under investigation some reports on ms s side here
1
428,509
12,412,692,053
IssuesEvent
2020-05-22 11:06:09
ahmedkaludi/accelerated-mobile-pages
https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages
closed
FlexMLS® IDX has a conflict with AMP
[Priority: HIGH] bug
Whenever the plugin(FlexMLS® IDX) is activated "link rel=canonical' is missing or incorrect." error pops up on the AMP validator test. The issue has been tested on Local Host as well. Reference: https://secure.helpscout.net/conversation/1171607964/131108?folderId=673452
1.0
FlexMLS® IDX has a conflict with AMP - Whenever the plugin(FlexMLS® IDX) is activated "link rel=canonical' is missing or incorrect." error pops up on the AMP validator test. The issue has been tested on Local Host as well. Reference: https://secure.helpscout.net/conversation/1171607964/131108?folderId=673452
non_build
flexmls® idx has a conflict with amp whenever the plugin flexmls® idx is activated link rel canonical is missing or incorrect error pops up on the amp validator test the issue has been tested on local host as well reference
0
190,662
22,149,147,841
IssuesEvent
2022-06-03 14:59:56
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Security Solution] UI is distorted in the 'Alerts' tab on expanding the toggle on the 'Detection and Response' Dashboard in the mobile screen mode
bug triage_needed impact:medium Team: SecuritySolution v8.3.0 v8.4.0
**Describe the bug:** - UI is distorted in the 'Alerts' tab on expanding the toggle on the 'Detection and Response' Dashboard in the mobile screen mode **Kibana/Elasticsearch Stack version:** ``` Branch: Main branch Commit : c1081c677d475670ff55926fb8f355deafa5c038 ``` **Pre-requisite:** - Elasticsearch should be up and running - Kibana should be up and running - Use command (Fn+F12) and click on the mobile resolution (538 X 675) **Steps to reproduce:** 1. Navigate to Kibana→ Advanced settings. 2. Search for Enable Grouped navigation toggle. 3. Enable Grouped navigation toggle. 4. Navigate to Security > Overview. 5. All the navigation under security is grouped. 6. Navigate to Dashboards > Detection and Response 7. Expand the toggle of 'Alerts' tab **Current behavior:** - UI is distorted in the 'Alerts' tab on expanding the toggle on the 'Detection and Response' Dashboard in the mobile screen mode **Expected behavior:** - UI should not be distorted in the 'Alerts' tab on expanding the toggle on the 'Detection and Response' Dashboard in the mobile screen mode **Screenshots:** https://user-images.githubusercontent.com/97870262/171143824-bb288b0f-5d2f-4c0e-ab3b-e838ddf0bace.mp4
True
[Security Solution] UI is distorted in the 'Alerts' tab on expanding the toggle on the 'Detection and Response' Dashboard in the mobile screen mode - **Describe the bug:** - UI is distorted in the 'Alerts' tab on expanding the toggle on the 'Detection and Response' Dashboard in the mobile screen mode **Kibana/Elasticsearch Stack version:** ``` Branch: Main branch Commit : c1081c677d475670ff55926fb8f355deafa5c038 ``` **Pre-requisite:** - Elasticsearch should be up and running - Kibana should be up and running - Use command (Fn+F12) and click on the mobile resolution (538 X 675) **Steps to reproduce:** 1. Navigate to Kibana→ Advanced settings. 2. Search for Enable Grouped navigation toggle. 3. Enable Grouped navigation toggle. 4. Navigate to Security > Overview. 5. All the navigation under security is grouped. 6. Navigate to Dashboards > Detection and Response 7. Expand the toggle of 'Alerts' tab **Current behavior:** - UI is distorted in the 'Alerts' tab on expanding the toggle on the 'Detection and Response' Dashboard in the mobile screen mode **Expected behavior:** - UI should not be distorted in the 'Alerts' tab on expanding the toggle on the 'Detection and Response' Dashboard in the mobile screen mode **Screenshots:** https://user-images.githubusercontent.com/97870262/171143824-bb288b0f-5d2f-4c0e-ab3b-e838ddf0bace.mp4
non_build
ui is distorted in the alerts tab on expanding the toggle on the detection and response dashboard in the mobile screen mode describe the bug ui is distorted in the alerts tab on expanding the toggle on the detection and response dashboard in the mobile screen mode kibana elasticsearch stack version branch main branch commit pre requisite elasticsearch should be up and running kibana should be up and running use command fn and click on the mobile resolution x steps to reproduce navigate to kibana→ advanced settings search for enable grouped navigation toggle enable grouped navigation toggle navigate to security overview all the navigation under security is grouped navigate to dashboards detection and response expand the toggle of alerts tab current behavior ui is distorted in the alerts tab on expanding the toggle on the detection and response dashboard in the mobile screen mode expected behavior ui should not be distorted in the alerts tab on expanding the toggle on the detection and response dashboard in the mobile screen mode screenshots
0
335,548
10,155,147,682
IssuesEvent
2019-08-06 09:38:46
our-city-app/mobicage-ios-client
https://api.github.com/repos/our-city-app/mobicage-ios-client
opened
Show items on a map
priority_major type_feature
- [ ] Define the HomeScreenItem to open a map - [ ] Load a map and show the items
1.0
Show items on a map - - [ ] Define the HomeScreenItem to open a map - [ ] Load a map and show the items
non_build
show items on a map define the homescreenitem to open a map load a map and show the items
0
76,803
21,572,516,294
IssuesEvent
2022-05-02 09:58:13
Crocoblock/suggestions
https://api.github.com/repos/Crocoblock/suggestions
closed
Load more feature isn't loading the current query
JetWooBuilder
Hello! I use the load more infinite scroll on my product category page and it was working fine until I came across this bug today. When I scroll the page to load more products, the products grid widget loads products from all the categories, not respecting the current query. 1. For example, this page is for the "dolls" category: https://lojahappytoys.com.br/categoria-produto/bonecas/ 2. Scroll to the end of the page, and the widget will start loading more products. 3. You will see that the whole products grid changes, suddenly its not a dolls category page anymore, because even the products above the load changes. This is a major issue, that's affecting the navigation of my store. I am using JetEngine version 2.11.7, JetSmartFilters version 2.3.11, JetWooBuilder version 1.12.3 and the Load More plugin version 1.0.0. Thank you in advance.
1.0
Load more feature isn't loading the current query - Hello! I use the load more infinite scroll on my product category page and it was working fine until I came across this bug today. When I scroll the page to load more products, the products grid widget loads products from all the categories, not respecting the current query. 1. For example, this page is for the "dolls" category: https://lojahappytoys.com.br/categoria-produto/bonecas/ 2. Scroll to the end of the page, and the widget will start loading more products. 3. You will see that the whole products grid changes, suddenly its not a dolls category page anymore, because even the products above the load changes. This is a major issue, that's affecting the navigation of my store. I am using JetEngine version 2.11.7, JetSmartFilters version 2.3.11, JetWooBuilder version 1.12.3 and the Load More plugin version 1.0.0. Thank you in advance.
build
load more feature isn t loading the current query hello i use the load more infinite scroll on my product category page and it was working fine until i came across this bug today when i scroll the page to load more products the products grid widget loads products from all the categories not respecting the current query for example this page is for the dolls category scroll to the end of the page and the widget will start loading more products you will see that the whole products grid changes suddenly its not a dolls category page anymore because even the products above the load changes this is a major issue that s affecting the navigation of my store i am using jetengine version jetsmartfilters version jetwoobuilder version and the load more plugin version thank you in advance
1
3,154
3,081,741,519
IssuesEvent
2015-08-23 01:46:51
JuliaLang/julia
https://api.github.com/repos/JuliaLang/julia
closed
Unable to build with LLVM release candidate (LLVM_VER=3.7.0rc2)
build
If you set `LLVM_VER=svn`, as in the instructions to be able to use https://github.com/Keno/Cxx.jl, LLVM doesn't compile, due to some compiler error (see https://groups.google.com/forum/#!topic/julia-dev/51zRgNO3fNI). I then tried to use the latest 3.7 release candidate, which is at: `llvm.org/pre-releases/3.7.0/rc2/llvm-3.7.0rc2.src.tar.xz`, however it seems like the Julia make files can only handle releases or svn, but not release candidates (I'd submit a PR, but I'm not good at make file magic!)
1.0
Unable to build with LLVM release candidate (LLVM_VER=3.7.0rc2) - If you set `LLVM_VER=svn`, as in the instructions to be able to use https://github.com/Keno/Cxx.jl, LLVM doesn't compile, due to some compiler error (see https://groups.google.com/forum/#!topic/julia-dev/51zRgNO3fNI). I then tried to use the latest 3.7 release candidate, which is at: `llvm.org/pre-releases/3.7.0/rc2/llvm-3.7.0rc2.src.tar.xz`, however it seems like the Julia make files can only handle releases or svn, but not release candidates (I'd submit a PR, but I'm not good at make file magic!)
build
unable to build with llvm release candidate llvm ver if you set llvm ver svn as in the instructions to be able to use llvm doesn t compile due to some compiler error see i then tried to use the latest release candidate which is at llvm org pre releases llvm src tar xz however it seems like the julia make files can only handle releases or svn but not release candidates i d submit a pr but i m not good at make file magic
1
292,582
8,959,708,288
IssuesEvent
2019-01-28 01:03:13
razzius/hms-weave
https://api.github.com/repos/razzius/hms-weave
closed
Student log in directs to faculty log in interface
bug high priority
When I logged in from the student view, it directed me to the faculty log in interface, with the tutorial video present. The student log in view ideally should only say: Confirm email verification Successfully logged in as XXX@XXX.edu. In fact, clicking on "Create profile" links to the faculty profile creation view. Which is not what we want students to be doing. Clicking on "HMS Weave" at the top then correctly links to the student view. ![image](https://user-images.githubusercontent.com/43801169/51791398-39b92700-2170-11e9-8b11-80f9a62c04fc.png)
1.0
Student log in directs to faculty log in interface - When I logged in from the student view, it directed me to the faculty log in interface, with the tutorial video present. The student log in view ideally should only say: Confirm email verification Successfully logged in as XXX@XXX.edu. In fact, clicking on "Create profile" links to the faculty profile creation view. Which is not what we want students to be doing. Clicking on "HMS Weave" at the top then correctly links to the student view. ![image](https://user-images.githubusercontent.com/43801169/51791398-39b92700-2170-11e9-8b11-80f9a62c04fc.png)
non_build
student log in directs to faculty log in interface when i logged in from the student view it directed me to the faculty log in interface with the tutorial video present the student log in view ideally should only say confirm email verification successfully logged in as xxx xxx edu in fact clicking on create profile links to the faculty profile creation view which is not what we want students to be doing clicking on hms weave at the top then correctly links to the student view
0
65,702
8,831,822,159
IssuesEvent
2019-01-04 01:00:03
OfficeDev/office-ui-fabric-react
https://api.github.com/repos/OfficeDev/office-ui-fabric-react
closed
Styling utilities are, contrary to the docs, not utilizing memoization, leading to unnecessary style recomputations
Needs: Author Feedback Type: Documentation Type: Question ❔ no-recent-activity
<!-- Use this template for bug/error reporting only --> ### Bug Report - __Package version(s)__: 6.22.0 #### Priorities and help requested: Are you willing to submit a PR to fix? No Requested priority: Normal #### Describe the issue: A lot of the styling utilities of Fabric are designed to minimize (or avoid) unnecessary recomputations. For example, `getClassNames` is ([according to the docs](https://github.com/OfficeDev/office-ui-fabric-react/wiki/Component-Styling#styled-higher-order-component)) memoized. However, I just implemented a really simple custom component (see codepen link below) and realized that styles are always recomputed, even though the props don't change. #### Actual behavior: Passing the same styleProps n times causes my `getStyles` function to get called n times. #### Expected behavior: Passing the same styleProps causes my `getStyles` function to get called once ### If applicable, please provide a codepen repro: https://codepen.io/nimelrian/pen/LBVGKV
1.0
Styling utilities are, contrary to the docs, not utilizing memoization, leading to unnecessary style recomputations - <!-- Use this template for bug/error reporting only --> ### Bug Report - __Package version(s)__: 6.22.0 #### Priorities and help requested: Are you willing to submit a PR to fix? No Requested priority: Normal #### Describe the issue: A lot of the styling utilities of Fabric are designed to minimize (or avoid) unnecessary recomputations. For example, `getClassNames` is ([according to the docs](https://github.com/OfficeDev/office-ui-fabric-react/wiki/Component-Styling#styled-higher-order-component)) memoized. However, I just implemented a really simple custom component (see codepen link below) and realized that styles are always recomputed, even though the props don't change. #### Actual behavior: Passing the same styleProps n times causes my `getStyles` function to get called n times. #### Expected behavior: Passing the same styleProps causes my `getStyles` function to get called once ### If applicable, please provide a codepen repro: https://codepen.io/nimelrian/pen/LBVGKV
non_build
styling utilities are contrary to the docs not utilizing memoization leading to unnecessary style recomputations bug report package version s priorities and help requested are you willing to submit a pr to fix no requested priority normal describe the issue a lot of the styling utilities of fabric are designed to minimize or avoid unnecessary recomputations for example getclassnames is memoized however i just implemented a really simple custom component see codepen link below and realized that styles are always recomputed even though the props don t change actual behavior passing the same styleprops n times causes my getstyles function to get called n times expected behavior passing the same styleprops causes my getstyles function to get called once if applicable please provide a codepen repro
0
198,503
15,711,155,197
IssuesEvent
2021-03-27 05:41:26
wadep23/robot-gladiators
https://api.github.com/repos/wadep23/robot-gladiators
closed
Initial Game Functionality
documentation
**Description** _Must Have_ - Build a game where a player's robot can fight another robot until one of them loses. - If the enemy-robot loses first, the player's robot will move on to fight another enemy robot. _Features_ - The players robot's name can be dynamically created by the player through the browser. - The player is given the option to skip the fight by paying a penalty fee, or continue with the fight.
1.0
Initial Game Functionality - **Description** _Must Have_ - Build a game where a player's robot can fight another robot until one of them loses. - If the enemy-robot loses first, the player's robot will move on to fight another enemy robot. _Features_ - The players robot's name can be dynamically created by the player through the browser. - The player is given the option to skip the fight by paying a penalty fee, or continue with the fight.
non_build
initial game functionality description must have build a game where a player s robot can fight another robot until one of them loses if the enemy robot loses first the player s robot will move on to fight another enemy robot features the players robot s name can be dynamically created by the player through the browser the player is given the option to skip the fight by paying a penalty fee or continue with the fight
0
380,098
11,253,843,963
IssuesEvent
2020-01-11 19:05:50
GeyserMC/Geyser
https://api.github.com/repos/GeyserMC/Geyser
closed
Baby Zombies show as adults in bedrock, entities don't show items in hand, armor, etc
Confirmed Bug Priority: Low Work in Progress
**Describe the bug** Baby mobs only show as adults. Items do not show in main hand, off hand, armour, etc **To Reproduce** Steps to reproduce the behavior: 1. Spawn an baby mob **Expected behavior** I was expecting the mob to be an baby instead of an adult. I was also expecting mobs to have items in their main hand. **Screenshots** <img src="https://steamuserimages-a.akamaihd.net/ugc/793127482998002373/813EC9DE7F692E37333E42201458B7652F95EFE1/"> <img src="https://steamuserimages-a.akamaihd.net/ugc/793127482998005615/F2B0D03F054C63ABB5A099D376E419C00D63B12C/"> **Geyser version** Jenkins **Bedrock version** 1.13 **Additional Info** Tamed entities do not show their tamed textures and data in bedrock.
1.0
Baby Zombies show as adults in bedrock, entities don't show items in hand, armor, etc - **Describe the bug** Baby mobs only show as adults. Items do not show in main hand, off hand, armour, etc **To Reproduce** Steps to reproduce the behavior: 1. Spawn an baby mob **Expected behavior** I was expecting the mob to be an baby instead of an adult. I was also expecting mobs to have items in their main hand. **Screenshots** <img src="https://steamuserimages-a.akamaihd.net/ugc/793127482998002373/813EC9DE7F692E37333E42201458B7652F95EFE1/"> <img src="https://steamuserimages-a.akamaihd.net/ugc/793127482998005615/F2B0D03F054C63ABB5A099D376E419C00D63B12C/"> **Geyser version** Jenkins **Bedrock version** 1.13 **Additional Info** Tamed entities do not show their tamed textures and data in bedrock.
non_build
baby zombies show as adults in bedrock entities don t show items in hand armor etc describe the bug baby mobs only show as adults items do not show in main hand off hand armour etc to reproduce steps to reproduce the behavior spawn an baby mob expected behavior i was expecting the mob to be an baby instead of an adult i was also expecting mobs to have items in their main hand screenshots img src img src geyser version jenkins bedrock version additional info tamed entities do not show their tamed textures and data in bedrock
0
53,361
13,158,466,954
IssuesEvent
2020-08-10 14:22:33
aristanetworks/ansible-avd
https://api.github.com/repos/aristanetworks/ansible-avd
closed
Implement variable to allow user to define root directory for documentation and intended
avd tools role: build_output_folders role: eos_cli_config_gen role: eos_l3ls_evpn state: in-progress state: testing type: code quality type: enhancement
**Issue Type** - Role enhancement **Is your feature request related to a problem? Please describe.** When implementing molecule unit-test, script execution generates a large number of warnings related to all generated files part of inventory folder ```shell [WARNING]: Skipping unexpected key (Ethernet8) in group (ethernet_interfaces), only "vars", "children" and "hosts" are valid [WARNING]: Unable to parse /Users/tgrimonet/.cache/molecule/avd/avd- build/inventory/intended/structured_configs/cvp as an inventory source ``` It does not break CI process but make it hard to read. Here is an example: [CI runs in fork](https://github.com/titom73/ansible-avd/runs/866782241?check_suite_focus=true#step:4:232) **Describe the solution you'd like** Roles might use a custom key to allow specific location for generated files and set to `{{inventory_dir}}` by default. An example is already present in [__arista.avd.build_output_folders__](https://github.com/aristanetworks/ansible-avd/blob/devel/ansible_collections/arista/avd/roles/build_output_folders/defaults/main.yml#L5) **Describe alternatives you've considered** Tried to avoid warning from molecule with no success **Additional context** Only related to CI implementation.
1.0
Implement variable to allow user to define root directory for documentation and intended - **Issue Type** - Role enhancement **Is your feature request related to a problem? Please describe.** When implementing molecule unit-test, script execution generates a large number of warnings related to all generated files part of inventory folder ```shell [WARNING]: Skipping unexpected key (Ethernet8) in group (ethernet_interfaces), only "vars", "children" and "hosts" are valid [WARNING]: Unable to parse /Users/tgrimonet/.cache/molecule/avd/avd- build/inventory/intended/structured_configs/cvp as an inventory source ``` It does not break CI process but make it hard to read. Here is an example: [CI runs in fork](https://github.com/titom73/ansible-avd/runs/866782241?check_suite_focus=true#step:4:232) **Describe the solution you'd like** Roles might use a custom key to allow specific location for generated files and set to `{{inventory_dir}}` by default. An example is already present in [__arista.avd.build_output_folders__](https://github.com/aristanetworks/ansible-avd/blob/devel/ansible_collections/arista/avd/roles/build_output_folders/defaults/main.yml#L5) **Describe alternatives you've considered** Tried to avoid warning from molecule with no success **Additional context** Only related to CI implementation.
build
implement variable to allow user to define root directory for documentation and intended issue type role enhancement is your feature request related to a problem please describe when implementing molecule unit test script execution generates a large number of warnings related to all generated files part of inventory folder shell skipping unexpected key in group ethernet interfaces only vars children and hosts are valid unable to parse users tgrimonet cache molecule avd avd build inventory intended structured configs cvp as an inventory source it does not break ci process but make it hard to read here is an example describe the solution you d like roles might use a custom key to allow specific location for generated files and set to inventory dir by default an example is already present in describe alternatives you ve considered tried to avoid warning from molecule with no success additional context only related to ci implementation
1
61,912
15,105,758,923
IssuesEvent
2021-02-08 13:28:43
thoth-station/python
https://api.github.com/repos/thoth-station/python
closed
Failed to update dependencies to their latest version
bot human_intervention_required kind/bug sig/build
Automatic dependency update failed for the current master with SHA 396dacc4beddcfc9828ad1f0110e169d62b02443. The automatic dependency management cannot continue. Please fix errors reported bellow. ##### Command ``` $ pipenv lock ``` <details> <summary>Standard output</summary> ``` ``` </details> <details> <summary>Standard error</summary> ``` Locking [dev-packages] dependencies... Building requirements...  Resolving dependencies...  FAIL Traceback (most recent call last): File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connection.py", line 169, in _new_conn conn = connection.create_connection( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/util/connection.py", line 73, in create_connection for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): File "/usr/lib64/python3.8/socket.py", line 918, in getaddrinfo for res in _socket.getaddrinfo(host, port, family, type, proto, flags): socket.gaierror: [Errno -2] Name or service not known During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connectionpool.py", line 699, in urlopen httplib_response = self._make_request( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connectionpool.py", line 382, in _make_request self._validate_conn(conn) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connectionpool.py", line 1010, in _validate_conn conn.connect() File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connection.py", line 353, in connect conn = self._new_conn() File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connection.py", line 181, in _new_conn raise NewConnectionError( urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7ff4b19a2b50>: Failed to establish a new connection: [Errno -2] Name or service not known During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/requests/adapters.py", line 439, in send resp = conn.urlopen( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connectionpool.py", line 755, in urlopen retries = retries.increment( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/util/retry.py", line 573, in increment raise MaxRetryError(_pool, url, error or ResponseError(cause)) urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='pypi.org', port=443): Max retries exceeded with url: /pypi/py/json (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7ff4b19a2b50>: Failed to establish a new connection: [Errno -2] Name or service not known')) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/resolver.py", line 764, in <module> main() File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/resolver.py", line 758, in main _main(parsed.pre, parsed.clear, parsed.verbose, parsed.system, parsed.write, File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/resolver.py", line 741, in _main resolve_packages(pre, clear, verbose, system, write, requirements_dir, packages, dev) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/resolver.py", line 702, in resolve_packages results, resolver = resolve( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/resolver.py", line 684, in resolve return resolve_deps( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1395, in resolve_deps results, hashes, markers_lookup, resolver, skipped = actually_resolve_deps( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1111, in actually_resolve_deps results = resolver.clean_results() File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1017, in clean_results collected_hashes = self.collect_hashes(ireq) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 900, in collect_hashes r = session.get(pkg_url, timeout=10) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/requests/sessions.py", line 555, in get return self.request('GET', url, **kwargs) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/requests/sessions.py", line 542, in request resp = self.send(prep, **send_kwargs) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/requests/sessions.py", line 655, in send r = adapter.send(request, **kwargs) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/requests/adapters.py", line 516, in send raise ConnectionError(e, request=request) requests.exceptions.ConnectionError: HTTPSConnectionPool(host='pypi.org', port=443): Max retries exceeded with url: /pypi/py/json (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7ff4b19a2b50>: Failed to establish a new connection: [Errno -2] Name or service not known')) Traceback (most recent call last): File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1131, in create_spinner yield sp File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1338, in venv_resolve_deps c = resolve(cmd, sp) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1176, in resolve sys.exit(c.return_code) SystemExit: 1  ``` </details> <details> <summary>Environment details</summary> ``` Kebechet version: 1.2.3 Python version: 3.8.3 Platform: Linux-4.18.0-193.29.1.el8_2.x86_64-x86_64-with-glibc2.2.5 pipenv version: pipenv, version 2020.11.15 ``` </details> <details> <summary>Dependency graph</summary> ``` aiohttp==3.7.3 - async-timeout [required: >=3.0,<4.0, installed: 3.0.1] - attrs [required: >=17.3.0, installed: 20.3.0] - chardet [required: >=2.0,<4.0, installed: 3.0.4] - multidict [required: >=4.5,<7.0, installed: 5.1.0] - typing-extensions [required: >=3.6.5, installed: 3.7.4.3] - yarl [required: >=1.0,<2.0, installed: 1.6.3] - idna [required: >=2.0, installed: 2.10] - multidict [required: >=4.0, installed: 5.1.0] beautifulsoup4==4.6.3 flexmock==0.10.4 lxml==4.6.2 pyelftools==0.27 pytest-asyncio==0.14.0 - pytest [required: >=5.4.0, installed: 6.2.1] - attrs [required: >=19.2.0, installed: 20.3.0] - iniconfig [required: Any, installed: 1.1.1] - packaging [required: Any, installed: 20.8] - pyparsing [required: >=2.0.2, installed: 2.4.7] - pluggy [required: >=0.12,<1.0.0a1, installed: 0.13.1] - py [required: >=1.8.2, installed: 1.10.0] - toml [required: Any, installed: 0.10.2] pytest-cov==2.10.1 - coverage [required: >=4.4, installed: 5.3.1] - pytest [required: >=4.6, installed: 6.2.1] - attrs [required: >=19.2.0, installed: 20.3.0] - iniconfig [required: Any, installed: 1.1.1] - packaging [required: Any, installed: 20.8] - pyparsing [required: >=2.0.2, installed: 2.4.7] - pluggy [required: >=0.12,<1.0.0a1, installed: 0.13.1] - py [required: >=1.8.2, installed: 1.10.0] - toml [required: Any, installed: 0.10.2] pytest-mypy==0.8.0 - attrs [required: >=19.0, installed: 20.3.0] - filelock [required: >=3.0, installed: 3.0.12] - mypy [required: >=0.700, installed: 0.790] - mypy-extensions [required: >=0.4.3,<0.5.0, installed: 0.4.3] - typed-ast [required: >=1.4.0,<1.5.0, installed: 1.4.2] - typing-extensions [required: >=3.7.4, installed: 3.7.4.3] - pytest [required: >=3.5, installed: 6.2.1] - attrs [required: >=19.2.0, installed: 20.3.0] - iniconfig [required: Any, installed: 1.1.1] - packaging [required: Any, installed: 20.8] - pyparsing [required: >=2.0.2, installed: 2.4.7] - pluggy [required: >=0.12,<1.0.0a1, installed: 0.13.1] - py [required: >=1.8.2, installed: 1.10.0] - toml [required: Any, installed: 0.10.2] pytest-timeout==1.4.2 - pytest [required: >=3.6.0, installed: 6.2.1] - attrs [required: >=19.2.0, installed: 20.3.0] - iniconfig [required: Any, installed: 1.1.1] - packaging [required: Any, installed: 20.8] - pyparsing [required: >=2.0.2, installed: 2.4.7] - pluggy [required: >=0.12,<1.0.0a1, installed: 0.13.1] - py [required: >=1.8.2, installed: 1.10.0] - toml [required: Any, installed: 0.10.2] thoth-analyzer==0.1.8 - click [required: Any, installed: 7.1.2] - delegator.py [required: >=0.1.0, installed: 0.1.1] - pexpect [required: >=4.1.0, installed: 4.8.0] - ptyprocess [required: >=0.5, installed: 0.7.0] - distro [required: Any, installed: 1.5.0] - requests [required: Any, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - thoth-common [required: Any, installed: 0.21.3] - argo-workflows [required: >=3.5,<4, installed: 3.6.1] - kubernetes [required: >=11.0.0, installed: 11.0.0] - certifi [required: >=14.05.14, installed: 2020.12.5] - google-auth [required: >=1.0.1, installed: 1.24.0] - cachetools [required: >=2.0.0,<5.0, installed: 4.2.0] - pyasn1-modules [required: >=0.2.1, installed: 0.2.8] - pyasn1 [required: >=0.4.6,<0.5.0, installed: 0.4.8] - rsa [required: >=3.1.4,<5, installed: 4.7] - pyasn1 [required: >=0.1.3, installed: 0.4.8] - setuptools [required: >=40.3.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - python-dateutil [required: >=2.5.3, installed: 2.8.1] - six [required: >=1.5, installed: 1.15.0] - pyyaml [required: >=3.12, installed: 5.3.1] - requests [required: Any, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - requests-oauthlib [required: Any, installed: 1.3.0] - oauthlib [required: >=3.0.0, installed: 3.1.0] - requests [required: >=2.0.0, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - setuptools [required: >=21.0.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - urllib3 [required: >=1.24.2, installed: 1.26.2] - websocket-client [required: >=0.32.0,!=0.42.*,!=0.41.*,!=0.40.0, installed: 0.57.0] - six [required: Any, installed: 1.15.0] - mock [required: Any, installed: 4.0.3] - attrdict [required: Any, installed: 2.0.1] - six [required: Any, installed: 1.15.0] - attrs [required: Any, installed: 20.3.0] - daiquiri [required: Any, installed: 3.0.0] - python-json-logger [required: Any, installed: 2.0.1] - jsonformatter [required: Any, installed: 0.3.1] - kubernetes [required: ~=11.0.0, installed: 11.0.0] - certifi [required: >=14.05.14, installed: 2020.12.5] - google-auth [required: >=1.0.1, installed: 1.24.0] - cachetools [required: >=2.0.0,<5.0, installed: 4.2.0] - pyasn1-modules [required: >=0.2.1, installed: 0.2.8] - pyasn1 [required: >=0.4.6,<0.5.0, installed: 0.4.8] - rsa [required: >=3.1.4,<5, installed: 4.7] - pyasn1 [required: >=0.1.3, installed: 0.4.8] - setuptools [required: >=40.3.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - python-dateutil [required: >=2.5.3, installed: 2.8.1] - six [required: >=1.5, installed: 1.15.0] - pyyaml [required: >=3.12, installed: 5.3.1] - requests [required: Any, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - requests-oauthlib [required: Any, installed: 1.3.0] - oauthlib [required: >=3.0.0, installed: 3.1.0] - requests [required: >=2.0.0, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - setuptools [required: >=21.0.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - urllib3 [required: >=1.24.2, installed: 1.26.2] - websocket-client [required: >=0.32.0,!=0.42.*,!=0.41.*,!=0.40.0, installed: 0.57.0] - six [required: Any, installed: 1.15.0] - openshift [required: Any, installed: 0.11.2] - jinja2 [required: Any, installed: 2.11.2] - MarkupSafe [required: >=0.23, installed: 1.1.1] - kubernetes [required: ~=11.0.0, installed: 11.0.0] - certifi [required: >=14.05.14, installed: 2020.12.5] - google-auth [required: >=1.0.1, installed: 1.24.0] - cachetools [required: >=2.0.0,<5.0, installed: 4.2.0] - pyasn1-modules [required: >=0.2.1, installed: 0.2.8] - pyasn1 [required: >=0.4.6,<0.5.0, installed: 0.4.8] - rsa [required: >=3.1.4,<5, installed: 4.7] - pyasn1 [required: >=0.1.3, installed: 0.4.8] - setuptools [required: >=40.3.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - python-dateutil [required: >=2.5.3, installed: 2.8.1] - six [required: >=1.5, installed: 1.15.0] - pyyaml [required: >=3.12, installed: 5.3.1] - requests [required: Any, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - requests-oauthlib [required: Any, installed: 1.3.0] - oauthlib [required: >=3.0.0, installed: 3.1.0] - requests [required: >=2.0.0, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - setuptools [required: >=21.0.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - urllib3 [required: >=1.24.2, installed: 1.26.2] - websocket-client [required: >=0.32.0,!=0.42.*,!=0.41.*,!=0.40.0, installed: 0.57.0] - six [required: Any, installed: 1.15.0] - python-string-utils [required: Any, installed: 1.0.0] - ruamel.yaml [required: >=0.15, installed: 0.16.12] - ruamel.yaml.clib [required: >=0.1.2, installed: 0.2.2] - six [required: Any, installed: 1.15.0] - pyyaml [required: Any, installed: 5.3.1] - requests [required: Any, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - rfc5424-logging-handler [required: Any, installed: 1.4.3] - pytz [required: Any, installed: 2020.5] - tzlocal [required: Any, installed: 2.1] - pytz [required: Any, installed: 2020.5] - sentry-sdk [required: Any, installed: 0.19.5] - certifi [required: Any, installed: 2020.12.5] - urllib3 [required: >=1.10.0, installed: 1.26.2] ``` </details> ##### Notes For more information, see [Pipfile](https://raw.githubusercontent.com/thoth-station/python/master/Pipfile) and [Pipfile.lock](https://raw.githubusercontent.com/thoth-station/python/master/Pipfile.lock). Once this issue is resolved, the issue will be automatically closed by bot.
1.0
Failed to update dependencies to their latest version - Automatic dependency update failed for the current master with SHA 396dacc4beddcfc9828ad1f0110e169d62b02443. The automatic dependency management cannot continue. Please fix errors reported bellow. ##### Command ``` $ pipenv lock ``` <details> <summary>Standard output</summary> ``` ``` </details> <details> <summary>Standard error</summary> ``` Locking [dev-packages] dependencies... Building requirements...  Resolving dependencies...  FAIL Traceback (most recent call last): File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connection.py", line 169, in _new_conn conn = connection.create_connection( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/util/connection.py", line 73, in create_connection for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): File "/usr/lib64/python3.8/socket.py", line 918, in getaddrinfo for res in _socket.getaddrinfo(host, port, family, type, proto, flags): socket.gaierror: [Errno -2] Name or service not known During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connectionpool.py", line 699, in urlopen httplib_response = self._make_request( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connectionpool.py", line 382, in _make_request self._validate_conn(conn) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connectionpool.py", line 1010, in _validate_conn conn.connect() File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connection.py", line 353, in connect conn = self._new_conn() File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connection.py", line 181, in _new_conn raise NewConnectionError( urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x7ff4b19a2b50>: Failed to establish a new connection: [Errno -2] Name or service not known During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/requests/adapters.py", line 439, in send resp = conn.urlopen( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/connectionpool.py", line 755, in urlopen retries = retries.increment( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/urllib3/util/retry.py", line 573, in increment raise MaxRetryError(_pool, url, error or ResponseError(cause)) urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='pypi.org', port=443): Max retries exceeded with url: /pypi/py/json (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7ff4b19a2b50>: Failed to establish a new connection: [Errno -2] Name or service not known')) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/resolver.py", line 764, in <module> main() File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/resolver.py", line 758, in main _main(parsed.pre, parsed.clear, parsed.verbose, parsed.system, parsed.write, File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/resolver.py", line 741, in _main resolve_packages(pre, clear, verbose, system, write, requirements_dir, packages, dev) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/resolver.py", line 702, in resolve_packages results, resolver = resolve( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/resolver.py", line 684, in resolve return resolve_deps( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1395, in resolve_deps results, hashes, markers_lookup, resolver, skipped = actually_resolve_deps( File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1111, in actually_resolve_deps results = resolver.clean_results() File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1017, in clean_results collected_hashes = self.collect_hashes(ireq) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 900, in collect_hashes r = session.get(pkg_url, timeout=10) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/requests/sessions.py", line 555, in get return self.request('GET', url, **kwargs) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/requests/sessions.py", line 542, in request resp = self.send(prep, **send_kwargs) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/requests/sessions.py", line 655, in send r = adapter.send(request, **kwargs) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/vendor/requests/adapters.py", line 516, in send raise ConnectionError(e, request=request) requests.exceptions.ConnectionError: HTTPSConnectionPool(host='pypi.org', port=443): Max retries exceeded with url: /pypi/py/json (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7ff4b19a2b50>: Failed to establish a new connection: [Errno -2] Name or service not known')) Traceback (most recent call last): File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1131, in create_spinner yield sp File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1338, in venv_resolve_deps c = resolve(cmd, sp) File "/home/user/.local/share/virtualenvs/user-5PivF_Od/lib/python3.8/site-packages/pipenv/utils.py", line 1176, in resolve sys.exit(c.return_code) SystemExit: 1  ``` </details> <details> <summary>Environment details</summary> ``` Kebechet version: 1.2.3 Python version: 3.8.3 Platform: Linux-4.18.0-193.29.1.el8_2.x86_64-x86_64-with-glibc2.2.5 pipenv version: pipenv, version 2020.11.15 ``` </details> <details> <summary>Dependency graph</summary> ``` aiohttp==3.7.3 - async-timeout [required: >=3.0,<4.0, installed: 3.0.1] - attrs [required: >=17.3.0, installed: 20.3.0] - chardet [required: >=2.0,<4.0, installed: 3.0.4] - multidict [required: >=4.5,<7.0, installed: 5.1.0] - typing-extensions [required: >=3.6.5, installed: 3.7.4.3] - yarl [required: >=1.0,<2.0, installed: 1.6.3] - idna [required: >=2.0, installed: 2.10] - multidict [required: >=4.0, installed: 5.1.0] beautifulsoup4==4.6.3 flexmock==0.10.4 lxml==4.6.2 pyelftools==0.27 pytest-asyncio==0.14.0 - pytest [required: >=5.4.0, installed: 6.2.1] - attrs [required: >=19.2.0, installed: 20.3.0] - iniconfig [required: Any, installed: 1.1.1] - packaging [required: Any, installed: 20.8] - pyparsing [required: >=2.0.2, installed: 2.4.7] - pluggy [required: >=0.12,<1.0.0a1, installed: 0.13.1] - py [required: >=1.8.2, installed: 1.10.0] - toml [required: Any, installed: 0.10.2] pytest-cov==2.10.1 - coverage [required: >=4.4, installed: 5.3.1] - pytest [required: >=4.6, installed: 6.2.1] - attrs [required: >=19.2.0, installed: 20.3.0] - iniconfig [required: Any, installed: 1.1.1] - packaging [required: Any, installed: 20.8] - pyparsing [required: >=2.0.2, installed: 2.4.7] - pluggy [required: >=0.12,<1.0.0a1, installed: 0.13.1] - py [required: >=1.8.2, installed: 1.10.0] - toml [required: Any, installed: 0.10.2] pytest-mypy==0.8.0 - attrs [required: >=19.0, installed: 20.3.0] - filelock [required: >=3.0, installed: 3.0.12] - mypy [required: >=0.700, installed: 0.790] - mypy-extensions [required: >=0.4.3,<0.5.0, installed: 0.4.3] - typed-ast [required: >=1.4.0,<1.5.0, installed: 1.4.2] - typing-extensions [required: >=3.7.4, installed: 3.7.4.3] - pytest [required: >=3.5, installed: 6.2.1] - attrs [required: >=19.2.0, installed: 20.3.0] - iniconfig [required: Any, installed: 1.1.1] - packaging [required: Any, installed: 20.8] - pyparsing [required: >=2.0.2, installed: 2.4.7] - pluggy [required: >=0.12,<1.0.0a1, installed: 0.13.1] - py [required: >=1.8.2, installed: 1.10.0] - toml [required: Any, installed: 0.10.2] pytest-timeout==1.4.2 - pytest [required: >=3.6.0, installed: 6.2.1] - attrs [required: >=19.2.0, installed: 20.3.0] - iniconfig [required: Any, installed: 1.1.1] - packaging [required: Any, installed: 20.8] - pyparsing [required: >=2.0.2, installed: 2.4.7] - pluggy [required: >=0.12,<1.0.0a1, installed: 0.13.1] - py [required: >=1.8.2, installed: 1.10.0] - toml [required: Any, installed: 0.10.2] thoth-analyzer==0.1.8 - click [required: Any, installed: 7.1.2] - delegator.py [required: >=0.1.0, installed: 0.1.1] - pexpect [required: >=4.1.0, installed: 4.8.0] - ptyprocess [required: >=0.5, installed: 0.7.0] - distro [required: Any, installed: 1.5.0] - requests [required: Any, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - thoth-common [required: Any, installed: 0.21.3] - argo-workflows [required: >=3.5,<4, installed: 3.6.1] - kubernetes [required: >=11.0.0, installed: 11.0.0] - certifi [required: >=14.05.14, installed: 2020.12.5] - google-auth [required: >=1.0.1, installed: 1.24.0] - cachetools [required: >=2.0.0,<5.0, installed: 4.2.0] - pyasn1-modules [required: >=0.2.1, installed: 0.2.8] - pyasn1 [required: >=0.4.6,<0.5.0, installed: 0.4.8] - rsa [required: >=3.1.4,<5, installed: 4.7] - pyasn1 [required: >=0.1.3, installed: 0.4.8] - setuptools [required: >=40.3.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - python-dateutil [required: >=2.5.3, installed: 2.8.1] - six [required: >=1.5, installed: 1.15.0] - pyyaml [required: >=3.12, installed: 5.3.1] - requests [required: Any, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - requests-oauthlib [required: Any, installed: 1.3.0] - oauthlib [required: >=3.0.0, installed: 3.1.0] - requests [required: >=2.0.0, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - setuptools [required: >=21.0.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - urllib3 [required: >=1.24.2, installed: 1.26.2] - websocket-client [required: >=0.32.0,!=0.42.*,!=0.41.*,!=0.40.0, installed: 0.57.0] - six [required: Any, installed: 1.15.0] - mock [required: Any, installed: 4.0.3] - attrdict [required: Any, installed: 2.0.1] - six [required: Any, installed: 1.15.0] - attrs [required: Any, installed: 20.3.0] - daiquiri [required: Any, installed: 3.0.0] - python-json-logger [required: Any, installed: 2.0.1] - jsonformatter [required: Any, installed: 0.3.1] - kubernetes [required: ~=11.0.0, installed: 11.0.0] - certifi [required: >=14.05.14, installed: 2020.12.5] - google-auth [required: >=1.0.1, installed: 1.24.0] - cachetools [required: >=2.0.0,<5.0, installed: 4.2.0] - pyasn1-modules [required: >=0.2.1, installed: 0.2.8] - pyasn1 [required: >=0.4.6,<0.5.0, installed: 0.4.8] - rsa [required: >=3.1.4,<5, installed: 4.7] - pyasn1 [required: >=0.1.3, installed: 0.4.8] - setuptools [required: >=40.3.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - python-dateutil [required: >=2.5.3, installed: 2.8.1] - six [required: >=1.5, installed: 1.15.0] - pyyaml [required: >=3.12, installed: 5.3.1] - requests [required: Any, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - requests-oauthlib [required: Any, installed: 1.3.0] - oauthlib [required: >=3.0.0, installed: 3.1.0] - requests [required: >=2.0.0, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - setuptools [required: >=21.0.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - urllib3 [required: >=1.24.2, installed: 1.26.2] - websocket-client [required: >=0.32.0,!=0.42.*,!=0.41.*,!=0.40.0, installed: 0.57.0] - six [required: Any, installed: 1.15.0] - openshift [required: Any, installed: 0.11.2] - jinja2 [required: Any, installed: 2.11.2] - MarkupSafe [required: >=0.23, installed: 1.1.1] - kubernetes [required: ~=11.0.0, installed: 11.0.0] - certifi [required: >=14.05.14, installed: 2020.12.5] - google-auth [required: >=1.0.1, installed: 1.24.0] - cachetools [required: >=2.0.0,<5.0, installed: 4.2.0] - pyasn1-modules [required: >=0.2.1, installed: 0.2.8] - pyasn1 [required: >=0.4.6,<0.5.0, installed: 0.4.8] - rsa [required: >=3.1.4,<5, installed: 4.7] - pyasn1 [required: >=0.1.3, installed: 0.4.8] - setuptools [required: >=40.3.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - python-dateutil [required: >=2.5.3, installed: 2.8.1] - six [required: >=1.5, installed: 1.15.0] - pyyaml [required: >=3.12, installed: 5.3.1] - requests [required: Any, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - requests-oauthlib [required: Any, installed: 1.3.0] - oauthlib [required: >=3.0.0, installed: 3.1.0] - requests [required: >=2.0.0, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - setuptools [required: >=21.0.0, installed: 51.0.0] - six [required: >=1.9.0, installed: 1.15.0] - urllib3 [required: >=1.24.2, installed: 1.26.2] - websocket-client [required: >=0.32.0,!=0.42.*,!=0.41.*,!=0.40.0, installed: 0.57.0] - six [required: Any, installed: 1.15.0] - python-string-utils [required: Any, installed: 1.0.0] - ruamel.yaml [required: >=0.15, installed: 0.16.12] - ruamel.yaml.clib [required: >=0.1.2, installed: 0.2.2] - six [required: Any, installed: 1.15.0] - pyyaml [required: Any, installed: 5.3.1] - requests [required: Any, installed: 2.25.1] - certifi [required: >=2017.4.17, installed: 2020.12.5] - chardet [required: >=3.0.2,<5, installed: 3.0.4] - idna [required: >=2.5,<3, installed: 2.10] - urllib3 [required: >=1.21.1,<1.27, installed: 1.26.2] - rfc5424-logging-handler [required: Any, installed: 1.4.3] - pytz [required: Any, installed: 2020.5] - tzlocal [required: Any, installed: 2.1] - pytz [required: Any, installed: 2020.5] - sentry-sdk [required: Any, installed: 0.19.5] - certifi [required: Any, installed: 2020.12.5] - urllib3 [required: >=1.10.0, installed: 1.26.2] ``` </details> ##### Notes For more information, see [Pipfile](https://raw.githubusercontent.com/thoth-station/python/master/Pipfile) and [Pipfile.lock](https://raw.githubusercontent.com/thoth-station/python/master/Pipfile.lock). Once this issue is resolved, the issue will be automatically closed by bot.
build
failed to update dependencies to their latest version automatic dependency update failed for the current master with sha the automatic dependency management cannot continue please fix errors reported bellow command pipenv lock standard output standard error locking dependencies building requirements  k resolving dependencies  k fail  ktraceback most recent call last file home user local share virtualenvs user od lib site packages pipenv vendor connection py line in new conn conn connection create connection file home user local share virtualenvs user od lib site packages pipenv vendor util connection py line in create connection for res in socket getaddrinfo host port family socket sock stream file usr socket py line in getaddrinfo for res in socket getaddrinfo host port family type proto flags socket gaierror name or service not known during handling of the above exception another exception occurred traceback most recent call last file home user local share virtualenvs user od lib site packages pipenv vendor connectionpool py line in urlopen httplib response self make request file home user local share virtualenvs user od lib site packages pipenv vendor connectionpool py line in make request self validate conn conn file home user local share virtualenvs user od lib site packages pipenv vendor connectionpool py line in validate conn conn connect file home user local share virtualenvs user od lib site packages pipenv vendor connection py line in connect conn self new conn file home user local share virtualenvs user od lib site packages pipenv vendor connection py line in new conn raise newconnectionerror exceptions newconnectionerror failed to establish a new connection name or service not known during handling of the above exception another exception occurred traceback most recent call last file home user local share virtualenvs user od lib site packages pipenv vendor requests adapters py line in send resp conn urlopen file home user local share virtualenvs user od lib site packages pipenv vendor connectionpool py line in urlopen retries retries increment file home user local share virtualenvs user od lib site packages pipenv vendor util retry py line in increment raise maxretryerror pool url error or responseerror cause exceptions maxretryerror httpsconnectionpool host pypi org port max retries exceeded with url pypi py json caused by newconnectionerror failed to establish a new connection name or service not known during handling of the above exception another exception occurred traceback most recent call last file home user local share virtualenvs user od lib site packages pipenv resolver py line in main file home user local share virtualenvs user od lib site packages pipenv resolver py line in main main parsed pre parsed clear parsed verbose parsed system parsed write file home user local share virtualenvs user od lib site packages pipenv resolver py line in main resolve packages pre clear verbose system write requirements dir packages dev file home user local share virtualenvs user od lib site packages pipenv resolver py line in resolve packages results resolver resolve file home user local share virtualenvs user od lib site packages pipenv resolver py line in resolve return resolve deps file home user local share virtualenvs user od lib site packages pipenv utils py line in resolve deps results hashes markers lookup resolver skipped actually resolve deps file home user local share virtualenvs user od lib site packages pipenv utils py line in actually resolve deps results resolver clean results file home user local share virtualenvs user od lib site packages pipenv utils py line in clean results collected hashes self collect hashes ireq file home user local share virtualenvs user od lib site packages pipenv utils py line in collect hashes r session get pkg url timeout file home user local share virtualenvs user od lib site packages pipenv vendor requests sessions py line in get return self request get url kwargs file home user local share virtualenvs user od lib site packages pipenv vendor requests sessions py line in request resp self send prep send kwargs file home user local share virtualenvs user od lib site packages pipenv vendor requests sessions py line in send r adapter send request kwargs file home user local share virtualenvs user od lib site packages pipenv vendor requests adapters py line in send raise connectionerror e request request requests exceptions connectionerror httpsconnectionpool host pypi org port max retries exceeded with url pypi py json caused by newconnectionerror failed to establish a new connection name or service not known traceback most recent call last file home user local share virtualenvs user od lib site packages pipenv utils py line in create spinner yield sp file home user local share virtualenvs user od lib site packages pipenv utils py line in venv resolve deps c resolve cmd sp file home user local share virtualenvs user od lib site packages pipenv utils py line in resolve sys exit c return code systemexit  k environment details kebechet version python version platform linux with pipenv version pipenv version dependency graph aiohttp async timeout attrs chardet multidict typing extensions yarl idna multidict flexmock lxml pyelftools pytest asyncio pytest attrs iniconfig packaging pyparsing pluggy py toml pytest cov coverage pytest attrs iniconfig packaging pyparsing pluggy py toml pytest mypy attrs filelock mypy mypy extensions typed ast typing extensions pytest attrs iniconfig packaging pyparsing pluggy py toml pytest timeout pytest attrs iniconfig packaging pyparsing pluggy py toml thoth analyzer click delegator py pexpect ptyprocess distro requests certifi chardet idna thoth common argo workflows kubernetes certifi google auth cachetools modules rsa setuptools six python dateutil six pyyaml requests certifi chardet idna requests oauthlib oauthlib requests certifi chardet idna setuptools six websocket client six mock attrdict six attrs daiquiri python json logger jsonformatter kubernetes certifi google auth cachetools modules rsa setuptools six python dateutil six pyyaml requests certifi chardet idna requests oauthlib oauthlib requests certifi chardet idna setuptools six websocket client six openshift markupsafe kubernetes certifi google auth cachetools modules rsa setuptools six python dateutil six pyyaml requests certifi chardet idna requests oauthlib oauthlib requests certifi chardet idna setuptools six websocket client six python string utils ruamel yaml ruamel yaml clib six pyyaml requests certifi chardet idna logging handler pytz tzlocal pytz sentry sdk certifi notes for more information see and once this issue is resolved the issue will be automatically closed by bot
1
62,695
15,341,469,723
IssuesEvent
2021-02-27 12:07:06
JDimproved/JDim
https://api.github.com/repos/JDimproved/JDim
closed
Meson 0.57を使ったとき test サブコマンドが失敗する
bug build
<!-- バグの報告ありがとうございます! 以下の項目は修正の手掛かり・助けになりますので記入をお願いいたします。 不具合なのかはっきりしない動作や挙動はDiscussionsで質問してみてください。 --> **バグの説明** <!-- 何が起こったか、エラーメッセージがあればそれも書く --> Meson 0.57の **test** サブコマンドでテストプログラムをビルドするとエラーメッセージが出て失敗します。 0.56ではエラーが出ずテストを実行できます。 edit: 0.57.1でもエラーが出て失敗しました。 エラーメッセージ ```sh [222/273] Compiling C++ object test/gtest_jdim.p/.._src_environment.cpp.o FAILED: test/gtest_jdim.p/.._src_environment.cpp.o ccache c++ -Itest/gtest_jdim.p -Itest -I../test -Isrc -I../src -I/usr/include/gtkmm-3.0 -I/usr/lib/x86_64-linux-gnu/gtkmm-3.0/include -I/usr/include/atkmm-1.6 -I/usr/include/gtk-3.0/unix-print -I/usr/include/gdkmm-3.0 -I/usr/lib/x86_64-linux-gnu/gdkmm-3.0/include -I/usr/include/giomm-2.4 -I/usr/lib/x86_64-linux-gnu/giomm-2.4/include -I/usr/include/pangomm-1.4 -I/usr/lib/x86_64-linux-gnu/pangomm-1.4/include -I/usr/include/glibmm-2.4 -I/usr/lib/x86_64-linux-gnu/glibmm-2.4/include -I/usr/include/gtk-3.0 -I/usr/include/at-spi2-atk/2.0 -I/usr/include/at-spi-2.0 -I/usr/include/dbus-1.0 -I/usr/lib/x86_64-linux-gnu/dbus-1.0/include -I/usr/include/gio-unix-2.0 -I/usr/include/cairo -I/usr/include/pango-1.0 -I/usr/include/fribidi -I/usr/include/harfbuzz -I/usr/include/atk-1.0 -I/usr/include/cairomm-1.0 -I/usr/lib/x86_64-linux-gnu/cairomm-1.0/include -I/usr/include/pixman-1 -I/usr/include/uuid -I/usr/include/freetype2 -I/usr/include/libpng16 -I/usr/include/sigc++-2.0 -I/usr/lib/x86_64-linux-gnu/sigc++-2.0/include -I/usr/include/gdk-pixbuf-2.0 -I/usr/include/libmount -I/usr/include/blkid -I/usr/include/glib-2.0 -I/usr/lib/x86_64-linux-gnu/glib-2.0/include -I/usr/include/p11-kit-1 -fdiagnostics-color=always -pipe -D_FILE_OFFSET_BITS=64 -Wall -Winvalid-pch -Wnon-virtual-dtor -Wextra -Wpedantic -std=c++1z -g -DHAVE_CONFIG_H=1 '-DGTK_DOMAIN="gtk30"' -Wno-unused-parameter -pthread -Wl,--start-group -lpthread -Wl,--end-group -DGTEST_HAS_PTHREAD=1 -MD -MQ test/gtest_jdim.p/.._src_environment.cpp.o -MF test/gtest_jdim.p/.._src_environment.cpp.o.d -o test/gtest_jdim.p/.._src_environment.cpp.o -c ../src/environment.cpp In file included from ../src/environment.cpp:10: ../src/jdversion.h:10:10: fatal error: buildinfo.h: そのようなファイルやディレクトリはありません 10 | #include "buildinfo.h" | ^~~~~~~~~~~~~ compilation terminated. ``` **再現の方法** <!-- バグの動作を再現する手順を書く --> Meson 0.57をインストールして端末から `meson test` を実行します。 ```sh $ meson --version 0.57.0 $ meson t057 (snip) $ meson test -C t057 (snip) # 上記のエラーメッセージが出る ``` **やりたかったこと・期待する結果** <!-- 本来の望ましい動作を書く --> testサブコマンドでテストのビルドと実行をする。 **スクリーンショット** <!-- 表示の問題はスクリーンショットがあれば分かりやすい --> なし **動作環境** <!-- クリップボードへのコピーを利用する --> ``` [バージョン] JDim 0.5.0-20210221(git:d9d1ddddf4) [ディストリ ] Ubuntu 20.10 (x86_64) [パッケージ] バイナリ/ソース( <配布元> ) [ DE/WM ] KDE [ gtkmm  ] 3.24.2 [ glibmm  ] 2.64.2 [ TLS lib ] GnuTLS 3.6.15 [ そ の 他 ] ``` **追加の情報** <!-- ユーザーエージェントなどが手掛かりとなる情報 --> **compile** サブコマンドでビルドするときはエラーが出ませんでした。 ```sh $ meson --version 0.57.0 $ meson b057 (snip) $ meson compile -C b057 ninja: Entering directory `b057' [318/318] Linking target test/gtest_jdim $ meson test -C b057 [1/3] Generating buildinfo.h with a custom command (wrapped by meson to capture output) 1/1 gtest tests OK 0.04s (snip) ``` Meson 0.57の更新で **test** サブコマンドはテストプログラムのみ再ビルドするように変更されたようです。 https://mesonbuild.com/Release-notes-for-0-57-0.html#meson-test-only-rebuilds-test-dependencies > Until now, meson test rebuilt the whole project independent of the requested tests and their dependencies. With this release, meson test will only rebuild what is needed for the tests or suites that will be run. > (snip) > However, this change could cause failures when upgrading to 0.57, if the dependencies are not specified correctly in meson.build.
1.0
Meson 0.57を使ったとき test サブコマンドが失敗する - <!-- バグの報告ありがとうございます! 以下の項目は修正の手掛かり・助けになりますので記入をお願いいたします。 不具合なのかはっきりしない動作や挙動はDiscussionsで質問してみてください。 --> **バグの説明** <!-- 何が起こったか、エラーメッセージがあればそれも書く --> Meson 0.57の **test** サブコマンドでテストプログラムをビルドするとエラーメッセージが出て失敗します。 0.56ではエラーが出ずテストを実行できます。 edit: 0.57.1でもエラーが出て失敗しました。 エラーメッセージ ```sh [222/273] Compiling C++ object test/gtest_jdim.p/.._src_environment.cpp.o FAILED: test/gtest_jdim.p/.._src_environment.cpp.o ccache c++ -Itest/gtest_jdim.p -Itest -I../test -Isrc -I../src -I/usr/include/gtkmm-3.0 -I/usr/lib/x86_64-linux-gnu/gtkmm-3.0/include -I/usr/include/atkmm-1.6 -I/usr/include/gtk-3.0/unix-print -I/usr/include/gdkmm-3.0 -I/usr/lib/x86_64-linux-gnu/gdkmm-3.0/include -I/usr/include/giomm-2.4 -I/usr/lib/x86_64-linux-gnu/giomm-2.4/include -I/usr/include/pangomm-1.4 -I/usr/lib/x86_64-linux-gnu/pangomm-1.4/include -I/usr/include/glibmm-2.4 -I/usr/lib/x86_64-linux-gnu/glibmm-2.4/include -I/usr/include/gtk-3.0 -I/usr/include/at-spi2-atk/2.0 -I/usr/include/at-spi-2.0 -I/usr/include/dbus-1.0 -I/usr/lib/x86_64-linux-gnu/dbus-1.0/include -I/usr/include/gio-unix-2.0 -I/usr/include/cairo -I/usr/include/pango-1.0 -I/usr/include/fribidi -I/usr/include/harfbuzz -I/usr/include/atk-1.0 -I/usr/include/cairomm-1.0 -I/usr/lib/x86_64-linux-gnu/cairomm-1.0/include -I/usr/include/pixman-1 -I/usr/include/uuid -I/usr/include/freetype2 -I/usr/include/libpng16 -I/usr/include/sigc++-2.0 -I/usr/lib/x86_64-linux-gnu/sigc++-2.0/include -I/usr/include/gdk-pixbuf-2.0 -I/usr/include/libmount -I/usr/include/blkid -I/usr/include/glib-2.0 -I/usr/lib/x86_64-linux-gnu/glib-2.0/include -I/usr/include/p11-kit-1 -fdiagnostics-color=always -pipe -D_FILE_OFFSET_BITS=64 -Wall -Winvalid-pch -Wnon-virtual-dtor -Wextra -Wpedantic -std=c++1z -g -DHAVE_CONFIG_H=1 '-DGTK_DOMAIN="gtk30"' -Wno-unused-parameter -pthread -Wl,--start-group -lpthread -Wl,--end-group -DGTEST_HAS_PTHREAD=1 -MD -MQ test/gtest_jdim.p/.._src_environment.cpp.o -MF test/gtest_jdim.p/.._src_environment.cpp.o.d -o test/gtest_jdim.p/.._src_environment.cpp.o -c ../src/environment.cpp In file included from ../src/environment.cpp:10: ../src/jdversion.h:10:10: fatal error: buildinfo.h: そのようなファイルやディレクトリはありません 10 | #include "buildinfo.h" | ^~~~~~~~~~~~~ compilation terminated. ``` **再現の方法** <!-- バグの動作を再現する手順を書く --> Meson 0.57をインストールして端末から `meson test` を実行します。 ```sh $ meson --version 0.57.0 $ meson t057 (snip) $ meson test -C t057 (snip) # 上記のエラーメッセージが出る ``` **やりたかったこと・期待する結果** <!-- 本来の望ましい動作を書く --> testサブコマンドでテストのビルドと実行をする。 **スクリーンショット** <!-- 表示の問題はスクリーンショットがあれば分かりやすい --> なし **動作環境** <!-- クリップボードへのコピーを利用する --> ``` [バージョン] JDim 0.5.0-20210221(git:d9d1ddddf4) [ディストリ ] Ubuntu 20.10 (x86_64) [パッケージ] バイナリ/ソース( <配布元> ) [ DE/WM ] KDE [ gtkmm  ] 3.24.2 [ glibmm  ] 2.64.2 [ TLS lib ] GnuTLS 3.6.15 [ そ の 他 ] ``` **追加の情報** <!-- ユーザーエージェントなどが手掛かりとなる情報 --> **compile** サブコマンドでビルドするときはエラーが出ませんでした。 ```sh $ meson --version 0.57.0 $ meson b057 (snip) $ meson compile -C b057 ninja: Entering directory `b057' [318/318] Linking target test/gtest_jdim $ meson test -C b057 [1/3] Generating buildinfo.h with a custom command (wrapped by meson to capture output) 1/1 gtest tests OK 0.04s (snip) ``` Meson 0.57の更新で **test** サブコマンドはテストプログラムのみ再ビルドするように変更されたようです。 https://mesonbuild.com/Release-notes-for-0-57-0.html#meson-test-only-rebuilds-test-dependencies > Until now, meson test rebuilt the whole project independent of the requested tests and their dependencies. With this release, meson test will only rebuild what is needed for the tests or suites that will be run. > (snip) > However, this change could cause failures when upgrading to 0.57, if the dependencies are not specified correctly in meson.build.
build
meson test サブコマンドが失敗する バグの報告ありがとうございます! 以下の項目は修正の手掛かり・助けになりますので記入をお願いいたします。 不具合なのかはっきりしない動作や挙動はdiscussionsで質問してみてください。 バグの説明 meson test サブコマンドでテストプログラムをビルドするとエラーメッセージが出て失敗します。 。 edit 。 エラーメッセージ sh compiling c object test gtest jdim p src environment cpp o failed test gtest jdim p src environment cpp o ccache c itest gtest jdim p itest i test isrc i src i usr include gtkmm i usr lib linux gnu gtkmm include i usr include atkmm i usr include gtk unix print i usr include gdkmm i usr lib linux gnu gdkmm include i usr include giomm i usr lib linux gnu giomm include i usr include pangomm i usr lib linux gnu pangomm include i usr include glibmm i usr lib linux gnu glibmm include i usr include gtk i usr include at atk i usr include at spi i usr include dbus i usr lib linux gnu dbus include i usr include gio unix i usr include cairo i usr include pango i usr include fribidi i usr include harfbuzz i usr include atk i usr include cairomm i usr lib linux gnu cairomm include i usr include pixman i usr include uuid i usr include i usr include i usr include sigc i usr lib linux gnu sigc include i usr include gdk pixbuf i usr include libmount i usr include blkid i usr include glib i usr lib linux gnu glib include i usr include kit fdiagnostics color always pipe d file offset bits wall winvalid pch wnon virtual dtor wextra wpedantic std c g dhave config h dgtk domain wno unused parameter pthread wl start group lpthread wl end group dgtest has pthread md mq test gtest jdim p src environment cpp o mf test gtest jdim p src environment cpp o d o test gtest jdim p src environment cpp o c src environment cpp in file included from src environment cpp src jdversion h fatal error buildinfo h そのようなファイルやディレクトリはありません include buildinfo h compilation terminated 再現の方法 meson meson test を実行します。 sh meson version meson snip meson test c snip 上記のエラーメッセージが出る やりたかったこと・期待する結果 testサブコマンドでテストのビルドと実行をする。 スクリーンショット なし 動作環境 jdim git ubuntu バイナリ ソース kde gnutls 追加の情報 compile サブコマンドでビルドするときはエラーが出ませんでした。 sh meson version meson snip meson compile c ninja entering directory linking target test gtest jdim meson test c generating buildinfo h with a custom command wrapped by meson to capture output gtest tests ok snip meson test サブコマンドはテストプログラムのみ再ビルドするように変更されたようです。 until now meson test rebuilt the whole project independent of the requested tests and their dependencies with this release meson test will only rebuild what is needed for the tests or suites that will be run snip however this change could cause failures when upgrading to if the dependencies are not specified correctly in meson build
1
88,946
25,544,574,448
IssuesEvent
2022-11-29 17:41:26
zowe/imperative
https://api.github.com/repos/zowe/imperative
closed
Degulpify sequential build tasks
build
We have sequential tasks using Gulp that don't need to be, and their performance could be improved by degulpifying them 🙂 For more complex tasks like parallel commands (watch = tsc + lint) or web help bundle generation, it's probably fine to keep using Gulp for now. _Originally posted by @t1m0thyj in https://github.com/zowe/imperative/pull/684#discussion_r740366764_
1.0
Degulpify sequential build tasks - We have sequential tasks using Gulp that don't need to be, and their performance could be improved by degulpifying them 🙂 For more complex tasks like parallel commands (watch = tsc + lint) or web help bundle generation, it's probably fine to keep using Gulp for now. _Originally posted by @t1m0thyj in https://github.com/zowe/imperative/pull/684#discussion_r740366764_
build
degulpify sequential build tasks we have sequential tasks using gulp that don t need to be and their performance could be improved by degulpifying them 🙂 for more complex tasks like parallel commands watch tsc lint or web help bundle generation it s probably fine to keep using gulp for now originally posted by in
1
11,129
4,892,077,318
IssuesEvent
2016-11-18 18:36:37
mozilla-mobile/prox
https://api.github.com/repos/mozilla-mobile/prox
opened
mheubusch@mozilla.com on build #274: Vet showing up as a category
1.0 buddybuild-feedback
Feedback from mheubusch@mozilla.com : Vet showing up as a category <img src="https://s3-us-west-2.amazonaws.com/buddybuild-screenshots/57f51f8bcaa70a01005412ee/582c90352d9fa90100991a16/981c8efa-5ed2-43db-85cd-84c7f05e98ac.jpg" width="33%" height="33%" /><table><tr><td>Created</td><td>Fri Nov 18 2016 18:36:30 GMT+0000 (UTC)</td></tr><tr><td>CFBundleShortVersion</td><td>1.0</td></tr><tr><td>CFBundleVersion</td><td>1</td></tr><tr><td>App uptime</td><td>54367.97</td></tr><tr><td>Build</td><td>274</td></tr><tr><td>Device type</td><td>iPhone 7</td></tr><tr><td>Device name</td><td>Michelle's iPhone</td></tr><tr><td>Screen size</td><td>375</td></tr><tr><td>Screen size</td><td>375px by 667px</td></tr><tr><td>Battery</td><td>18% Unplugged</td></tr><tr><td>Memory free</td><td>40 MB / 1426 MB</td></tr><tr><td>Disk free</td><td>103975 MB / 116315 MB</td></tr><tr><td>Network IP</td><td>0.0.0.0</td></tr></table> [Link to buddybuild feedback from build 274](https://dashboard.buddybuild.com/apps/57f51f8bcaa70a01005412ee/feedback?fid=582f4a2e05613401004913f4&bnum=274)
1.0
mheubusch@mozilla.com on build #274: Vet showing up as a category - Feedback from mheubusch@mozilla.com : Vet showing up as a category <img src="https://s3-us-west-2.amazonaws.com/buddybuild-screenshots/57f51f8bcaa70a01005412ee/582c90352d9fa90100991a16/981c8efa-5ed2-43db-85cd-84c7f05e98ac.jpg" width="33%" height="33%" /><table><tr><td>Created</td><td>Fri Nov 18 2016 18:36:30 GMT+0000 (UTC)</td></tr><tr><td>CFBundleShortVersion</td><td>1.0</td></tr><tr><td>CFBundleVersion</td><td>1</td></tr><tr><td>App uptime</td><td>54367.97</td></tr><tr><td>Build</td><td>274</td></tr><tr><td>Device type</td><td>iPhone 7</td></tr><tr><td>Device name</td><td>Michelle's iPhone</td></tr><tr><td>Screen size</td><td>375</td></tr><tr><td>Screen size</td><td>375px by 667px</td></tr><tr><td>Battery</td><td>18% Unplugged</td></tr><tr><td>Memory free</td><td>40 MB / 1426 MB</td></tr><tr><td>Disk free</td><td>103975 MB / 116315 MB</td></tr><tr><td>Network IP</td><td>0.0.0.0</td></tr></table> [Link to buddybuild feedback from build 274](https://dashboard.buddybuild.com/apps/57f51f8bcaa70a01005412ee/feedback?fid=582f4a2e05613401004913f4&bnum=274)
build
mheubusch mozilla com on build vet showing up as a category feedback from mheubusch mozilla com vet showing up as a category created fri nov gmt utc cfbundleshortversion cfbundleversion app uptime build device type iphone device name michelle s iphone screen size screen size by battery unplugged memory free mb mb disk free mb mb network ip
1
99,378
30,440,370,510
IssuesEvent
2023-07-15 02:07:20
google/mediapipe
https://api.github.com/repos/google/mediapipe
closed
Sporadic error starting up MediaPipe - attachToGLContext
type:build/install stat:awaiting response platform::android stale
Sometimes MediaPipe crashes on startup (~ 1 out of 10 times) with the following error in my Android app. android.graphics.SurfaceTexture in attachToGLContext at line 289 com.google.mediapipe.components.ExternalTextureConverter$RenderThread in setSurfaceTextureAndAttachToGLContext at line 15 com.google.mediapipe.components.ExternalTextureConverter in lambda$setSurfaceTextureAndAttachToGLContext$2 at line 3 com.google.mediapipe.components.ExternalTextureConverter in a at line 1 com.google.mediapipe.components.h in run at line 9 android.os.Handler in handleCallback at line 873 android.os.Handler in dispatchMessage at line 99 android.os.Looper in loop at line 193 com.google.mediapipe.glutil.GlThread in run at line 54 Is there any way to catch this error and prevent my app from crashing? I've added an error listener to ExternalTextureConverter temporarily to pass Exceptions back to my app but wandering if there was a better way that doesn't involve editing the library. Thank you
1.0
Sporadic error starting up MediaPipe - attachToGLContext - Sometimes MediaPipe crashes on startup (~ 1 out of 10 times) with the following error in my Android app. android.graphics.SurfaceTexture in attachToGLContext at line 289 com.google.mediapipe.components.ExternalTextureConverter$RenderThread in setSurfaceTextureAndAttachToGLContext at line 15 com.google.mediapipe.components.ExternalTextureConverter in lambda$setSurfaceTextureAndAttachToGLContext$2 at line 3 com.google.mediapipe.components.ExternalTextureConverter in a at line 1 com.google.mediapipe.components.h in run at line 9 android.os.Handler in handleCallback at line 873 android.os.Handler in dispatchMessage at line 99 android.os.Looper in loop at line 193 com.google.mediapipe.glutil.GlThread in run at line 54 Is there any way to catch this error and prevent my app from crashing? I've added an error listener to ExternalTextureConverter temporarily to pass Exceptions back to my app but wandering if there was a better way that doesn't involve editing the library. Thank you
build
sporadic error starting up mediapipe attachtoglcontext sometimes mediapipe crashes on startup out of times with the following error in my android app android graphics surfacetexture in attachtoglcontext at line com google mediapipe components externaltextureconverter renderthread in setsurfacetextureandattachtoglcontext at line com google mediapipe components externaltextureconverter in lambda setsurfacetextureandattachtoglcontext at line com google mediapipe components externaltextureconverter in a at line com google mediapipe components h in run at line android os handler in handlecallback at line android os handler in dispatchmessage at line android os looper in loop at line com google mediapipe glutil glthread in run at line is there any way to catch this error and prevent my app from crashing i ve added an error listener to externaltextureconverter temporarily to pass exceptions back to my app but wandering if there was a better way that doesn t involve editing the library thank you
1
11,974
5,116,410,509
IssuesEvent
2017-01-07 03:08:03
girder/girder
https://api.github.com/repos/girder/girder
opened
webpack ExtractTextPlugin is not pulling the CSS out of the JS module
bug build
All of the CSS is still inlined inside our .js files, in addition to being extracted into the .css files. We should find some way to remove this duplication, we might just be misusing ExtractTextPlugin. @sebastienbarre I banged my head on this for a while, but the docs for ExtractTextPlugin leave a lot to be desired, and I couldn't get it working... any experience with this issue?
1.0
webpack ExtractTextPlugin is not pulling the CSS out of the JS module - All of the CSS is still inlined inside our .js files, in addition to being extracted into the .css files. We should find some way to remove this duplication, we might just be misusing ExtractTextPlugin. @sebastienbarre I banged my head on this for a while, but the docs for ExtractTextPlugin leave a lot to be desired, and I couldn't get it working... any experience with this issue?
build
webpack extracttextplugin is not pulling the css out of the js module all of the css is still inlined inside our js files in addition to being extracted into the css files we should find some way to remove this duplication we might just be misusing extracttextplugin sebastienbarre i banged my head on this for a while but the docs for extracttextplugin leave a lot to be desired and i couldn t get it working any experience with this issue
1
266,045
28,298,878,046
IssuesEvent
2023-04-10 02:50:56
nidhi7598/linux-4.19.72
https://api.github.com/repos/nidhi7598/linux-4.19.72
closed
CVE-2021-3573 (Medium) detected in linuxlinux-4.19.254 - autoclosed
Mend: dependency security vulnerability
## CVE-2021-3573 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.254</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72/commit/10a8c99e4f60044163c159867bc6f5452c1c36e5">10a8c99e4f60044163c159867bc6f5452c1c36e5</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_sock.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_sock.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A use-after-free in function hci_sock_bound_ioctl() of the Linux kernel HCI subsystem was found in the way user calls ioct HCIUNBLOCKADDR or other way triggers race condition of the call hci_unregister_dev() together with one of the calls hci_sock_blacklist_add(), hci_sock_blacklist_del(), hci_get_conn_info(), hci_get_auth_info(). A privileged local user could use this flaw to crash the system or escalate their privileges on the system. This flaw affects the Linux kernel versions prior to 5.13-rc5. <p>Publish Date: 2021-08-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3573>CVE-2021-3573</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-3573">https://www.linuxkernelcves.com/cves/CVE-2021-3573</a></p> <p>Release Date: 2021-08-13</p> <p>Fix Resolution: v4.4.272, v4.9.272, v4.14.236, v4.19.194, v5.4.125, v5.10.43, v5.12.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3573 (Medium) detected in linuxlinux-4.19.254 - autoclosed - ## CVE-2021-3573 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.254</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72/commit/10a8c99e4f60044163c159867bc6f5452c1c36e5">10a8c99e4f60044163c159867bc6f5452c1c36e5</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_sock.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_sock.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A use-after-free in function hci_sock_bound_ioctl() of the Linux kernel HCI subsystem was found in the way user calls ioct HCIUNBLOCKADDR or other way triggers race condition of the call hci_unregister_dev() together with one of the calls hci_sock_blacklist_add(), hci_sock_blacklist_del(), hci_get_conn_info(), hci_get_auth_info(). A privileged local user could use this flaw to crash the system or escalate their privileges on the system. This flaw affects the Linux kernel versions prior to 5.13-rc5. <p>Publish Date: 2021-08-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3573>CVE-2021-3573</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-3573">https://www.linuxkernelcves.com/cves/CVE-2021-3573</a></p> <p>Release Date: 2021-08-13</p> <p>Fix Resolution: v4.4.272, v4.9.272, v4.14.236, v4.19.194, v5.4.125, v5.10.43, v5.12.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_build
cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net bluetooth hci sock c net bluetooth hci sock c vulnerability details a use after free in function hci sock bound ioctl of the linux kernel hci subsystem was found in the way user calls ioct hciunblockaddr or other way triggers race condition of the call hci unregister dev together with one of the calls hci sock blacklist add hci sock blacklist del hci get conn info hci get auth info a privileged local user could use this flaw to crash the system or escalate their privileges on the system this flaw affects the linux kernel versions prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
9,512
4,540,152,429
IssuesEvent
2016-09-09 13:48:13
zalando-incubator/atlas-ios
https://api.github.com/repos/zalando-incubator/atlas-ios
closed
raphael.koch@zalando.de on build #345: Got this error after selecting the billing address
1.0 buddybuild ui
Feedback from raphael.koch@zalando.de : Got this error after selecting the billing address <img src="https://s3-us-west-2.amazonaws.com/buddybuild-screenshots/57a305cb34a9450100595b71/57ccdfb0284b7f01006b3e5f/a5d5cbd0-533c-434a-8b38-28e360637b94.jpg" width="33%" height="33%" /><table><tr><td>Created</td><td>Wed Sep 07 2016 08:43:57 GMT+0000 (UTC)</td></tr><tr><td>CFBundleShortVersion</td><td>1.0</td></tr><tr><td>CFBundleVersion</td><td>20160905.345</td></tr><tr><td>App uptime</td><td>1054.744</td></tr><tr><td>Build</td><td>345</td></tr><tr><td>Device type</td><td>iPhone 6</td></tr><tr><td>Device name</td><td>Raphaels iPhone 7</td></tr><tr><td>Screen size</td><td>375</td></tr><tr><td>Screen size</td><td>375px by 667px</td></tr><tr><td>Battery</td><td>95% Unplugged</td></tr><tr><td>Memory free</td><td>97 MB / 2568 MB</td></tr><tr><td>Disk free</td><td>2292 MB / 56981 MB</td></tr><tr><td>Network IP</td><td>10.161.57.89</td></tr></table> [Link to buddybuild feedback from build 345](https://dashboard.buddybuild.com/apps/57a305cb34a9450100595b71/feedback?fid=57cfd34d38c63c01006b4f7e&bnum=345)
1.0
raphael.koch@zalando.de on build #345: Got this error after selecting the billing address - Feedback from raphael.koch@zalando.de : Got this error after selecting the billing address <img src="https://s3-us-west-2.amazonaws.com/buddybuild-screenshots/57a305cb34a9450100595b71/57ccdfb0284b7f01006b3e5f/a5d5cbd0-533c-434a-8b38-28e360637b94.jpg" width="33%" height="33%" /><table><tr><td>Created</td><td>Wed Sep 07 2016 08:43:57 GMT+0000 (UTC)</td></tr><tr><td>CFBundleShortVersion</td><td>1.0</td></tr><tr><td>CFBundleVersion</td><td>20160905.345</td></tr><tr><td>App uptime</td><td>1054.744</td></tr><tr><td>Build</td><td>345</td></tr><tr><td>Device type</td><td>iPhone 6</td></tr><tr><td>Device name</td><td>Raphaels iPhone 7</td></tr><tr><td>Screen size</td><td>375</td></tr><tr><td>Screen size</td><td>375px by 667px</td></tr><tr><td>Battery</td><td>95% Unplugged</td></tr><tr><td>Memory free</td><td>97 MB / 2568 MB</td></tr><tr><td>Disk free</td><td>2292 MB / 56981 MB</td></tr><tr><td>Network IP</td><td>10.161.57.89</td></tr></table> [Link to buddybuild feedback from build 345](https://dashboard.buddybuild.com/apps/57a305cb34a9450100595b71/feedback?fid=57cfd34d38c63c01006b4f7e&bnum=345)
build
raphael koch zalando de on build got this error after selecting the billing address feedback from raphael koch zalando de got this error after selecting the billing address created wed sep gmt utc cfbundleshortversion cfbundleversion app uptime build device type iphone device name raphaels iphone screen size screen size by battery unplugged memory free mb mb disk free mb mb network ip
1
28,195
32,028,950,152
IssuesEvent
2023-09-22 10:51:51
zaproxy/zaproxy
https://api.github.com/repos/zaproxy/zaproxy
opened
Handle snaps better
enhancement Usability add-on
See https://stackoverflow.com/questions/77156598/run-owasp-zap-zaproxy-on-ubuntu We can tell if ZAP is installed as a snap, and we should be able to check if chrome and firefox are as well. Automatically configuring ZAP to handle this case would be very useful and helpful for our users.
True
Handle snaps better - See https://stackoverflow.com/questions/77156598/run-owasp-zap-zaproxy-on-ubuntu We can tell if ZAP is installed as a snap, and we should be able to check if chrome and firefox are as well. Automatically configuring ZAP to handle this case would be very useful and helpful for our users.
non_build
handle snaps better see we can tell if zap is installed as a snap and we should be able to check if chrome and firefox are as well automatically configuring zap to handle this case would be very useful and helpful for our users
0
107,998
11,577,983,404
IssuesEvent
2020-02-21 15:07:38
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
Update feature toggle documentation
content-ia-team documentation
## User Story or Problem Statement As a developer, I need to understand feature toggles so I can test new functionality. ## Goal Developers should be able to use the documentation to understand what feature toggles are and how to use them. ## Tasks - Pull together existing feature toggle documentation - Obtain additional input and reviews from SMEs ## Acceptance Criteria - Signoff from SMEs
1.0
Update feature toggle documentation - ## User Story or Problem Statement As a developer, I need to understand feature toggles so I can test new functionality. ## Goal Developers should be able to use the documentation to understand what feature toggles are and how to use them. ## Tasks - Pull together existing feature toggle documentation - Obtain additional input and reviews from SMEs ## Acceptance Criteria - Signoff from SMEs
non_build
update feature toggle documentation user story or problem statement as a developer i need to understand feature toggles so i can test new functionality goal developers should be able to use the documentation to understand what feature toggles are and how to use them tasks pull together existing feature toggle documentation obtain additional input and reviews from smes acceptance criteria signoff from smes
0
16,484
6,206,810,693
IssuesEvent
2017-07-06 19:17:48
docker/compose
https://api.github.com/repos/docker/compose
closed
Support build --network
area/build kind/feature kind/parity
`docker build` supports `--network` option (note, this is different from `docker run`'s `--net` option). Request extending docker-compose yaml spec to include the following: ```yaml build: context: . network: host ``` Thanks.
1.0
Support build --network - `docker build` supports `--network` option (note, this is different from `docker run`'s `--net` option). Request extending docker-compose yaml spec to include the following: ```yaml build: context: . network: host ``` Thanks.
build
support build network docker build supports network option note this is different from docker run s net option request extending docker compose yaml spec to include the following yaml build context network host thanks
1
311,340
26,782,935,852
IssuesEvent
2023-01-31 23:02:26
ZcashFoundation/zebra
https://api.github.com/repos/ZcashFoundation/zebra
closed
Test the getblocktemplate RPC's response as a block template proposal in CI
A-rust C-enhancement S-needs-triage P-Low :snowflake: C-testing A-rpc
## Motivation The acceptance test could confirm that the `getblocktemplate` response in 'template' mode will be accepted if it has a valid solution by submitting the same data to the `getblocktemplate` method in 'proposal' mode. ### Design In the acceptance test that syncs until the tip and calls the `getblocktemplate` RPC method: - Deserialize the response - Convert the template into a block with empty `solution` and `nonce` fields - Hex-encode the resulting block - Call `getblocktemplate` in 'proposal' mode - Assert a "null" response indicating successful validation (similar to the `submitblock` test)
1.0
Test the getblocktemplate RPC's response as a block template proposal in CI - ## Motivation The acceptance test could confirm that the `getblocktemplate` response in 'template' mode will be accepted if it has a valid solution by submitting the same data to the `getblocktemplate` method in 'proposal' mode. ### Design In the acceptance test that syncs until the tip and calls the `getblocktemplate` RPC method: - Deserialize the response - Convert the template into a block with empty `solution` and `nonce` fields - Hex-encode the resulting block - Call `getblocktemplate` in 'proposal' mode - Assert a "null" response indicating successful validation (similar to the `submitblock` test)
non_build
test the getblocktemplate rpc s response as a block template proposal in ci motivation the acceptance test could confirm that the getblocktemplate response in template mode will be accepted if it has a valid solution by submitting the same data to the getblocktemplate method in proposal mode design in the acceptance test that syncs until the tip and calls the getblocktemplate rpc method deserialize the response convert the template into a block with empty solution and nonce fields hex encode the resulting block call getblocktemplate in proposal mode assert a null response indicating successful validation similar to the submitblock test
0
688,052
23,546,902,361
IssuesEvent
2022-08-21 08:47:39
docker-mailserver/docker-mailserver
https://api.github.com/repos/docker-mailserver/docker-mailserver
opened
Should i be worried about this email?
meta/needs triage meta/help wanted kind/question priority/low
# Subject <!-- Select one, remove the others --> Other ## Description <!-- When copy/pasting code, format the code with tripe backticks (`) ! --> This is off-topic, however, thought someone would be nice enough to look at this :) I've received a spam bounceback, but it originates from my own email (my domain has been replaced with `domain`) Heres the body of the email: ``` Hello, this is the mail server on packages13.dhl.com. I am sending you this message to inform you on the delivery status of a message you previously sent. Immediately below you will find a list of the affected recipients; also attached is a Delivery Status Notification (DSN) report in standard format, as well as the headers of the original message. <[signs@domain.com.au](mailto:signs@domain.com.au)> delivery failed; will not continue trying Reporting-MTA: dns;packages13.dhl.com X-PowerMTA-VirtualMTA: pmta-vmta13 Received-From-MTA: dns;DESKTOP-TCRDU4C (46.148.114.172) Arrival-Date: Sat, 20 Aug 2022 04:05:51 -0500 Final-Recipient: rfc822;[signs@domain.com.au](mailto:signs@domain.com.au) Action: failed Status: 5.7.23 (SPF validation failed) Remote-MTA: dns;mail.domain.com.au (myipv4) Diagnostic-Code: smtp;550 5.7.23 <[signs@domain.com.au](mailto:signs@domain.com.au)>: Recipient address rejected: Message rejected due to: SPF fail - not authorized. X-PowerMTA-BounceCategory: policy-related MIME-Version: 1.0 From: "Support" <[signs@domain.com.au](mailto:signs@domain.com.au)> To: [signs@domain.com.au](mailto:signs@domain.com.au) Priority: urgent Importance: high Date: 20 Aug 2022 10:05:51 +0100 Subject: Your membership Cancellation ! Content-Type: multipart/alternative; boundary=--boundary_3515_a99df6ff-539c-4cca-8aac-02dbffb86061 ``` and here are the headers: ``` Return-Path: <> Delivered-To: signs@domain.com.au Received: from mail.domain.services by mail.domain.services with LMTP id I7N/OfujAGNtFgAAbn96Gg (envelope-from <>) for <signs@domain.com.au>; Sat, 20 Aug 2022 19:06:03 +1000 Received: from localhost (Discovery [127.0.0.1]) by mail.domain.services (Postfix) with ESMTP id E48B932000005577FCA for <signs@domain.com.au>; Sat, 20 Aug 2022 19:06:03 +1000 (AEST) Received-SPF: None (no SPF record) identity=helo; client-ip=104.223.32.12; helo=packages13.dhl.com; envelope-from=<>; receiver=<UNKNOWN> Authentication-Results: mail.domain.services; dmarc=none (p=none dis=none) header.from=packages13.dhl.com Received: from packages13.dhl.com (unknown [104.223.32.12]) by mail.domain.services (Postfix) with ESMTPS id 115D732000005577FC0 for <signs@domain.com.au>; Sat, 20 Aug 2022 19:06:00 +1000 (AEST) Date: Sat, 20 Aug 2022 04:06:01 -0500 From: postmaster@packages13.dhl.com Subject: Delivery report To: signs@domain.com.au MIME-Version: 1.0 Content-Type: multipart/report; report-type=delivery-status; boundary="report6300A3F9@packages13.dhl.com" Message-Id: <20220820090603.E48B932000005577FCA@mail.domain.services> ``` from my vague understanding, someone has tried to send myself a spam email so it looks official, however, they must have been authenticated in order to get this far? is this something to worry about?
1.0
Should i be worried about this email? - # Subject <!-- Select one, remove the others --> Other ## Description <!-- When copy/pasting code, format the code with tripe backticks (`) ! --> This is off-topic, however, thought someone would be nice enough to look at this :) I've received a spam bounceback, but it originates from my own email (my domain has been replaced with `domain`) Heres the body of the email: ``` Hello, this is the mail server on packages13.dhl.com. I am sending you this message to inform you on the delivery status of a message you previously sent. Immediately below you will find a list of the affected recipients; also attached is a Delivery Status Notification (DSN) report in standard format, as well as the headers of the original message. <[signs@domain.com.au](mailto:signs@domain.com.au)> delivery failed; will not continue trying Reporting-MTA: dns;packages13.dhl.com X-PowerMTA-VirtualMTA: pmta-vmta13 Received-From-MTA: dns;DESKTOP-TCRDU4C (46.148.114.172) Arrival-Date: Sat, 20 Aug 2022 04:05:51 -0500 Final-Recipient: rfc822;[signs@domain.com.au](mailto:signs@domain.com.au) Action: failed Status: 5.7.23 (SPF validation failed) Remote-MTA: dns;mail.domain.com.au (myipv4) Diagnostic-Code: smtp;550 5.7.23 <[signs@domain.com.au](mailto:signs@domain.com.au)>: Recipient address rejected: Message rejected due to: SPF fail - not authorized. X-PowerMTA-BounceCategory: policy-related MIME-Version: 1.0 From: "Support" <[signs@domain.com.au](mailto:signs@domain.com.au)> To: [signs@domain.com.au](mailto:signs@domain.com.au) Priority: urgent Importance: high Date: 20 Aug 2022 10:05:51 +0100 Subject: Your membership Cancellation ! Content-Type: multipart/alternative; boundary=--boundary_3515_a99df6ff-539c-4cca-8aac-02dbffb86061 ``` and here are the headers: ``` Return-Path: <> Delivered-To: signs@domain.com.au Received: from mail.domain.services by mail.domain.services with LMTP id I7N/OfujAGNtFgAAbn96Gg (envelope-from <>) for <signs@domain.com.au>; Sat, 20 Aug 2022 19:06:03 +1000 Received: from localhost (Discovery [127.0.0.1]) by mail.domain.services (Postfix) with ESMTP id E48B932000005577FCA for <signs@domain.com.au>; Sat, 20 Aug 2022 19:06:03 +1000 (AEST) Received-SPF: None (no SPF record) identity=helo; client-ip=104.223.32.12; helo=packages13.dhl.com; envelope-from=<>; receiver=<UNKNOWN> Authentication-Results: mail.domain.services; dmarc=none (p=none dis=none) header.from=packages13.dhl.com Received: from packages13.dhl.com (unknown [104.223.32.12]) by mail.domain.services (Postfix) with ESMTPS id 115D732000005577FC0 for <signs@domain.com.au>; Sat, 20 Aug 2022 19:06:00 +1000 (AEST) Date: Sat, 20 Aug 2022 04:06:01 -0500 From: postmaster@packages13.dhl.com Subject: Delivery report To: signs@domain.com.au MIME-Version: 1.0 Content-Type: multipart/report; report-type=delivery-status; boundary="report6300A3F9@packages13.dhl.com" Message-Id: <20220820090603.E48B932000005577FCA@mail.domain.services> ``` from my vague understanding, someone has tried to send myself a spam email so it looks official, however, they must have been authenticated in order to get this far? is this something to worry about?
non_build
should i be worried about this email subject other description this is off topic however thought someone would be nice enough to look at this i ve received a spam bounceback but it originates from my own email my domain has been replaced with domain heres the body of the email hello this is the mail server on dhl com i am sending you this message to inform you on the delivery status of a message you previously sent immediately below you will find a list of the affected recipients also attached is a delivery status notification dsn report in standard format as well as the headers of the original message delivery failed will not continue trying reporting mta dns dhl com x powermta virtualmta pmta received from mta dns desktop arrival date sat aug final recipient mailto signs domain com au action failed status spf validation failed remote mta dns mail domain com au diagnostic code smtp recipient address rejected message rejected due to spf fail not authorized x powermta bouncecategory policy related mime version from support to mailto signs domain com au priority urgent importance high date aug subject your membership cancellation content type multipart alternative boundary boundary and here are the headers return path delivered to signs domain com au received from mail domain services by mail domain services with lmtp id envelope from for sat aug received from localhost discovery by mail domain services postfix with esmtp id for sat aug aest received spf none no spf record identity helo client ip helo dhl com envelope from receiver authentication results mail domain services dmarc none p none dis none header from dhl com received from dhl com unknown by mail domain services postfix with esmtps id for sat aug aest date sat aug from postmaster dhl com subject delivery report to signs domain com au mime version content type multipart report report type delivery status boundary dhl com message id from my vague understanding someone has tried to send myself a spam email so it looks official however they must have been authenticated in order to get this far is this something to worry about
0
9,043
7,781,970,740
IssuesEvent
2018-06-06 03:36:07
glauth/glauth
https://api.github.com/repos/glauth/glauth
closed
Setup Auto Deploys to Docker Hub
enhancement in progress infrastructure
_From @benyanke on April 19, 2018 0:45_ Docker should be a first class deployment tool. Need to setup docker tags to properly build on each successful build. As far as tags, stable/master goes to 'latest', versions go to 'vx.x', and 'vx' tags, and dev branch goes to 'nightly'. _Copied from original issue: nmcclain/glauth#11_
1.0
Setup Auto Deploys to Docker Hub - _From @benyanke on April 19, 2018 0:45_ Docker should be a first class deployment tool. Need to setup docker tags to properly build on each successful build. As far as tags, stable/master goes to 'latest', versions go to 'vx.x', and 'vx' tags, and dev branch goes to 'nightly'. _Copied from original issue: nmcclain/glauth#11_
non_build
setup auto deploys to docker hub from benyanke on april docker should be a first class deployment tool need to setup docker tags to properly build on each successful build as far as tags stable master goes to latest versions go to vx x and vx tags and dev branch goes to nightly copied from original issue nmcclain glauth
0
270,319
8,454,622,825
IssuesEvent
2018-10-21 05:49:55
EUCweb/BIS-F
https://api.github.com/repos/EUCweb/BIS-F
closed
XA/ XD 7.x Cache folder will be created
Priority: Low Status: In Progress Type: Bug
_From @matthias-schimm on March 12, 2018 20:20_ IF XA/XD 7.x is installed, the Cache folder will be created in PVSWriteCacheDisk\Citrix\Cache or C:\Windows\Logs\ _Copied from original issue: EUCweb/BIS-F#10_
1.0
XA/ XD 7.x Cache folder will be created - _From @matthias-schimm on March 12, 2018 20:20_ IF XA/XD 7.x is installed, the Cache folder will be created in PVSWriteCacheDisk\Citrix\Cache or C:\Windows\Logs\ _Copied from original issue: EUCweb/BIS-F#10_
non_build
xa xd x cache folder will be created from matthias schimm on march if xa xd x is installed the cache folder will be created in pvswritecachedisk citrix cache or c windows logs copied from original issue eucweb bis f
0
87,830
25,228,488,135
IssuesEvent
2022-11-14 17:45:19
adoptium/ci-jenkins-pipelines
https://api.github.com/repos/adoptium/ci-jenkins-pipelines
closed
Add support to have configurable tag when pulling down jenkins-helper repo
enhancement reproduciblebuild
https://github.com/adoptium/jenkins-helper is not often changed, but in order to make reproducible build, we need to have either tag or SHA1 when calling shared library Currently, we have code like ` library(identifier: 'openjdk-jenkins-helper@master')` or ` def JobHelper = context.library(identifier: 'openjdk-jenkins-helper@master').JobHelper` hardcoded in ci-jenkins-pipeline. since Jenkins plugin of library only support branch or tag not SHA1, we need to make tag on jenkins-helper and add support in ci-jenkins-pipeline to be able to config this tag but also support the normal case (use master branch as default) Related to https://github.com/adoptium/temurin-build/issues/3103
1.0
Add support to have configurable tag when pulling down jenkins-helper repo - https://github.com/adoptium/jenkins-helper is not often changed, but in order to make reproducible build, we need to have either tag or SHA1 when calling shared library Currently, we have code like ` library(identifier: 'openjdk-jenkins-helper@master')` or ` def JobHelper = context.library(identifier: 'openjdk-jenkins-helper@master').JobHelper` hardcoded in ci-jenkins-pipeline. since Jenkins plugin of library only support branch or tag not SHA1, we need to make tag on jenkins-helper and add support in ci-jenkins-pipeline to be able to config this tag but also support the normal case (use master branch as default) Related to https://github.com/adoptium/temurin-build/issues/3103
build
add support to have configurable tag when pulling down jenkins helper repo is not often changed but in order to make reproducible build we need to have either tag or when calling shared library currently we have code like library identifier openjdk jenkins helper master or def jobhelper context library identifier openjdk jenkins helper master jobhelper hardcoded in ci jenkins pipeline since jenkins plugin of library only support branch or tag not we need to make tag on jenkins helper and add support in ci jenkins pipeline to be able to config this tag but also support the normal case use master branch as default related to
1
98,562
30,004,265,107
IssuesEvent
2023-06-26 11:22:29
ballerina-platform/ballerina-standard-library
https://api.github.com/repos/ballerina-platform/ballerina-standard-library
closed
Update the Release Process Doc
Type/Improvement Area/Build Area/Docs
**Description:** Release process doc has to be updated with the information about how to handle breaking changes from the lang side.
1.0
Update the Release Process Doc - **Description:** Release process doc has to be updated with the information about how to handle breaking changes from the lang side.
build
update the release process doc description release process doc has to be updated with the information about how to handle breaking changes from the lang side
1
388,710
26,778,977,737
IssuesEvent
2023-01-31 19:27:59
lameRER/GeekBrains
https://api.github.com/repos/lameRER/GeekBrains
opened
Java-HomeWork-4
documentation
Урок 4. Хранение и обработка данных ч1: приоритетные коллекции Основной зал - задача 2. Задача 3 по желанию. Зал1,Зал3 - Задача 3 и HomeWork Зал2 - Задача 2 и Задача3. HomeWork - по желанию Кто не был на семинаре - Задача 1, Задача 2 - остальное по желанию. Задачи дублирую --> public class Task1 { // Дан Deque состоящий из последовательности цифр. // Необходимо проверить, что последовательность цифр является палиндромом public static void main(String[] args) { Deque<Integer> deque = new ArrayDeque<>(Arrays.asList(1,2,3,4,5,6)); } public boolean checkOn(Deque<Integer> deque){ return false; } } public class Task2 { //Даны два Deque представляющие два неотрицательных целых числа. Цифры хранятся в обратном порядке, // и каждый из их узлов содержит одну цифру. // Сложите два числа и верните сумму в виде связанного списка. public static void main(String[] args) { Deque<Integer> d1 = new ArrayDeque<>(Arrays.asList(1,2,3)); Deque<Integer> d2 = new ArrayDeque<>(Arrays.asList(5,4,7)); // result [6,6,0,1] } public Deque<Integer> sum(Deque<Integer> d1, Deque<Integer> d2) { return new ArrayDeque<>(); } } public class Task3 { //Дана строка содержащая только символы '(', ')', '{', '}', '[' и ']', определите, // является ли входная строка логически правильной. // Входная строка логически правильная, если: // 1) Открытые скобки должны быть закрыты скобками того же типа. // 2) Открытые скобки должны быть закрыты в правильном порядке. Каждая закрывающая скобка имеет соответствующую // открытую скобку того же типа. // ()[] = true // () = true // {[()]} = true // ()() = true // )()( = false public static void main(String[] args) { } public boolean validate(Deque<Integer> deque){ return false; } } public class Homework { //Даны два Deque представляющие два целых числа. Цифры хранятся в обратном порядке, // и каждый из их узлов содержит одну цифру. public static void main(String[] args) { Homework hw = new Homework(); hw.multiple(new ArrayDeque<>(Arrays.asList(5,2)), new ArrayDeque<>(Arrays.asList(4))); // result [0,0,1] hw.sum(new ArrayDeque<>(Arrays.asList(5,-2)), new ArrayDeque<>(Arrays.asList(5))); // result [0,-2] } // Умножьте два числа и верните произведение в виде связанного списка. public Deque<Integer> multiple(Deque<Integer> d1, Deque<Integer> d2){ return new ArrayDeque<>(); } // Сложите два числа и верните сумму в виде связанного списка. Одно или два числа должны быть отрицательными public Deque<Integer> sum(Deque<Integer> d1, Deque<Integer> d2){ return new ArrayDeque<>(); } }
1.0
Java-HomeWork-4 - Урок 4. Хранение и обработка данных ч1: приоритетные коллекции Основной зал - задача 2. Задача 3 по желанию. Зал1,Зал3 - Задача 3 и HomeWork Зал2 - Задача 2 и Задача3. HomeWork - по желанию Кто не был на семинаре - Задача 1, Задача 2 - остальное по желанию. Задачи дублирую --> public class Task1 { // Дан Deque состоящий из последовательности цифр. // Необходимо проверить, что последовательность цифр является палиндромом public static void main(String[] args) { Deque<Integer> deque = new ArrayDeque<>(Arrays.asList(1,2,3,4,5,6)); } public boolean checkOn(Deque<Integer> deque){ return false; } } public class Task2 { //Даны два Deque представляющие два неотрицательных целых числа. Цифры хранятся в обратном порядке, // и каждый из их узлов содержит одну цифру. // Сложите два числа и верните сумму в виде связанного списка. public static void main(String[] args) { Deque<Integer> d1 = new ArrayDeque<>(Arrays.asList(1,2,3)); Deque<Integer> d2 = new ArrayDeque<>(Arrays.asList(5,4,7)); // result [6,6,0,1] } public Deque<Integer> sum(Deque<Integer> d1, Deque<Integer> d2) { return new ArrayDeque<>(); } } public class Task3 { //Дана строка содержащая только символы '(', ')', '{', '}', '[' и ']', определите, // является ли входная строка логически правильной. // Входная строка логически правильная, если: // 1) Открытые скобки должны быть закрыты скобками того же типа. // 2) Открытые скобки должны быть закрыты в правильном порядке. Каждая закрывающая скобка имеет соответствующую // открытую скобку того же типа. // ()[] = true // () = true // {[()]} = true // ()() = true // )()( = false public static void main(String[] args) { } public boolean validate(Deque<Integer> deque){ return false; } } public class Homework { //Даны два Deque представляющие два целых числа. Цифры хранятся в обратном порядке, // и каждый из их узлов содержит одну цифру. public static void main(String[] args) { Homework hw = new Homework(); hw.multiple(new ArrayDeque<>(Arrays.asList(5,2)), new ArrayDeque<>(Arrays.asList(4))); // result [0,0,1] hw.sum(new ArrayDeque<>(Arrays.asList(5,-2)), new ArrayDeque<>(Arrays.asList(5))); // result [0,-2] } // Умножьте два числа и верните произведение в виде связанного списка. public Deque<Integer> multiple(Deque<Integer> d1, Deque<Integer> d2){ return new ArrayDeque<>(); } // Сложите два числа и верните сумму в виде связанного списка. Одно или два числа должны быть отрицательными public Deque<Integer> sum(Deque<Integer> d1, Deque<Integer> d2){ return new ArrayDeque<>(); } }
non_build
java homework урок хранение и обработка данных приоритетные коллекции основной зал задача задача по желанию задача и homework задача и homework по желанию кто не был на семинаре задача задача остальное по желанию задачи дублирую public class дан deque состоящий из последовательности цифр необходимо проверить что последовательность цифр является палиндромом public static void main string args deque deque new arraydeque arrays aslist public boolean checkon deque deque return false public class даны два deque представляющие два неотрицательных целых числа цифры хранятся в обратном порядке и каждый из их узлов содержит одну цифру сложите два числа и верните сумму в виде связанного списка public static void main string args deque new arraydeque arrays aslist deque new arraydeque arrays aslist result public deque sum deque deque return new arraydeque public class дана строка содержащая только символы определите является ли входная строка логически правильной входная строка логически правильная если открытые скобки должны быть закрыты скобками того же типа открытые скобки должны быть закрыты в правильном порядке каждая закрывающая скобка имеет соответствующую открытую скобку того же типа true true true true false public static void main string args public boolean validate deque deque return false public class homework даны два deque представляющие два целых числа цифры хранятся в обратном порядке и каждый из их узлов содержит одну цифру public static void main string args homework hw new homework hw multiple new arraydeque arrays aslist new arraydeque arrays aslist result hw sum new arraydeque arrays aslist new arraydeque arrays aslist result умножьте два числа и верните произведение в виде связанного списка public deque multiple deque deque return new arraydeque сложите два числа и верните сумму в виде связанного списка одно или два числа должны быть отрицательными public deque sum deque deque return new arraydeque
0
163,895
25,893,398,426
IssuesEvent
2022-12-14 20:01:37
Expensify/App
https://api.github.com/repos/Expensify/App
opened
[Core Branding] Update default bank account icon
Design
We recently updated our icons for the core branding project, but one icon that slipped my mind was the default icon we use for bank accounts. <img width="380" alt="image" src="https://user-images.githubusercontent.com/2319350/207701564-12250847-619c-4fd5-b304-5ea94946c9f5.png"> Let's update that with a new one like this: <img width="246" alt="image" src="https://user-images.githubusercontent.com/2319350/207701935-ea4eb975-097f-40f6-b8c3-2759b7d28646.png"> File is here: [bank-account.svg.zip](https://github.com/Expensify/App/files/10231232/bank-account.svg.zip) cc @stitesExpensify - any reason why a proper bank logo (Citi or SVB Private in my case) isn't being populated here? I vaguely recall you working on this in the past cc @grgia @Luke9389 in case one of you wants to tackle this one quickly!
1.0
[Core Branding] Update default bank account icon - We recently updated our icons for the core branding project, but one icon that slipped my mind was the default icon we use for bank accounts. <img width="380" alt="image" src="https://user-images.githubusercontent.com/2319350/207701564-12250847-619c-4fd5-b304-5ea94946c9f5.png"> Let's update that with a new one like this: <img width="246" alt="image" src="https://user-images.githubusercontent.com/2319350/207701935-ea4eb975-097f-40f6-b8c3-2759b7d28646.png"> File is here: [bank-account.svg.zip](https://github.com/Expensify/App/files/10231232/bank-account.svg.zip) cc @stitesExpensify - any reason why a proper bank logo (Citi or SVB Private in my case) isn't being populated here? I vaguely recall you working on this in the past cc @grgia @Luke9389 in case one of you wants to tackle this one quickly!
non_build
update default bank account icon we recently updated our icons for the core branding project but one icon that slipped my mind was the default icon we use for bank accounts img width alt image src let s update that with a new one like this img width alt image src file is here cc stitesexpensify any reason why a proper bank logo citi or svb private in my case isn t being populated here i vaguely recall you working on this in the past cc grgia in case one of you wants to tackle this one quickly
0
45,919
11,758,786,006
IssuesEvent
2020-03-13 16:01:47
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Official builds losing connection with machine pool
area-Infrastructure-libraries blocking-official-build untriaged
The error will be in the following form: > We stopped hearing from agent NetCoreInternal-Pool 19. FR issue https://github.com/dotnet/core-eng/issues/9300
1.0
Official builds losing connection with machine pool - The error will be in the following form: > We stopped hearing from agent NetCoreInternal-Pool 19. FR issue https://github.com/dotnet/core-eng/issues/9300
build
official builds losing connection with machine pool the error will be in the following form we stopped hearing from agent netcoreinternal pool fr issue
1
91,700
26,466,872,137
IssuesEvent
2023-01-17 01:13:47
orbeon/orbeon-forms
https://api.github.com/repos/orbeon/orbeon-forms
opened
Configurable date format per field/form
Module: Form Builder Type: RFE Area: XBL Components
Following #5524/#5621, we should do something similar for date fields. Here are the aspects of date format we need to be able to control, based on the formats we currently support: - month first or day first - separator: `/`, `.`, `-` - ability to pad the month or day to 2 digits We currently don't support (I think): - putting the year first (like for the ISO format) - probably desirable - setting a padding/rounding of the year - might not be desirable
1.0
Configurable date format per field/form - Following #5524/#5621, we should do something similar for date fields. Here are the aspects of date format we need to be able to control, based on the formats we currently support: - month first or day first - separator: `/`, `.`, `-` - ability to pad the month or day to 2 digits We currently don't support (I think): - putting the year first (like for the ISO format) - probably desirable - setting a padding/rounding of the year - might not be desirable
build
configurable date format per field form following we should do something similar for date fields here are the aspects of date format we need to be able to control based on the formats we currently support month first or day first separator ability to pad the month or day to digits we currently don t support i think putting the year first like for the iso format probably desirable setting a padding rounding of the year might not be desirable
1
21,725
7,060,527,446
IssuesEvent
2018-01-05 09:12:20
armbian/build
https://api.github.com/repos/armbian/build
closed
Build script won't run on "unsupported" operating system
build scripts question
I've tried to compile Armbian on my Deepin machine, and the script complained about it not being supported, and the build script didn't start. How about instead of stopping the script, display a big ol' message like "You are running a unsupported host system, this is not recommended or supported by us, your mileage may vary. If you encounter any bugs/something doesn't work, do not report it." A quick change in lib/general.sh to make it display a warning instead of exit with error made the script work just fine for me. I could submit it as a pull request, or perhaps, you don't like that kind of solution, a mechanism to detect a debian-based distro could be implemented?
1.0
Build script won't run on "unsupported" operating system - I've tried to compile Armbian on my Deepin machine, and the script complained about it not being supported, and the build script didn't start. How about instead of stopping the script, display a big ol' message like "You are running a unsupported host system, this is not recommended or supported by us, your mileage may vary. If you encounter any bugs/something doesn't work, do not report it." A quick change in lib/general.sh to make it display a warning instead of exit with error made the script work just fine for me. I could submit it as a pull request, or perhaps, you don't like that kind of solution, a mechanism to detect a debian-based distro could be implemented?
build
build script won t run on unsupported operating system i ve tried to compile armbian on my deepin machine and the script complained about it not being supported and the build script didn t start how about instead of stopping the script display a big ol message like you are running a unsupported host system this is not recommended or supported by us your mileage may vary if you encounter any bugs something doesn t work do not report it a quick change in lib general sh to make it display a warning instead of exit with error made the script work just fine for me i could submit it as a pull request or perhaps you don t like that kind of solution a mechanism to detect a debian based distro could be implemented
1
327,616
24,145,058,366
IssuesEvent
2022-09-21 17:58:22
fleetdm/fleet
https://api.github.com/repos/fleetdm/fleet
closed
Document how Fleet determines online/offline
:improve documentation
Soon, the Fleet UI will define online hosts as hosts that will respond to a live query. The Fleet UI will define offline hosts as hosts that will not respond to a live query because they may be turned off or not connected to the internet. ### Goal Add documentation that explains how Fleet determines online and offline status. ### How? - [ ] Proposed solution is to add a section to the [01-Using-Fleet/FAQ.md](https://github.com/fleetdm/fleet/blob/main/docs/01-Using-Fleet/FAQ.md) PR with the doc changes: https://github.com/fleetdm/fleet/pull/7706
1.0
Document how Fleet determines online/offline - Soon, the Fleet UI will define online hosts as hosts that will respond to a live query. The Fleet UI will define offline hosts as hosts that will not respond to a live query because they may be turned off or not connected to the internet. ### Goal Add documentation that explains how Fleet determines online and offline status. ### How? - [ ] Proposed solution is to add a section to the [01-Using-Fleet/FAQ.md](https://github.com/fleetdm/fleet/blob/main/docs/01-Using-Fleet/FAQ.md) PR with the doc changes: https://github.com/fleetdm/fleet/pull/7706
non_build
document how fleet determines online offline soon the fleet ui will define online hosts as hosts that will respond to a live query the fleet ui will define offline hosts as hosts that will not respond to a live query because they may be turned off or not connected to the internet goal add documentation that explains how fleet determines online and offline status how proposed solution is to add a section to the pr with the doc changes
0
191,115
14,593,289,639
IssuesEvent
2020-12-19 21:58:32
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
mhausenblas/burry.sh: vendor/golang.org/x/crypto/ssh/kex_test.go; 29 LoC
fresh small test vendored
Found a possible issue in [mhausenblas/burry.sh](https://www.github.com/mhausenblas/burry.sh) at [vendor/golang.org/x/crypto/ssh/kex_test.go](https://github.com/mhausenblas/burry.sh/blob/18268370d8aaf62c62bc83eb5596b6736fa640e8/vendor/golang.org/x/crypto/ssh/kex_test.go#L21-L49) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable kex used in defer or goroutine at line 28 [Click here to see the code in its original context.](https://github.com/mhausenblas/burry.sh/blob/18268370d8aaf62c62bc83eb5596b6736fa640e8/vendor/golang.org/x/crypto/ssh/kex_test.go#L21-L49) <details> <summary>Click here to show the 29 line(s) of Go which triggered the analyzer.</summary> ```go for name, kex := range kexAlgoMap { a, b := memPipe() s := make(chan kexResultErr, 1) c := make(chan kexResultErr, 1) var magics handshakeMagics go func() { r, e := kex.Client(a, rand.Reader, &magics) a.Close() c <- kexResultErr{r, e} }() go func() { r, e := kex.Server(b, rand.Reader, &magics, testSigners["ecdsa"]) b.Close() s <- kexResultErr{r, e} }() clientRes := <-c serverRes := <-s if clientRes.err != nil { t.Errorf("client: %v", clientRes.err) } if serverRes.err != nil { t.Errorf("server: %v", serverRes.err) } if !reflect.DeepEqual(clientRes.result, serverRes.result) { t.Errorf("kex %q: mismatch %#v, %#v", name, clientRes.result, serverRes.result) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 18268370d8aaf62c62bc83eb5596b6736fa640e8
1.0
mhausenblas/burry.sh: vendor/golang.org/x/crypto/ssh/kex_test.go; 29 LoC - Found a possible issue in [mhausenblas/burry.sh](https://www.github.com/mhausenblas/burry.sh) at [vendor/golang.org/x/crypto/ssh/kex_test.go](https://github.com/mhausenblas/burry.sh/blob/18268370d8aaf62c62bc83eb5596b6736fa640e8/vendor/golang.org/x/crypto/ssh/kex_test.go#L21-L49) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable kex used in defer or goroutine at line 28 [Click here to see the code in its original context.](https://github.com/mhausenblas/burry.sh/blob/18268370d8aaf62c62bc83eb5596b6736fa640e8/vendor/golang.org/x/crypto/ssh/kex_test.go#L21-L49) <details> <summary>Click here to show the 29 line(s) of Go which triggered the analyzer.</summary> ```go for name, kex := range kexAlgoMap { a, b := memPipe() s := make(chan kexResultErr, 1) c := make(chan kexResultErr, 1) var magics handshakeMagics go func() { r, e := kex.Client(a, rand.Reader, &magics) a.Close() c <- kexResultErr{r, e} }() go func() { r, e := kex.Server(b, rand.Reader, &magics, testSigners["ecdsa"]) b.Close() s <- kexResultErr{r, e} }() clientRes := <-c serverRes := <-s if clientRes.err != nil { t.Errorf("client: %v", clientRes.err) } if serverRes.err != nil { t.Errorf("server: %v", serverRes.err) } if !reflect.DeepEqual(clientRes.result, serverRes.result) { t.Errorf("kex %q: mismatch %#v, %#v", name, clientRes.result, serverRes.result) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 18268370d8aaf62c62bc83eb5596b6736fa640e8
non_build
mhausenblas burry sh vendor golang org x crypto ssh kex test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable kex used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for name kex range kexalgomap a b mempipe s make chan kexresulterr c make chan kexresulterr var magics handshakemagics go func r e kex client a rand reader magics a close c kexresulterr r e go func r e kex server b rand reader magics testsigners b close s kexresulterr r e clientres c serverres s if clientres err nil t errorf client v clientres err if serverres err nil t errorf server v serverres err if reflect deepequal clientres result serverres result t errorf kex q mismatch v v name clientres result serverres result leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
436,113
30,538,023,268
IssuesEvent
2023-07-19 18:55:48
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
GitHub Access Changes
platform-content-team documentation-support pw-footer-feedback
### Description It has been shared with me that some teams have trouble using our GitHub access information when teams are moving into the VA.gov space for the first time. These teams do not have a program manager who has access to GitHub so they cannot create a request for access because those managers don't have a GH account either. We need additional information that details what to do when our step 1 isn't possible. The GitHub Handbook provides the best information for this case, so I have linked it below. The GitHub handbook is the single source of truth on this, so it may be best just to link there. ### Relevant URLs [Platform Website link ](https://depo-platform-documentation.scrollhelp.site/getting-started/request-access-to-tools#Requestaccesstotools-GitHub)[GitHub Handbook ](https://department-of-veterans-affairs.github.io/github-handbook/guides/onboarding/getting-access) ### Which type of team are you on? (Platform team, VFS team, or Leadership) VA Leadership
1.0
GitHub Access Changes - ### Description It has been shared with me that some teams have trouble using our GitHub access information when teams are moving into the VA.gov space for the first time. These teams do not have a program manager who has access to GitHub so they cannot create a request for access because those managers don't have a GH account either. We need additional information that details what to do when our step 1 isn't possible. The GitHub Handbook provides the best information for this case, so I have linked it below. The GitHub handbook is the single source of truth on this, so it may be best just to link there. ### Relevant URLs [Platform Website link ](https://depo-platform-documentation.scrollhelp.site/getting-started/request-access-to-tools#Requestaccesstotools-GitHub)[GitHub Handbook ](https://department-of-veterans-affairs.github.io/github-handbook/guides/onboarding/getting-access) ### Which type of team are you on? (Platform team, VFS team, or Leadership) VA Leadership
non_build
github access changes description it has been shared with me that some teams have trouble using our github access information when teams are moving into the va gov space for the first time these teams do not have a program manager who has access to github so they cannot create a request for access because those managers don t have a gh account either we need additional information that details what to do when our step isn t possible the github handbook provides the best information for this case so i have linked it below the github handbook is the single source of truth on this so it may be best just to link there relevant urls platform website link handbook which type of team are you on platform team vfs team or leadership va leadership
0
66,236
16,568,730,282
IssuesEvent
2021-05-30 00:41:44
enslit/bbbs-team-6
https://api.github.com/repos/enslit/bbbs-team-6
closed
Защищенный роут
build enhancement
Реализовать защищенный роут. При переходе на этот роут, должен показываться попап с формой авторизации. После успешной авторизации нужно перейти на запрашиваемую страницу
1.0
Защищенный роут - Реализовать защищенный роут. При переходе на этот роут, должен показываться попап с формой авторизации. После успешной авторизации нужно перейти на запрашиваемую страницу
build
защищенный роут реализовать защищенный роут при переходе на этот роут должен показываться попап с формой авторизации после успешной авторизации нужно перейти на запрашиваемую страницу
1
540,149
15,801,614,964
IssuesEvent
2021-04-03 05:46:04
PyCQA/flake8
https://api.github.com/repos/PyCQA/flake8
closed
support for "local plugins" (configured explicitly instead of via entry points)
feature:accepted priority:medium
In GitLab by @carljm on Jul 31, 2017, 14:45 We have a number of repo-specific lint rules we'd like to apply. It's important that engineers working in the repo can add or update lint rules with low friction, so we want the plugin code in the same repo that is being checked. It's possible to achieve this by adding a setup.py to a flake8 plugin subdirectory and modifying requirements files to separately editable-install it, but it's a bit unpleasant: generates egg-info metadata that has to be committed or ignored, takes extra time on pip install because editable installs are always reinstalled, requires a full directory with at least three files in it, where otherwise you'd probably just need a single module. It would be really nice to have a way to configure a plugin purely in `setup.cfg`, without having to involve setuptools at all. Pytest is an example of a project that uses setuptools entry points, but also provides a way to configure local plugins without the use of setuptools.
1.0
support for "local plugins" (configured explicitly instead of via entry points) - In GitLab by @carljm on Jul 31, 2017, 14:45 We have a number of repo-specific lint rules we'd like to apply. It's important that engineers working in the repo can add or update lint rules with low friction, so we want the plugin code in the same repo that is being checked. It's possible to achieve this by adding a setup.py to a flake8 plugin subdirectory and modifying requirements files to separately editable-install it, but it's a bit unpleasant: generates egg-info metadata that has to be committed or ignored, takes extra time on pip install because editable installs are always reinstalled, requires a full directory with at least three files in it, where otherwise you'd probably just need a single module. It would be really nice to have a way to configure a plugin purely in `setup.cfg`, without having to involve setuptools at all. Pytest is an example of a project that uses setuptools entry points, but also provides a way to configure local plugins without the use of setuptools.
non_build
support for local plugins configured explicitly instead of via entry points in gitlab by carljm on jul we have a number of repo specific lint rules we d like to apply it s important that engineers working in the repo can add or update lint rules with low friction so we want the plugin code in the same repo that is being checked it s possible to achieve this by adding a setup py to a plugin subdirectory and modifying requirements files to separately editable install it but it s a bit unpleasant generates egg info metadata that has to be committed or ignored takes extra time on pip install because editable installs are always reinstalled requires a full directory with at least three files in it where otherwise you d probably just need a single module it would be really nice to have a way to configure a plugin purely in setup cfg without having to involve setuptools at all pytest is an example of a project that uses setuptools entry points but also provides a way to configure local plugins without the use of setuptools
0
130,220
27,633,140,762
IssuesEvent
2023-03-10 12:29:38
Regalis11/Barotrauma
https://api.github.com/repos/Regalis11/Barotrauma
closed
Overlapping map markers make both markers text unreadable
Bug Code Low prio
- [x] I have searched the issue tracker to check if the issue has already been reported. **Description** when multiple map markers are at the same angle, the text overlaps, resulting in NONE of the markers texts being readable **Steps To Reproduce** have 2 map markers in approximately the same direction, so that their text boxes overlap on the nav terminal **Version** 0.15 **Additional information** it's been going on for multiple versions maybe fix is to increase opaqness?
1.0
Overlapping map markers make both markers text unreadable - - [x] I have searched the issue tracker to check if the issue has already been reported. **Description** when multiple map markers are at the same angle, the text overlaps, resulting in NONE of the markers texts being readable **Steps To Reproduce** have 2 map markers in approximately the same direction, so that their text boxes overlap on the nav terminal **Version** 0.15 **Additional information** it's been going on for multiple versions maybe fix is to increase opaqness?
non_build
overlapping map markers make both markers text unreadable i have searched the issue tracker to check if the issue has already been reported description when multiple map markers are at the same angle the text overlaps resulting in none of the markers texts being readable steps to reproduce have map markers in approximately the same direction so that their text boxes overlap on the nav terminal version additional information it s been going on for multiple versions maybe fix is to increase opaqness
0
140,496
32,008,153,191
IssuesEvent
2023-09-21 16:04:40
HMIS/LSASampleCode
https://api.github.com/repos/HMIS/LSASampleCode
closed
Sample Code Step 4.5: issue with [Availability] and [ESBedType]
Sample code
Hi @MollyMcEvilley , I the spec (both LSA and CSV Export) [Availability] and [ESBedType] columns are mandatory for ProjectType = (0, 1): ![image](https://github.com/HMIS/LSASampleCode/assets/28144855/6e628847-1632-4236-97b0-44a2a225163d) In the Sample Code [Availability] and [ESBedType] columns are selected for ProjectType = 1 only: ![image](https://github.com/HMIS/LSASampleCode/assets/28144855/7cd62e27-d276-41c3-b9f9-9d13b088034f) Could you please take a look? Thank you! Natalie
1.0
Sample Code Step 4.5: issue with [Availability] and [ESBedType] - Hi @MollyMcEvilley , I the spec (both LSA and CSV Export) [Availability] and [ESBedType] columns are mandatory for ProjectType = (0, 1): ![image](https://github.com/HMIS/LSASampleCode/assets/28144855/6e628847-1632-4236-97b0-44a2a225163d) In the Sample Code [Availability] and [ESBedType] columns are selected for ProjectType = 1 only: ![image](https://github.com/HMIS/LSASampleCode/assets/28144855/7cd62e27-d276-41c3-b9f9-9d13b088034f) Could you please take a look? Thank you! Natalie
non_build
sample code step issue with and hi mollymcevilley i the spec both lsa and csv export and columns are mandatory for projecttype in the sample code and columns are selected for projecttype only could you please take a look thank you natalie
0
92,885
26,794,563,521
IssuesEvent
2023-02-01 10:55:09
r5py/r5py
https://api.github.com/repos/r5py/r5py
closed
black 23.1.0 introduces breaking changes
bug build system
On the updated runners that use the python:3 docker image, our code does not pass linting, as black updated its style for 2023 last night. This has not yet arrived in conda, but is in pip (so the GitHub Action runners check for the new style) see https://github.com/psf/black/releases/tag/23.1.0 this is the one that affects our code: https://github.com/psf/black/pull/2945 , see https://github.com/r5py/r5py/actions/runs/4062655304/jobs/6994094010
1.0
black 23.1.0 introduces breaking changes - On the updated runners that use the python:3 docker image, our code does not pass linting, as black updated its style for 2023 last night. This has not yet arrived in conda, but is in pip (so the GitHub Action runners check for the new style) see https://github.com/psf/black/releases/tag/23.1.0 this is the one that affects our code: https://github.com/psf/black/pull/2945 , see https://github.com/r5py/r5py/actions/runs/4062655304/jobs/6994094010
build
black introduces breaking changes on the updated runners that use the python docker image our code does not pass linting as black updated its style for last night this has not yet arrived in conda but is in pip so the github action runners check for the new style see this is the one that affects our code see
1
176,411
21,411,032,133
IssuesEvent
2022-04-22 05:58:58
pazhanivel07/frameworks_base_Aosp10_r33
https://api.github.com/repos/pazhanivel07/frameworks_base_Aosp10_r33
opened
CVE-2021-0653 (Medium) detected in baseandroid-10.0.0_r46
security vulnerability
## CVE-2021-0653 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r46</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_base_Aosp10_r33/commit/d0a412c03562493a433dc7e698ff88ab06a3468a">d0a412c03562493a433dc7e698ff88ab06a3468a</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/core/java/com/android/server/net/NetworkPolicyManagerService.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In enqueueNotification of NetworkPolicyManagerService.java, there is a possible way to retrieve a trackable identifier due to a missing permission check. This could lead to local information disclosure with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10 Android-11 Android-9Android ID: A-177931370 <p>Publish Date: 2021-12-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0653>CVE-2021-0653</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2021-11-01">https://source.android.com/security/bulletin/2021-11-01</a></p> <p>Release Date: 2021-12-15</p> <p>Fix Resolution: android-11.0.0_r46</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-0653 (Medium) detected in baseandroid-10.0.0_r46 - ## CVE-2021-0653 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r46</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_base_Aosp10_r33/commit/d0a412c03562493a433dc7e698ff88ab06a3468a">d0a412c03562493a433dc7e698ff88ab06a3468a</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/core/java/com/android/server/net/NetworkPolicyManagerService.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In enqueueNotification of NetworkPolicyManagerService.java, there is a possible way to retrieve a trackable identifier due to a missing permission check. This could lead to local information disclosure with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10 Android-11 Android-9Android ID: A-177931370 <p>Publish Date: 2021-12-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0653>CVE-2021-0653</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2021-11-01">https://source.android.com/security/bulletin/2021-11-01</a></p> <p>Release Date: 2021-12-15</p> <p>Fix Resolution: android-11.0.0_r46</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_build
cve medium detected in baseandroid cve medium severity vulnerability vulnerable library baseandroid android framework classes and services library home page a href found in head commit a href found in base branch main vulnerable source files services core java com android server net networkpolicymanagerservice java vulnerability details in enqueuenotification of networkpolicymanagerservice java there is a possible way to retrieve a trackable identifier due to a missing permission check this could lead to local information disclosure with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with whitesource
0
472,296
13,622,438,624
IssuesEvent
2020-09-24 03:37:31
TerryCavanagh/diceydungeons.com
https://api.github.com/repos/TerryCavanagh/diceydungeons.com
closed
Gadget creation should let you look at your inventory and the map
High Priority reported in v0.7.1
When creating a gadget, you should be able to look at the rest of your equipment and inventory (to know what else you have) and the map (to know the upcoming enemies).
1.0
Gadget creation should let you look at your inventory and the map - When creating a gadget, you should be able to look at the rest of your equipment and inventory (to know what else you have) and the map (to know the upcoming enemies).
non_build
gadget creation should let you look at your inventory and the map when creating a gadget you should be able to look at the rest of your equipment and inventory to know what else you have and the map to know the upcoming enemies
0
30,614
8,568,037,511
IssuesEvent
2018-11-10 17:40:14
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
compile from source fails
stat:awaiting response type:build/install
**System information** - Ubuntu 18.04 - From source / r1.12 - TensorFlow version: - Python 3.6.6 - Inside virtualenv - Bazel 0.18.0 - gcc 6.4.0 - Cuda 9 / Cudnn 7 **The problem** Building from source, I end up with: ``` ERROR: /home/dev/tensorflow/tensorflow/core/BUILD:319:1: undeclared inclusion(s) in rule '//tensorflow/core:platform_base': this rule is missing dependency declarations for the following files included by 'tensorflow/core/platform/env_time.cc': '/usr/lib/gcc/x86_64-linux-gnu/6/include/stdint.h' '/usr/lib/gcc/x86_64-linux-gnu/6/include/stddef.h' '/usr/lib/gcc/x86_64-linux-gnu/6/include/stdarg.h' Target //tensorflow/tools/pip_package:build_pip_package failed to build Use --verbose_failures to see the command lines of failed build steps. INFO: Elapsed time: 0.460s, Critical Path: 0.26s INFO: 0 processes. FAILED: Build did NOT complete successfully ``` This is when I run: `bazel build --config=opt --config=cuda //tensorflow/tools/pip_package:build_pip_package` Also note that if I repeat the command several times, I don't always get the same error message. After I do `bazel clean` I get: ``` ERROR: /home/.cache/bazel/_bazel_nnnnn/75c0d842b4eca8fbdb48dc37e31275de/external/protobuf_archive/BUILD:659:1: C++ compilation of rule '@protobuf_archive//:python/google/protobuf/pyext/_message.so' failed (Exit 1) In file included from external/protobuf_archive/python/google/protobuf/pyext/map_container.cc:33:0: external/protobuf_archive/python/google/protobuf/pyext/map_container.h:34:20: fatal error: Python.h: No such file or directory #include <Python.h> ^ compilation terminated. ```
1.0
compile from source fails - **System information** - Ubuntu 18.04 - From source / r1.12 - TensorFlow version: - Python 3.6.6 - Inside virtualenv - Bazel 0.18.0 - gcc 6.4.0 - Cuda 9 / Cudnn 7 **The problem** Building from source, I end up with: ``` ERROR: /home/dev/tensorflow/tensorflow/core/BUILD:319:1: undeclared inclusion(s) in rule '//tensorflow/core:platform_base': this rule is missing dependency declarations for the following files included by 'tensorflow/core/platform/env_time.cc': '/usr/lib/gcc/x86_64-linux-gnu/6/include/stdint.h' '/usr/lib/gcc/x86_64-linux-gnu/6/include/stddef.h' '/usr/lib/gcc/x86_64-linux-gnu/6/include/stdarg.h' Target //tensorflow/tools/pip_package:build_pip_package failed to build Use --verbose_failures to see the command lines of failed build steps. INFO: Elapsed time: 0.460s, Critical Path: 0.26s INFO: 0 processes. FAILED: Build did NOT complete successfully ``` This is when I run: `bazel build --config=opt --config=cuda //tensorflow/tools/pip_package:build_pip_package` Also note that if I repeat the command several times, I don't always get the same error message. After I do `bazel clean` I get: ``` ERROR: /home/.cache/bazel/_bazel_nnnnn/75c0d842b4eca8fbdb48dc37e31275de/external/protobuf_archive/BUILD:659:1: C++ compilation of rule '@protobuf_archive//:python/google/protobuf/pyext/_message.so' failed (Exit 1) In file included from external/protobuf_archive/python/google/protobuf/pyext/map_container.cc:33:0: external/protobuf_archive/python/google/protobuf/pyext/map_container.h:34:20: fatal error: Python.h: No such file or directory #include <Python.h> ^ compilation terminated. ```
build
compile from source fails system information ubuntu from source tensorflow version python inside virtualenv bazel gcc cuda cudnn the problem building from source i end up with error home dev tensorflow tensorflow core build undeclared inclusion s in rule tensorflow core platform base this rule is missing dependency declarations for the following files included by tensorflow core platform env time cc usr lib gcc linux gnu include stdint h usr lib gcc linux gnu include stddef h usr lib gcc linux gnu include stdarg h target tensorflow tools pip package build pip package failed to build use verbose failures to see the command lines of failed build steps info elapsed time critical path info processes failed build did not complete successfully this is when i run bazel build config opt config cuda tensorflow tools pip package build pip package also note that if i repeat the command several times i don t always get the same error message after i do bazel clean i get error home cache bazel bazel nnnnn external protobuf archive build c compilation of rule protobuf archive python google protobuf pyext message so failed exit in file included from external protobuf archive python google protobuf pyext map container cc external protobuf archive python google protobuf pyext map container h fatal error python h no such file or directory include compilation terminated
1
651,733
21,486,601,990
IssuesEvent
2022-04-27 00:33:59
rstudio/gt
https://api.github.com/repos/rstudio/gt
closed
Add the `fmt_partsper()` function
Difficulty: [2] Intermediate Effort: [3] High Priority: [2] Medium Type: ★ Enhancement
There is a fmt_percent formatting function in gt but, in the same spirit, it would be lovely to have a an additional general function that does parts-per notation (eg. ppm, ppb, ppt, ppmV, per mille and per ten thousand, to name a few). These types of units are pretty common in the sciences and so this function would definitely get a lot of use from the scientific community.
1.0
Add the `fmt_partsper()` function - There is a fmt_percent formatting function in gt but, in the same spirit, it would be lovely to have a an additional general function that does parts-per notation (eg. ppm, ppb, ppt, ppmV, per mille and per ten thousand, to name a few). These types of units are pretty common in the sciences and so this function would definitely get a lot of use from the scientific community.
non_build
add the fmt partsper function there is a fmt percent formatting function in gt but in the same spirit it would be lovely to have a an additional general function that does parts per notation eg ppm ppb ppt ppmv per mille and per ten thousand to name a few these types of units are pretty common in the sciences and so this function would definitely get a lot of use from the scientific community
0
58,424
14,383,997,988
IssuesEvent
2020-12-02 09:52:50
googleapis/java-aiplatform
https://api.github.com/repos/googleapis/java-aiplatform
closed
aiplatform.GetModelEvaluationSliceSampleTest: testGetModelEvaluationSliceSample failed
buildcop: issue priority: p1 type: bug
This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: 783580731c9699266ca79b0ed50b035696921abf buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e0bccb45-0981-42b1-a4f8-be10b05c5627), [Sponge](http://sponge2/e0bccb45-0981-42b1-a4f8-be10b05c5627) status: failed <details><summary>Test output</summary><br><pre>com.google.api.gax.rpc.NotFoundException: io.grpc.StatusRuntimeException: NOT_FOUND: The Model does not exist. at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:45) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1041) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1215) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:983) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:771) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:464) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:428) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:461) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:617) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:803) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:782) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Suppressed: com.google.api.gax.rpc.AsyncTaskException: Asynchronous task failed at com.google.api.gax.rpc.ApiExceptions.callAndTranslateApiException(ApiExceptions.java:57) at com.google.api.gax.rpc.UnaryCallable.call(UnaryCallable.java:112) at com.google.cloud.aiplatform.v1beta1.ModelServiceClient.getModelEvaluationSlice(ModelServiceClient.java:1179) at com.google.cloud.aiplatform.v1beta1.ModelServiceClient.getModelEvaluationSlice(ModelServiceClient.java:1132) at aiplatform.GetModelEvaluationSliceSample.getModelEvaluationSliceSample(GetModelEvaluationSliceSample.java:55) at aiplatform.GetModelEvaluationSliceSampleTest.testGetModelEvaluationSliceSample(GetModelEvaluationSliceSampleTest.java:71) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: io.grpc.StatusRuntimeException: NOT_FOUND: The Model does not exist. at io.grpc.Status.asRuntimeException(Status.java:533) ... 17 more </pre></details>
1.0
aiplatform.GetModelEvaluationSliceSampleTest: testGetModelEvaluationSliceSample failed - This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: 783580731c9699266ca79b0ed50b035696921abf buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e0bccb45-0981-42b1-a4f8-be10b05c5627), [Sponge](http://sponge2/e0bccb45-0981-42b1-a4f8-be10b05c5627) status: failed <details><summary>Test output</summary><br><pre>com.google.api.gax.rpc.NotFoundException: io.grpc.StatusRuntimeException: NOT_FOUND: The Model does not exist. at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:45) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1041) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1215) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:983) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:771) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:464) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:428) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:461) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:617) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:803) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:782) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Suppressed: com.google.api.gax.rpc.AsyncTaskException: Asynchronous task failed at com.google.api.gax.rpc.ApiExceptions.callAndTranslateApiException(ApiExceptions.java:57) at com.google.api.gax.rpc.UnaryCallable.call(UnaryCallable.java:112) at com.google.cloud.aiplatform.v1beta1.ModelServiceClient.getModelEvaluationSlice(ModelServiceClient.java:1179) at com.google.cloud.aiplatform.v1beta1.ModelServiceClient.getModelEvaluationSlice(ModelServiceClient.java:1132) at aiplatform.GetModelEvaluationSliceSample.getModelEvaluationSliceSample(GetModelEvaluationSliceSample.java:55) at aiplatform.GetModelEvaluationSliceSampleTest.testGetModelEvaluationSliceSample(GetModelEvaluationSliceSampleTest.java:71) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: io.grpc.StatusRuntimeException: NOT_FOUND: The Model does not exist. at io.grpc.Status.asRuntimeException(Status.java:533) ... 17 more </pre></details>
build
aiplatform getmodelevaluationslicesampletest testgetmodelevaluationslicesample failed this test failed to configure my behavior see if i m commenting on this issue too often add the buildcop quiet label and i will stop commenting commit buildurl status failed test output com google api gax rpc notfoundexception io grpc statusruntimeexception not found the model does not exist at com google api gax rpc apiexceptionfactory createexception apiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcexceptioncallable exceptiontransformingfuture onfailure grpcexceptioncallable java at com google api core apifutures onfailure apifutures java at com google common util concurrent futures callbacklistener run futures java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at io grpc stub clientcalls grpcfuture setexception clientcalls java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc internal delayedclientcall delayedlistener run delayedclientcall java at io grpc internal delayedclientcall delayedlistener delayorexecute delayedclientcall java at io grpc internal delayedclientcall delayedlistener onclose delayedclientcall java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask access scheduledthreadpoolexecutor java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java suppressed com google api gax rpc asynctaskexception asynchronous task failed at com google api gax rpc apiexceptions callandtranslateapiexception apiexceptions java at com google api gax rpc unarycallable call unarycallable java at com google cloud aiplatform modelserviceclient getmodelevaluationslice modelserviceclient java at com google cloud aiplatform modelserviceclient getmodelevaluationslice modelserviceclient java at aiplatform getmodelevaluationslicesample getmodelevaluationslicesample getmodelevaluationslicesample java at aiplatform getmodelevaluationslicesampletest testgetmodelevaluationslicesample getmodelevaluationslicesampletest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire execute java at org apache maven surefire executewithrerun java at org apache maven surefire executetestset java at org apache maven surefire invoke java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by io grpc statusruntimeexception not found the model does not exist at io grpc status asruntimeexception status java more
1
544,613
15,895,032,023
IssuesEvent
2021-04-11 12:33:00
arcomage/arcomage-hd
https://api.github.com/repos/arcomage/arcomage-hd
closed
Locally save settings (incl. lang, volume) data
enhancement medium-priority
Maybe with localStorage in web app, and local files in Electron Desktop version
1.0
Locally save settings (incl. lang, volume) data - Maybe with localStorage in web app, and local files in Electron Desktop version
non_build
locally save settings incl lang volume data maybe with localstorage in web app and local files in electron desktop version
0
72,501
19,293,336,220
IssuesEvent
2021-12-12 06:31:53
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
How to build tensorflow lite C api for Linux on a MacOS machine?
stat:awaiting response type:build/install stalled comp:lite subtype:macOS
Hi thanks for the lib! I need to build tensorflow lite C api for Linux on a MacOS machine. If I do `bazel build -c opt //tensorflow/lite/c:tensorflowlite_c`, I get a `.dylib` that is of mach-o (macos) format instead of a `.so` that is of linux format. Thus I wonder what should I do? Thanks! I have tried: `--cpu=linux` and its friends (no luck). `--config-linux` (also no use).
1.0
How to build tensorflow lite C api for Linux on a MacOS machine? - Hi thanks for the lib! I need to build tensorflow lite C api for Linux on a MacOS machine. If I do `bazel build -c opt //tensorflow/lite/c:tensorflowlite_c`, I get a `.dylib` that is of mach-o (macos) format instead of a `.so` that is of linux format. Thus I wonder what should I do? Thanks! I have tried: `--cpu=linux` and its friends (no luck). `--config-linux` (also no use).
build
how to build tensorflow lite c api for linux on a macos machine hi thanks for the lib i need to build tensorflow lite c api for linux on a macos machine if i do bazel build c opt tensorflow lite c tensorflowlite c i get a dylib that is of mach o macos format instead of a so that is of linux format thus i wonder what should i do thanks i have tried cpu linux and its friends no luck config linux also no use
1
76,261
7,523,211,272
IssuesEvent
2018-04-12 23:38:05
CougsInSpace/CougSat1-Hardware
https://api.github.com/repos/CougsInSpace/CougSat1-Hardware
opened
Prototype Germination Enclosure
CAD Germination Structure Vibration Testing
As stated above. This includes: -Making the enclosure airtight. -Mounting all of the following components: - -Deployment System - -CO2 Sensor - -Aluminized Mylar (buy) - -Blue and red LEDS. Small ones B. - -Pass through. - -Camera (Bradley) - -Nutrient substraites (cotton balls, plant vitamins), - -Pressure sensor - -Heating resistor - -Temperature. -Attaching the enclosure to the Cubesat. -Wiring the components to systems -Programming the components
1.0
Prototype Germination Enclosure - As stated above. This includes: -Making the enclosure airtight. -Mounting all of the following components: - -Deployment System - -CO2 Sensor - -Aluminized Mylar (buy) - -Blue and red LEDS. Small ones B. - -Pass through. - -Camera (Bradley) - -Nutrient substraites (cotton balls, plant vitamins), - -Pressure sensor - -Heating resistor - -Temperature. -Attaching the enclosure to the Cubesat. -Wiring the components to systems -Programming the components
non_build
prototype germination enclosure as stated above this includes making the enclosure airtight mounting all of the following components deployment system sensor aluminized mylar buy blue and red leds small ones b pass through camera bradley nutrient substraites cotton balls plant vitamins pressure sensor heating resistor temperature attaching the enclosure to the cubesat wiring the components to systems programming the components
0
482,303
13,904,180,296
IssuesEvent
2020-10-20 08:16:04
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
opened
"Context does not exist" setting up SAML SSO
Priority/Normal Type/Bug
### Description: Cannot log into devportal using SAML SSO with shibboleth. I think the problem comes from the session cookies path which is set to "/devportal" in the browser. Since the SAMLResponse from shibboleth is being browser POSTed to "/commonauth" the cookies are blocked by browser. For example, Chrome says: "This cookie was blocked because its path was not an exact match for or a superdirectory of the request url's path". In this situation the SAMLResponse comes into APIM with NO essential cookies CLIENT_ID and JSESSIONID, what can explain the "empty context" error related. ### Steps to reproduce: - add an "idp provider" and configure to call an external IdP - make SP "devportal" to use this created IdP - try to sign into devportal, you will get an error. In the logs: `TID: [-1234] [] [2020-10-19 15:26:30,598] ERROR {org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultRequestCoordinator} - Context does not exist. Probably due to invalidated cache. Requested client: 130.104.240.14, URI :POST:/commonauth, User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/86.0.4240.80 Safari/537.36 , Referer: https://idp.com/ TID: [-1234] [] [2020-10-19 15:26:30,601] WARN {org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultRequestCoordinator} - Authentication context is null, redirect parameter filtering will not be done for null TID: [-1] [] [2020-10-19 15:26:30,621] ERROR {org.wso2.carbon.tomcat.ext.valves.CarbonContextCreatorValve} - Could not handle request: null java.lang.NullPointerException ` ### Affected Product Version: APIM 3.2.0 ### Environment details (with versions): - OS: linux - Client: MacOS (firefox and chrome) - Env (Docker/K8s): debian --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
1.0
"Context does not exist" setting up SAML SSO - ### Description: Cannot log into devportal using SAML SSO with shibboleth. I think the problem comes from the session cookies path which is set to "/devportal" in the browser. Since the SAMLResponse from shibboleth is being browser POSTed to "/commonauth" the cookies are blocked by browser. For example, Chrome says: "This cookie was blocked because its path was not an exact match for or a superdirectory of the request url's path". In this situation the SAMLResponse comes into APIM with NO essential cookies CLIENT_ID and JSESSIONID, what can explain the "empty context" error related. ### Steps to reproduce: - add an "idp provider" and configure to call an external IdP - make SP "devportal" to use this created IdP - try to sign into devportal, you will get an error. In the logs: `TID: [-1234] [] [2020-10-19 15:26:30,598] ERROR {org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultRequestCoordinator} - Context does not exist. Probably due to invalidated cache. Requested client: 130.104.240.14, URI :POST:/commonauth, User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/86.0.4240.80 Safari/537.36 , Referer: https://idp.com/ TID: [-1234] [] [2020-10-19 15:26:30,601] WARN {org.wso2.carbon.identity.application.authentication.framework.handler.request.impl.DefaultRequestCoordinator} - Authentication context is null, redirect parameter filtering will not be done for null TID: [-1] [] [2020-10-19 15:26:30,621] ERROR {org.wso2.carbon.tomcat.ext.valves.CarbonContextCreatorValve} - Could not handle request: null java.lang.NullPointerException ` ### Affected Product Version: APIM 3.2.0 ### Environment details (with versions): - OS: linux - Client: MacOS (firefox and chrome) - Env (Docker/K8s): debian --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
non_build
context does not exist setting up saml sso description cannot log into devportal using saml sso with shibboleth i think the problem comes from the session cookies path which is set to devportal in the browser since the samlresponse from shibboleth is being browser posted to commonauth the cookies are blocked by browser for example chrome says this cookie was blocked because its path was not an exact match for or a superdirectory of the request url s path in this situation the samlresponse comes into apim with no essential cookies client id and jsessionid what can explain the empty context error related steps to reproduce add an idp provider and configure to call an external idp make sp devportal to use this created idp try to sign into devportal you will get an error in the logs tid error org carbon identity application authentication framework handler request impl defaultrequestcoordinator context does not exist probably due to invalidated cache requested client uri post commonauth user agent mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari referer tid warn org carbon identity application authentication framework handler request impl defaultrequestcoordinator authentication context is null redirect parameter filtering will not be done for null tid error org carbon tomcat ext valves carboncontextcreatorvalve could not handle request null java lang nullpointerexception affected product version apim environment details with versions os linux client macos firefox and chrome env docker debian optional fields related issues suggested labels suggested assignees
0
617,638
19,401,604,615
IssuesEvent
2021-12-19 09:27:35
zvladn7/news-searcher
https://api.github.com/repos/zvladn7/news-searcher
closed
Исправление багов
bug backend high priority
1. Корректно находить заголовок для результата поиска. Необходимо в результате из индекса находить строку вхождения запроса в полный текст для корректного отображения заголовка результата поиска 2. Исправить возвращаемый заголовок результата. Сделать максимальную длину заголовку 200 символов. Если заголовок больше 200 символов, то вместо последних трех символов вставлять троеточие. ("заголо...") 3. запрос similar должен возвращать массив строк, а не массив объектов. Пример ответа: ``` [ "title result", "title result2", ... ] ``` 4. Количество возвращаемых результатов на 1 страницу должно быть 10 штук, сейчас возвращается по 5 штук на 1 страницу ![image](https://user-images.githubusercontent.com/43933761/144676396-9eae2006-0567-4047-8f63-d65aa435f0a2.png) 5. Не работает пагинация для запроса изображений. 6. При запросе изображений не должны возвращаться элементы с пустым `imageUrl` 7. На каждую страницу запроса изображений должно возвращаться 20 элементов, сейчас возвращается 10. ![image](https://user-images.githubusercontent.com/43933761/144677263-2f1946ef-fb95-413b-8ecb-91d491e35ee9.png)
1.0
Исправление багов - 1. Корректно находить заголовок для результата поиска. Необходимо в результате из индекса находить строку вхождения запроса в полный текст для корректного отображения заголовка результата поиска 2. Исправить возвращаемый заголовок результата. Сделать максимальную длину заголовку 200 символов. Если заголовок больше 200 символов, то вместо последних трех символов вставлять троеточие. ("заголо...") 3. запрос similar должен возвращать массив строк, а не массив объектов. Пример ответа: ``` [ "title result", "title result2", ... ] ``` 4. Количество возвращаемых результатов на 1 страницу должно быть 10 штук, сейчас возвращается по 5 штук на 1 страницу ![image](https://user-images.githubusercontent.com/43933761/144676396-9eae2006-0567-4047-8f63-d65aa435f0a2.png) 5. Не работает пагинация для запроса изображений. 6. При запросе изображений не должны возвращаться элементы с пустым `imageUrl` 7. На каждую страницу запроса изображений должно возвращаться 20 элементов, сейчас возвращается 10. ![image](https://user-images.githubusercontent.com/43933761/144677263-2f1946ef-fb95-413b-8ecb-91d491e35ee9.png)
non_build
исправление багов корректно находить заголовок для результата поиска необходимо в результате из индекса находить строку вхождения запроса в полный текст для корректного отображения заголовка результата поиска исправить возвращаемый заголовок результата сделать максимальную длину заголовку символов если заголовок больше символов то вместо последних трех символов вставлять троеточие заголо запрос similar должен возвращать массив строк а не массив объектов пример ответа title result title количество возвращаемых результатов на страницу должно быть штук сейчас возвращается по штук на страницу не работает пагинация для запроса изображений при запросе изображений не должны возвращаться элементы с пустым imageurl на каждую страницу запроса изображений должно возвращаться элементов сейчас возвращается
0
142,644
13,036,761,095
IssuesEvent
2020-07-28 12:48:19
oracle-terraform-modules/terraform-oci-oke
https://api.github.com/repos/oracle-terraform-modules/terraform-oci-oke
opened
Remove disable_auto_retries in provider in quickstart guide
bug documentation
<!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version and Provider Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are using a local copy of the Terraform Oracle Cloud Infrastructure Provider, run the plugin directly to get the version: `<path-to-plugin>/terraform-provider-oci` If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ### Affected Resource(s) <!--- Please list the affected resources and data sources. For example, "oci_core_vcn". ---> ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl # Copy-paste your Terraform configurations here - for large Terraform configs, # please use a service like Dropbox and share a link to the ZIP file. # Please remove any sensitive information from configuration files before sharing them. ``` ### Debug Output <!--- Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist. To obtain the debug output, see the [Verbose logging for OCI Terraform Provider](https://www.terraform.io/docs/providers/oci/guides/troubleshooting.html#verbose-logging-for-oci-terraform-provider). Github Gist: https://gist.github.com/ ---> ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. Github Gist: https://gist.github.com/ ---> ### Expected Behavior <!--- What should have happened? ---> ### Actual Behavior <!--- What actually happened? ---> ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` ### Important Factoids <!--- Is there anything atypical about your environment that we should know? For example: Is the issue specific to a region? ---> ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example: --->
1.0
Remove disable_auto_retries in provider in quickstart guide - <!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version and Provider Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are using a local copy of the Terraform Oracle Cloud Infrastructure Provider, run the plugin directly to get the version: `<path-to-plugin>/terraform-provider-oci` If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ### Affected Resource(s) <!--- Please list the affected resources and data sources. For example, "oci_core_vcn". ---> ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl # Copy-paste your Terraform configurations here - for large Terraform configs, # please use a service like Dropbox and share a link to the ZIP file. # Please remove any sensitive information from configuration files before sharing them. ``` ### Debug Output <!--- Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist. To obtain the debug output, see the [Verbose logging for OCI Terraform Provider](https://www.terraform.io/docs/providers/oci/guides/troubleshooting.html#verbose-logging-for-oci-terraform-provider). Github Gist: https://gist.github.com/ ---> ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. Github Gist: https://gist.github.com/ ---> ### Expected Behavior <!--- What should have happened? ---> ### Actual Behavior <!--- What actually happened? ---> ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` ### Important Factoids <!--- Is there anything atypical about your environment that we should know? For example: Is the issue specific to a region? ---> ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example: --->
non_build
remove disable auto retries in provider in quickstart guide please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version and provider version please run terraform v to show the terraform core version and provider version s if you are using a local copy of the terraform oracle cloud infrastructure provider run the plugin directly to get the version terraform provider oci if you are not running the latest version of terraform or the provider please upgrade because your issue may have already been fixed affected resource s terraform configuration files hcl copy paste your terraform configurations here for large terraform configs please use a service like dropbox and share a link to the zip file please remove any sensitive information from configuration files before sharing them debug output please provide a link to a github gist containing the complete debug output please do not paste the debug output in the issue just paste a link to the gist to obtain the debug output see the github gist panic output if terraform produced a panic please provide a link to a github gist containing the output of the crash log github gist expected behavior actual behavior steps to reproduce terraform apply important factoids references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor documentation for example
0
70,777
18,272,323,052
IssuesEvent
2021-10-04 14:58:10
ME-ON1/envoy
https://api.github.com/repos/ME-ON1/envoy
opened
Newer release available `com_github_cyan4973_xxhash`: v0.8.0 (current: v0.7.3)
dependencies area/build no stalebot
Package Name: com_github_cyan4973_xxhash Current Version: v0.7.3@$2020-03-05 15:54:16 Available Version: v0.8.0@2020-07-27 18:57:06 Upstream link: https://github.com/Cyan4973/xxHash
1.0
Newer release available `com_github_cyan4973_xxhash`: v0.8.0 (current: v0.7.3) - Package Name: com_github_cyan4973_xxhash Current Version: v0.7.3@$2020-03-05 15:54:16 Available Version: v0.8.0@2020-07-27 18:57:06 Upstream link: https://github.com/Cyan4973/xxHash
build
newer release available com github xxhash current package name com github xxhash current version available version upstream link
1
13,593
4,738,312,142
IssuesEvent
2016-10-20 03:31:13
hoodiehq/camp
https://api.github.com/repos/hoodiehq/camp
closed
[Hoodie Account Server API] api.sessions.remove -> rejects if options is not passed
Code ready
### 🎃💻👕 Hacktoberfest: Trick or Treat! If you haven’t yet, sign up for [Hacktoberfest](https://hacktoberfest.digitalocean.com/) to earn an exclusive T-Shirt. Plus I’m sure we can teach you a cool trick or two in the process ### 🤔 What you will need to know You should have worked with JavaScript, Node.js and Testing. If you haven’t yet, we recommend the [JavaScript Track on exercism.io](http://exercism.io/languages/javascript) ### 🐞 The Bug ```js var AccountApi = require('@hoodie/account-server-api') var PouchDB = require('pouchdb-core').plugin(require('pouchdb-adapter-memory')).plugin(require('pouchdb-mapreduce')) var api = new AccountApi({ PouchDB: PouchDB, secret: 'secret123' }) api.sessions.remove('sessionid').catch(function (error) { // { [not_found: missing] // status: 404, // name: 'not_found', // message: 'missing', // error: true, // reason: 'missing' } }) ``` We prepared a test to reproduce this issue at [test/unit/sessions/remove-test.js#L25](https://github.com/hoodiehq/hoodie-account-server-api/blob/4d49fb56eb887df38e3b40a658daf8f43368eaa4/test/unit/sessions/remove-test.js#L25-L26) ### :dart: The goal Make the test pass :) ### :clipboard: Step by Step If this is your first, welcome :tada: :smile: [Here is a great tutorial](https://egghead.io/series/how-to-contribute-to-an-open-source-project-on-github) on how to send a pull request using the terminal. - [x] 🙋 **Claim this issue**: Comment below (or assign yourself and continue at step 4 :) Please 🙏 only claim if you want to start working on it during the event. Once claimed we add you as contributor to this repository. - [x] 👌 **Accept our invitation** to this repository. Once accepted, assign yourself to this repository - [x] 👓 **Please review** our [Code of Conduct](http://hood.ie/code-of-conduct/) In a nutshell: be patient and actively kind with us 😊 - [x] 🔄 **replace** the `up for grabs` label with `in progress`. - [x] 🗜 [**Setup the repository locally**](https://github.com/hoodiehq/hoodie-account-server-api#testing) - [x] Remove the comment and the `, {skip: true}` at [test/unit/sessions/remove-test.js#L25](https://github.com/hoodiehq/hoodie-account-server-api/blob/4d49fb56eb887df38e3b40a658daf8f43368eaa4/test/unit/sessions/remove-test.js#L25-L26). Tests should fail now. - [x] **Commit the change** with `test: sessions.remove('sessionid')` and push it either to your fork or push your new branch. - [x] 🔀 **Start a Pull Request**. Mention `closes hoodiehq/camp#58` in the description. - [x] 🏁 **Done** :+1: Replace the `in progress` label with `ready`. Ask in comments for a review :) ### 🤔❓ Questions Ping us in the [Hoodie Chat](http://hood.ie/chat/) or on [Twitter](https://twitter.com/hoodiehq/)
1.0
[Hoodie Account Server API] api.sessions.remove -> rejects if options is not passed - ### 🎃💻👕 Hacktoberfest: Trick or Treat! If you haven’t yet, sign up for [Hacktoberfest](https://hacktoberfest.digitalocean.com/) to earn an exclusive T-Shirt. Plus I’m sure we can teach you a cool trick or two in the process ### 🤔 What you will need to know You should have worked with JavaScript, Node.js and Testing. If you haven’t yet, we recommend the [JavaScript Track on exercism.io](http://exercism.io/languages/javascript) ### 🐞 The Bug ```js var AccountApi = require('@hoodie/account-server-api') var PouchDB = require('pouchdb-core').plugin(require('pouchdb-adapter-memory')).plugin(require('pouchdb-mapreduce')) var api = new AccountApi({ PouchDB: PouchDB, secret: 'secret123' }) api.sessions.remove('sessionid').catch(function (error) { // { [not_found: missing] // status: 404, // name: 'not_found', // message: 'missing', // error: true, // reason: 'missing' } }) ``` We prepared a test to reproduce this issue at [test/unit/sessions/remove-test.js#L25](https://github.com/hoodiehq/hoodie-account-server-api/blob/4d49fb56eb887df38e3b40a658daf8f43368eaa4/test/unit/sessions/remove-test.js#L25-L26) ### :dart: The goal Make the test pass :) ### :clipboard: Step by Step If this is your first, welcome :tada: :smile: [Here is a great tutorial](https://egghead.io/series/how-to-contribute-to-an-open-source-project-on-github) on how to send a pull request using the terminal. - [x] 🙋 **Claim this issue**: Comment below (or assign yourself and continue at step 4 :) Please 🙏 only claim if you want to start working on it during the event. Once claimed we add you as contributor to this repository. - [x] 👌 **Accept our invitation** to this repository. Once accepted, assign yourself to this repository - [x] 👓 **Please review** our [Code of Conduct](http://hood.ie/code-of-conduct/) In a nutshell: be patient and actively kind with us 😊 - [x] 🔄 **replace** the `up for grabs` label with `in progress`. - [x] 🗜 [**Setup the repository locally**](https://github.com/hoodiehq/hoodie-account-server-api#testing) - [x] Remove the comment and the `, {skip: true}` at [test/unit/sessions/remove-test.js#L25](https://github.com/hoodiehq/hoodie-account-server-api/blob/4d49fb56eb887df38e3b40a658daf8f43368eaa4/test/unit/sessions/remove-test.js#L25-L26). Tests should fail now. - [x] **Commit the change** with `test: sessions.remove('sessionid')` and push it either to your fork or push your new branch. - [x] 🔀 **Start a Pull Request**. Mention `closes hoodiehq/camp#58` in the description. - [x] 🏁 **Done** :+1: Replace the `in progress` label with `ready`. Ask in comments for a review :) ### 🤔❓ Questions Ping us in the [Hoodie Chat](http://hood.ie/chat/) or on [Twitter](https://twitter.com/hoodiehq/)
non_build
api sessions remove rejects if options is not passed 🎃💻👕 hacktoberfest trick or treat if you haven’t yet sign up for to earn an exclusive t shirt plus i’m sure we can teach you a cool trick or two in the process 🤔 what you will need to know you should have worked with javascript node js and testing if you haven’t yet we recommend the 🐞 the bug js var accountapi require hoodie account server api var pouchdb require pouchdb core plugin require pouchdb adapter memory plugin require pouchdb mapreduce var api new accountapi pouchdb pouchdb secret api sessions remove sessionid catch function error status name not found message missing error true reason missing we prepared a test to reproduce this issue at dart the goal make the test pass clipboard step by step if this is your first welcome tada smile on how to send a pull request using the terminal 🙋 claim this issue comment below or assign yourself and continue at step please 🙏 only claim if you want to start working on it during the event once claimed we add you as contributor to this repository 👌 accept our invitation to this repository once accepted assign yourself to this repository 👓 please review our in a nutshell be patient and actively kind with us 😊 🔄 replace the up for grabs label with in progress 🗜 remove the comment and the skip true at tests should fail now commit the change with test sessions remove sessionid and push it either to your fork or push your new branch 🔀 start a pull request mention closes hoodiehq camp in the description 🏁 done replace the in progress label with ready ask in comments for a review 🤔❓ questions ping us in the or on
0
164,765
13,960,033,985
IssuesEvent
2020-10-24 19:07:15
threedesigns/printNC
https://api.github.com/repos/threedesigns/printNC
opened
Printed Parts -> Community Mods -> README
documentation enhancement
Create a readme with a table that describes all of the community mods, who developed them, and why you might use the mod.
1.0
Printed Parts -> Community Mods -> README - Create a readme with a table that describes all of the community mods, who developed them, and why you might use the mod.
non_build
printed parts community mods readme create a readme with a table that describes all of the community mods who developed them and why you might use the mod
0
2,226
5,074,220,774
IssuesEvent
2016-12-27 13:11:01
DynareTeam/dynare
https://api.github.com/repos/DynareTeam/dynare
closed
Allow adding auxiliary variables like Ramsey multipliers to var_list_
preprocessor
The auxiliary variables are endogenous variables like every other variable. A call like `ramsey_policy(instruments=(i),irf=13,planner_discount=betta,periods=200) x pi MULT_1;` would be suficient to display IRFs for the multiplier 1. However, the preprocessor does not allow adding `MULT_1` to the variable list, because: `Unknown symbol: MULT_1` We should allow adding any variable present in `M_.endo_names` to the `var_list_`. @houtanb Could you do this, please? Related to http://www.dynare.org/phpBB3/viewtopic.php?f=1&t=12117
1.0
Allow adding auxiliary variables like Ramsey multipliers to var_list_ - The auxiliary variables are endogenous variables like every other variable. A call like `ramsey_policy(instruments=(i),irf=13,planner_discount=betta,periods=200) x pi MULT_1;` would be suficient to display IRFs for the multiplier 1. However, the preprocessor does not allow adding `MULT_1` to the variable list, because: `Unknown symbol: MULT_1` We should allow adding any variable present in `M_.endo_names` to the `var_list_`. @houtanb Could you do this, please? Related to http://www.dynare.org/phpBB3/viewtopic.php?f=1&t=12117
non_build
allow adding auxiliary variables like ramsey multipliers to var list the auxiliary variables are endogenous variables like every other variable a call like ramsey policy instruments i irf planner discount betta periods x pi mult would be suficient to display irfs for the multiplier however the preprocessor does not allow adding mult to the variable list because unknown symbol mult we should allow adding any variable present in m endo names to the var list houtanb could you do this please related to
0
35,788
14,881,267,370
IssuesEvent
2021-01-20 10:16:20
LiskHQ/lisk-service
https://api.github.com/repos/LiskHQ/lisk-service
closed
Retrieve peers
service/core service/gateway type: improvement
### Description Make sure Lisk Service supports peer retrieval, using the same techniques that work in the previous versions of the SDK. ### Motivation - Random access and search by those properties is required by the UI ### Acceptance Criteria - Peer retrieval is possible - Gateway is updated with the API version 2 ### Additional information Make sure the following search criteria are supported: Property | Source endpoints | UI? | Comment -- | -- | -- | -- ip | n/a | Yes |   networkVersion | n/a | No | New in SDKv5 state | n/a | Yes |   height | n/a | Yes |   Enable sorting by those properties: Property | UI? | Comment -- | -- | -- height | Yes |   networkVersion | Yes | New in SDKv5 Use the following endpoints: `node.getConnectedPeers, node.getDisconnectedPeers` The following data can be stored in-memory. #### Endpoints - HTTP `/api/v2/peers` - RPC `get.peers` #### Request parameters Parameter | Type | Validation | Default | Comment -- | -- | -- | -- | -- ip | String | `/^(?:(?:25[0-5]\|2[0-4][0-9]\|[01]?[0-9][0-9]?)\.){3}(?:25[0-5]\|2[0-4][0-9]\|[01]?[0-9][0-9]?)$/` | *(empty)* |   networkVersion | String | `/^(0\|[1-9]\d*)\.(0\|[1-9]\d*)\.(0\|[1-9]\d*)(-(0\|[1-9]\d*\|\d*[a-zA-Z-][0-9a-zA-Z-]*)(\.(0\|[1-9]\d*\|\d*[a-zA-Z-][0-9a-zA-Z-]*))*)?(\+[0-9a-zA-Z-]+(\.[0-9a-zA-Z-]+)*)?$/ ` | *(empty)* |   state | Array of Strings | `[“connected”, “disconnected”, ”any”]` | connected |   height | Number |` <1;+Inf>` | *(empty)* |   limit | Number | `<1;100>` | 10 |   offset | Number | `<0;+Inf>` | 0 |   sort | String | `[“height:asc”, “height:desc”, “networkVersion:asc”, ”networkVersion:desc”]` | “height:desc” |   #### Response example 200 OK ``` { "data": [ { "ip": "127.0.0.1", "port": 4000, "networkVersion": "2.0", "state": ”connected”, "height": 8350681, "networkIdentifier": "258974416d58533227c6a3da1b6333f0541b06c65b41e45cf31926847a3db1ea", "location": { "countryCode": "DE", "countryName": "Germany", "hostname": "host.210.239.23.62.rev.coltfrance.com", "ip": "210.239.23.62", } } ], "meta": { "count": 100, "offset": 25, "total": 43749 }, "links": {} } ``` 400 Bad Request ``` { "error": true, "message": "Unknown input parameter(s): <param_name>" } ``` 404 Not Found ``` { "error": true, "message": "Account <account_id> not found." } ```
2.0
Retrieve peers - ### Description Make sure Lisk Service supports peer retrieval, using the same techniques that work in the previous versions of the SDK. ### Motivation - Random access and search by those properties is required by the UI ### Acceptance Criteria - Peer retrieval is possible - Gateway is updated with the API version 2 ### Additional information Make sure the following search criteria are supported: Property | Source endpoints | UI? | Comment -- | -- | -- | -- ip | n/a | Yes |   networkVersion | n/a | No | New in SDKv5 state | n/a | Yes |   height | n/a | Yes |   Enable sorting by those properties: Property | UI? | Comment -- | -- | -- height | Yes |   networkVersion | Yes | New in SDKv5 Use the following endpoints: `node.getConnectedPeers, node.getDisconnectedPeers` The following data can be stored in-memory. #### Endpoints - HTTP `/api/v2/peers` - RPC `get.peers` #### Request parameters Parameter | Type | Validation | Default | Comment -- | -- | -- | -- | -- ip | String | `/^(?:(?:25[0-5]\|2[0-4][0-9]\|[01]?[0-9][0-9]?)\.){3}(?:25[0-5]\|2[0-4][0-9]\|[01]?[0-9][0-9]?)$/` | *(empty)* |   networkVersion | String | `/^(0\|[1-9]\d*)\.(0\|[1-9]\d*)\.(0\|[1-9]\d*)(-(0\|[1-9]\d*\|\d*[a-zA-Z-][0-9a-zA-Z-]*)(\.(0\|[1-9]\d*\|\d*[a-zA-Z-][0-9a-zA-Z-]*))*)?(\+[0-9a-zA-Z-]+(\.[0-9a-zA-Z-]+)*)?$/ ` | *(empty)* |   state | Array of Strings | `[“connected”, “disconnected”, ”any”]` | connected |   height | Number |` <1;+Inf>` | *(empty)* |   limit | Number | `<1;100>` | 10 |   offset | Number | `<0;+Inf>` | 0 |   sort | String | `[“height:asc”, “height:desc”, “networkVersion:asc”, ”networkVersion:desc”]` | “height:desc” |   #### Response example 200 OK ``` { "data": [ { "ip": "127.0.0.1", "port": 4000, "networkVersion": "2.0", "state": ”connected”, "height": 8350681, "networkIdentifier": "258974416d58533227c6a3da1b6333f0541b06c65b41e45cf31926847a3db1ea", "location": { "countryCode": "DE", "countryName": "Germany", "hostname": "host.210.239.23.62.rev.coltfrance.com", "ip": "210.239.23.62", } } ], "meta": { "count": 100, "offset": 25, "total": 43749 }, "links": {} } ``` 400 Bad Request ``` { "error": true, "message": "Unknown input parameter(s): <param_name>" } ``` 404 Not Found ``` { "error": true, "message": "Account <account_id> not found." } ```
non_build
retrieve peers description make sure lisk service supports peer retrieval using the same techniques that work in the previous versions of the sdk motivation random access and search by those properties is required by the ui acceptance criteria peer retrieval is possible gateway is updated with the api version additional information make sure the following search criteria are supported property source endpoints ui comment ip n a yes   networkversion n a no new in state n a yes   height n a yes   enable sorting by those properties property ui comment height yes   networkversion yes new in use the following endpoints node getconnectedpeers node getdisconnectedpeers the following data can be stored in memory endpoints http api peers rpc get peers request parameters parameter type validation default comment ip string empty   networkversion string d d d d d d d empty   state array of strings connected   height number empty   limit number   offset number   sort string “height desc”   response example ok data ip port networkversion state ”connected” height networkidentifier location countrycode de countryname germany hostname host rev coltfrance com ip meta count offset total links bad request error true message unknown input parameter s not found error true message account not found
0
124,216
16,597,053,292
IssuesEvent
2021-06-01 14:36:06
readthedocs/sphinx-hoverxref
https://api.github.com/repos/readthedocs/sphinx-hoverxref
opened
Figure it out a good way to remove `title=` attribute on intersphinx nodes
Improvement Needed: design decision
I originally removed the `node['reftitle']` attribute because it makes the browser show the default tooltip (browser's built-in) and immediately after that our tooltip was shown. This behavior is a little confusing and it's not _styled_ in my opinion. That's why I removed it. ![Screenshot_2021-06-01_16-34-21](https://user-images.githubusercontent.com/244656/120341408-23e73e00-c2f7-11eb-8c16-ec27a62c675c.png) However, removing this attribute made other extensions break (`sphinx.ext.inheritance_diagram`) which expects that attribute exists. I went back in [`0a75b73` (#86)](https://github.com/readthedocs/sphinx-hoverxref/pull/86/commits/0a75b73d1fd7969dc5c838fb0093e130de774dd3) and I'm opening this issue to find a better solution to avoid this. We could: 1. define it as `node['reftitle'] = ''` * it makes the browser to not show the built-in tooltip * I'm not sure about the accessibility implication it has * is the HTML still valid? It will generate a `<a title="">` tag 2. use javascript to remove all the `title=` from tags with `.hoverxref .external` on document load 3. other ideas? See https://github.com/readthedocs/sphinx-hoverxref/pull/86#issuecomment-852136211
1.0
Figure it out a good way to remove `title=` attribute on intersphinx nodes - I originally removed the `node['reftitle']` attribute because it makes the browser show the default tooltip (browser's built-in) and immediately after that our tooltip was shown. This behavior is a little confusing and it's not _styled_ in my opinion. That's why I removed it. ![Screenshot_2021-06-01_16-34-21](https://user-images.githubusercontent.com/244656/120341408-23e73e00-c2f7-11eb-8c16-ec27a62c675c.png) However, removing this attribute made other extensions break (`sphinx.ext.inheritance_diagram`) which expects that attribute exists. I went back in [`0a75b73` (#86)](https://github.com/readthedocs/sphinx-hoverxref/pull/86/commits/0a75b73d1fd7969dc5c838fb0093e130de774dd3) and I'm opening this issue to find a better solution to avoid this. We could: 1. define it as `node['reftitle'] = ''` * it makes the browser to not show the built-in tooltip * I'm not sure about the accessibility implication it has * is the HTML still valid? It will generate a `<a title="">` tag 2. use javascript to remove all the `title=` from tags with `.hoverxref .external` on document load 3. other ideas? See https://github.com/readthedocs/sphinx-hoverxref/pull/86#issuecomment-852136211
non_build
figure it out a good way to remove title attribute on intersphinx nodes i originally removed the node attribute because it makes the browser show the default tooltip browser s built in and immediately after that our tooltip was shown this behavior is a little confusing and it s not styled in my opinion that s why i removed it however removing this attribute made other extensions break sphinx ext inheritance diagram which expects that attribute exists i went back in and i m opening this issue to find a better solution to avoid this we could define it as node it makes the browser to not show the built in tooltip i m not sure about the accessibility implication it has is the html still valid it will generate a tag use javascript to remove all the title from tags with hoverxref external on document load other ideas see
0
66,335
16,594,196,075
IssuesEvent
2021-06-01 11:30:36
arrayfire/arrayfire
https://api.github.com/repos/arrayfire/arrayfire
opened
installation with vcpkg
build
<!-- `./vcpkg install arrayfire` on a osx fails --> Description =========== ``` Computing installation plan... The following packages will be built and installed: arrayfire[core,cpu,unified]:x64-osx -> 3.7.3#3 Detecting compiler hash for triplet x64-osx... Could not locate cached archive: /Users/amir/.cache/vcpkg/archives/40/408499aa90b7888f9fb2d84caee57320dfa08452.zip Starting package 1/1: arrayfire:x64-osx Building package arrayfire[core,cpu,unified]:x64-osx... -- Using cached /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-arrayfire-59ac7b980d1ae124aae914fb29cbf086c948954d.tar.gz -- Cleaning sources at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/86c948954d-c996b9f837.clean. Use --editable to skip cleaning for the packages you specify. -- Extracting source /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-arrayfire-59ac7b980d1ae124aae914fb29cbf086c948954d.tar.gz -- Applying patch build.patch -- Using source at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/86c948954d-c996b9f837.clean -- Using cached /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-threads-b666773940269179f19ef11c8f1eb77005e85d9a.tar.gz -- Cleaning sources at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/7005e85d9a-64aebcce4b.clean. Use --editable to skip cleaning for the packages you specify. -- Extracting source /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-threads-b666773940269179f19ef11c8f1eb77005e85d9a.tar.gz -- Using source at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/7005e85d9a-64aebcce4b.clean -- Using cached /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-forge-1a0f0cb6371a8c8053ab5eb7cbe3039c95132389.tar.gz -- Cleaning sources at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/9c95132389-511398ace8.clean. Use --editable to skip cleaning for the packages you specify. -- Extracting source /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-forge-1a0f0cb6371a8c8053ab5eb7cbe3039c95132389.tar.gz -- Using source at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/9c95132389-511398ace8.clean -- Configuring x64-osx-dbg -- Configuring x64-osx-rel -- Building x64-osx-dbg CMake Error at scripts/cmake/vcpkg_execute_build_process.cmake:146 (message): Command failed: /Users/amir/projects/github-code/alaki/vcpkg/downloads/tools/cmake-3.20.2-osx/cmake-3.20.2-macos-universal/CMake.app/Contents/bin/cmake --build . --config Debug --target install -- -v -j9 Working Directory: /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/x64-osx-dbg See logs for more information: /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/install-x64-osx-dbg-err.log Call Stack (most recent call first): scripts/cmake/vcpkg_build_cmake.cmake:104 (vcpkg_execute_build_process) scripts/cmake/vcpkg_install_cmake.cmake:44 (vcpkg_build_cmake) ports/arrayfire/portfile.cmake:76 (vcpkg_install_cmake) scripts/ports.cmake:141 (include) Error: Building package arrayfire:x64-osx failed with: BUILD_FAILED Please ensure you're using the latest portfiles with `./vcpkg update`, then submit an issue at https://github.com/Microsoft/vcpkg/issues including: Package: arrayfire:x64-osx Vcpkg version: 2021-05-05-unknownhash Additionally, attach any relevant sections from the log files above. ``` Build Environment ----------------- Compiler version: clang 12.0 Operating system: osx Build environment: vcpkg CMake variables: 3.19.1
1.0
installation with vcpkg - <!-- `./vcpkg install arrayfire` on a osx fails --> Description =========== ``` Computing installation plan... The following packages will be built and installed: arrayfire[core,cpu,unified]:x64-osx -> 3.7.3#3 Detecting compiler hash for triplet x64-osx... Could not locate cached archive: /Users/amir/.cache/vcpkg/archives/40/408499aa90b7888f9fb2d84caee57320dfa08452.zip Starting package 1/1: arrayfire:x64-osx Building package arrayfire[core,cpu,unified]:x64-osx... -- Using cached /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-arrayfire-59ac7b980d1ae124aae914fb29cbf086c948954d.tar.gz -- Cleaning sources at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/86c948954d-c996b9f837.clean. Use --editable to skip cleaning for the packages you specify. -- Extracting source /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-arrayfire-59ac7b980d1ae124aae914fb29cbf086c948954d.tar.gz -- Applying patch build.patch -- Using source at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/86c948954d-c996b9f837.clean -- Using cached /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-threads-b666773940269179f19ef11c8f1eb77005e85d9a.tar.gz -- Cleaning sources at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/7005e85d9a-64aebcce4b.clean. Use --editable to skip cleaning for the packages you specify. -- Extracting source /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-threads-b666773940269179f19ef11c8f1eb77005e85d9a.tar.gz -- Using source at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/7005e85d9a-64aebcce4b.clean -- Using cached /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-forge-1a0f0cb6371a8c8053ab5eb7cbe3039c95132389.tar.gz -- Cleaning sources at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/9c95132389-511398ace8.clean. Use --editable to skip cleaning for the packages you specify. -- Extracting source /Users/amir/projects/github-code/alaki/vcpkg/downloads/arrayfire-forge-1a0f0cb6371a8c8053ab5eb7cbe3039c95132389.tar.gz -- Using source at /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/src/9c95132389-511398ace8.clean -- Configuring x64-osx-dbg -- Configuring x64-osx-rel -- Building x64-osx-dbg CMake Error at scripts/cmake/vcpkg_execute_build_process.cmake:146 (message): Command failed: /Users/amir/projects/github-code/alaki/vcpkg/downloads/tools/cmake-3.20.2-osx/cmake-3.20.2-macos-universal/CMake.app/Contents/bin/cmake --build . --config Debug --target install -- -v -j9 Working Directory: /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/x64-osx-dbg See logs for more information: /Users/amir/projects/github-code/alaki/vcpkg/buildtrees/arrayfire/install-x64-osx-dbg-err.log Call Stack (most recent call first): scripts/cmake/vcpkg_build_cmake.cmake:104 (vcpkg_execute_build_process) scripts/cmake/vcpkg_install_cmake.cmake:44 (vcpkg_build_cmake) ports/arrayfire/portfile.cmake:76 (vcpkg_install_cmake) scripts/ports.cmake:141 (include) Error: Building package arrayfire:x64-osx failed with: BUILD_FAILED Please ensure you're using the latest portfiles with `./vcpkg update`, then submit an issue at https://github.com/Microsoft/vcpkg/issues including: Package: arrayfire:x64-osx Vcpkg version: 2021-05-05-unknownhash Additionally, attach any relevant sections from the log files above. ``` Build Environment ----------------- Compiler version: clang 12.0 Operating system: osx Build environment: vcpkg CMake variables: 3.19.1
build
installation with vcpkg vcpkg install arrayfire on a osx fails description computing installation plan the following packages will be built and installed arrayfire osx detecting compiler hash for triplet osx could not locate cached archive users amir cache vcpkg archives zip starting package arrayfire osx building package arrayfire osx using cached users amir projects github code alaki vcpkg downloads arrayfire arrayfire tar gz cleaning sources at users amir projects github code alaki vcpkg buildtrees arrayfire src clean use editable to skip cleaning for the packages you specify extracting source users amir projects github code alaki vcpkg downloads arrayfire arrayfire tar gz applying patch build patch using source at users amir projects github code alaki vcpkg buildtrees arrayfire src clean using cached users amir projects github code alaki vcpkg downloads arrayfire threads tar gz cleaning sources at users amir projects github code alaki vcpkg buildtrees arrayfire src clean use editable to skip cleaning for the packages you specify extracting source users amir projects github code alaki vcpkg downloads arrayfire threads tar gz using source at users amir projects github code alaki vcpkg buildtrees arrayfire src clean using cached users amir projects github code alaki vcpkg downloads arrayfire forge tar gz cleaning sources at users amir projects github code alaki vcpkg buildtrees arrayfire src clean use editable to skip cleaning for the packages you specify extracting source users amir projects github code alaki vcpkg downloads arrayfire forge tar gz using source at users amir projects github code alaki vcpkg buildtrees arrayfire src clean configuring osx dbg configuring osx rel building osx dbg cmake error at scripts cmake vcpkg execute build process cmake message command failed users amir projects github code alaki vcpkg downloads tools cmake osx cmake macos universal cmake app contents bin cmake build config debug target install v working directory users amir projects github code alaki vcpkg buildtrees arrayfire osx dbg see logs for more information users amir projects github code alaki vcpkg buildtrees arrayfire install osx dbg err log call stack most recent call first scripts cmake vcpkg build cmake cmake vcpkg execute build process scripts cmake vcpkg install cmake cmake vcpkg build cmake ports arrayfire portfile cmake vcpkg install cmake scripts ports cmake include error building package arrayfire osx failed with build failed please ensure you re using the latest portfiles with vcpkg update then submit an issue at including package arrayfire osx vcpkg version unknownhash additionally attach any relevant sections from the log files above build environment compiler version clang operating system osx build environment vcpkg cmake variables
1
73,308
19,651,885,532
IssuesEvent
2022-01-10 08:18:52
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
some firmware packages timestamped with 1980 in iso_minimal
0.kind: bug 6.topic: reproducible builds 2.status: stale
I tried to reproduce building nixos-minimal-21.05.596.3a2e0c36e79-x86_64-linux.iso . When I mount the `nix-store.squashfs` inside the `iso`, the contents appear to be the same (nix-hash `99d5f8add79590e77015c51bb7101dac`), but the `nix-store.squashfs` files themselves differ. The first difference seems to be near the start of the file: ``` 3,6c3,6 < 00000020 91 16 7e 05 0c 00 00 00 ca e0 b9 26 00 00 00 00 |..~........&....| < 00000030 c2 e0 b9 26 00 00 00 00 ff ff ff ff ff ff ff ff |...&............| < 00000040 ba b8 9d 26 00 00 00 00 b6 c4 a9 26 00 00 00 00 |...&.......&....| < 00000050 9e 07 b8 26 00 00 00 00 74 dd b9 26 00 00 00 00 |...&....t..&....| --- > 00000020 91 16 da 05 0c 00 00 00 f2 e0 b9 26 00 00 00 00 |...........&....| > 00000030 ea e0 b9 26 00 00 00 00 ff ff ff ff ff ff ff ff |...&............| > 00000040 ba b8 9d 26 00 00 00 00 12 c5 a9 26 00 00 00 00 |...&.......&....| > 00000050 12 08 b8 26 00 00 00 00 9c dd b9 26 00 00 00 00 |...&.......&....| ``` I think this is using `mksquashfs` from https://github.com/plougher/squashfs-tools as invoked from `nixos/lib/make-squashfs.nix`, so it should have the fix from #114454, but I haven't double-checked yet. /cc @plougher
1.0
some firmware packages timestamped with 1980 in iso_minimal - I tried to reproduce building nixos-minimal-21.05.596.3a2e0c36e79-x86_64-linux.iso . When I mount the `nix-store.squashfs` inside the `iso`, the contents appear to be the same (nix-hash `99d5f8add79590e77015c51bb7101dac`), but the `nix-store.squashfs` files themselves differ. The first difference seems to be near the start of the file: ``` 3,6c3,6 < 00000020 91 16 7e 05 0c 00 00 00 ca e0 b9 26 00 00 00 00 |..~........&....| < 00000030 c2 e0 b9 26 00 00 00 00 ff ff ff ff ff ff ff ff |...&............| < 00000040 ba b8 9d 26 00 00 00 00 b6 c4 a9 26 00 00 00 00 |...&.......&....| < 00000050 9e 07 b8 26 00 00 00 00 74 dd b9 26 00 00 00 00 |...&....t..&....| --- > 00000020 91 16 da 05 0c 00 00 00 f2 e0 b9 26 00 00 00 00 |...........&....| > 00000030 ea e0 b9 26 00 00 00 00 ff ff ff ff ff ff ff ff |...&............| > 00000040 ba b8 9d 26 00 00 00 00 12 c5 a9 26 00 00 00 00 |...&.......&....| > 00000050 12 08 b8 26 00 00 00 00 9c dd b9 26 00 00 00 00 |...&.......&....| ``` I think this is using `mksquashfs` from https://github.com/plougher/squashfs-tools as invoked from `nixos/lib/make-squashfs.nix`, so it should have the fix from #114454, but I haven't double-checked yet. /cc @plougher
build
some firmware packages timestamped with in iso minimal i tried to reproduce building nixos minimal linux iso when i mount the nix store squashfs inside the iso the contents appear to be the same nix hash but the nix store squashfs files themselves differ the first difference seems to be near the start of the file ca ff ff ff ff ff ff ff ff ba dd t da ea ff ff ff ff ff ff ff ff ba dd i think this is using mksquashfs from as invoked from nixos lib make squashfs nix so it should have the fix from but i haven t double checked yet cc plougher
1
43,395
17,595,466,987
IssuesEvent
2021-08-17 04:02:43
tuna/issues
https://api.github.com/repos/tuna/issues
opened
北外Fedora镜像出现HTTP 403错误
Service Issue
<!-- 请使用此模板来报告 bug,并尽可能多地提供信息。 Please use this template while reporting a bug and provide as much info as possible. --> #### 发生了什么(What happened) `sudo dnf update` 时,出现403报错,但是我手动单独更新报错的那几个包时,又可以更新, 此后再执行 `sudo dnf update` 又会报错,周而复始 :recycle: #### 期望的现象(What you expected to happen) `sudo dnf update` 应该毫无违和感的运行下去 #### 如何重现(How to reproduce it) 不确定,报错看起来是随机出现的 #### 其他事项(Anything else we need to know) - 该问题是否被之前的 issue 提出过: #### 您的环境(Environment) - 操作系统(OS Version):Fedora workstation 33 - 浏览器(如果适用)(Browser version, if applicable): - 其他(Others):ISP: 中国联通 ---- <details> <summary>当时DNF部分输出</summary> ``` 事务概要 ==================================================================================== 安装 10 软件包 升级 310 软件包 移除 1 软件包 总下载:574 M 确定吗?[y/N]: Y 下载软件包: (1/320): python3-packaging-20.4-2.fc33.noarch.rpm 180 kB/s | 67 kB 00:00 (2/320): libsodium-1.0.18-6.fc33.x86_64.rpm 413 kB/s | 166 kB 00:00 [MIRROR] python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/releases/33/Everything/x86_64/os/Packages/p/python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) (3/320): python3-pyparsing-2.4.7-4.fc33.noarch.rpm 338 kB/s | 149 kB 00:00 [MIRROR] kernel-devel-5.13.9-100.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/updates/33/Everything/x86_64/Packages/k/kernel-devel-5.13.9-100.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [MIRROR] python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/releases/33/Everything/x86_64/os/Packages/p/python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [MIRROR] kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/updates/33/Everything/x86_64/Packages/k/kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [MIRROR] kernel-devel-5.13.9-100.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/updates/33/Everything/x86_64/Packages/k/kernel-devel-5.13.9-100.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [MIRROR] python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/releases/33/Everything/x86_64/os/Packages/p/python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [MIRROR] kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/updates/33/Everything/x86_64/Packages/k/kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [FAILED] kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm: No more mirrors to try - All mirrors were already tried without success (5-6/320): python3-sim 0% [ ] 685 kB/s | 2.4 MB 14:13 ETA 下载的软件包保存在缓存中,直到下次成功执行事务。 您可以通过执行 'dnf clean packages' 删除软件包缓存。 错误:下载软件包出错 : Cannot download Packages/k/kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm: All mirrors were tried ``` </details>
1.0
北外Fedora镜像出现HTTP 403错误 - <!-- 请使用此模板来报告 bug,并尽可能多地提供信息。 Please use this template while reporting a bug and provide as much info as possible. --> #### 发生了什么(What happened) `sudo dnf update` 时,出现403报错,但是我手动单独更新报错的那几个包时,又可以更新, 此后再执行 `sudo dnf update` 又会报错,周而复始 :recycle: #### 期望的现象(What you expected to happen) `sudo dnf update` 应该毫无违和感的运行下去 #### 如何重现(How to reproduce it) 不确定,报错看起来是随机出现的 #### 其他事项(Anything else we need to know) - 该问题是否被之前的 issue 提出过: #### 您的环境(Environment) - 操作系统(OS Version):Fedora workstation 33 - 浏览器(如果适用)(Browser version, if applicable): - 其他(Others):ISP: 中国联通 ---- <details> <summary>当时DNF部分输出</summary> ``` 事务概要 ==================================================================================== 安装 10 软件包 升级 310 软件包 移除 1 软件包 总下载:574 M 确定吗?[y/N]: Y 下载软件包: (1/320): python3-packaging-20.4-2.fc33.noarch.rpm 180 kB/s | 67 kB 00:00 (2/320): libsodium-1.0.18-6.fc33.x86_64.rpm 413 kB/s | 166 kB 00:00 [MIRROR] python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/releases/33/Everything/x86_64/os/Packages/p/python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) (3/320): python3-pyparsing-2.4.7-4.fc33.noarch.rpm 338 kB/s | 149 kB 00:00 [MIRROR] kernel-devel-5.13.9-100.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/updates/33/Everything/x86_64/Packages/k/kernel-devel-5.13.9-100.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [MIRROR] python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/releases/33/Everything/x86_64/os/Packages/p/python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [MIRROR] kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/updates/33/Everything/x86_64/Packages/k/kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [MIRROR] kernel-devel-5.13.9-100.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/updates/33/Everything/x86_64/Packages/k/kernel-devel-5.13.9-100.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [MIRROR] python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/releases/33/Everything/x86_64/os/Packages/p/python3-simpleaudio-1.0.4-2.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [MIRROR] kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm: Status code: 403 for https://mirrors.bfsu.edu.cn/fedora/updates/33/Everything/x86_64/Packages/k/kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm (IP: 2001:da8:20f:4435:4adf:37ff:fe55:2840) [FAILED] kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm: No more mirrors to try - All mirrors were already tried without success (5-6/320): python3-sim 0% [ ] 685 kB/s | 2.4 MB 14:13 ETA 下载的软件包保存在缓存中,直到下次成功执行事务。 您可以通过执行 'dnf clean packages' 删除软件包缓存。 错误:下载软件包出错 : Cannot download Packages/k/kernel-modules-extra-5.13.9-100.fc33.x86_64.rpm: All mirrors were tried ``` </details>
non_build
北外fedora镜像出现http 请使用此模板来报告 bug,并尽可能多地提供信息。 please use this template while reporting a bug and provide as much info as possible 发生了什么(what happened) sudo dnf update 时, ,但是我手动单独更新报错的那几个包时,又可以更新, 此后再执行 sudo dnf update 又会报错,周而复始 recycle 期望的现象(what you expected to happen) sudo dnf update 应该毫无违和感的运行下去 如何重现(how to reproduce it) 不确定,报错看起来是随机出现的 其他事项(anything else we need to know) 该问题是否被之前的 issue 提出过: 您的环境(environment) 操作系统(os version):fedora workstation 浏览器(如果适用)(browser version if applicable): 其他(others):isp 中国联通 当时dnf部分输出 事务概要 安装 软件包 升级 软件包 移除 软件包 总下载: m 确定吗? : y 下载软件包: packaging noarch rpm kb s kb libsodium rpm kb s kb simpleaudio rpm status code for ip pyparsing noarch rpm kb s kb kernel devel rpm status code for ip simpleaudio rpm status code for ip kernel modules extra rpm status code for ip kernel devel rpm status code for ip simpleaudio rpm status code for ip kernel modules extra rpm status code for ip kernel modules extra rpm no more mirrors to try all mirrors were already tried without success sim kb s mb eta 下载的软件包保存在缓存中,直到下次成功执行事务。 您可以通过执行 dnf clean packages 删除软件包缓存。 错误:下载软件包出错 cannot download packages k kernel modules extra rpm all mirrors were tried
0
5,762
3,653,744,423
IssuesEvent
2016-02-17 09:19:26
DevExpress/testcafe
https://api.github.com/repos/DevExpress/testcafe
closed
Update `testcafe-hammerhead` to v6.0.0
!IMPORTANT! AREA: build SYSTEM: hammerhead TYPE: enhancement
There is a breaking change in the https://github.com/DevExpress/testcafe-hammerhead/issues/410 PR. We should modify our code according to this.
1.0
Update `testcafe-hammerhead` to v6.0.0 - There is a breaking change in the https://github.com/DevExpress/testcafe-hammerhead/issues/410 PR. We should modify our code according to this.
build
update testcafe hammerhead to there is a breaking change in the pr we should modify our code according to this
1
510,976
14,851,234,316
IssuesEvent
2021-01-18 06:29:39
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.google.com - site is not usable
browser-firefox-mobile engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/65674 --> **URL**: https://www.google.com/ **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Yes Other **Problem type**: Site is not usable **Description**: Unable to login **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.google.com - site is not usable - <!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/65674 --> **URL**: https://www.google.com/ **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Yes Other **Problem type**: Site is not usable **Description**: Unable to login **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_build
site is not usable url browser version firefox mobile operating system android tested another browser yes other problem type site is not usable description unable to login steps to reproduce browser configuration none from with ❤️
0
139,866
31,799,970,324
IssuesEvent
2023-09-13 10:25:41
octomation/go-tool
https://api.github.com/repos/octomation/go-tool
closed
go: extended support of go version since 1.18
type: feature scope: code impact: medium effort: easy
**Motivation:** extract more information for "versioning". https://github.com/octomation/go-tool/blob/87401abb06ada0ee8b388b98945f1e88f35c74bf/main.go#L30-L35 See https://go.dev/doc/go1.18#go-command, > ### go version > > The go command now embeds version control information in binaries. It includes the currently checked-out revision, commit time, and a flag indicating whether edited or untracked files are present. Version control information is embedded if the go command is invoked in a directory within a Git, Mercurial, Fossil, or Bazaar repository, and the main package and its containing main module are in the same repository. This information may be omitted using the flag -buildvcs=false. > > Additionally, the go command embeds information about the build, including build and tool tags (set with -tags), compiler, assembler, and linker flags (like -gcflags), whether cgo was enabled, and if it was, the values of the cgo environment variables (like CGO_CFLAGS). Both VCS and build information may be read together with module information using go version -m file or runtime/debug.ReadBuildInfo (for the currently running binary) or the new [debug/buildinfo](https://go.dev/doc/go1.18#debug/buildinfo) package. > > The underlying data format of the embedded build information can change with new go releases, so an older version of go may not handle the build information produced with a newer version of go. To read the version information from a binary built with go 1.18, use the go version command and the debug/buildinfo package from go 1.18+.
1.0
go: extended support of go version since 1.18 - **Motivation:** extract more information for "versioning". https://github.com/octomation/go-tool/blob/87401abb06ada0ee8b388b98945f1e88f35c74bf/main.go#L30-L35 See https://go.dev/doc/go1.18#go-command, > ### go version > > The go command now embeds version control information in binaries. It includes the currently checked-out revision, commit time, and a flag indicating whether edited or untracked files are present. Version control information is embedded if the go command is invoked in a directory within a Git, Mercurial, Fossil, or Bazaar repository, and the main package and its containing main module are in the same repository. This information may be omitted using the flag -buildvcs=false. > > Additionally, the go command embeds information about the build, including build and tool tags (set with -tags), compiler, assembler, and linker flags (like -gcflags), whether cgo was enabled, and if it was, the values of the cgo environment variables (like CGO_CFLAGS). Both VCS and build information may be read together with module information using go version -m file or runtime/debug.ReadBuildInfo (for the currently running binary) or the new [debug/buildinfo](https://go.dev/doc/go1.18#debug/buildinfo) package. > > The underlying data format of the embedded build information can change with new go releases, so an older version of go may not handle the build information produced with a newer version of go. To read the version information from a binary built with go 1.18, use the go version command and the debug/buildinfo package from go 1.18+.
non_build
go extended support of go version since motivation extract more information for versioning see go version the go command now embeds version control information in binaries it includes the currently checked out revision commit time and a flag indicating whether edited or untracked files are present version control information is embedded if the go command is invoked in a directory within a git mercurial fossil or bazaar repository and the main package and its containing main module are in the same repository this information may be omitted using the flag buildvcs false additionally the go command embeds information about the build including build and tool tags set with tags compiler assembler and linker flags like gcflags whether cgo was enabled and if it was the values of the cgo environment variables like cgo cflags both vcs and build information may be read together with module information using go version m file or runtime debug readbuildinfo for the currently running binary or the new package the underlying data format of the embedded build information can change with new go releases so an older version of go may not handle the build information produced with a newer version of go to read the version information from a binary built with go use the go version command and the debug buildinfo package from go
0
599,529
18,276,427,363
IssuesEvent
2021-10-04 19:25:03
0perationPrivacy/VoIP
https://api.github.com/repos/0perationPrivacy/VoIP
closed
Icons Missing on Mobile
bug priority/p1
**Platform:** [Mobile] **Browser:** [ Brave] **App Version:** [13 Commits] **Provider:** [Telnyx, Twilio, Both] ### Describe the bug Compose Message Icon is missing on mobile. On brave browser on mobile I have on Night Mode (Makes Websites dark pages, easy on the eyes) Turning off the shield does nothing. If you click the invisible icon the 'compose message' opens Ps: This number is deleted and had nothing tied to me ever. Don't care if the number is public. :) **To Reproduce** 1. Go to Appearance in Brave 2. Enable Night Mode (Experimental) 3. Go to the website 4. Icon gone **Screenshots** Desktop: ![a](https://user-images.githubusercontent.com/90421861/132931830-ed1171b4-61e7-48d0-93de-c6cea647e6a0.png) Moblie: ![webapp](https://user-images.githubusercontent.com/90421861/132931833-e0a3b5a6-0eb2-4e71-b88d-95616386cf38.png)
1.0
Icons Missing on Mobile - **Platform:** [Mobile] **Browser:** [ Brave] **App Version:** [13 Commits] **Provider:** [Telnyx, Twilio, Both] ### Describe the bug Compose Message Icon is missing on mobile. On brave browser on mobile I have on Night Mode (Makes Websites dark pages, easy on the eyes) Turning off the shield does nothing. If you click the invisible icon the 'compose message' opens Ps: This number is deleted and had nothing tied to me ever. Don't care if the number is public. :) **To Reproduce** 1. Go to Appearance in Brave 2. Enable Night Mode (Experimental) 3. Go to the website 4. Icon gone **Screenshots** Desktop: ![a](https://user-images.githubusercontent.com/90421861/132931830-ed1171b4-61e7-48d0-93de-c6cea647e6a0.png) Moblie: ![webapp](https://user-images.githubusercontent.com/90421861/132931833-e0a3b5a6-0eb2-4e71-b88d-95616386cf38.png)
non_build
icons missing on mobile platform browser app version provider describe the bug compose message icon is missing on mobile on brave browser on mobile i have on night mode makes websites dark pages easy on the eyes turning off the shield does nothing if you click the invisible icon the compose message opens ps this number is deleted and had nothing tied to me ever don t care if the number is public to reproduce go to appearance in brave enable night mode experimental go to the website icon gone screenshots desktop moblie
0
97,984
29,171,761,092
IssuesEvent
2023-05-19 02:55:18
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
closed
[FEA] Bump up cudf, JNI, private version to 23.06.0-SNAPSHOT
build
**Is your feature request related to a problem? Please describe.** Currently we update plugin version to 23.06.0-SNAPSHOT in branch-23.06 to unblock development for new release cudf, JNI, private dependencies version will stay at 23.04 until we have new 23.06 artifacts available Create this issue to remind us later
1.0
[FEA] Bump up cudf, JNI, private version to 23.06.0-SNAPSHOT - **Is your feature request related to a problem? Please describe.** Currently we update plugin version to 23.06.0-SNAPSHOT in branch-23.06 to unblock development for new release cudf, JNI, private dependencies version will stay at 23.04 until we have new 23.06 artifacts available Create this issue to remind us later
build
bump up cudf jni private version to snapshot is your feature request related to a problem please describe currently we update plugin version to snapshot in branch to unblock development for new release cudf jni private dependencies version will stay at until we have new artifacts available create this issue to remind us later
1
43,480
11,233,998,915
IssuesEvent
2020-01-09 03:24:20
linewalks/MDwalks-UI
https://api.github.com/repos/linewalks/MDwalks-UI
closed
build 후에 version 이 실행되게
Build
## 기능 요청이 문제와 연관이 있습니까? 설명해주세요 tag 적용 후 build 가 되는 경우가 있습니다. ## 원하는 솔루션에 대한 설명 이건 수동으로 되어 생기는 실수로 version 명령어에 build 를 추가하여 해결 할 수 있습니다
1.0
build 후에 version 이 실행되게 - ## 기능 요청이 문제와 연관이 있습니까? 설명해주세요 tag 적용 후 build 가 되는 경우가 있습니다. ## 원하는 솔루션에 대한 설명 이건 수동으로 되어 생기는 실수로 version 명령어에 build 를 추가하여 해결 할 수 있습니다
build
build 후에 version 이 실행되게 기능 요청이 문제와 연관이 있습니까 설명해주세요 tag 적용 후 build 가 되는 경우가 있습니다 원하는 솔루션에 대한 설명 이건 수동으로 되어 생기는 실수로 version 명령어에 build 를 추가하여 해결 할 수 있습니다
1
182,529
14,139,559,153
IssuesEvent
2020-11-10 09:59:40
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
closed
Intermittent UI test failure - SmokeTest.verifyPageMainMenuItemsListInPortraitNormalModeTest
Feature:MainMenu disabled-test eng:ui-test intermittent-test
### Firebase Test Run: https://console.firebase.google.com/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7596046872173049036/executions/bs.f180e42384712dd/testcases/2/test-cases ### Stacktrace: ### Build: 7/26 Flaky test: failed 1 out of 2 tries. cc: @TejaswiKarasani
3.0
Intermittent UI test failure - SmokeTest.verifyPageMainMenuItemsListInPortraitNormalModeTest - ### Firebase Test Run: https://console.firebase.google.com/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/7596046872173049036/executions/bs.f180e42384712dd/testcases/2/test-cases ### Stacktrace: ### Build: 7/26 Flaky test: failed 1 out of 2 tries. cc: @TejaswiKarasani
non_build
intermittent ui test failure smoketest verifypagemainmenuitemslistinportraitnormalmodetest firebase test run stacktrace build flaky test failed out of tries cc tejaswikarasani
0
49,577
26,217,095,184
IssuesEvent
2023-01-04 11:55:56
keras-team/keras
https://api.github.com/repos/keras-team/keras
closed
Significant slowdown in Keras model.fit() on simple problem when using validation data
type:bug/performance stat:awaiting response from contributor stalled
**System information**. - Have I written custom code: No - OS Platform and Distribution: Ubuntu 18.04 - TensorFlow installed from: Binary - TensorFlow version: 2.11.0 - Python version: 3.10.8 - Bazel version: N/A - GPU model and memory: NVIDIA Quadro GV100 (32 GB) - Exact command to reproduce: See below. **Describe the problem**. Per instructions I received from @mohantym, this is a cross-post to this repository of the issue tensorflow/tensorflow#58828, which I opened yesterday. The issue documents a significant slowdown in training speed when using Keras model.fit() with the validation_data parameter set. On my hardware, I see a slowdown of approximately 10x. On Colab, @mohantym sees a slowdown of only 3.5x, but this still seems excessive to me given that the validation data is only 1/8 the size of the training data. I would not expect it to take 3.5x time to process 12.5% more data per epoch. (In fact, I see the same slowdown even if I reduce the validation set size to a single example, so it's a 3.5x to 10x speed penalty for an 0.05% increase in the amount of data to be processed.) For further details, including code that can be used to reproduce the problem, please see the aforementioned issue. **Describe the current behavior**. A 3.5x-10x slowdown during training using model.fit() with validation_data set. **Describe the expected behavior**. Some slowdown (validation isn't free)---but a whole lot less than 3.5x-10x. **Standalone code to reproduce the issue**. See the aforementioned issue. **Source code / logs**. See the aforementioned issue.
True
Significant slowdown in Keras model.fit() on simple problem when using validation data - **System information**. - Have I written custom code: No - OS Platform and Distribution: Ubuntu 18.04 - TensorFlow installed from: Binary - TensorFlow version: 2.11.0 - Python version: 3.10.8 - Bazel version: N/A - GPU model and memory: NVIDIA Quadro GV100 (32 GB) - Exact command to reproduce: See below. **Describe the problem**. Per instructions I received from @mohantym, this is a cross-post to this repository of the issue tensorflow/tensorflow#58828, which I opened yesterday. The issue documents a significant slowdown in training speed when using Keras model.fit() with the validation_data parameter set. On my hardware, I see a slowdown of approximately 10x. On Colab, @mohantym sees a slowdown of only 3.5x, but this still seems excessive to me given that the validation data is only 1/8 the size of the training data. I would not expect it to take 3.5x time to process 12.5% more data per epoch. (In fact, I see the same slowdown even if I reduce the validation set size to a single example, so it's a 3.5x to 10x speed penalty for an 0.05% increase in the amount of data to be processed.) For further details, including code that can be used to reproduce the problem, please see the aforementioned issue. **Describe the current behavior**. A 3.5x-10x slowdown during training using model.fit() with validation_data set. **Describe the expected behavior**. Some slowdown (validation isn't free)---but a whole lot less than 3.5x-10x. **Standalone code to reproduce the issue**. See the aforementioned issue. **Source code / logs**. See the aforementioned issue.
non_build
significant slowdown in keras model fit on simple problem when using validation data system information have i written custom code no os platform and distribution ubuntu tensorflow installed from binary tensorflow version python version bazel version n a gpu model and memory nvidia quadro gb exact command to reproduce see below describe the problem per instructions i received from mohantym this is a cross post to this repository of the issue tensorflow tensorflow which i opened yesterday the issue documents a significant slowdown in training speed when using keras model fit with the validation data parameter set on my hardware i see a slowdown of approximately on colab mohantym sees a slowdown of only but this still seems excessive to me given that the validation data is only the size of the training data i would not expect it to take time to process more data per epoch in fact i see the same slowdown even if i reduce the validation set size to a single example so it s a to speed penalty for an increase in the amount of data to be processed for further details including code that can be used to reproduce the problem please see the aforementioned issue describe the current behavior a slowdown during training using model fit with validation data set describe the expected behavior some slowdown validation isn t free but a whole lot less than standalone code to reproduce the issue see the aforementioned issue source code logs see the aforementioned issue
0
100,244
12,510,882,704
IssuesEvent
2020-06-02 19:29:42
patternfly/patternfly-design
https://api.github.com/repos/patternfly/patternfly-design
opened
Review Topology View Contribution
Interaction design New Feature Visual Design
Review proposed topology contribution to provide design input, @mceledonia @LHinson feel free to add more detail here.
2.0
Review Topology View Contribution - Review proposed topology contribution to provide design input, @mceledonia @LHinson feel free to add more detail here.
non_build
review topology view contribution review proposed topology contribution to provide design input mceledonia lhinson feel free to add more detail here
0
89,063
25,571,732,703
IssuesEvent
2022-11-30 18:15:30
dealii/dealii
https://api.github.com/repos/dealii/dealii
closed
Cannot build with CUDA enabled and C++17
Build system GPU C++17
Hey all, I can't compile the library with CUDA enabled and C++17. Somehow it looks like we inject `-std=c++14` in the cuda flags and `nvcc` does not like to have both c++14 and c++17. Unfortunately, I can't find where the `-std=c++14` flag is added. You can see in the [detailed.log](https://github.com/dealii/dealii/files/5335799/detailed.log) that the CUDA flags are correct: ``` # DEAL_II_CUDA_FLAGS = -arch=sm_70 -std=c++17 # DEAL_II_CUDA_FLAGS_RELEASE = # DEAL_II_CUDA_FLAGS_DEBUG = -G ``` but the CUDA flags that are used are different (from [source/lac/CMakeFiles/obj_lac_debug.dir/flags.make](https://github.com/dealii/dealii/files/5335828/source-lac-CMakeFiles-obj_lac_debug.dir-flags.make.txt)) ``` -arch=sm_70 -std=c++17 -G -std=c++14 ``` I obviously tried to grep for `-std=c++14` but I couldn't find it. Looking at [this](https://github.com/dealii/dealii/blob/master/cmake/configure/configure_1_cuda.cmake#L151) I would expect `-G` to be the last flag. Any idea how to debug this?
1.0
Cannot build with CUDA enabled and C++17 - Hey all, I can't compile the library with CUDA enabled and C++17. Somehow it looks like we inject `-std=c++14` in the cuda flags and `nvcc` does not like to have both c++14 and c++17. Unfortunately, I can't find where the `-std=c++14` flag is added. You can see in the [detailed.log](https://github.com/dealii/dealii/files/5335799/detailed.log) that the CUDA flags are correct: ``` # DEAL_II_CUDA_FLAGS = -arch=sm_70 -std=c++17 # DEAL_II_CUDA_FLAGS_RELEASE = # DEAL_II_CUDA_FLAGS_DEBUG = -G ``` but the CUDA flags that are used are different (from [source/lac/CMakeFiles/obj_lac_debug.dir/flags.make](https://github.com/dealii/dealii/files/5335828/source-lac-CMakeFiles-obj_lac_debug.dir-flags.make.txt)) ``` -arch=sm_70 -std=c++17 -G -std=c++14 ``` I obviously tried to grep for `-std=c++14` but I couldn't find it. Looking at [this](https://github.com/dealii/dealii/blob/master/cmake/configure/configure_1_cuda.cmake#L151) I would expect `-G` to be the last flag. Any idea how to debug this?
build
cannot build with cuda enabled and c hey all i can t compile the library with cuda enabled and c somehow it looks like we inject std c in the cuda flags and nvcc does not like to have both c and c unfortunately i can t find where the std c flag is added you can see in the that the cuda flags are correct deal ii cuda flags arch sm std c deal ii cuda flags release deal ii cuda flags debug g but the cuda flags that are used are different from arch sm std c g std c i obviously tried to grep for std c but i couldn t find it looking at i would expect g to be the last flag any idea how to debug this
1
358,983
10,652,464,824
IssuesEvent
2019-10-17 12:40:40
AY1920S1-CS2103T-F14-3/main
https://api.github.com/repos/AY1920S1-CS2103T-F14-3/main
closed
Add conflict-checking between engagements
Type.Enhancement priority.High
Two engagements cannot be scheduled at overlapping time intervals
1.0
Add conflict-checking between engagements - Two engagements cannot be scheduled at overlapping time intervals
non_build
add conflict checking between engagements two engagements cannot be scheduled at overlapping time intervals
0
157,320
12,370,222,666
IssuesEvent
2020-05-18 16:26:29
istio/istio
https://api.github.com/repos/istio/istio
closed
TLS origination for egress traffic gets 503
area/networking community/testing days
**Bug description** I am following the user guide and testing this step https://preliminary.istio.io/docs/tasks/traffic-management/egress/egress-tls-origination/#tls-origination-for-egress-traffic $kubectl exec -it $SOURCE_POD -c sleep -- curl -sL -o /dev/null -D - http://edition.cnn.com/politics HTTP/1.1 503 Service Unavailable date: Wed, 13 May 2020 18:32:28 GMT server: envoy content-length: 0 curl to https succeeds $kubectl exec -it $SOURCE_POD -c sleep -- curl -sL -o /dev/null -D - https://edition.cnn.com/politics HTTP/2 200 content-type: text/html; charset=utf-8 x-servedbyhost: ::ffff:127.0.0.1 **Expected behavior** curl command should return 200 **Steps to reproduce the bug** **Version (include the output of `istioctl version --remote` and `kubectl version` and `helm version` if you used Helm)** $istioctl version client version: 1.6.0-beta.1 control plane version: 1.6.0-beta.1 data plane version: 1.6.0-beta.1 (3 proxies) **How was Istio installed?** $istioctl manifest apply **Environment where bug was observed (cloud vendor, OS, etc)** GKE
1.0
TLS origination for egress traffic gets 503 - **Bug description** I am following the user guide and testing this step https://preliminary.istio.io/docs/tasks/traffic-management/egress/egress-tls-origination/#tls-origination-for-egress-traffic $kubectl exec -it $SOURCE_POD -c sleep -- curl -sL -o /dev/null -D - http://edition.cnn.com/politics HTTP/1.1 503 Service Unavailable date: Wed, 13 May 2020 18:32:28 GMT server: envoy content-length: 0 curl to https succeeds $kubectl exec -it $SOURCE_POD -c sleep -- curl -sL -o /dev/null -D - https://edition.cnn.com/politics HTTP/2 200 content-type: text/html; charset=utf-8 x-servedbyhost: ::ffff:127.0.0.1 **Expected behavior** curl command should return 200 **Steps to reproduce the bug** **Version (include the output of `istioctl version --remote` and `kubectl version` and `helm version` if you used Helm)** $istioctl version client version: 1.6.0-beta.1 control plane version: 1.6.0-beta.1 data plane version: 1.6.0-beta.1 (3 proxies) **How was Istio installed?** $istioctl manifest apply **Environment where bug was observed (cloud vendor, OS, etc)** GKE
non_build
tls origination for egress traffic gets bug description i am following the user guide and testing this step kubectl exec it source pod c sleep curl sl o dev null d http service unavailable date wed may gmt server envoy content length curl to https succeeds kubectl exec it source pod c sleep curl sl o dev null d http content type text html charset utf x servedbyhost ffff expected behavior curl command should return steps to reproduce the bug version include the output of istioctl version remote and kubectl version and helm version if you used helm istioctl version client version beta control plane version beta data plane version beta proxies how was istio installed istioctl manifest apply environment where bug was observed cloud vendor os etc gke
0
310,353
26,711,648,687
IssuesEvent
2023-01-28 01:18:22
opentibiabr/canary
https://api.github.com/repos/opentibiabr/canary
closed
Changing loot type to party hunt leader and disconect character!
Type: Bug Priority: High Status: Pending Test
### Priority High ### Area - [X] Datapack - [X] Source - [ ] Map - [ ] Other ### What happened? just changing loot type. ![image](https://user-images.githubusercontent.com/82473579/201743530-41583e18-f538-46b1-ad43-d641de80cfc3.png) ### What OS are you seeing the problem on? Linux, Windows ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
Changing loot type to party hunt leader and disconect character! - ### Priority High ### Area - [X] Datapack - [X] Source - [ ] Map - [ ] Other ### What happened? just changing loot type. ![image](https://user-images.githubusercontent.com/82473579/201743530-41583e18-f538-46b1-ad43-d641de80cfc3.png) ### What OS are you seeing the problem on? Linux, Windows ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
non_build
changing loot type to party hunt leader and disconect character priority high area datapack source map other what happened just changing loot type what os are you seeing the problem on linux windows code of conduct i agree to follow this project s code of conduct
0
91,690
26,463,299,948
IssuesEvent
2023-01-16 20:05:09
sandermvanvliet/RoadCaptain
https://api.github.com/repos/sandermvanvliet/RoadCaptain
closed
Update documentation
documentation runner routebuilder
Current readme is out of date as it hasn’t been updated since 0.2.0 or so. Should set up github pages or something similar on the new domain.
1.0
Update documentation - Current readme is out of date as it hasn’t been updated since 0.2.0 or so. Should set up github pages or something similar on the new domain.
build
update documentation current readme is out of date as it hasn’t been updated since or so should set up github pages or something similar on the new domain
1
138,250
18,773,430,101
IssuesEvent
2021-11-07 08:38:18
sultanabubaker/gradle-simple-project
https://api.github.com/repos/sultanabubaker/gradle-simple-project
closed
CVE-2017-5929 (High) detected in logback-classic-1.1.3.jar - autoclosed
security vulnerability
## CVE-2017-5929 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>logback-classic-1.1.3.jar</b></p></summary> <p>logback-classic module</p> <p>Library home page: <a href="http://logback.qos.ch">http://logback.qos.ch</a></p> <p>Path to dependency file: gradle-simple-project/build.gradle</p> <p>Path to vulnerable library: modules-2/files-2.1/ch.qos.logback/logback-classic/1.1.3/d90276fff414f06cb375f2057f6778cd63c6082f/logback-classic-1.1.3.jar</p> <p> Dependency Hierarchy: - :x: **logback-classic-1.1.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/gradle-simple-project/commit/ecccd58338e1dd8f9ce34bb4ff8bd3cc77eb6af0">ecccd58338e1dd8f9ce34bb4ff8bd3cc77eb6af0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> QOS.ch Logback before 1.2.0 has a serialization vulnerability affecting the SocketServer and ServerSocketReceiver components. <p>Publish Date: 2017-03-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5929>CVE-2017-5929</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5929">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5929</a></p> <p>Release Date: 2017-03-13</p> <p>Fix Resolution: ch.qos.logback:logback-core:1.2.0;ch.qos.logback:logback-access:1.2.0;ch.qos.logback:logback-classic:1.2.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"ch.qos.logback","packageName":"logback-classic","packageVersion":"1.1.3","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"ch.qos.logback:logback-classic:1.1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ch.qos.logback:logback-core:1.2.0;ch.qos.logback:logback-access:1.2.0;ch.qos.logback:logback-classic:1.2.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-5929","vulnerabilityDetails":"QOS.ch Logback before 1.2.0 has a serialization vulnerability affecting the SocketServer and ServerSocketReceiver components.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5929","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2017-5929 (High) detected in logback-classic-1.1.3.jar - autoclosed - ## CVE-2017-5929 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>logback-classic-1.1.3.jar</b></p></summary> <p>logback-classic module</p> <p>Library home page: <a href="http://logback.qos.ch">http://logback.qos.ch</a></p> <p>Path to dependency file: gradle-simple-project/build.gradle</p> <p>Path to vulnerable library: modules-2/files-2.1/ch.qos.logback/logback-classic/1.1.3/d90276fff414f06cb375f2057f6778cd63c6082f/logback-classic-1.1.3.jar</p> <p> Dependency Hierarchy: - :x: **logback-classic-1.1.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/gradle-simple-project/commit/ecccd58338e1dd8f9ce34bb4ff8bd3cc77eb6af0">ecccd58338e1dd8f9ce34bb4ff8bd3cc77eb6af0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> QOS.ch Logback before 1.2.0 has a serialization vulnerability affecting the SocketServer and ServerSocketReceiver components. <p>Publish Date: 2017-03-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5929>CVE-2017-5929</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5929">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5929</a></p> <p>Release Date: 2017-03-13</p> <p>Fix Resolution: ch.qos.logback:logback-core:1.2.0;ch.qos.logback:logback-access:1.2.0;ch.qos.logback:logback-classic:1.2.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"ch.qos.logback","packageName":"logback-classic","packageVersion":"1.1.3","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"ch.qos.logback:logback-classic:1.1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ch.qos.logback:logback-core:1.2.0;ch.qos.logback:logback-access:1.2.0;ch.qos.logback:logback-classic:1.2.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-5929","vulnerabilityDetails":"QOS.ch Logback before 1.2.0 has a serialization vulnerability affecting the SocketServer and ServerSocketReceiver components.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-5929","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_build
cve high detected in logback classic jar autoclosed cve high severity vulnerability vulnerable library logback classic jar logback classic module library home page a href path to dependency file gradle simple project build gradle path to vulnerable library modules files ch qos logback logback classic logback classic jar dependency hierarchy x logback classic jar vulnerable library found in head commit a href found in base branch master vulnerability details qos ch logback before has a serialization vulnerability affecting the socketserver and serversocketreceiver components publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ch qos logback logback core ch qos logback logback access ch qos logback logback classic isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree ch qos logback logback classic isminimumfixversionavailable true minimumfixversion ch qos logback logback core ch qos logback logback access ch qos logback logback classic basebranches vulnerabilityidentifier cve vulnerabilitydetails qos ch logback before has a serialization vulnerability affecting the socketserver and serversocketreceiver components vulnerabilityurl
0
108
2,509,128,754
IssuesEvent
2015-01-13 11:09:12
mediumorg/medium
https://api.github.com/repos/mediumorg/medium
opened
Migrate build files to CoffeeScript
build CoffeeScript help-wanted
The build system would look a lot nicer if it were running on coffeescript because it is quite declarative in nature. Thus, it would be great to migrate existing JavaScript such that new development could be done using coffee. [This article](http://pem-musing.blogspot.de/2014/02/a-gulp-of-coffee-your-gulpfile-in.html) illustrates how gulp files look when written in coffeescript
1.0
Migrate build files to CoffeeScript - The build system would look a lot nicer if it were running on coffeescript because it is quite declarative in nature. Thus, it would be great to migrate existing JavaScript such that new development could be done using coffee. [This article](http://pem-musing.blogspot.de/2014/02/a-gulp-of-coffee-your-gulpfile-in.html) illustrates how gulp files look when written in coffeescript
build
migrate build files to coffeescript the build system would look a lot nicer if it were running on coffeescript because it is quite declarative in nature thus it would be great to migrate existing javascript such that new development could be done using coffee illustrates how gulp files look when written in coffeescript
1
77,587
21,900,832,751
IssuesEvent
2022-05-20 13:16:31
nokeedev/gradle-native
https://api.github.com/repos/nokeedev/gradle-native
opened
Support Xcode scheme and workspace in build adapter
a:enhancement in:xcode-build-adapter
For repositories that use Xcode workspace, we need to pass the workspace into the projects as well as use schemes instead of targets. It's also important to use Xcode workspace when using cocoapod as the project and pods are in two different projects which are _linked_ using a workspace.
1.0
Support Xcode scheme and workspace in build adapter - For repositories that use Xcode workspace, we need to pass the workspace into the projects as well as use schemes instead of targets. It's also important to use Xcode workspace when using cocoapod as the project and pods are in two different projects which are _linked_ using a workspace.
build
support xcode scheme and workspace in build adapter for repositories that use xcode workspace we need to pass the workspace into the projects as well as use schemes instead of targets it s also important to use xcode workspace when using cocoapod as the project and pods are in two different projects which are linked using a workspace
1
54,760
13,447,220,501
IssuesEvent
2020-09-08 13:58:56
ApolloAuto/apollo
https://api.github.com/repos/ApolloAuto/apollo
closed
ERROR: cannot load '//tools:cuda_library.bzl': no such file
Good first issue Module: Build Type: Help wanted
Hi, See following: ``` $ ./apollo.sh config [INFO] Apollo Environment Settings: [INFO] APOLLO_ROOT_DIR: /apollo [INFO] APOLLO_CACHE_DIR: /apollo/.cache [INFO] APOLLO_IN_DOCKER: true [INFO] APOLLO_VERSION: master-2020-09-08-2aaa67e179 [INFO] DOCKER_IMG: dev-x86_64-18.04-20200823_0534 [INFO] APOLLO_ENV: STAGE=dev USE_ESD_CAN=false [INFO] USE_GPU: USE_GPU_HOST=1 USE_GPU_TARGET=1 $ ./apollo.sh build_gpu [INFO] Apollo Environment Settings: [INFO] APOLLO_ROOT_DIR: /apollo [INFO] APOLLO_CACHE_DIR: /apollo/.cache [INFO] APOLLO_IN_DOCKER: true [INFO] APOLLO_VERSION: master-2020-09-08-2aaa67e179 [INFO] DOCKER_IMG: dev-x86_64-18.04-20200823_0534 [INFO] APOLLO_ENV: STAGE=dev USE_ESD_CAN=false [INFO] USE_GPU: USE_GPU_HOST=1 USE_GPU_TARGET=1 [ OK ] Running GPU build on x86_64 platform. [WARNING] ESD CAN library supplied by ESD Electronics doesn't exist. [WARNING] If you need ESD CAN, please refer to: [WARNING] third_party/can_card_library/esd_can/README.md [INFO] Build Overview: [INFO] USE_GPU: 1 [ 0 for CPU, 1 for GPU ] [INFO] Bazel Options: --config=gpu [INFO] Build Targets: //modules/... union //cyber/... [INFO] Disabled: except //modules/drivers/canbus/can_client/esd/... (23:09:04) INFO: Current date is 2020-09-07 (23:09:04) ERROR: cannot load '//tools:cuda_library.bzl': no such file (23:09:04) INFO: Elapsed time: 0.481s (23:09:04) INFO: 0 processes. (23:09:04) FAILED: Build did NOT complete successfully (172 packages loaded) currently loading: modules/map/pnc_map ... (3 packages) Fetching @local_config_cuda; fetching ``` Any hints for this problem? Regards, Hy
1.0
ERROR: cannot load '//tools:cuda_library.bzl': no such file - Hi, See following: ``` $ ./apollo.sh config [INFO] Apollo Environment Settings: [INFO] APOLLO_ROOT_DIR: /apollo [INFO] APOLLO_CACHE_DIR: /apollo/.cache [INFO] APOLLO_IN_DOCKER: true [INFO] APOLLO_VERSION: master-2020-09-08-2aaa67e179 [INFO] DOCKER_IMG: dev-x86_64-18.04-20200823_0534 [INFO] APOLLO_ENV: STAGE=dev USE_ESD_CAN=false [INFO] USE_GPU: USE_GPU_HOST=1 USE_GPU_TARGET=1 $ ./apollo.sh build_gpu [INFO] Apollo Environment Settings: [INFO] APOLLO_ROOT_DIR: /apollo [INFO] APOLLO_CACHE_DIR: /apollo/.cache [INFO] APOLLO_IN_DOCKER: true [INFO] APOLLO_VERSION: master-2020-09-08-2aaa67e179 [INFO] DOCKER_IMG: dev-x86_64-18.04-20200823_0534 [INFO] APOLLO_ENV: STAGE=dev USE_ESD_CAN=false [INFO] USE_GPU: USE_GPU_HOST=1 USE_GPU_TARGET=1 [ OK ] Running GPU build on x86_64 platform. [WARNING] ESD CAN library supplied by ESD Electronics doesn't exist. [WARNING] If you need ESD CAN, please refer to: [WARNING] third_party/can_card_library/esd_can/README.md [INFO] Build Overview: [INFO] USE_GPU: 1 [ 0 for CPU, 1 for GPU ] [INFO] Bazel Options: --config=gpu [INFO] Build Targets: //modules/... union //cyber/... [INFO] Disabled: except //modules/drivers/canbus/can_client/esd/... (23:09:04) INFO: Current date is 2020-09-07 (23:09:04) ERROR: cannot load '//tools:cuda_library.bzl': no such file (23:09:04) INFO: Elapsed time: 0.481s (23:09:04) INFO: 0 processes. (23:09:04) FAILED: Build did NOT complete successfully (172 packages loaded) currently loading: modules/map/pnc_map ... (3 packages) Fetching @local_config_cuda; fetching ``` Any hints for this problem? Regards, Hy
build
error cannot load tools cuda library bzl no such file hi see following apollo sh config apollo environment settings apollo root dir apollo apollo cache dir apollo cache apollo in docker true apollo version master docker img dev apollo env stage dev use esd can false use gpu use gpu host use gpu target apollo sh build gpu apollo environment settings apollo root dir apollo apollo cache dir apollo cache apollo in docker true apollo version master docker img dev apollo env stage dev use esd can false use gpu use gpu host use gpu target running gpu build on platform esd can library supplied by esd electronics doesn t exist if you need esd can please refer to third party can card library esd can readme md build overview use gpu bazel options config gpu build targets modules union cyber disabled except modules drivers canbus can client esd info current date is error cannot load tools cuda library bzl no such file info elapsed time info processes failed build did not complete successfully packages loaded currently loading modules map pnc map packages fetching local config cuda fetching any hints for this problem regards hy
1
431,087
30,217,550,182
IssuesEvent
2023-07-05 16:39:48
cloudflare/cloudflare-docs
https://api.github.com/repos/cloudflare/cloudflare-docs
closed
Pages missing from "Available Notifications"
documentation content:edit
### Which Cloudflare product does this pertain to? Fundamentals ### Existing documentation URL(s) https://developers.cloudflare.com/fundamentals/notifications/notification-available/ ### What changes are you suggesting? We need to add Pages to this list. We have a few event types available: ![image](https://github.com/cloudflare/cloudflare-docs/assets/8492901/0cf2de65-c31d-4105-b5a6-62cedbb06a85) and you can pick which Pages projects trigger these events and send notifications. cc: @deadlypants1973 ### Additional information _No response_
1.0
Pages missing from "Available Notifications" - ### Which Cloudflare product does this pertain to? Fundamentals ### Existing documentation URL(s) https://developers.cloudflare.com/fundamentals/notifications/notification-available/ ### What changes are you suggesting? We need to add Pages to this list. We have a few event types available: ![image](https://github.com/cloudflare/cloudflare-docs/assets/8492901/0cf2de65-c31d-4105-b5a6-62cedbb06a85) and you can pick which Pages projects trigger these events and send notifications. cc: @deadlypants1973 ### Additional information _No response_
non_build
pages missing from available notifications which cloudflare product does this pertain to fundamentals existing documentation url s what changes are you suggesting we need to add pages to this list we have a few event types available and you can pick which pages projects trigger these events and send notifications cc additional information no response
0
271,660
29,659,331,463
IssuesEvent
2023-06-10 01:17:12
pazhanivel07/linux-4.19.72
https://api.github.com/repos/pazhanivel07/linux-4.19.72
closed
CVE-2022-26966 (Medium) detected in linuxlinux-4.19.83 - autoclosed
Mend: dependency security vulnerability
## CVE-2022-26966 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.83</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/linux-4.19.72/commit/ce28e4f7a922d93d9b737061ae46827305c8c30a">ce28e4f7a922d93d9b737061ae46827305c8c30a</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/usb/sr9700.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/usb/sr9700.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.16.12. drivers/net/usb/sr9700.c allows attackers to obtain sensitive information from heap memory via crafted frame lengths from a device. <p>Publish Date: 2022-03-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-26966>CVE-2022-26966</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-26966">https://www.linuxkernelcves.com/cves/CVE-2022-26966</a></p> <p>Release Date: 2022-03-12</p> <p>Fix Resolution: v4.9.304,v4.14.269,v4.19.232,v5.4.182,v5.10.103,v5.15.26,v5.16.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-26966 (Medium) detected in linuxlinux-4.19.83 - autoclosed - ## CVE-2022-26966 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.83</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/linux-4.19.72/commit/ce28e4f7a922d93d9b737061ae46827305c8c30a">ce28e4f7a922d93d9b737061ae46827305c8c30a</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/usb/sr9700.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/usb/sr9700.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.16.12. drivers/net/usb/sr9700.c allows attackers to obtain sensitive information from heap memory via crafted frame lengths from a device. <p>Publish Date: 2022-03-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-26966>CVE-2022-26966</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-26966">https://www.linuxkernelcves.com/cves/CVE-2022-26966</a></p> <p>Release Date: 2022-03-12</p> <p>Fix Resolution: v4.9.304,v4.14.269,v4.19.232,v5.4.182,v5.10.103,v5.15.26,v5.16.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_build
cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files drivers net usb c drivers net usb c vulnerability details an issue was discovered in the linux kernel before drivers net usb c allows attackers to obtain sensitive information from heap memory via crafted frame lengths from a device publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
382,544
26,502,934,583
IssuesEvent
2023-01-18 11:42:36
eclipse-ee4j/nosql
https://api.github.com/repos/eclipse-ee4j/nosql
opened
A deep review of Jakarta NoSQL after the only mapping decision
documentation
After we decided to move Jakarta NoSQL only with the Mapping side and the regular part, we removed a considerable amount of code keeping: The template interfaces Mapping annotations The goal would be to review and check the following points: - Check if the README file is updated. - Check if the spec has been updated. - Check if the JavaDoc is ready. Also, you can say that all feedbacks are welcome.
1.0
A deep review of Jakarta NoSQL after the only mapping decision - After we decided to move Jakarta NoSQL only with the Mapping side and the regular part, we removed a considerable amount of code keeping: The template interfaces Mapping annotations The goal would be to review and check the following points: - Check if the README file is updated. - Check if the spec has been updated. - Check if the JavaDoc is ready. Also, you can say that all feedbacks are welcome.
non_build
a deep review of jakarta nosql after the only mapping decision after we decided to move jakarta nosql only with the mapping side and the regular part we removed a considerable amount of code keeping the template interfaces mapping annotations the goal would be to review and check the following points check if the readme file is updated check if the spec has been updated check if the javadoc is ready also you can say that all feedbacks are welcome
0
126,427
4,995,444,199
IssuesEvent
2016-12-09 10:10:20
emoncms/MyHomeEnergyPlanner
https://api.github.com/repos/emoncms/MyHomeEnergyPlanner
opened
empty libraries
For release Medium priority
When you apply a fabric measure,if the library is empty, it tries to display walls but the list of leasure is empty. The reason is that the library is empty no that there is a bug. Anyway adding a message saying _This library is empty_ would look nicer
1.0
empty libraries - When you apply a fabric measure,if the library is empty, it tries to display walls but the list of leasure is empty. The reason is that the library is empty no that there is a bug. Anyway adding a message saying _This library is empty_ would look nicer
non_build
empty libraries when you apply a fabric measure if the library is empty it tries to display walls but the list of leasure is empty the reason is that the library is empty no that there is a bug anyway adding a message saying this library is empty would look nicer
0
142,194
11,458,024,392
IssuesEvent
2020-02-07 01:48:24
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
[Flaky Test] Stage
kind/flake priority/critical-urgent sig/testing
**Which jobs are flaking**: pull-kubernetes-e2e-gce pull-kubernetes-e2e-gce-rbe **Which test(s) are flaking**: Stage **Testgrid link**: https://testgrid.k8s.io/presubmits-kubernetes-nonblocking#pull-kubernetes-e2e-gce-rbe&include-filter-by-regex=Overall%7CBuild%7CStage https://testgrid.k8s.io/presubmits-kubernetes-blocking#pull-kubernetes-e2e-gce&include-filter-by-regex=Overall%7CBuild%7CStage **Reason for failure**: This isn't technically a test that's failing, it's part of the build->stage->up->test cycle used by our e2e jobs. There are two main clusters of failures that I can tell. One is timeouts. The other looks like: ``` - Hashing and copying public release artifacts to gs://kubernetes-release-pull/ci/pull-kubernetes-e2e-gce/v1.18.0-alpha.2.14+1b7738cc4ebdbe: $TEST_TMPDIR defined: output root default is '/bazel-scratch/.cache/bazel' and max_idle_secs default is '15'. Server terminated abruptly (error code: 14, error message: '', log file: '/bazel-scratch/.cache/bazel/_bazel_prow/cae228f2a89ef5ee47c2085e441a3561/server/jvm.out') Signal ERR caught! Traceback (line function script): 208 main /home/prow/go/src/k8s.io/release/push-build.sh Exiting... ``` Example -gce jobs: - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/85861/pull-kubernetes-e2e-gce/1219647945268269056 (#85861) - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/85846/pull-kubernetes-e2e-gce/1219646811342376960 (#85846) https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/87422/pull-kubernetes-e2e-gce/1219612336432615426 (#87422) - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/87372/pull-kubernetes-e2e-gce/1219266675761745921 (#87372) Example -gce-rbe jobs: - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/87435/pull-kubernetes-e2e-gce-rbe/1219718868767870977 - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/87266/pull-kubernetes-e2e-gce-rbe/1219655997224652803 - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/86408/pull-kubernetes-e2e-gce-rbe/1219620136827228160 (source gist https://gist.github.com/spiffxp/ed54d006e630b8fd8126ddfe67d9dc1e) **Anything else we need to know**: - Triage link for the job that actually reports to PRs: https://storage.googleapis.com/k8s-gubernator/triage/index.html?pr=1&job=gce%24&test=Stage - [The line where the crash is happening]( https://github.com/kubernetes/release/blob/f59f6ba61ac5f225cde91ff0fc179a0a6f762841/lib/releaselib.sh#L1148) ends up running `bazel run //:push-build $gcs_stage $gcs_destination` - Some googling for "bazel error code 14" leads me to believe that bazel (or its jvm) is getting killed due to OOM. [This job requests 6Gi of memory](https://github.com/kubernetes/test-infra/blob/b2471685eed6a7d063d7e1e19032282bb33679db/config/jobs/kubernetes/sig-cloud-provider/gcp/gcp-gce.yaml#L59-L61) which is a value that's been cargo culted around as far as I can tell. I was unable to find the decision that led to this value. I'm wondering if a bump to 8Gi of memory would help reduce occurrences of this flake. - I'm referencing the -rbe jobs to point out that this happens even if we use RBE, but they're not blocking
1.0
[Flaky Test] Stage - **Which jobs are flaking**: pull-kubernetes-e2e-gce pull-kubernetes-e2e-gce-rbe **Which test(s) are flaking**: Stage **Testgrid link**: https://testgrid.k8s.io/presubmits-kubernetes-nonblocking#pull-kubernetes-e2e-gce-rbe&include-filter-by-regex=Overall%7CBuild%7CStage https://testgrid.k8s.io/presubmits-kubernetes-blocking#pull-kubernetes-e2e-gce&include-filter-by-regex=Overall%7CBuild%7CStage **Reason for failure**: This isn't technically a test that's failing, it's part of the build->stage->up->test cycle used by our e2e jobs. There are two main clusters of failures that I can tell. One is timeouts. The other looks like: ``` - Hashing and copying public release artifacts to gs://kubernetes-release-pull/ci/pull-kubernetes-e2e-gce/v1.18.0-alpha.2.14+1b7738cc4ebdbe: $TEST_TMPDIR defined: output root default is '/bazel-scratch/.cache/bazel' and max_idle_secs default is '15'. Server terminated abruptly (error code: 14, error message: '', log file: '/bazel-scratch/.cache/bazel/_bazel_prow/cae228f2a89ef5ee47c2085e441a3561/server/jvm.out') Signal ERR caught! Traceback (line function script): 208 main /home/prow/go/src/k8s.io/release/push-build.sh Exiting... ``` Example -gce jobs: - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/85861/pull-kubernetes-e2e-gce/1219647945268269056 (#85861) - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/85846/pull-kubernetes-e2e-gce/1219646811342376960 (#85846) https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/87422/pull-kubernetes-e2e-gce/1219612336432615426 (#87422) - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/87372/pull-kubernetes-e2e-gce/1219266675761745921 (#87372) Example -gce-rbe jobs: - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/87435/pull-kubernetes-e2e-gce-rbe/1219718868767870977 - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/87266/pull-kubernetes-e2e-gce-rbe/1219655997224652803 - https://prow.k8s.io/view/gcs/kubernetes-jenkins/pr-logs/pull/86408/pull-kubernetes-e2e-gce-rbe/1219620136827228160 (source gist https://gist.github.com/spiffxp/ed54d006e630b8fd8126ddfe67d9dc1e) **Anything else we need to know**: - Triage link for the job that actually reports to PRs: https://storage.googleapis.com/k8s-gubernator/triage/index.html?pr=1&job=gce%24&test=Stage - [The line where the crash is happening]( https://github.com/kubernetes/release/blob/f59f6ba61ac5f225cde91ff0fc179a0a6f762841/lib/releaselib.sh#L1148) ends up running `bazel run //:push-build $gcs_stage $gcs_destination` - Some googling for "bazel error code 14" leads me to believe that bazel (or its jvm) is getting killed due to OOM. [This job requests 6Gi of memory](https://github.com/kubernetes/test-infra/blob/b2471685eed6a7d063d7e1e19032282bb33679db/config/jobs/kubernetes/sig-cloud-provider/gcp/gcp-gce.yaml#L59-L61) which is a value that's been cargo culted around as far as I can tell. I was unable to find the decision that led to this value. I'm wondering if a bump to 8Gi of memory would help reduce occurrences of this flake. - I'm referencing the -rbe jobs to point out that this happens even if we use RBE, but they're not blocking
non_build
stage which jobs are flaking pull kubernetes gce pull kubernetes gce rbe which test s are flaking stage testgrid link reason for failure this isn t technically a test that s failing it s part of the build stage up test cycle used by our jobs there are two main clusters of failures that i can tell one is timeouts the other looks like hashing and copying public release artifacts to gs kubernetes release pull ci pull kubernetes gce alpha test tmpdir defined output root default is bazel scratch cache bazel and max idle secs default is server terminated abruptly error code error message log file bazel scratch cache bazel bazel prow server jvm out signal err caught traceback line function script main home prow go src io release push build sh exiting example gce jobs example gce rbe jobs source gist anything else we need to know triage link for the job that actually reports to prs ends up running bazel run push build gcs stage gcs destination some googling for bazel error code leads me to believe that bazel or its jvm is getting killed due to oom which is a value that s been cargo culted around as far as i can tell i was unable to find the decision that led to this value i m wondering if a bump to of memory would help reduce occurrences of this flake i m referencing the rbe jobs to point out that this happens even if we use rbe but they re not blocking
0
691,806
23,711,429,049
IssuesEvent
2022-08-30 08:11:56
EthanThatOneKid/acmcsuf.com
https://api.github.com/repos/EthanThatOneKid/acmcsuf.com
closed
Need AI Team Button
bug HIGH PRIORITY
### What happened? Right now there is no AI Team hyperlink button on the front page when you scroll down - only Algo, Create and Dev. We also need to add a short description for the Teams page.
1.0
Need AI Team Button - ### What happened? Right now there is no AI Team hyperlink button on the front page when you scroll down - only Algo, Create and Dev. We also need to add a short description for the Teams page.
non_build
need ai team button what happened right now there is no ai team hyperlink button on the front page when you scroll down only algo create and dev we also need to add a short description for the teams page
0
6,150
3,343,531,577
IssuesEvent
2015-11-15 15:46:45
zeqing-guo/algorithms-study
https://api.github.com/repos/zeqing-guo/algorithms-study
opened
Leetcode-171: Excel Sheet Column Number
Leetcode Math
# Excel Sheet Column Number ## Description Related to question Excel Sheet Column Title Given a column title as appear in an Excel sheet, return its corresponding column number. For example: ``` A -> 1 B -> 2 C -> 3 ... Z -> 26 AA -> 27 AB -> 28 ``` # My Solution 代码的run time是2 ms,时间复杂度是O(n),空间复杂度是O(1)。 ```java public class Solution { public int titleToNumber(String s) { if (s == null) { return 0; } int sum = 0; int len = s.length(); for (int i = 0; i < len; ++i) { sum = (s.charAt(i) - 'A' + 1) + sum * 26; } return sum; } } ``` ## Analysis 太简单,没什么好说的。
1.0
Leetcode-171: Excel Sheet Column Number - # Excel Sheet Column Number ## Description Related to question Excel Sheet Column Title Given a column title as appear in an Excel sheet, return its corresponding column number. For example: ``` A -> 1 B -> 2 C -> 3 ... Z -> 26 AA -> 27 AB -> 28 ``` # My Solution 代码的run time是2 ms,时间复杂度是O(n),空间复杂度是O(1)。 ```java public class Solution { public int titleToNumber(String s) { if (s == null) { return 0; } int sum = 0; int len = s.length(); for (int i = 0; i < len; ++i) { sum = (s.charAt(i) - 'A' + 1) + sum * 26; } return sum; } } ``` ## Analysis 太简单,没什么好说的。
non_build
leetcode excel sheet column number excel sheet column number description related to question excel sheet column title given a column title as appear in an excel sheet return its corresponding column number for example a b c z aa ab my solution 代码的run ms,时间复杂度是o n ,空间复杂度是o 。 java public class solution public int titletonumber string s if s null return int sum int len s length for int i i len i sum s charat i a sum return sum analysis 太简单,没什么好说的。
0
97,412
28,262,074,214
IssuesEvent
2023-04-07 00:45:51
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
Add a feature to allow for multiple yaml suite test case clients
>feature :Delivery/Build Team:Delivery
Ideally, we would be able to run each yaml test with a different cluster setup, but that's currently limited by only supporting a single yaml test suite client. @mark-vieira suggested we may able to support multiple clients that have a method to filter which yaml tests are run. This would also be great for running the same tests under different cluster conditions (single node vs multi node, etc).
1.0
Add a feature to allow for multiple yaml suite test case clients - Ideally, we would be able to run each yaml test with a different cluster setup, but that's currently limited by only supporting a single yaml test suite client. @mark-vieira suggested we may able to support multiple clients that have a method to filter which yaml tests are run. This would also be great for running the same tests under different cluster conditions (single node vs multi node, etc).
build
add a feature to allow for multiple yaml suite test case clients ideally we would be able to run each yaml test with a different cluster setup but that s currently limited by only supporting a single yaml test suite client mark vieira suggested we may able to support multiple clients that have a method to filter which yaml tests are run this would also be great for running the same tests under different cluster conditions single node vs multi node etc
1
7,369
3,965,282,047
IssuesEvent
2016-05-03 07:29:13
Carthage/Carthage
https://api.github.com/repos/Carthage/Carthage
closed
`xcodebuild -list` will hang if it requires auth
bug build wontfix
Sometimes (depending on your machine configuration), when attempting to run `carthage build`, carthage will immediately throw an error saying "{Project} has no shared schemes". In this case, it was Alamofire. It appears that the issue is due to a hang while requesting authorization. The workaround was to manually run `xcodebuild -list -project Carthage/Checkouts/Alamofire/Alamofire.xcodeproj`, which prompted for a password, and allowed us to build everything successfully. My best guess is that it's coming from this [function](https://github.com/Carthage/Carthage/blob/1ee40ca0dc982abbf88b79f35cefd5e67fa62198/Source/CarthageKit/Xcode.swift#L213), and timing out. Just wanted to bring this to your attention, and see if there are any known fixes or changes that could be made to ensure this doesn't happen anymore. Thanks!
1.0
`xcodebuild -list` will hang if it requires auth - Sometimes (depending on your machine configuration), when attempting to run `carthage build`, carthage will immediately throw an error saying "{Project} has no shared schemes". In this case, it was Alamofire. It appears that the issue is due to a hang while requesting authorization. The workaround was to manually run `xcodebuild -list -project Carthage/Checkouts/Alamofire/Alamofire.xcodeproj`, which prompted for a password, and allowed us to build everything successfully. My best guess is that it's coming from this [function](https://github.com/Carthage/Carthage/blob/1ee40ca0dc982abbf88b79f35cefd5e67fa62198/Source/CarthageKit/Xcode.swift#L213), and timing out. Just wanted to bring this to your attention, and see if there are any known fixes or changes that could be made to ensure this doesn't happen anymore. Thanks!
build
xcodebuild list will hang if it requires auth sometimes depending on your machine configuration when attempting to run carthage build carthage will immediately throw an error saying project has no shared schemes in this case it was alamofire it appears that the issue is due to a hang while requesting authorization the workaround was to manually run xcodebuild list project carthage checkouts alamofire alamofire xcodeproj which prompted for a password and allowed us to build everything successfully my best guess is that it s coming from this and timing out just wanted to bring this to your attention and see if there are any known fixes or changes that could be made to ensure this doesn t happen anymore thanks
1
171,208
13,223,687,378
IssuesEvent
2020-08-17 17:41:35
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
Allow to declare group variables in dynamic inventory (aws_ec2)
affects_2.10 bot_closed feature inventory support:core test
##### SUMMARY At the moment there is no way for declaring variables in a dynamic inventory other than throught "compose". This covers a lot of cases however, would be useful having a way to define group vars easily in a "keyed_groups" or even "groups". ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME aws_ec2.py ##### ADDITIONAL INFORMATION I think something like the next examples can give us a lot of power when defining groups and specifically in situations where groups are at the same level (ansible_group_priority). an example for "groups" could be: ```yaml groups: a_group: tags.CustomTag == "foo" b_group: tags.CustomTag == "bar" group_vars: a_group: ansible_group_priority: 10 custom_variable: "hello world" ``` an example for keyed groups: ```yaml keyed_groups: - prefix: 'c' key: ec2_region group_vars: ansible_group_priority: 0 - prefix: 'b' key: ec2_architecture group_vars: foo: "{{ tags.CustomArchName }}" - prefix: 'a' key: tags.Role group_vars: project: "{{ tags.Project }}" env: production ansible_group_priority: 10 ```
1.0
Allow to declare group variables in dynamic inventory (aws_ec2) - ##### SUMMARY At the moment there is no way for declaring variables in a dynamic inventory other than throught "compose". This covers a lot of cases however, would be useful having a way to define group vars easily in a "keyed_groups" or even "groups". ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME aws_ec2.py ##### ADDITIONAL INFORMATION I think something like the next examples can give us a lot of power when defining groups and specifically in situations where groups are at the same level (ansible_group_priority). an example for "groups" could be: ```yaml groups: a_group: tags.CustomTag == "foo" b_group: tags.CustomTag == "bar" group_vars: a_group: ansible_group_priority: 10 custom_variable: "hello world" ``` an example for keyed groups: ```yaml keyed_groups: - prefix: 'c' key: ec2_region group_vars: ansible_group_priority: 0 - prefix: 'b' key: ec2_architecture group_vars: foo: "{{ tags.CustomArchName }}" - prefix: 'a' key: tags.Role group_vars: project: "{{ tags.Project }}" env: production ansible_group_priority: 10 ```
non_build
allow to declare group variables in dynamic inventory aws summary at the moment there is no way for declaring variables in a dynamic inventory other than throught compose this covers a lot of cases however would be useful having a way to define group vars easily in a keyed groups or even groups issue type feature idea component name aws py additional information i think something like the next examples can give us a lot of power when defining groups and specifically in situations where groups are at the same level ansible group priority an example for groups could be yaml groups a group tags customtag foo b group tags customtag bar group vars a group ansible group priority custom variable hello world an example for keyed groups yaml keyed groups prefix c key region group vars ansible group priority prefix b key architecture group vars foo tags customarchname prefix a key tags role group vars project tags project env production ansible group priority
0
670,152
22,677,713,272
IssuesEvent
2022-07-04 07:00:28
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.udemy.com - video or audio doesn't play
browser-firefox-mobile priority-important engine-gecko
<!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/106833 --> **URL**: https://www.udemy.com/course/certified-kubernetes-application-developer/learn/lecture/12299412#overview **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Firefox **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: udemy.com videos are not playing in vr headset. <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.udemy.com - video or audio doesn't play - <!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/106833 --> **URL**: https://www.udemy.com/course/certified-kubernetes-application-developer/learn/lecture/12299412#overview **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Firefox **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: udemy.com videos are not playing in vr headset. <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_build
video or audio doesn t play url browser version firefox mobile operating system android tested another browser yes firefox problem type video or audio doesn t play description the video or audio does not play steps to reproduce udemy com videos are not playing in vr headset browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true from with ❤️
0
37,504
10,021,891,387
IssuesEvent
2019-07-16 15:30:07
angular/angular
https://api.github.com/repos/angular/angular
closed
Docs: update polyfill doc in Browser Support page
comp: docs docsarea: build & deploy freq2: medium severity3: broken state: has PR type: bug/fix
**Description** Polyfills doc needs to be updated for version 8. The inclusion of the "Mandatory" es6 polyfills section for IE and friends is no longer a manual step (as in previous versions when we had to add multiple core-js imports). See https://unpkg.com/ccbl-gfx-angular-test@0.0.6/src/polyfills.ts See https://github.com/angular/angular/pull/31115#issuecomment-503301271 ### What's the affected URL?** https://angular.io/guide/browser-support ### Expected vs Actual Behavior** Remove manual steps, refer to https://angular.io/guide/deployment#differential-loading
1.0
Docs: update polyfill doc in Browser Support page - **Description** Polyfills doc needs to be updated for version 8. The inclusion of the "Mandatory" es6 polyfills section for IE and friends is no longer a manual step (as in previous versions when we had to add multiple core-js imports). See https://unpkg.com/ccbl-gfx-angular-test@0.0.6/src/polyfills.ts See https://github.com/angular/angular/pull/31115#issuecomment-503301271 ### What's the affected URL?** https://angular.io/guide/browser-support ### Expected vs Actual Behavior** Remove manual steps, refer to https://angular.io/guide/deployment#differential-loading
build
docs update polyfill doc in browser support page description polyfills doc needs to be updated for version the inclusion of the mandatory polyfills section for ie and friends is no longer a manual step as in previous versions when we had to add multiple core js imports see see what s the affected url expected vs actual behavior remove manual steps refer to
1
33,847
4,863,725,341
IssuesEvent
2016-11-14 16:08:48
openshift/origin
https://api.github.com/repos/openshift/origin
opened
Multiple test fails
kind/test-flake
https://ci.openshift.redhat.com/jenkins/job/test_pr_origin_extended/793/ Test Result (11 failures / +10) Extended.[k8s.io] Garbage collector [Feature:GarbageCollector] should delete pods created by rc when not orphaning Extended.[networking] OVS generic should add and remove flows when pods are added and removed Extended.[imageapis] openshift limit range admission should deny a push of built image exceeding openshift.io/Image limit Extended.[k8s.io] Deployment RollingUpdateDeployment should scale up and down in the right order Extended.[networking] OVS generic should add and remove flows when nodes are added and removed Extended.[k8s.io] Federation secrets [Feature:Federation12] Secret objects should be created and deleted successfully Extended.[image_ecosystem][jenkins] schedule jobs on pod slaves use of jenkins with kubernetes plugin by creating slave from existing builder and adding it to Jenkins master Extended.[k8s.io] Deployment deployment should label adopted RSs and pods Extended.[k8s.io] Deployment RollingUpdateDeployment should delete old pods and create new ones Extended.[k8s.io] Deployment RecreateDeployment should delete old pods and create new ones
1.0
Multiple test fails - https://ci.openshift.redhat.com/jenkins/job/test_pr_origin_extended/793/ Test Result (11 failures / +10) Extended.[k8s.io] Garbage collector [Feature:GarbageCollector] should delete pods created by rc when not orphaning Extended.[networking] OVS generic should add and remove flows when pods are added and removed Extended.[imageapis] openshift limit range admission should deny a push of built image exceeding openshift.io/Image limit Extended.[k8s.io] Deployment RollingUpdateDeployment should scale up and down in the right order Extended.[networking] OVS generic should add and remove flows when nodes are added and removed Extended.[k8s.io] Federation secrets [Feature:Federation12] Secret objects should be created and deleted successfully Extended.[image_ecosystem][jenkins] schedule jobs on pod slaves use of jenkins with kubernetes plugin by creating slave from existing builder and adding it to Jenkins master Extended.[k8s.io] Deployment deployment should label adopted RSs and pods Extended.[k8s.io] Deployment RollingUpdateDeployment should delete old pods and create new ones Extended.[k8s.io] Deployment RecreateDeployment should delete old pods and create new ones
non_build
multiple test fails test result failures extended garbage collector should delete pods created by rc when not orphaning extended ovs generic should add and remove flows when pods are added and removed extended openshift limit range admission should deny a push of built image exceeding openshift io image limit extended deployment rollingupdatedeployment should scale up and down in the right order extended ovs generic should add and remove flows when nodes are added and removed extended federation secrets secret objects should be created and deleted successfully extended schedule jobs on pod slaves use of jenkins with kubernetes plugin by creating slave from existing builder and adding it to jenkins master extended deployment deployment should label adopted rss and pods extended deployment rollingupdatedeployment should delete old pods and create new ones extended deployment recreatedeployment should delete old pods and create new ones
0
429,299
30,034,425,828
IssuesEvent
2023-06-27 11:50:09
PaddlePaddle/PaddleNLP
https://api.github.com/repos/PaddlePaddle/PaddleNLP
opened
[Docs]:
documentation
### 软件环境 ```Markdown - paddlepaddle: - paddlepaddle-gpu: - paddlenlp: ``` ### 详细描述 ```Markdown 请问nlp gpt2模型345m的默认配置训练step为500000,训练完后,其loss标准如何判定,能否提供一个cuda的loss收敛的参照数据 ```
1.0
[Docs]: - ### 软件环境 ```Markdown - paddlepaddle: - paddlepaddle-gpu: - paddlenlp: ``` ### 详细描述 ```Markdown 请问nlp gpt2模型345m的默认配置训练step为500000,训练完后,其loss标准如何判定,能否提供一个cuda的loss收敛的参照数据 ```
non_build
软件环境 markdown paddlepaddle paddlepaddle gpu paddlenlp 详细描述 markdown 请问nlp ,训练完后,其loss标准如何判定,能否提供一个cuda的loss收敛的参照数据
0
92,228
26,622,572,290
IssuesEvent
2023-01-24 12:22:28
JuliaPackaging/Yggdrasil
https://api.github.com/repos/JuliaPackaging/Yggdrasil
closed
Request: Slate
builder request :pray:
https://icl.utk.edu/slate/ Quoting: The ultimate objective of SLATE is to replace the venerable Scalable Linear Algebra PACKage (ScaLAPACK) library, which has become the industry standard for dense linear algebra operations in distributed memory environments. However, after two decades of operation, ScaLAPACK is past the end of its lifecycle and overdue for a replacement, as it can hardly be retrofitted to support hardware accelerators, which are an integral part of today’s HPC hardware infrastructure.
1.0
Request: Slate - https://icl.utk.edu/slate/ Quoting: The ultimate objective of SLATE is to replace the venerable Scalable Linear Algebra PACKage (ScaLAPACK) library, which has become the industry standard for dense linear algebra operations in distributed memory environments. However, after two decades of operation, ScaLAPACK is past the end of its lifecycle and overdue for a replacement, as it can hardly be retrofitted to support hardware accelerators, which are an integral part of today’s HPC hardware infrastructure.
build
request slate quoting the ultimate objective of slate is to replace the venerable scalable linear algebra package scalapack library which has become the industry standard for dense linear algebra operations in distributed memory environments however after two decades of operation scalapack is past the end of its lifecycle and overdue for a replacement as it can hardly be retrofitted to support hardware accelerators which are an integral part of today’s hpc hardware infrastructure
1
814,423
30,507,195,287
IssuesEvent
2023-07-18 17:48:35
ClangBuiltLinux/linux
https://api.github.com/repos/ClangBuiltLinux/linux
closed
KASAN_HW_TAGS might be incompatible with LTO
[BUG] Untriaged low priority [FEATURE] LTO [BUG] linux (android)
android seems to have this patch https://android-review.git.corp.google.com/c/kernel/common/+/1621951 but Lee Jones is reporting that enabling the two is crashing LLD: ``` AR kernel/built-in.a GEN .version CHK include/generated/compile.h GEN .tmp_initcalls.lds GEN .tmp_symversions.lds LTO vmlinux.o inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location %226 = call i8* @memcpy(i8* %26, i8* %14, i64 128) inlinable function call in a function with debug info must have a !dbg location %67 = call i8* @memcpy(i8* %66, i8* %8, i64 24) inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location %494 = call i8* @memset(i8* %418, i32 0, i64 16) inlinable function call in a function with debug info must have a !dbg location %84 = call i8* @memcpy(i8* %83, i8* %66, i64 12) inlinable function call in a function with debug info must have a !dbg location %11 = call i8* @memcpy(i8* %10, i8* %8, i64 24) inlinable function call in a function with debug info must have a !dbg location %24 = call i8* @memcpy(i8* %21, i8* %23, i64 40) inlinable function call in a function with debug info must have a !dbg location %32 = call i8* @memcpy(i8* %31, i8* %22, i64 12) inlinable function call in a function with debug info must have a !dbg location %37 = call i8* @memset(i8* %36, i32 0, i64 12) inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location %35 = call i8* @memset(i8* %21, i32 0, i64 16) inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() LLVM ERROR: Broken module found, compilation aborted! PLEASE submit a bug report to https://github.com/android-ndk/ndk/issues and include the crash backtrace. Stack dump: 0. Program arguments: ld.lld -m elf_x86_64 -mllvm -import-instr-limit=5 -z noexecstack -r -o vmlinux.o -T .tmp_initcalls.lds -T .tmp_symversions.lds --whole-archive arch/x86/kernel/head_64.o arch/x86/kernel/head64.o arch/x86/kernel/ebda.o arch/x86/kernel/platform-quirks.o init/built-in.a usr/built-in.a arch/x86/built-in.a kernel/built-in.a certs/built-in.a mm/built-in.a fs/built-in.a ipc/built-in.a security/built-in.a crypto/built-in.a block/built-in.a io_uring/built-in.a lib/built-in.a arch/x86/lib/built-in.a lib/lib.a arch/x86/lib/lib.a drivers/built-in.a sound/built-in.a net/built-in.a virt/built-in.a arch/x86/pci/built-in.a arch/x86/power/built-in.a --no-whole-archive --start-group --end-group #0 0x0000000002a060f8 llvm::sys::PrintStackTrace(llvm::raw_ostream&, int) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2a060f8) #1 0x0000000002a05260 llvm::sys::RunSignalHandlers() (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2a05260) #2 0x0000000002a0654a (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2a0654a) #3 0x00007f18f34f1f90 (/lib/x86_64-linux-gnu/libc.so.6+0x3bf90) #4 0x00007f18f3540ccc __pthread_kill_implementation ./nptl/./nptl/pthread_kill.c:44:76 #5 0x00007f18f34f1ef2 raise ./signal/../sysdeps/posix/raise.c:27:6 #6 0x00007f18f34dc472 abort ./stdlib/./stdlib/abort.c:81:7 #7 0x0000000001f42e22 llvm::report_fatal_error(llvm::Twine const&, bool) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x1f42e22) #8 0x0000000001ca7785 (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x1ca7785) #9 0x0000000002707fbf (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2707fbf) #10 0x0000000002707f2d (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2707f2d) #11 0x0000000002515c5f llvm::PassManager<llvm::Module, llvm::AnalysisManager<llvm::Module> >::run(llvm::Module&, llvm::AnalysisManager<llvm::Module>&) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2515c5f) #12 0x00000000022c0ea5 (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x22c0ea5) #13 0x00000000022bc1c9 llvm::lto::opt(llvm::lto::Config const&, llvm::TargetMachine*, unsigned int, llvm::Module&, bool, llvm::ModuleSummaryIndex*, llvm::ModuleSummaryIndex const*, std::__1::vector<unsigned char, std::__1::allocator<unsigned char> > const&) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x22bc1c9) #14 0x00000000027c4b2b llvm::lto::backend(llvm::lto::Config const&, std::__1::function<llvm::Expected<std::__1::unique_ptr<llvm::CachedFileStream, std::__1::default_delete<llvm::CachedFileStream> > > (unsigned int)>, unsigned int, llvm::Module&, llvm::ModuleSummaryIndex&) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x27c4b2b) #15 0x00000000027c3a3c llvm::lto::LTO::runRegularLTO(std::__1::function<llvm::Expected<std::__1::unique_ptr<llvm::CachedFileStream, std::__1::default_delete<llvm::CachedFileStream> > > (unsigned int)>) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x27c3a3c) #16 0x00000000027c226e llvm::lto::LTO::run(std::__1::function<llvm::Expected<std::__1::unique_ptr<llvm::CachedFileStream, std::__1::default_delete<llvm::CachedFileStream> > > (unsigned int)>, std::__1::function<llvm::Expected<std::__1::function<llvm::Expected<std::__1::unique_ptr<llvm::CachedFileStream, std::__1::default_delete<llvm::CachedFileStream> > > (unsigned int)> > (unsigned int, llvm::StringRef)>) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x27c226e) #17 0x00000000027c0654 lld::elf::BitcodeCompiler::compile() (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x27c0654) #18 0x00000000027bcfc0 void lld::elf::LinkerDriver::compileBitcodeFiles<llvm::object::ELFType<(llvm::support::endianness)1, true> >(bool) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x27bcfc0) #19 0x00000000020ca24f void lld::elf::LinkerDriver::link<llvm::object::ELFType<(llvm::support::endianness)1, true> >(llvm::opt::InputArgList&) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x20ca24f) #20 0x00000000020b27f7 lld::elf::LinkerDriver::linkerMain(llvm::ArrayRef<char const*>) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x20b27f7) #21 0x00000000020b1c45 lld::elf::link(llvm::ArrayRef<char const*>, llvm::raw_ostream&, llvm::raw_ostream&, bool, bool) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x20b1c45) #22 0x00000000018faac5 (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x18faac5) #23 0x00000000026d9fa2 main (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x26d9fa2) #24 0x00007f18f34dd18a __libc_start_call_main ./csu/../sysdeps/nptl/libc_start_call_main.h:74:3 #25 0x00007f18f34dd245 call_init ./csu/../csu/libc-start.c:128:20 #26 0x00007f18f34dd245 __libc_start_main ./csu/../csu/libc-start.c:368:5 #27 0x00000000029d2189 _start (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x29d2189) /usr/local/google/home/joneslee/projects/google/android13-5.15/common/scripts/link-vmlinux.sh: line 72: 439407 Aborted (core dumped) ${LD} ${KBUILD_LDFLAGS} -r -o ${1} ${lds} ${objects} make: *** [/usr/local/google/home/joneslee/projects/google/android13-5.15/common/Makefile:1283: vmlinux] Error 134 ```
1.0
KASAN_HW_TAGS might be incompatible with LTO - android seems to have this patch https://android-review.git.corp.google.com/c/kernel/common/+/1621951 but Lee Jones is reporting that enabling the two is crashing LLD: ``` AR kernel/built-in.a GEN .version CHK include/generated/compile.h GEN .tmp_initcalls.lds GEN .tmp_symversions.lds LTO vmlinux.o inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location %226 = call i8* @memcpy(i8* %26, i8* %14, i64 128) inlinable function call in a function with debug info must have a !dbg location %67 = call i8* @memcpy(i8* %66, i8* %8, i64 24) inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location %494 = call i8* @memset(i8* %418, i32 0, i64 16) inlinable function call in a function with debug info must have a !dbg location %84 = call i8* @memcpy(i8* %83, i8* %66, i64 12) inlinable function call in a function with debug info must have a !dbg location %11 = call i8* @memcpy(i8* %10, i8* %8, i64 24) inlinable function call in a function with debug info must have a !dbg location %24 = call i8* @memcpy(i8* %21, i8* %23, i64 40) inlinable function call in a function with debug info must have a !dbg location %32 = call i8* @memcpy(i8* %31, i8* %22, i64 12) inlinable function call in a function with debug info must have a !dbg location %37 = call i8* @memset(i8* %36, i32 0, i64 12) inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location %35 = call i8* @memset(i8* %21, i32 0, i64 16) inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() inlinable function call in a function with debug info must have a !dbg location call void @__asan_handle_no_return() LLVM ERROR: Broken module found, compilation aborted! PLEASE submit a bug report to https://github.com/android-ndk/ndk/issues and include the crash backtrace. Stack dump: 0. Program arguments: ld.lld -m elf_x86_64 -mllvm -import-instr-limit=5 -z noexecstack -r -o vmlinux.o -T .tmp_initcalls.lds -T .tmp_symversions.lds --whole-archive arch/x86/kernel/head_64.o arch/x86/kernel/head64.o arch/x86/kernel/ebda.o arch/x86/kernel/platform-quirks.o init/built-in.a usr/built-in.a arch/x86/built-in.a kernel/built-in.a certs/built-in.a mm/built-in.a fs/built-in.a ipc/built-in.a security/built-in.a crypto/built-in.a block/built-in.a io_uring/built-in.a lib/built-in.a arch/x86/lib/built-in.a lib/lib.a arch/x86/lib/lib.a drivers/built-in.a sound/built-in.a net/built-in.a virt/built-in.a arch/x86/pci/built-in.a arch/x86/power/built-in.a --no-whole-archive --start-group --end-group #0 0x0000000002a060f8 llvm::sys::PrintStackTrace(llvm::raw_ostream&, int) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2a060f8) #1 0x0000000002a05260 llvm::sys::RunSignalHandlers() (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2a05260) #2 0x0000000002a0654a (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2a0654a) #3 0x00007f18f34f1f90 (/lib/x86_64-linux-gnu/libc.so.6+0x3bf90) #4 0x00007f18f3540ccc __pthread_kill_implementation ./nptl/./nptl/pthread_kill.c:44:76 #5 0x00007f18f34f1ef2 raise ./signal/../sysdeps/posix/raise.c:27:6 #6 0x00007f18f34dc472 abort ./stdlib/./stdlib/abort.c:81:7 #7 0x0000000001f42e22 llvm::report_fatal_error(llvm::Twine const&, bool) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x1f42e22) #8 0x0000000001ca7785 (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x1ca7785) #9 0x0000000002707fbf (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2707fbf) #10 0x0000000002707f2d (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2707f2d) #11 0x0000000002515c5f llvm::PassManager<llvm::Module, llvm::AnalysisManager<llvm::Module> >::run(llvm::Module&, llvm::AnalysisManager<llvm::Module>&) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x2515c5f) #12 0x00000000022c0ea5 (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x22c0ea5) #13 0x00000000022bc1c9 llvm::lto::opt(llvm::lto::Config const&, llvm::TargetMachine*, unsigned int, llvm::Module&, bool, llvm::ModuleSummaryIndex*, llvm::ModuleSummaryIndex const*, std::__1::vector<unsigned char, std::__1::allocator<unsigned char> > const&) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x22bc1c9) #14 0x00000000027c4b2b llvm::lto::backend(llvm::lto::Config const&, std::__1::function<llvm::Expected<std::__1::unique_ptr<llvm::CachedFileStream, std::__1::default_delete<llvm::CachedFileStream> > > (unsigned int)>, unsigned int, llvm::Module&, llvm::ModuleSummaryIndex&) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x27c4b2b) #15 0x00000000027c3a3c llvm::lto::LTO::runRegularLTO(std::__1::function<llvm::Expected<std::__1::unique_ptr<llvm::CachedFileStream, std::__1::default_delete<llvm::CachedFileStream> > > (unsigned int)>) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x27c3a3c) #16 0x00000000027c226e llvm::lto::LTO::run(std::__1::function<llvm::Expected<std::__1::unique_ptr<llvm::CachedFileStream, std::__1::default_delete<llvm::CachedFileStream> > > (unsigned int)>, std::__1::function<llvm::Expected<std::__1::function<llvm::Expected<std::__1::unique_ptr<llvm::CachedFileStream, std::__1::default_delete<llvm::CachedFileStream> > > (unsigned int)> > (unsigned int, llvm::StringRef)>) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x27c226e) #17 0x00000000027c0654 lld::elf::BitcodeCompiler::compile() (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x27c0654) #18 0x00000000027bcfc0 void lld::elf::LinkerDriver::compileBitcodeFiles<llvm::object::ELFType<(llvm::support::endianness)1, true> >(bool) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x27bcfc0) #19 0x00000000020ca24f void lld::elf::LinkerDriver::link<llvm::object::ELFType<(llvm::support::endianness)1, true> >(llvm::opt::InputArgList&) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x20ca24f) #20 0x00000000020b27f7 lld::elf::LinkerDriver::linkerMain(llvm::ArrayRef<char const*>) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x20b27f7) #21 0x00000000020b1c45 lld::elf::link(llvm::ArrayRef<char const*>, llvm::raw_ostream&, llvm::raw_ostream&, bool, bool) (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x20b1c45) #22 0x00000000018faac5 (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x18faac5) #23 0x00000000026d9fa2 main (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x26d9fa2) #24 0x00007f18f34dd18a __libc_start_call_main ./csu/../sysdeps/nptl/libc_start_call_main.h:74:3 #25 0x00007f18f34dd245 call_init ./csu/../csu/libc-start.c:128:20 #26 0x00007f18f34dd245 __libc_start_main ./csu/../csu/libc-start.c:368:5 #27 0x00000000029d2189 _start (/usr/local/google/home/joneslee/projects/google/android13-5.15/prebuilts/clang/host/linux-x86/clang-r450784e/bin/lld+0x29d2189) /usr/local/google/home/joneslee/projects/google/android13-5.15/common/scripts/link-vmlinux.sh: line 72: 439407 Aborted (core dumped) ${LD} ${KBUILD_LDFLAGS} -r -o ${1} ${lds} ${objects} make: *** [/usr/local/google/home/joneslee/projects/google/android13-5.15/common/Makefile:1283: vmlinux] Error 134 ```
non_build
kasan hw tags might be incompatible with lto android seems to have this patch but lee jones is reporting that enabling the two is crashing lld ar kernel built in a gen version chk include generated compile h gen tmp initcalls lds gen tmp symversions lds lto vmlinux o inlinable function call in a function with debug info must have a dbg location call void asan handle no return inlinable function call in a function with debug info must have a dbg location call void asan handle no return inlinable function call in a function with debug info must have a dbg location call memcpy inlinable function call in a function with debug info must have a dbg location call memcpy inlinable function call in a function with debug info must have a dbg location call void asan handle no return inlinable function call in a function with debug info must have a dbg location call void asan handle no return inlinable function call in a function with debug info must have a dbg location call void asan handle no return inlinable function call in a function with debug info must have a dbg location call memset inlinable function call in a function with debug info must have a dbg location call memcpy inlinable function call in a function with debug info must have a dbg location call memcpy inlinable function call in a function with debug info must have a dbg location call memcpy inlinable function call in a function with debug info must have a dbg location call memcpy inlinable function call in a function with debug info must have a dbg location call memset inlinable function call in a function with debug info must have a dbg location call void asan handle no return inlinable function call in a function with debug info must have a dbg location call memset inlinable function call in a function with debug info must have a dbg location call void asan handle no return inlinable function call in a function with debug info must have a dbg location call void asan handle no return inlinable function call in a function with debug info must have a dbg location call void asan handle no return inlinable function call in a function with debug info must have a dbg location call void asan handle no return inlinable function call in a function with debug info must have a dbg location call void asan handle no return llvm error broken module found compilation aborted please submit a bug report to and include the crash backtrace stack dump program arguments ld lld m elf mllvm import instr limit z noexecstack r o vmlinux o t tmp initcalls lds t tmp symversions lds whole archive arch kernel head o arch kernel o arch kernel ebda o arch kernel platform quirks o init built in a usr built in a arch built in a kernel built in a certs built in a mm built in a fs built in a ipc built in a security built in a crypto built in a block built in a io uring built in a lib built in a arch lib built in a lib lib a arch lib lib a drivers built in a sound built in a net built in a virt built in a arch pci built in a arch power built in a no whole archive start group end group llvm sys printstacktrace llvm raw ostream int usr local google home joneslee projects google prebuilts clang host linux clang bin lld llvm sys runsignalhandlers usr local google home joneslee projects google prebuilts clang host linux clang bin lld usr local google home joneslee projects google prebuilts clang host linux clang bin lld lib linux gnu libc so pthread kill implementation nptl nptl pthread kill c raise signal sysdeps posix raise c abort stdlib stdlib abort c llvm report fatal error llvm twine const bool usr local google home joneslee projects google prebuilts clang host linux clang bin lld usr local google home joneslee projects google prebuilts clang host linux clang bin lld usr local google home joneslee projects google prebuilts clang host linux clang bin lld usr local google home joneslee projects google prebuilts clang host linux clang bin lld llvm passmanager run llvm module llvm analysismanager usr local google home joneslee projects google prebuilts clang host linux clang bin lld usr local google home joneslee projects google prebuilts clang host linux clang bin lld llvm lto opt llvm lto config const llvm targetmachine unsigned int llvm module bool llvm modulesummaryindex llvm modulesummaryindex const std vector const usr local google home joneslee projects google prebuilts clang host linux clang bin lld llvm lto backend llvm lto config const std function unsigned int unsigned int llvm module llvm modulesummaryindex usr local google home joneslee projects google prebuilts clang host linux clang bin lld llvm lto lto runregularlto std function unsigned int usr local google home joneslee projects google prebuilts clang host linux clang bin lld llvm lto lto run std function unsigned int std function unsigned int unsigned int llvm stringref usr local google home joneslee projects google prebuilts clang host linux clang bin lld lld elf bitcodecompiler compile usr local google home joneslee projects google prebuilts clang host linux clang bin lld void lld elf linkerdriver compilebitcodefiles bool usr local google home joneslee projects google prebuilts clang host linux clang bin lld void lld elf linkerdriver link llvm opt inputarglist usr local google home joneslee projects google prebuilts clang host linux clang bin lld lld elf linkerdriver linkermain llvm arrayref usr local google home joneslee projects google prebuilts clang host linux clang bin lld lld elf link llvm arrayref llvm raw ostream llvm raw ostream bool bool usr local google home joneslee projects google prebuilts clang host linux clang bin lld usr local google home joneslee projects google prebuilts clang host linux clang bin lld main usr local google home joneslee projects google prebuilts clang host linux clang bin lld libc start call main csu sysdeps nptl libc start call main h call init csu csu libc start c libc start main csu csu libc start c start usr local google home joneslee projects google prebuilts clang host linux clang bin lld usr local google home joneslee projects google common scripts link vmlinux sh line aborted core dumped ld kbuild ldflags r o lds objects make error
0
98,570
30,006,542,440
IssuesEvent
2023-06-26 12:50:02
Joe-Collins-1986/BitWise-DRF
https://api.github.com/repos/Joe-Collins-1986/BitWise-DRF
closed
USER STORY: ADD LINKS
EPIC - API Build MUST HAVE
As a **frontend developer** I can **access the links add endpoint** so that **I can allow users to add links to their own articles.** **EPIC** API Build **Tasks:** - [x] View – perform create to set owner to current user **Acceptance Criteria:** - [x] API accessible via URL - [x] Links added as expected
1.0
USER STORY: ADD LINKS - As a **frontend developer** I can **access the links add endpoint** so that **I can allow users to add links to their own articles.** **EPIC** API Build **Tasks:** - [x] View – perform create to set owner to current user **Acceptance Criteria:** - [x] API accessible via URL - [x] Links added as expected
build
user story add links as a frontend developer i can access the links add endpoint so that i can allow users to add links to their own articles epic api build tasks view – perform create to set owner to current user acceptance criteria api accessible via url links added as expected
1