Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
39,075
5,037,388,063
IssuesEvent
2016-12-17 16:50:13
eloipuertas/ES2016F
https://api.github.com/repos/eloipuertas/ES2016F
closed
[A] Animation model saruman
Animations Design Team-A
3.21 As a graphic designer, I want a set of animations for Saruman, to give movement to the character. [Low priority]
1.0
[A] Animation model saruman - 3.21 As a graphic designer, I want a set of animations for Saruman, to give movement to the character. [Low priority]
non_process
animation model saruman as a graphic designer i want a set of animations for saruman to give movement to the character
0
4,320
7,214,021,250
IssuesEvent
2018-02-08 00:00:22
lvergergsk/BibGallery-FrontEnd
https://api.github.com/repos/lvergergsk/BibGallery-FrontEnd
closed
Using SQL to check free space for a user
Data Processing
## Oracle * Possible way 1 ``` SELECT ts.tablespace_name, TO_CHAR(SUM(NVL(fs.bytes,0))/1024/1024, '99,999,990.99') AS MB_FREE FROM user_free_space fs, user_tablespaces ts, user_users us WHERE fs.tablespace_name(+) = ts.tablespace_name AND ts.tablespace_name(+) = us.default_tablespace GROUP BY ts.tablespace_name; ``` * Possible way 2 ``` SELECT table_name as Table_Name, row_cnt as Row_Count, SUM(mb) as Size_MB FROM (SELECT in_tbl.table_name, to_number(extractvalue(xmltype(dbms_xmlgen.getxml('select count(*) c from ' ||ut.table_name)),'/ROWSET/ROW/C')) AS row_cnt , mb FROM (SELECT CASE WHEN lob_tables IS NULL THEN table_name WHEN lob_tables IS NOT NULL THEN lob_tables END AS table_name , mb FROM (SELECT ul.table_name AS lob_tables, us.segment_name AS table_name , us.bytes/1024/1024 MB FROM user_segments us LEFT JOIN user_lobs ul ON us.segment_name = ul.segment_name ) ) in_tbl INNER JOIN user_tables ut ON in_tbl.table_name = ut.table_name ) GROUP BY table_name, row_cnt ORDER BY 3 DESC; ``` * Possible way 3 ``` SELECT SUM(bytes) FROM user_segments ``` ## MySQL ``` select table_schema, sum((data_length+index_length)/1024/1024) AS MB from information_schema.tables group by 1; ```
1.0
Using SQL to check free space for a user - ## Oracle * Possible way 1 ``` SELECT ts.tablespace_name, TO_CHAR(SUM(NVL(fs.bytes,0))/1024/1024, '99,999,990.99') AS MB_FREE FROM user_free_space fs, user_tablespaces ts, user_users us WHERE fs.tablespace_name(+) = ts.tablespace_name AND ts.tablespace_name(+) = us.default_tablespace GROUP BY ts.tablespace_name; ``` * Possible way 2 ``` SELECT table_name as Table_Name, row_cnt as Row_Count, SUM(mb) as Size_MB FROM (SELECT in_tbl.table_name, to_number(extractvalue(xmltype(dbms_xmlgen.getxml('select count(*) c from ' ||ut.table_name)),'/ROWSET/ROW/C')) AS row_cnt , mb FROM (SELECT CASE WHEN lob_tables IS NULL THEN table_name WHEN lob_tables IS NOT NULL THEN lob_tables END AS table_name , mb FROM (SELECT ul.table_name AS lob_tables, us.segment_name AS table_name , us.bytes/1024/1024 MB FROM user_segments us LEFT JOIN user_lobs ul ON us.segment_name = ul.segment_name ) ) in_tbl INNER JOIN user_tables ut ON in_tbl.table_name = ut.table_name ) GROUP BY table_name, row_cnt ORDER BY 3 DESC; ``` * Possible way 3 ``` SELECT SUM(bytes) FROM user_segments ``` ## MySQL ``` select table_schema, sum((data_length+index_length)/1024/1024) AS MB from information_schema.tables group by 1; ```
process
using sql to check free space for a user oracle possible way select ts tablespace name to char sum nvl fs bytes as mb free from user free space fs user tablespaces ts user users us where fs tablespace name ts tablespace name and ts tablespace name us default tablespace group by ts tablespace name possible way select table name as table name row cnt as row count sum mb as size mb from select in tbl table name to number extractvalue xmltype dbms xmlgen getxml select count c from ut table name rowset row c as row cnt mb from select case when lob tables is null then table name when lob tables is not null then lob tables end as table name mb from select ul table name as lob tables us segment name as table name us bytes mb from user segments us left join user lobs ul on us segment name ul segment name in tbl inner join user tables ut on in tbl table name ut table name group by table name row cnt order by desc possible way select sum bytes from user segments mysql select table schema sum data length index length as mb from information schema tables group by
1
6,905
10,056,518,350
IssuesEvent
2019-07-22 09:20:31
CymChad/BaseRecyclerViewAdapterHelper
https://api.github.com/repos/CymChad/BaseRecyclerViewAdapterHelper
closed
树形列表删除bug
processing
![image](https://user-images.githubusercontent.com/24364422/43033030-5aa7a424-8cf5-11e8-81b6-5db85d9c3dfd.png) 在树形列表中,如果一级目录下没有子目录时,在调用BaseMultiItemQuickAdapter中remove(int position),删除,会导致程序崩溃。你使用你写的列子也是这样。 ![image](https://user-images.githubusercontent.com/24364422/43033085-3bf98064-8cf6-11e8-864e-ef230a14fd64.png) ![image](https://user-images.githubusercontent.com/24364422/43033093-4f3d20c2-8cf6-11e8-9d5b-f07f14625c58.png)
1.0
树形列表删除bug - ![image](https://user-images.githubusercontent.com/24364422/43033030-5aa7a424-8cf5-11e8-81b6-5db85d9c3dfd.png) 在树形列表中,如果一级目录下没有子目录时,在调用BaseMultiItemQuickAdapter中remove(int position),删除,会导致程序崩溃。你使用你写的列子也是这样。 ![image](https://user-images.githubusercontent.com/24364422/43033085-3bf98064-8cf6-11e8-864e-ef230a14fd64.png) ![image](https://user-images.githubusercontent.com/24364422/43033093-4f3d20c2-8cf6-11e8-9d5b-f07f14625c58.png)
process
树形列表删除bug 在树形列表中,如果一级目录下没有子目录时,在调用basemultiitemquickadapter中remove int position 删除,会导致程序崩溃。你使用你写的列子也是这样。
1
262,022
8,249,166,046
IssuesEvent
2018-09-11 20:42:16
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
opened
[studio] Non-site members have access to site and full permissions
bug priority: high
### Expected behavior Non-site members shouldn't have access to sites ### Actual behavior Everyone has full access to everything (seems that all groups are being returned for a user). ### Steps to reproduce the problem * Create a user and a site, but don't assign the user to any groups * Login as that user * See the site ### Log/stack trace (use https://gist.github.com) ### Specs #### Version ``` Studio Version Number: 3.1.0-SNAPSHOT-0a0946 Build Number: 0a0946d689773516fda8075b4058b38e67f8920c Build Date/Time: 09-11-2018 13:57:45 -0400 ``` #### OS Any #### Browser Any
1.0
[studio] Non-site members have access to site and full permissions - ### Expected behavior Non-site members shouldn't have access to sites ### Actual behavior Everyone has full access to everything (seems that all groups are being returned for a user). ### Steps to reproduce the problem * Create a user and a site, but don't assign the user to any groups * Login as that user * See the site ### Log/stack trace (use https://gist.github.com) ### Specs #### Version ``` Studio Version Number: 3.1.0-SNAPSHOT-0a0946 Build Number: 0a0946d689773516fda8075b4058b38e67f8920c Build Date/Time: 09-11-2018 13:57:45 -0400 ``` #### OS Any #### Browser Any
non_process
non site members have access to site and full permissions expected behavior non site members shouldn t have access to sites actual behavior everyone has full access to everything seems that all groups are being returned for a user steps to reproduce the problem create a user and a site but don t assign the user to any groups login as that user see the site log stack trace use specs version studio version number snapshot build number build date time os any browser any
0
44,091
13,048,237,902
IssuesEvent
2020-07-29 12:11:54
jgeraigery/imhotep
https://api.github.com/repos/jgeraigery/imhotep
opened
WS-2018-0125 (Medium) detected in jackson-core-2.2.3.jar, jackson-core-2.6.7.jar
security vulnerability
## WS-2018-0125 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-core-2.2.3.jar</b>, <b>jackson-core-2.6.7.jar</b></p></summary> <p> <details><summary><b>jackson-core-2.2.3.jar</b></p></summary> <p>Core Jackson abstractions, basic JSON streaming API implementation</p> <p>Path to dependency file: /tmp/ws-scm/imhotep/imhotep-archive/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-core/2.2.3/jackson-core-2.2.3.jar</p> <p> Dependency Hierarchy: - hadoop-client-2.6.0-cdh5.4.11.jar (Root Library) - hadoop-aws-2.6.0-cdh5.4.11.jar - jackson-databind-2.2.3.jar - :x: **jackson-core-2.2.3.jar** (Vulnerable Library) </details> <details><summary><b>jackson-core-2.6.7.jar</b></p></summary> <p>Core Jackson abstractions, basic JSON streaming API implementation</p> <p>Library home page: <a href="https://github.com/FasterXML/jackson-core">https://github.com/FasterXML/jackson-core</a></p> <p>Path to dependency file: /tmp/ws-scm/imhotep/imhotep-server/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-core/2.6.7/jackson-core-2.6.7.jar</p> <p> Dependency Hierarchy: - jackson-databind-2.6.7.1.jar (Root Library) - :x: **jackson-core-2.6.7.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/imhotep/commit/4432df39a5fc652b4512ad35a6db8f1a3776b771">4432df39a5fc652b4512ad35a6db8f1a3776b771</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> OutOfMemoryError when writing BigDecimal In Jackson Core before version 2.7.7. When enabled the WRITE_BIGDECIMAL_AS_PLAIN setting, Jackson will attempt to write out the whole number, no matter how large the exponent. <p>Publish Date: 2016-08-25 <p>URL: <a href=https://github.com/FasterXML/jackson-core/issues/315>WS-2018-0125</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-core/releases/tag/jackson-core-2.7.7">https://github.com/FasterXML/jackson-core/releases/tag/jackson-core-2.7.7</a></p> <p>Release Date: 2016-08-25</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-core:2.7.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-core","packageVersion":"2.2.3","isTransitiveDependency":true,"dependencyTree":"org.apache.hadoop:hadoop-client:2.6.0-cdh5.4.11;org.apache.hadoop:hadoop-aws:2.6.0-cdh5.4.11;com.fasterxml.jackson.core:jackson-databind:2.2.3;com.fasterxml.jackson.core:jackson-core:2.2.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-core:2.7.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-core","packageVersion":"2.6.7","isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.7.1;com.fasterxml.jackson.core:jackson-core:2.6.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-core:2.7.7"}],"vulnerabilityIdentifier":"WS-2018-0125","vulnerabilityDetails":"OutOfMemoryError when writing BigDecimal In Jackson Core before version 2.7.7.\nWhen enabled the WRITE_BIGDECIMAL_AS_PLAIN setting, Jackson will attempt to write out the whole number, no matter how large the exponent.","vulnerabilityUrl":"https://github.com/FasterXML/jackson-core/issues/315","cvss2Severity":"medium","cvss2Score":"5.5","extraData":{}}</REMEDIATE> -->
True
WS-2018-0125 (Medium) detected in jackson-core-2.2.3.jar, jackson-core-2.6.7.jar - ## WS-2018-0125 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-core-2.2.3.jar</b>, <b>jackson-core-2.6.7.jar</b></p></summary> <p> <details><summary><b>jackson-core-2.2.3.jar</b></p></summary> <p>Core Jackson abstractions, basic JSON streaming API implementation</p> <p>Path to dependency file: /tmp/ws-scm/imhotep/imhotep-archive/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-core/2.2.3/jackson-core-2.2.3.jar</p> <p> Dependency Hierarchy: - hadoop-client-2.6.0-cdh5.4.11.jar (Root Library) - hadoop-aws-2.6.0-cdh5.4.11.jar - jackson-databind-2.2.3.jar - :x: **jackson-core-2.2.3.jar** (Vulnerable Library) </details> <details><summary><b>jackson-core-2.6.7.jar</b></p></summary> <p>Core Jackson abstractions, basic JSON streaming API implementation</p> <p>Library home page: <a href="https://github.com/FasterXML/jackson-core">https://github.com/FasterXML/jackson-core</a></p> <p>Path to dependency file: /tmp/ws-scm/imhotep/imhotep-server/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-core/2.6.7/jackson-core-2.6.7.jar</p> <p> Dependency Hierarchy: - jackson-databind-2.6.7.1.jar (Root Library) - :x: **jackson-core-2.6.7.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/imhotep/commit/4432df39a5fc652b4512ad35a6db8f1a3776b771">4432df39a5fc652b4512ad35a6db8f1a3776b771</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> OutOfMemoryError when writing BigDecimal In Jackson Core before version 2.7.7. When enabled the WRITE_BIGDECIMAL_AS_PLAIN setting, Jackson will attempt to write out the whole number, no matter how large the exponent. <p>Publish Date: 2016-08-25 <p>URL: <a href=https://github.com/FasterXML/jackson-core/issues/315>WS-2018-0125</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-core/releases/tag/jackson-core-2.7.7">https://github.com/FasterXML/jackson-core/releases/tag/jackson-core-2.7.7</a></p> <p>Release Date: 2016-08-25</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-core:2.7.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-core","packageVersion":"2.2.3","isTransitiveDependency":true,"dependencyTree":"org.apache.hadoop:hadoop-client:2.6.0-cdh5.4.11;org.apache.hadoop:hadoop-aws:2.6.0-cdh5.4.11;com.fasterxml.jackson.core:jackson-databind:2.2.3;com.fasterxml.jackson.core:jackson-core:2.2.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-core:2.7.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-core","packageVersion":"2.6.7","isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.7.1;com.fasterxml.jackson.core:jackson-core:2.6.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-core:2.7.7"}],"vulnerabilityIdentifier":"WS-2018-0125","vulnerabilityDetails":"OutOfMemoryError when writing BigDecimal In Jackson Core before version 2.7.7.\nWhen enabled the WRITE_BIGDECIMAL_AS_PLAIN setting, Jackson will attempt to write out the whole number, no matter how large the exponent.","vulnerabilityUrl":"https://github.com/FasterXML/jackson-core/issues/315","cvss2Severity":"medium","cvss2Score":"5.5","extraData":{}}</REMEDIATE> -->
non_process
ws medium detected in jackson core jar jackson core jar ws medium severity vulnerability vulnerable libraries jackson core jar jackson core jar jackson core jar core jackson abstractions basic json streaming api implementation path to dependency file tmp ws scm imhotep imhotep archive pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson core jackson core jar dependency hierarchy hadoop client jar root library hadoop aws jar jackson databind jar x jackson core jar vulnerable library jackson core jar core jackson abstractions basic json streaming api implementation library home page a href path to dependency file tmp ws scm imhotep imhotep server pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson core jackson core jar dependency hierarchy jackson databind jar root library x jackson core jar vulnerable library found in head commit a href vulnerability details outofmemoryerror when writing bigdecimal in jackson core before version when enabled the write bigdecimal as plain setting jackson will attempt to write out the whole number no matter how large the exponent publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson core isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails outofmemoryerror when writing bigdecimal in jackson core before version nwhen enabled the write bigdecimal as plain setting jackson will attempt to write out the whole number no matter how large the exponent vulnerabilityurl
0
11,872
14,673,243,470
IssuesEvent
2020-12-30 12:38:14
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] UI Issues in Overview screen
Bug P2 Process: Tested dev iOS
Steps: 1. Launch the app 2. Observe the below issues in Overview screen Actual result 1. 'Sign in' text is displayed 2. Pagination button is missing above 'Get Started' button Expected result 1. 'Sign In' text should be displayed 2. Pagination button should be displayed above 'Get Started' button ![iOS2](https://user-images.githubusercontent.com/60386291/103291036-ecefc200-4a10-11eb-95cc-2f918a962f8e.png)
1.0
[iOS] UI Issues in Overview screen - Steps: 1. Launch the app 2. Observe the below issues in Overview screen Actual result 1. 'Sign in' text is displayed 2. Pagination button is missing above 'Get Started' button Expected result 1. 'Sign In' text should be displayed 2. Pagination button should be displayed above 'Get Started' button ![iOS2](https://user-images.githubusercontent.com/60386291/103291036-ecefc200-4a10-11eb-95cc-2f918a962f8e.png)
process
ui issues in overview screen steps launch the app observe the below issues in overview screen actual result sign in text is displayed pagination button is missing above get started button expected result sign in text should be displayed pagination button should be displayed above get started button
1
678,661
23,205,980,180
IssuesEvent
2022-08-02 05:22:44
phetsims/axon
https://api.github.com/repos/phetsims/axon
closed
Can we get rid of getListenerCount?
priority:2-high dev:typescript
From https://github.com/phetsims/axon/issues/402, @marlitas and I would like to remove getListenerCount from the Emitter and Property interfaces. Current usages seem to only be in tests. Can we get rid of the tests? If not, perhaps subclass and make that method public?
1.0
Can we get rid of getListenerCount? - From https://github.com/phetsims/axon/issues/402, @marlitas and I would like to remove getListenerCount from the Emitter and Property interfaces. Current usages seem to only be in tests. Can we get rid of the tests? If not, perhaps subclass and make that method public?
non_process
can we get rid of getlistenercount from marlitas and i would like to remove getlistenercount from the emitter and property interfaces current usages seem to only be in tests can we get rid of the tests if not perhaps subclass and make that method public
0
172,550
27,297,208,425
IssuesEvent
2023-02-23 21:28:58
AlaskaAirlines/auro-nav
https://api.github.com/repos/AlaskaAirlines/auro-nav
closed
Left navigation design
auro-nav Type: Design
# Blueprint [Documentation & research ](https://www.figma.com/file/e5SFMd5WwEB27iG2rcdPcU/Navigation?node-id=1%3A9088) [Desktop design](https://www.figma.com/file/vgBapdyc1pqZGONvOOvmcJ/Left-Navigation?node-id=1908%3A8814&t=7L24WZtyFeK5ziZf-1) [Mobile design](https://www.figma.com/file/vgBapdyc1pqZGONvOOvmcJ/Left-Navigation?node-id=1952%3A35184&t=7L24WZtyFeK5ziZf-1) ## Outline tasks - [x] anatomy - [x] color - [x] typography - [x] layout - [x] spacing - [x] animation/behavior - [x] variants - [x] states (hover, focus, active, focus-visible) - [x] a11y ## Optional - [x] Competitive analysis - [x] research - [x] site audit - [x] usage audit - [x] inspirational work
1.0
Left navigation design - # Blueprint [Documentation & research ](https://www.figma.com/file/e5SFMd5WwEB27iG2rcdPcU/Navigation?node-id=1%3A9088) [Desktop design](https://www.figma.com/file/vgBapdyc1pqZGONvOOvmcJ/Left-Navigation?node-id=1908%3A8814&t=7L24WZtyFeK5ziZf-1) [Mobile design](https://www.figma.com/file/vgBapdyc1pqZGONvOOvmcJ/Left-Navigation?node-id=1952%3A35184&t=7L24WZtyFeK5ziZf-1) ## Outline tasks - [x] anatomy - [x] color - [x] typography - [x] layout - [x] spacing - [x] animation/behavior - [x] variants - [x] states (hover, focus, active, focus-visible) - [x] a11y ## Optional - [x] Competitive analysis - [x] research - [x] site audit - [x] usage audit - [x] inspirational work
non_process
left navigation design blueprint documentation research outline tasks anatomy color typography layout spacing animation behavior variants states hover focus active focus visible optional competitive analysis research site audit usage audit inspirational work
0
77,453
3,506,387,130
IssuesEvent
2016-01-08 06:22:01
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
closed
A RAID of the 5 into the team, can enter the 40 game player (BB #498)
migrated Priority: Medium Type: Bug
This issue was migrated from bitbucket. **Original Reporter:** **Original Date:** 24.02.2014 12:42:32 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** invalid **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/498 <hr> A RAID of the 5 into the team, can enter the 40 game player I'm from china, and my english is not good.
1.0
A RAID of the 5 into the team, can enter the 40 game player (BB #498) - This issue was migrated from bitbucket. **Original Reporter:** **Original Date:** 24.02.2014 12:42:32 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** invalid **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/498 <hr> A RAID of the 5 into the team, can enter the 40 game player I'm from china, and my english is not good.
non_process
a raid of the into the team can enter the game player bb this issue was migrated from bitbucket original reporter original date gmt original priority major original type bug original state invalid direct link a raid of the into the team can enter the game player i m from china and my english is not good
0
51,966
7,739,946,667
IssuesEvent
2018-05-28 18:32:09
ProjectEvergreen/project-evergreen
https://api.github.com/repos/ProjectEvergreen/project-evergreen
closed
Create a basic Todo App example
documentation enhancement todo-mvc website
based [TodoMVC](http://todomvc.com/) <img width="621" alt="screen shot 2018-05-22 at 3 40 20 pm" src="https://user-images.githubusercontent.com/895923/40385989-7bd92968-5dd6-11e8-9cf0-c18a6f4bed5f.png"> 1. Basic workflows 1. Basic CRUD functionality 1. Some basic styles / examples of things like Web Components, CSS Grid, etc
1.0
Create a basic Todo App example - based [TodoMVC](http://todomvc.com/) <img width="621" alt="screen shot 2018-05-22 at 3 40 20 pm" src="https://user-images.githubusercontent.com/895923/40385989-7bd92968-5dd6-11e8-9cf0-c18a6f4bed5f.png"> 1. Basic workflows 1. Basic CRUD functionality 1. Some basic styles / examples of things like Web Components, CSS Grid, etc
non_process
create a basic todo app example based img width alt screen shot at pm src basic workflows basic crud functionality some basic styles examples of things like web components css grid etc
0
350,016
10,477,244,674
IssuesEvent
2019-09-23 20:27:36
wherebyus/general-tasks
https://api.github.com/repos/wherebyus/general-tasks
closed
events planner beta
Added After Sprint Planning Priority: High Product: Events Type: Bug UX: Validated
## Feature or problem can't enter date-says invalid date ## UX Validation Validated ### Suggested priority High ### Stakeholders *Submitted:* cristina349 ### Definition of done How will we know when this feature is complete? ### Subtasks A detailed list of changes that need to be made or subtasks. One checkbox per. - [ ] Brew the coffee ## Developer estimate To help the team accurately estimate the complexity of this task, take a moment to walk through this list and estimate each item. At the end, you can total the estimates and round to the nearest prime number. If any of these are at a `5` or higher, or if the total is above a `5`, consider breaking this issue into multiple smaller issues. - [ ] Changes to the database () - [ ] Changes to the API () - [ ] Testing Changes to the API () - [ ] Changes to Application Code () - [ ] Adding or updating unit tests () - [ ] Local developer testing () ### Total developer estimate: 0 ## Additional estimate - [ ] Code review () - [ ] QA Testing () - [ ] Stakeholder Sign-off () - [ ] Deploy to Production () ### Total additional estimate: ## QA Notes Detailed instructions for testing, one checkbox per test to be completed. ### Contextual tests - [ ] Accessibility check - [ ] Cross-browser check (Edge, Chrome, Firefox) - [ ] Responsive check
1.0
events planner beta - ## Feature or problem can't enter date-says invalid date ## UX Validation Validated ### Suggested priority High ### Stakeholders *Submitted:* cristina349 ### Definition of done How will we know when this feature is complete? ### Subtasks A detailed list of changes that need to be made or subtasks. One checkbox per. - [ ] Brew the coffee ## Developer estimate To help the team accurately estimate the complexity of this task, take a moment to walk through this list and estimate each item. At the end, you can total the estimates and round to the nearest prime number. If any of these are at a `5` or higher, or if the total is above a `5`, consider breaking this issue into multiple smaller issues. - [ ] Changes to the database () - [ ] Changes to the API () - [ ] Testing Changes to the API () - [ ] Changes to Application Code () - [ ] Adding or updating unit tests () - [ ] Local developer testing () ### Total developer estimate: 0 ## Additional estimate - [ ] Code review () - [ ] QA Testing () - [ ] Stakeholder Sign-off () - [ ] Deploy to Production () ### Total additional estimate: ## QA Notes Detailed instructions for testing, one checkbox per test to be completed. ### Contextual tests - [ ] Accessibility check - [ ] Cross-browser check (Edge, Chrome, Firefox) - [ ] Responsive check
non_process
events planner beta feature or problem can t enter date says invalid date ux validation validated suggested priority high stakeholders submitted definition of done how will we know when this feature is complete subtasks a detailed list of changes that need to be made or subtasks one checkbox per brew the coffee developer estimate to help the team accurately estimate the complexity of this task take a moment to walk through this list and estimate each item at the end you can total the estimates and round to the nearest prime number if any of these are at a or higher or if the total is above a consider breaking this issue into multiple smaller issues changes to the database changes to the api testing changes to the api changes to application code adding or updating unit tests local developer testing total developer estimate additional estimate code review qa testing stakeholder sign off deploy to production total additional estimate qa notes detailed instructions for testing one checkbox per test to be completed contextual tests accessibility check cross browser check edge chrome firefox responsive check
0
12,854
15,239,399,279
IssuesEvent
2021-02-19 04:22:27
topcoder-platform/community-app
https://api.github.com/repos/topcoder-platform/community-app
opened
Extract Member Skills History for past challenges
ShapeupProcess challenge- recommender-tool enhancement
Extract Member Skills History for past challenges.
1.0
Extract Member Skills History for past challenges - Extract Member Skills History for past challenges.
process
extract member skills history for past challenges extract member skills history for past challenges
1
665,014
22,296,061,695
IssuesEvent
2022-06-13 01:48:25
TencentBlueKing/bk-iam-saas
https://api.github.com/repos/TencentBlueKing/bk-iam-saas
closed
[RBAC] open api: 校验用户是否某个用户组的成员
Type: Enhancement Layer: Backend Priority: High Size: S backlog
支持批量用户组,批量数有限制 在目前的接口新增, 但是需要考虑, 大表查询校验关系`存在` ----------------- 需要重新梳理现在的接口列表, 如果蓝盾切换, 去掉依赖, 那么以前的部分接口可以下掉?
1.0
[RBAC] open api: 校验用户是否某个用户组的成员 - 支持批量用户组,批量数有限制 在目前的接口新增, 但是需要考虑, 大表查询校验关系`存在` ----------------- 需要重新梳理现在的接口列表, 如果蓝盾切换, 去掉依赖, 那么以前的部分接口可以下掉?
non_process
open api 校验用户是否某个用户组的成员 支持批量用户组,批量数有限制 在目前的接口新增 但是需要考虑 大表查询校验关系 存在 需要重新梳理现在的接口列表 如果蓝盾切换 去掉依赖 那么以前的部分接口可以下掉
0
217,391
16,855,762,320
IssuesEvent
2021-06-21 06:23:37
pingcap/tidb
https://api.github.com/repos/pingcap/tidb
closed
DATA_RACE:runtime.mapassign_fast64() failed
component/test
DATA_RACE:runtime.mapassign_fast64() ``` [2020-11-12T05:22:56.562Z] WARNING: DATA RACE [2020-11-12T05:22:56.562Z] runtime.mapassign_fast64() [2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/map_fast64.go:92 +0x0 [2020-11-12T05:22:56.562Z] github.com/pingcap/parser/terror.ErrClass.NewStdErr() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/parser@v0.0.0-20201109022253-d384bee1451e/terror/terror.go:163 +0x19a [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/hint.checkInsertStmtHintDuplicated() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/dbterror/terror.go:55 +0x4be [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/hint.ExtractTableHintsFromStmtNode() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/hint/hint_processor.go:76 +0x108 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/planner.Optimize() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/planner/optimize.go:108 +0x1aa [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*Compiler).Compile() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/compiler.go:62 +0x2fa [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.(*session).ExecuteStmt() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1207 +0x270 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).Exec() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:170 +0x2f1 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).MustExec() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:216 +0x91 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session_test.(*testSessionSuite2).TestStmtHints() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session_test.go:3097 +0xd1f [2020-11-12T05:22:56.562Z] runtime.call32() [2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/asm_amd64.s:539 +0x3a [2020-11-12T05:22:56.562Z] reflect.Value.Call() [2020-11-12T05:22:56.562Z] /usr/local/go/src/reflect/value.go:321 +0xd3 [2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkTest.func1() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:850 +0x9aa [2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkCall.func1() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:739 +0x113 [2020-11-12T05:22:56.562Z] [2020-11-12T05:22:56.562Z] Previous write at 0x00c0001e0420 by goroutine 456: [2020-11-12T05:22:56.562Z] runtime.mapassign_fast64() [2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/map_fast64.go:92 +0x0 [2020-11-12T05:22:56.562Z] github.com/pingcap/parser/terror.ErrClass.NewStdErr() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/parser@v0.0.0-20201109022253-d384bee1451e/terror/terror.go:163 +0x19a [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).handleErr() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:300 +0xa54 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).fastEvalRow() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:373 +0x569 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).fastEvalRow-fm() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:359 +0xaa [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.insertRows() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:233 +0x3b6 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertExec).Next() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert.go:288 +0x117 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.Next() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/executor.go:268 +0x27d [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).handleNoDelayExecutor() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:522 +0x38e [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).handleNoDelay() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:404 +0x254 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).Exec() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:354 +0x3f6 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.runStmt() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1285 +0x2c1 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.(*session).ExecuteStmt() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1229 +0xa57 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).Exec() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:170 +0x2f1 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session_test.(*testSessionSuite).TestPrepareZero() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session_test.go:1021 +0x209 [2020-11-12T05:22:56.562Z] runtime.call32() [2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/asm_amd64.s:539 +0x3a [2020-11-12T05:22:56.562Z] reflect.Value.Call() [2020-11-12T05:22:56.562Z] /usr/local/go/src/reflect/value.go:321 +0xd3 [2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkTest.func1() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:850 +0x9aa [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall.func1() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:739 +0x113 [2020-11-12T05:22:56.563Z] [2020-11-12T05:22:56.563Z] Goroutine 459 (running) created at: [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:734 +0x4a3 [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkTest() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:832 +0x1b9 [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).doRun() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:666 +0x13a [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).asyncRun.func1() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:650 +0xf7 [2020-11-12T05:22:56.563Z] [2020-11-12T05:22:56.563Z] Goroutine 456 (running) created at: [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:734 +0x4a3 [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkTest() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:832 +0x1b9 [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).doRun() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:666 +0x13a [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).asyncRun.func1() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:650 +0xf7 [2020-11-12T05:22:56.563Z] ================== ``` Latest failed builds: https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/57742/display/redirect
1.0
DATA_RACE:runtime.mapassign_fast64() failed - DATA_RACE:runtime.mapassign_fast64() ``` [2020-11-12T05:22:56.562Z] WARNING: DATA RACE [2020-11-12T05:22:56.562Z] runtime.mapassign_fast64() [2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/map_fast64.go:92 +0x0 [2020-11-12T05:22:56.562Z] github.com/pingcap/parser/terror.ErrClass.NewStdErr() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/parser@v0.0.0-20201109022253-d384bee1451e/terror/terror.go:163 +0x19a [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/hint.checkInsertStmtHintDuplicated() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/dbterror/terror.go:55 +0x4be [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/hint.ExtractTableHintsFromStmtNode() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/hint/hint_processor.go:76 +0x108 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/planner.Optimize() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/planner/optimize.go:108 +0x1aa [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*Compiler).Compile() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/compiler.go:62 +0x2fa [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.(*session).ExecuteStmt() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1207 +0x270 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).Exec() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:170 +0x2f1 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).MustExec() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:216 +0x91 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session_test.(*testSessionSuite2).TestStmtHints() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session_test.go:3097 +0xd1f [2020-11-12T05:22:56.562Z] runtime.call32() [2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/asm_amd64.s:539 +0x3a [2020-11-12T05:22:56.562Z] reflect.Value.Call() [2020-11-12T05:22:56.562Z] /usr/local/go/src/reflect/value.go:321 +0xd3 [2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkTest.func1() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:850 +0x9aa [2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkCall.func1() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:739 +0x113 [2020-11-12T05:22:56.562Z] [2020-11-12T05:22:56.562Z] Previous write at 0x00c0001e0420 by goroutine 456: [2020-11-12T05:22:56.562Z] runtime.mapassign_fast64() [2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/map_fast64.go:92 +0x0 [2020-11-12T05:22:56.562Z] github.com/pingcap/parser/terror.ErrClass.NewStdErr() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/parser@v0.0.0-20201109022253-d384bee1451e/terror/terror.go:163 +0x19a [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).handleErr() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:300 +0xa54 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).fastEvalRow() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:373 +0x569 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).fastEvalRow-fm() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:359 +0xaa [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.insertRows() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:233 +0x3b6 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertExec).Next() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert.go:288 +0x117 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.Next() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/executor.go:268 +0x27d [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).handleNoDelayExecutor() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:522 +0x38e [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).handleNoDelay() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:404 +0x254 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).Exec() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:354 +0x3f6 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.runStmt() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1285 +0x2c1 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.(*session).ExecuteStmt() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1229 +0xa57 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).Exec() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:170 +0x2f1 [2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session_test.(*testSessionSuite).TestPrepareZero() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session_test.go:1021 +0x209 [2020-11-12T05:22:56.562Z] runtime.call32() [2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/asm_amd64.s:539 +0x3a [2020-11-12T05:22:56.562Z] reflect.Value.Call() [2020-11-12T05:22:56.562Z] /usr/local/go/src/reflect/value.go:321 +0xd3 [2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkTest.func1() [2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:850 +0x9aa [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall.func1() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:739 +0x113 [2020-11-12T05:22:56.563Z] [2020-11-12T05:22:56.563Z] Goroutine 459 (running) created at: [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:734 +0x4a3 [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkTest() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:832 +0x1b9 [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).doRun() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:666 +0x13a [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).asyncRun.func1() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:650 +0xf7 [2020-11-12T05:22:56.563Z] [2020-11-12T05:22:56.563Z] Goroutine 456 (running) created at: [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:734 +0x4a3 [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkTest() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:832 +0x1b9 [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).doRun() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:666 +0x13a [2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).asyncRun.func1() [2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:650 +0xf7 [2020-11-12T05:22:56.563Z] ================== ``` Latest failed builds: https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/57742/display/redirect
non_process
data race runtime mapassign failed data race runtime mapassign warning data race runtime mapassign usr local go src runtime map go github com pingcap parser terror errclass newstderr home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap parser terror terror go github com pingcap tidb util hint checkinsertstmthintduplicated home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util dbterror terror go github com pingcap tidb util hint extracttablehintsfromstmtnode home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util hint hint processor go github com pingcap tidb planner optimize home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb planner optimize go github com pingcap tidb executor compiler compile home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor compiler go github com pingcap tidb session session executestmt home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb session session go github com pingcap tidb util testkit testkit exec home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util testkit testkit go github com pingcap tidb util testkit testkit mustexec home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util testkit testkit go github com pingcap tidb session test teststmthints home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb session session test go runtime usr local go src runtime asm s reflect value call usr local go src reflect value go github com pingcap check suiterunner forktest home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner forkcall home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go previous write at by goroutine runtime mapassign usr local go src runtime map go github com pingcap parser terror errclass newstderr home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap parser terror terror go github com pingcap tidb executor insertvalues handleerr home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor insert common go github com pingcap tidb executor insertvalues fastevalrow home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor insert common go github com pingcap tidb executor insertvalues fastevalrow fm home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor insert common go github com pingcap tidb executor insertrows home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor insert common go github com pingcap tidb executor insertexec next home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor insert go github com pingcap tidb executor next home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor executor go github com pingcap tidb executor execstmt handlenodelayexecutor home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor adapter go github com pingcap tidb executor execstmt handlenodelay home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor adapter go github com pingcap tidb executor execstmt exec home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor adapter go github com pingcap tidb session runstmt home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb session session go github com pingcap tidb session session executestmt home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb session session go github com pingcap tidb util testkit testkit exec home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util testkit testkit go github com pingcap tidb session test testsessionsuite testpreparezero home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb session session test go runtime usr local go src runtime asm s reflect value call usr local go src reflect value go github com pingcap check suiterunner forktest home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner forkcall home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go goroutine running created at github com pingcap check suiterunner forkcall home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner forktest home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner dorun home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner asyncrun home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go goroutine running created at github com pingcap check suiterunner forkcall home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner forktest home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner dorun home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner asyncrun home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go latest failed builds
0
19,847
26,247,923,079
IssuesEvent
2023-01-05 16:40:57
python/cpython
https://api.github.com/repos/python/cpython
closed
Insecure MD5 usage in Multiprocessing.connection
type-feature type-security expert-multiprocessing
# Feature or enhancement Remove insecure use of md5 in Multiprocessing.connection # Pitch We discovered uses off the md5 hash, which has been proven insecure for more than a decade, in the Multiprocessing.connection library in the methods `deliver_challenge` and `answer_challenge`. This usage was apparently added in 2013 since the default implicit hashing mode for `hmac.new` was deprecated at that time. `hmac.new` previously defaulted to MD5 if a hashing algorithm was not specified. The 2013 change brings to code back to consistency with its prior use, but that use is insecure. It should be trivial to change the two uses in this library to a SHA2/3 secure hashing function (e.g., SHA512). Failure to update the hashing algorithm may require organizations to fully cease use of the Multiprocessing library or components of the library to meet industry security standards with respect to acceptable uses of hashing algorithms. <!-- gh-linked-prs --> ### Linked PRs * gh-100772 <!-- /gh-linked-prs -->
1.0
Insecure MD5 usage in Multiprocessing.connection - # Feature or enhancement Remove insecure use of md5 in Multiprocessing.connection # Pitch We discovered uses off the md5 hash, which has been proven insecure for more than a decade, in the Multiprocessing.connection library in the methods `deliver_challenge` and `answer_challenge`. This usage was apparently added in 2013 since the default implicit hashing mode for `hmac.new` was deprecated at that time. `hmac.new` previously defaulted to MD5 if a hashing algorithm was not specified. The 2013 change brings to code back to consistency with its prior use, but that use is insecure. It should be trivial to change the two uses in this library to a SHA2/3 secure hashing function (e.g., SHA512). Failure to update the hashing algorithm may require organizations to fully cease use of the Multiprocessing library or components of the library to meet industry security standards with respect to acceptable uses of hashing algorithms. <!-- gh-linked-prs --> ### Linked PRs * gh-100772 <!-- /gh-linked-prs -->
process
insecure usage in multiprocessing connection feature or enhancement remove insecure use of in multiprocessing connection pitch we discovered uses off the hash which has been proven insecure for more than a decade in the multiprocessing connection library in the methods deliver challenge and answer challenge this usage was apparently added in since the default implicit hashing mode for hmac new was deprecated at that time hmac new previously defaulted to if a hashing algorithm was not specified the change brings to code back to consistency with its prior use but that use is insecure it should be trivial to change the two uses in this library to a secure hashing function e g failure to update the hashing algorithm may require organizations to fully cease use of the multiprocessing library or components of the library to meet industry security standards with respect to acceptable uses of hashing algorithms linked prs gh
1
3,809
6,795,273,248
IssuesEvent
2017-11-01 15:12:12
coala/projects
https://api.github.com/repos/coala/projects
opened
Relicense text
process/pending_review
AGPL is not ideal for large chunks of text. I suggest that we use CC-BY-SA 4.0
1.0
Relicense text - AGPL is not ideal for large chunks of text. I suggest that we use CC-BY-SA 4.0
process
relicense text agpl is not ideal for large chunks of text i suggest that we use cc by sa
1
167,307
26,484,054,017
IssuesEvent
2023-01-17 16:37:54
influxdata/ui
https://api.github.com/repos/influxdata/ui
closed
Script Editor SQL Adjustments
needs/design team/automation release/marty-23.01
**Task Description** <p></p> We need to do some UI Re-Design to make the script editor SQL experience better: <p></p> 1 \- Separate the Bucket Selection from the rest of the schema browser so it is more obvious to the user that the bucket selection is required, and that the schema browser is informational only \(see design from Julia\)\. <p></p> 2 \- Change the text wording to indicate that the user needs to select a database\/bucket first\. See design from Julia\. <p></p> 3 \- When a user runs a SQL query without a bucket selected, the user will be shown an error message that tells them they must select a database\. <p></p> 4 \- The raw text that is in the editor should go away once the user clicks into the box and starts typing \(they shouldn't have to delete it manually\)\. <p></p> 5 \- Remove the time range selection as it will not apply to SQL queries right now\. ### Figma https://www.figma.com/file/0qAntPk5LVangAHguWT74X/Query-Experience-Project?node-id=2651%3A112467&t=tGT6F3oSYXF2KSZz-1
1.0
Script Editor SQL Adjustments - **Task Description** <p></p> We need to do some UI Re-Design to make the script editor SQL experience better: <p></p> 1 \- Separate the Bucket Selection from the rest of the schema browser so it is more obvious to the user that the bucket selection is required, and that the schema browser is informational only \(see design from Julia\)\. <p></p> 2 \- Change the text wording to indicate that the user needs to select a database\/bucket first\. See design from Julia\. <p></p> 3 \- When a user runs a SQL query without a bucket selected, the user will be shown an error message that tells them they must select a database\. <p></p> 4 \- The raw text that is in the editor should go away once the user clicks into the box and starts typing \(they shouldn't have to delete it manually\)\. <p></p> 5 \- Remove the time range selection as it will not apply to SQL queries right now\. ### Figma https://www.figma.com/file/0qAntPk5LVangAHguWT74X/Query-Experience-Project?node-id=2651%3A112467&t=tGT6F3oSYXF2KSZz-1
non_process
script editor sql adjustments task description we need to do some ui re design to make the script editor sql experience better separate the bucket selection from the rest of the schema browser so it is more obvious to the user that the bucket selection is required and that the schema browser is informational only see design from julia change the text wording to indicate that the user needs to select a database bucket first see design from julia when a user runs a sql query without a bucket selected the user will be shown an error message that tells them they must select a database the raw text that is in the editor should go away once the user clicks into the box and starts typing they shouldn t have to delete it manually remove the time range selection as it will not apply to sql queries right now figma
0
139,823
11,287,214,652
IssuesEvent
2020-01-16 03:32:31
opentracing-contrib/java-specialagent
https://api.github.com/repos/opentracing-contrib/java-specialagent
closed
`spring-kafka` test failing after refactor
.25 bug test
@malafeev, would you be able to help me with a test? I have refactored the module structure under `/test`, and for some reason the `spring-kafka` test is failing. Could you take a look at it? The work is in [`circleci` branch](https://github.com/opentracing-contrib/java-specialagent/blob/circleci/test/spring-kafka/spring-kafka-2.3.3/pom.xml), and here's the log of the spring-kafka tests: https://api.travis-ci.org/v3/job/637261261/log.txt
1.0
`spring-kafka` test failing after refactor - @malafeev, would you be able to help me with a test? I have refactored the module structure under `/test`, and for some reason the `spring-kafka` test is failing. Could you take a look at it? The work is in [`circleci` branch](https://github.com/opentracing-contrib/java-specialagent/blob/circleci/test/spring-kafka/spring-kafka-2.3.3/pom.xml), and here's the log of the spring-kafka tests: https://api.travis-ci.org/v3/job/637261261/log.txt
non_process
spring kafka test failing after refactor malafeev would you be able to help me with a test i have refactored the module structure under test and for some reason the spring kafka test is failing could you take a look at it the work is in and here s the log of the spring kafka tests
0
130,758
12,462,194,340
IssuesEvent
2020-05-28 08:29:47
adriens/covid19-action-plan-nc
https://api.github.com/repos/adriens/covid19-action-plan-nc
closed
Scénario de test du tableau de bord pour crowd testing
PRODUCTION Tableau de Bord documentation enhancement
# Contexte La qualité de notre application dépend de: 1 - La justesse des données: ie. sont-t-elles bien alignées avec ce qui est annoncé et la réalité du terrain ? 2 La qualité de l'affichage: typos, texte impossible à lire, taille d'image pas adaptée, graphique gros gros/petit En Nouvelle-Calédonie a été lancée une plateforme de crowdtesting basée sur le blockchain # Ressources - Article [dédié sur le Blog Hightest](https://hightest.nc/blog/posts/le-crowdtesting-met-il-en-danger-les-testeurs-professionnels ) # Scénarios envisagés ## Dernières données Si au vu des données du jour, les chiffres de correspondent pas. ## Dates - Vérifier que le nom du jour et le numéro matchent bien (ie. que ce jour existe bien dans l'année en cours) ## Typo Toute erreur de typo: orthographe, grammaire, qui nuit à la qualité de la lectre du site
1.0
Scénario de test du tableau de bord pour crowd testing - # Contexte La qualité de notre application dépend de: 1 - La justesse des données: ie. sont-t-elles bien alignées avec ce qui est annoncé et la réalité du terrain ? 2 La qualité de l'affichage: typos, texte impossible à lire, taille d'image pas adaptée, graphique gros gros/petit En Nouvelle-Calédonie a été lancée une plateforme de crowdtesting basée sur le blockchain # Ressources - Article [dédié sur le Blog Hightest](https://hightest.nc/blog/posts/le-crowdtesting-met-il-en-danger-les-testeurs-professionnels ) # Scénarios envisagés ## Dernières données Si au vu des données du jour, les chiffres de correspondent pas. ## Dates - Vérifier que le nom du jour et le numéro matchent bien (ie. que ce jour existe bien dans l'année en cours) ## Typo Toute erreur de typo: orthographe, grammaire, qui nuit à la qualité de la lectre du site
non_process
scénario de test du tableau de bord pour crowd testing contexte la qualité de notre application dépend de la justesse des données ie sont t elles bien alignées avec ce qui est annoncé et la réalité du terrain la qualité de l affichage typos texte impossible à lire taille d image pas adaptée graphique gros gros petit en nouvelle calédonie a été lancée une plateforme de crowdtesting basée sur le blockchain ressources article scénarios envisagés dernières données si au vu des données du jour les chiffres de correspondent pas dates vérifier que le nom du jour et le numéro matchent bien ie que ce jour existe bien dans l année en cours typo toute erreur de typo orthographe grammaire qui nuit à la qualité de la lectre du site
0
55,544
13,639,169,004
IssuesEvent
2020-09-25 10:32:28
astropy/astropy
https://api.github.com/repos/astropy/astropy
opened
Astropy does not build on MacOSX with xcode 12.0.0
build
<!-- This comments are hidden when you submit the issue, so you do not need to remove them! --> <!-- Please be sure to check out our contributing guidelines, https://github.com/astropy/astropy/blob/master/CONTRIBUTING.md . Please be sure to check out our code of conduct, https://github.com/astropy/astropy/blob/master/CODE_OF_CONDUCT.md . --> <!-- Please have a search on our GitHub repository to see if a similar issue has already been posted. If a similar issue is closed, have a quick look to see if you are satisfied by the resolution. If not please go ahead and open an issue! --> <!-- Please check that the development version still produces the same bug. You can install development version with pip install git+https://github.com/astropy/astropy command. --> ### Description <!-- Provide a general description of the bug. --> When I try to build astropy on my Mac (Mojave + Xcode 12.0) it fails with the following error: ``` cextern/cfitsio/lib/group.c:5664:8: error: implicit declaration of function 'getcwd' is invalid in C99 [-Werror,-Wimplicit-function-declaration] if (!getcwd(buff,FLEN_FILENAME)) ^ 1 warning and 1 error generated. error: command 'gcc' failed with exit status 1 (astropy) ➜ astropy-tmp git:(master) gcc --version Configured with: --prefix=/Applications/Xcode.app/Contents/Developer/usr --with-gxx-include-dir=/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX.sdk/usr/include/c++/4.2.1 Apple clang version 12.0.0 (clang-1200.0.32.2) Target: x86_64-apple-darwin19.5.0 Thread model: posix InstalledDir: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin ``` Xcode 12.0 was just released Sep. 16 and I'm guessing that it was automatically updated on my Mac since I did not request an update. Astropy is apparently not alone, e.g. https://gitlab.com/graphviz/graphviz/-/issues/1826. There is a long slack thread with helpful inputs from @saimn and @manodeep here: https://astropy.slack.com/archives/C067V74GK/p1600984862007700 A workaround is: ``` CFLAGS=-Wno-error=implicit-function-declaration pip install -e . ``` ### Expected behavior <!-- What did you expect to happen. --> Astropy builds and runs tests. ### Actual behavior <!-- What actually happened. --> <!-- Was the output confusing or poorly described? --> Build failed as shown. ### Steps to Reproduce <!-- Ideally a code example could be provided so we can run it ourselves. --> <!-- If you are pasting code, use triple backticks (```) around your code snippet. --> <!-- If necessary, sanitize your screen output to be pasted so you do not reveal secrets like tokens and passwords. --> 1. Checkout astropy at master, `git clean -fxd` 2. `tox -e test` or `pip install -e .` ### System Details <!-- Even if you do not think this is necessary, it is useful information for the maintainers. Please run the following snippet and paste the output below: import platform; print(platform.platform()) import sys; print("Python", sys.version) import numpy; print("Numpy", numpy.__version__) import astropy; print("astropy", astropy.__version__) import scipy; print("Scipy", scipy.__version__) import matplotlib; print("Matplotlib", matplotlib.__version__) --> ``` Darwin-19.5.0-x86_64-i386-64bit Python 3.7.7 (default, Mar 26 2020, 10:32:53) [Clang 4.0.1 (tags/RELEASE_401/final)] Numpy 1.18.1 astropy 4.2.dev715+gfaccb8b41 Scipy 1.4.1 Matplotlib 3.1.3 ```
1.0
Astropy does not build on MacOSX with xcode 12.0.0 - <!-- This comments are hidden when you submit the issue, so you do not need to remove them! --> <!-- Please be sure to check out our contributing guidelines, https://github.com/astropy/astropy/blob/master/CONTRIBUTING.md . Please be sure to check out our code of conduct, https://github.com/astropy/astropy/blob/master/CODE_OF_CONDUCT.md . --> <!-- Please have a search on our GitHub repository to see if a similar issue has already been posted. If a similar issue is closed, have a quick look to see if you are satisfied by the resolution. If not please go ahead and open an issue! --> <!-- Please check that the development version still produces the same bug. You can install development version with pip install git+https://github.com/astropy/astropy command. --> ### Description <!-- Provide a general description of the bug. --> When I try to build astropy on my Mac (Mojave + Xcode 12.0) it fails with the following error: ``` cextern/cfitsio/lib/group.c:5664:8: error: implicit declaration of function 'getcwd' is invalid in C99 [-Werror,-Wimplicit-function-declaration] if (!getcwd(buff,FLEN_FILENAME)) ^ 1 warning and 1 error generated. error: command 'gcc' failed with exit status 1 (astropy) ➜ astropy-tmp git:(master) gcc --version Configured with: --prefix=/Applications/Xcode.app/Contents/Developer/usr --with-gxx-include-dir=/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX.sdk/usr/include/c++/4.2.1 Apple clang version 12.0.0 (clang-1200.0.32.2) Target: x86_64-apple-darwin19.5.0 Thread model: posix InstalledDir: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin ``` Xcode 12.0 was just released Sep. 16 and I'm guessing that it was automatically updated on my Mac since I did not request an update. Astropy is apparently not alone, e.g. https://gitlab.com/graphviz/graphviz/-/issues/1826. There is a long slack thread with helpful inputs from @saimn and @manodeep here: https://astropy.slack.com/archives/C067V74GK/p1600984862007700 A workaround is: ``` CFLAGS=-Wno-error=implicit-function-declaration pip install -e . ``` ### Expected behavior <!-- What did you expect to happen. --> Astropy builds and runs tests. ### Actual behavior <!-- What actually happened. --> <!-- Was the output confusing or poorly described? --> Build failed as shown. ### Steps to Reproduce <!-- Ideally a code example could be provided so we can run it ourselves. --> <!-- If you are pasting code, use triple backticks (```) around your code snippet. --> <!-- If necessary, sanitize your screen output to be pasted so you do not reveal secrets like tokens and passwords. --> 1. Checkout astropy at master, `git clean -fxd` 2. `tox -e test` or `pip install -e .` ### System Details <!-- Even if you do not think this is necessary, it is useful information for the maintainers. Please run the following snippet and paste the output below: import platform; print(platform.platform()) import sys; print("Python", sys.version) import numpy; print("Numpy", numpy.__version__) import astropy; print("astropy", astropy.__version__) import scipy; print("Scipy", scipy.__version__) import matplotlib; print("Matplotlib", matplotlib.__version__) --> ``` Darwin-19.5.0-x86_64-i386-64bit Python 3.7.7 (default, Mar 26 2020, 10:32:53) [Clang 4.0.1 (tags/RELEASE_401/final)] Numpy 1.18.1 astropy 4.2.dev715+gfaccb8b41 Scipy 1.4.1 Matplotlib 3.1.3 ```
non_process
astropy does not build on macosx with xcode this comments are hidden when you submit the issue so you do not need to remove them please be sure to check out our contributing guidelines please be sure to check out our code of conduct please have a search on our github repository to see if a similar issue has already been posted if a similar issue is closed have a quick look to see if you are satisfied by the resolution if not please go ahead and open an issue please check that the development version still produces the same bug you can install development version with pip install git command description when i try to build astropy on my mac mojave xcode it fails with the following error cextern cfitsio lib group c error implicit declaration of function getcwd is invalid in if getcwd buff flen filename warning and error generated error command gcc failed with exit status astropy ➜ astropy tmp git master gcc version configured with prefix applications xcode app contents developer usr with gxx include dir applications xcode app contents developer platforms macosx platform developer sdks macosx sdk usr include c apple clang version clang target apple thread model posix installeddir applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin xcode was just released sep and i m guessing that it was automatically updated on my mac since i did not request an update astropy is apparently not alone e g there is a long slack thread with helpful inputs from saimn and manodeep here a workaround is cflags wno error implicit function declaration pip install e expected behavior astropy builds and runs tests actual behavior build failed as shown steps to reproduce if you are pasting code use triple backticks around your code snippet if necessary sanitize your screen output to be pasted so you do not reveal secrets like tokens and passwords checkout astropy at master git clean fxd tox e test or pip install e system details even if you do not think this is necessary it is useful information for the maintainers please run the following snippet and paste the output below import platform print platform platform import sys print python sys version import numpy print numpy numpy version import astropy print astropy astropy version import scipy print scipy scipy version import matplotlib print matplotlib matplotlib version darwin python default mar numpy astropy scipy matplotlib
0
11,195
13,957,701,492
IssuesEvent
2020-10-24 08:13:18
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
NO: Question regarding the INSPIRE Geoportal Discovery Service Register
Geoportal Harvesting process NO - Norway
From: Lars Inge Arnevik Sent: 05 June 2018 14:04:23 (UTC+01:00) Brussels, Copenhagen, Madrid, Paris To: inspire-geoportal@jrc.ec.europa.eu Subject: Question regarding the INSPIRE Geoportal Discovery Service Register Hi! I wonder when the next harvesting from Geonorge will take place? I think last harvest was 31. of May and I guess it was interrupted because of our monthly maintenance. We are working on improving the metadata and it is motivating to see the results from the error report J Best regards Lars-Inge
1.0
NO: Question regarding the INSPIRE Geoportal Discovery Service Register - From: Lars Inge Arnevik Sent: 05 June 2018 14:04:23 (UTC+01:00) Brussels, Copenhagen, Madrid, Paris To: inspire-geoportal@jrc.ec.europa.eu Subject: Question regarding the INSPIRE Geoportal Discovery Service Register Hi! I wonder when the next harvesting from Geonorge will take place? I think last harvest was 31. of May and I guess it was interrupted because of our monthly maintenance. We are working on improving the metadata and it is motivating to see the results from the error report J Best regards Lars-Inge
process
no question regarding the inspire geoportal discovery service register from lars inge arnevik sent june utc brussels copenhagen madrid paris to inspire geoportal jrc ec europa eu subject question regarding the inspire geoportal discovery service register hi i wonder when the next harvesting from geonorge will take place i think last harvest was of may and i guess it was interrupted because of our monthly maintenance we are working on improving the metadata and it is motivating to see the results from the error report j best regards lars inge
1
884
2,582,912,635
IssuesEvent
2015-02-15 19:52:34
code-cracker/code-cracker
https://api.github.com/repos/code-cracker/code-cracker
closed
Introduce field from constructor
3 - Done analyzer C# code-fix enhancement
This is a refactoring. The idea is that on a constructor parameter you CTRL DOT and get a "Introduce field" dialog, that will create a field and assign the value from the parameter to the field. So this: ````csharp class Foo { public Foo(string bar) { } } ```` Becomes: ````csharp class Foo { private readonly string bar; public Foo(string bar) { this.bar = bar; } } ```` If the field is already assigned then no diagnostic is offered. If the field is already present, it is just assigned to. If the type does not match (e.g. parameter is `string`, existing field is `int`, a new field with `1` postfixed is added, like that: ````csharp class Foo { private int bar; public Foo(string bar) { } } ```` Becomes: ````csharp class Foo { private int bar; private readonly string bar1; public Foo(string bar) { this.bar1 = bar; } } ```` Diagnostic Id: `CC0071` Severity: `Hidden` (refactoring) Category: `Refactoring`
1.0
Introduce field from constructor - This is a refactoring. The idea is that on a constructor parameter you CTRL DOT and get a "Introduce field" dialog, that will create a field and assign the value from the parameter to the field. So this: ````csharp class Foo { public Foo(string bar) { } } ```` Becomes: ````csharp class Foo { private readonly string bar; public Foo(string bar) { this.bar = bar; } } ```` If the field is already assigned then no diagnostic is offered. If the field is already present, it is just assigned to. If the type does not match (e.g. parameter is `string`, existing field is `int`, a new field with `1` postfixed is added, like that: ````csharp class Foo { private int bar; public Foo(string bar) { } } ```` Becomes: ````csharp class Foo { private int bar; private readonly string bar1; public Foo(string bar) { this.bar1 = bar; } } ```` Diagnostic Id: `CC0071` Severity: `Hidden` (refactoring) Category: `Refactoring`
non_process
introduce field from constructor this is a refactoring the idea is that on a constructor parameter you ctrl dot and get a introduce field dialog that will create a field and assign the value from the parameter to the field so this csharp class foo public foo string bar becomes csharp class foo private readonly string bar public foo string bar this bar bar if the field is already assigned then no diagnostic is offered if the field is already present it is just assigned to if the type does not match e g parameter is string existing field is int a new field with postfixed is added like that csharp class foo private int bar public foo string bar becomes csharp class foo private int bar private readonly string public foo string bar this bar diagnostic id severity hidden refactoring category refactoring
0
58,635
6,612,413,518
IssuesEvent
2017-09-20 03:38:37
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
StorageClass.ReclaimPolicy, PV.MountOptions, and SC.MountOptions need e2e tests
kind/cleanup kind/e2e-test-failure milestone-labels-complete priority/important-soon sig/storage
https://github.com/kubernetes/kubernetes/pull/47987 for 1.8. /assign
1.0
StorageClass.ReclaimPolicy, PV.MountOptions, and SC.MountOptions need e2e tests - https://github.com/kubernetes/kubernetes/pull/47987 for 1.8. /assign
non_process
storageclass reclaimpolicy pv mountoptions and sc mountoptions need tests for assign
0
20,434
27,098,832,376
IssuesEvent
2023-02-15 06:42:23
alibaba/MNN
https://api.github.com/repos/alibaba/MNN
closed
How can initialize image and rect?
question cv/ImageProcess
Hi, I think if I will use tracker model by mnn, I have to initialize image and ROI rect but I don't know how to this. Exactly, I'd like to implement like this https://github.com/dongfangduoshou123/DaSiamRPN-Caffe2/blob/master/main.cpp#L46 Please let me know about this. Thank you.
1.0
How can initialize image and rect? - Hi, I think if I will use tracker model by mnn, I have to initialize image and ROI rect but I don't know how to this. Exactly, I'd like to implement like this https://github.com/dongfangduoshou123/DaSiamRPN-Caffe2/blob/master/main.cpp#L46 Please let me know about this. Thank you.
process
how can initialize image and rect hi i think if i will use tracker model by mnn i have to initialize image and roi rect but i don t know how to this exactly i d like to implement like this please let me know about this thank you
1
18,115
24,146,454,150
IssuesEvent
2022-09-21 19:13:02
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
`@bazel_tools`'s `unix_cc_configure.bzl` breaks when other link opts contain `lld`
P4 type: support / not a bug (process) team-Rules-CPP
## Problem When creating [`@local_config_cc` during autoconfiguration](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/cc_configure.bzl#L125), `unix_cc_configure.bzl`'s [`_find_linker_path`](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L174-L218) is used to parse [`-v` command line output from a compiler invocation](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L197) to figure out if [`lld`/`gold` are present](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L421-L422) and what their paths are. The parsing essentially looks for the [first space separated occurrence of the linker name](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L208-L210) in the [first line containing the linker name](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L205-L207) in the verbose output and then [strips spaces and quotes from it](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L213). The [compiler invocation](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L186-L198) is essentially: `$CXX -xc - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=$linker`. All together that's essentially: ```bash $CXX -xc++ - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=$linker -v \ |& grep $linker \ | xargs -n 1 \ | grep $linker \ | tr -d ' ''"'"'" \ | head -1 ``` The problem is that this can potentially isolate things other than the path to the linker. I ran into this issue when using a [nix-shell](https://nixos.org/manual/nix/stable/command-ref/nix-shell.html) with `lldb` present; `nix` by default [wraps](https://github.com/NixOS/nixpkgs/tree/master/pkgs/build-support/cc-wrapper) its compilers (including the compiler it exposes as `$CC` and `$CXX` which Bazel's cc toolchain autoconfiguration picks up by default to make `@local_config_cc`) in a shell script that adds the contents of `$NIX_LDFLAGS` (and other env vars) to the actual compiler invocation. In its verbose output, `clang` prints out the actual compiler invocation; `_find_linker_path` interprets the `-L` include that `nix`'s wrapper adds to the invocation for `lldb`'s library files to be the linker path. This manifests in compile errors on `cc_binary`s that look like: ``` INFO: Found 1 target... ERROR: .../BUILD.bazel:127:10: Linking <some_target> [for host] failed: (Exit 1): clang failed: error executing command /nix/store/q52j3nyvc8947za806109xrxaz4dqdzf-clang-wrapper-13.0.0/bin/clang @bazel-out/host/bin/external/some/target.params Use --sandbox_debug to see verbose messages from the sandbox clang-13: error: invalid linker name in argument '-fuse-ld=/nix/store/izgkyvmb4m35pary1blnpypa9l9j059y-lldb-13.0.0-dev/include' ``` ## To Reproduce ### Using a `nix-shell` In `nix-shell -p bazelisk -p lldb -p llvmPackages_13.bintools -p llvmPackages_13.clang`: ```bash cd $(mktemp -d) touch WORKSPACE echo "int main() { return 3; }" >> main.cc echo "5.0.0" >> .bazelversion cat <<EOF > BUILD.bazel cc_binary( name = "test", srcs = ["main.cc"], ) EOF export BAZEL_USE_CPP_ONLY_TOOLCHAIN=1 bazelisk build --action_env=CC=clang --action_env=CXX=clang++ //:test -s ```` Note the use of `bazelisk` to get Bazel 5.0+; [this commit](https://github.com/bazelbuild/bazel/commit/00e30ca5968d42b4a1e42327fa683debc1063b89) introduces the use of `lld` in autoconfiguration and it landed in Bazel 5.0. Also note that this is currently not reproducible on macOS on arm64, [even with `BAZEL_USE_CPP_ONLY_TOOLCHAIN=1`](https://github.com/bazelbuild/bazel/blob/be21f194ae00d1a21e8e36a6fbb30be9e449086c/tools/cpp/cc_configure.bzl#L121-L125), because of a [bug in `nix`'s cc-wrapper](https://github.com/NixOS/nixpkgs/issues/154203) that causes it to error when attempting to sign the output "file" (`/dev/null`). Thus, `lld` and `gold` are just not used by Bazel in these cases since the compiler invocation fails. ### The Verbose Output Alternatively, here is the output of `clang++ -xc++ - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=lld -v` in such a shell: <details> <summary>Click to expand (grep for "lld")</summary> ``` clang version 13.0.0 Target: x86_64-unknown-linux-gnu Thread model: posix InstalledDir: /nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin Found candidate GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0 Found candidate GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0 Selected GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0 Candidate multilib: .;@m64 Selected multilib: .;@m64 "/nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin/clang-13" -cc1 -triple x86_64-unknown-linux-gnu -emit-obj --mrelax-relocations -disable-free -disable-llvm-verifier -discard-value-names -main-file-name - -mrelocation-model pic -pic-level 2 -fhalf-no-semantic-interposition -mframe-pointer=none -fmath-errno -fno-rounding-math -mconstructor-aliases -munwind-tables -target-cpu x86-64 -tune-cpu generic -debugger-tuning=gdb -v -fcoverage-compilation-dir=/run/user/12064113/tmp.sqsgMmvY2k -nostdsysteminc -resource-dir /nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root -idirafter /nix/store/sr2bm5pp15dhb80w16vrwyhgh40kd4yf-glibc-2.33-59-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0 -isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu -D _FORTIFY_SOURCE=2 -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0 -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/x86_64-unknown-linux-gnu -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/backward -internal-isystem /nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root/include -O2 -Wformat -Wformat-security -Werror=format-security -fdeprecated-macro -fdebug-compilation-dir=/run/user/12064113/tmp.sqsgMmvY2k -ferror-limit 19 -fwrapv -stack-protector 2 -stack-protector-buffer-size 4 -fgnuc-version=4.2.1 -fcxx-exceptions -fexceptions -fcolor-diagnostics -vectorize-loops -vectorize-slp -faddrsig -D__GCC_HAVE_DWARF2_CFI_ASM=1 -o /run/user/12064113/--158e23.o -x c++ - clang -cc1 version 13.0.0 based upon LLVM 13.0.0 default target x86_64-unknown-linux-gnu ignoring duplicate directory "/nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include" ignoring duplicate directory "/nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include" ignoring duplicate directory "/nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include" ignoring duplicate directory "/nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include" ignoring duplicate directory "/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0" ignoring duplicate directory "/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu" #include "..." search starts here: #include <...> search starts here: /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0 /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/backward /nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root/include /nix/store/sr2bm5pp15dhb80w16vrwyhgh40kd4yf-glibc-2.33-59-dev/include End of search list. "/nix/store/kcg1vn8a2xf0fzd789fq16mww4wqvg0n-llvm-binutils-wrapper-13.0.0/bin/ld.lld" -z relro --hash-style=gnu --hash-style=both --enable-new-dtags --eh-frame-hdr -m elf_x86_64 -o /dev/null /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crt1.o /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crti.o /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0/crtbegin.o -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0 -L/nix/store/lwkpcfphv5nwymi0bvnvfga6q4p72v8i-gcc-10.3.0-lib/x86_64-unknown-linux-gnu/lib -L/nix/store/svzwv5yhcp03rysmc0lh5922a6f7qgp8-clang-13.0.0-lib/lib -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0 -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../lib64 -L/lib/../lib64 -L/usr/lib/../lib64 -L/nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin/../lib -L/lib -L/usr/lib -dynamic-linker=/nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/ld-linux-x86-64.so.2 /run/user/12064113/--158e23.o --start-lib --end-lib -rpath /nix/store/4ibh63jd2lzx03mrhjn9km8ihha9wq55-shell/lib64 -rpath /nix/store/4ibh63jd2lzx03mrhjn9km8ihha9wq55-shell/lib -lstdc++ -lm -lgcc_s -lgcc -lc -lgcc_s -lgcc /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0/crtend.o /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crtn.o ``` </details> ## Potential Solutions A trivial fix would be to search for the linker name plus a space or end quote (i.e. `lld `/`lld"`/`lld'`, `gold `/`gold"`/`gold'`) or, for `lld` specifically, to search for `.lld` (afaik all the flavors of `lld` *for unix platforms* are `<something>.lld`: `ld64.lld`, `ld.lld`). This _seems_ like it'd fix this issue without relying on compiler-specific output. Another potential solution is to only use the _last_ line of the verbose output. However, I am not sure we can rely on this always being the linker invocation, especially for compilers that aren't `clang`. Ideally we'd just ask the compiler for the path to the linker but I don't know of any way to do so that's well supported across different compilers. If any of these solutions are deemed a reasonable fix (or if a better path forward is proposed), I'm happy to make a PR. ## Workarounds To anyone facing a similar issue, my workaround for now was to simply wrap the `lldb` derivation and have it use another name: ```nix with nixpkgs; stdenvNoCC.mkDerivation { pname = "llvm-debugger"; inherit (lldb) version outputs; lldb_outputs = lldb.all; src = "${lldb}"; disallowedReferences = [lldb]; buildPhase = '' i=0 lldb_outs=($lldb_outputs) outs=($outputs) for o in "''${outs[@]}"; do echo "[$o]" "''${lldb_outs[$i]}" to "''${!o}" # can't symlink because then the paths used for `LDFLAGS` still # contain "lldb": cp -R "''${lldb_outs[$i]}" "''${!o}" : $((++i)) done # Remove some lingering references to the actual lldb package so that it's # not registered as a runtime dep of this package. scrub() { local i=0 for o in "''${outs[@]}"; do substituteInPlace "$1" \ --replace "''${lldb_outs[$i]}" "''${!o}" : $((++i)) done } scrub "$out/bin/lldb" scrub "$dev/nix-support/propagated-build-inputs" ''; installPhase = "true"; } # Alternatively the below will work too but will cause you to rebuild `lldb` from source: # nixpkgs.lldb.overrideDerivation (o: { name = "llvm-debugger"; }) ```
1.0
`@bazel_tools`'s `unix_cc_configure.bzl` breaks when other link opts contain `lld` - ## Problem When creating [`@local_config_cc` during autoconfiguration](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/cc_configure.bzl#L125), `unix_cc_configure.bzl`'s [`_find_linker_path`](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L174-L218) is used to parse [`-v` command line output from a compiler invocation](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L197) to figure out if [`lld`/`gold` are present](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L421-L422) and what their paths are. The parsing essentially looks for the [first space separated occurrence of the linker name](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L208-L210) in the [first line containing the linker name](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L205-L207) in the verbose output and then [strips spaces and quotes from it](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L213). The [compiler invocation](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L186-L198) is essentially: `$CXX -xc - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=$linker`. All together that's essentially: ```bash $CXX -xc++ - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=$linker -v \ |& grep $linker \ | xargs -n 1 \ | grep $linker \ | tr -d ' ''"'"'" \ | head -1 ``` The problem is that this can potentially isolate things other than the path to the linker. I ran into this issue when using a [nix-shell](https://nixos.org/manual/nix/stable/command-ref/nix-shell.html) with `lldb` present; `nix` by default [wraps](https://github.com/NixOS/nixpkgs/tree/master/pkgs/build-support/cc-wrapper) its compilers (including the compiler it exposes as `$CC` and `$CXX` which Bazel's cc toolchain autoconfiguration picks up by default to make `@local_config_cc`) in a shell script that adds the contents of `$NIX_LDFLAGS` (and other env vars) to the actual compiler invocation. In its verbose output, `clang` prints out the actual compiler invocation; `_find_linker_path` interprets the `-L` include that `nix`'s wrapper adds to the invocation for `lldb`'s library files to be the linker path. This manifests in compile errors on `cc_binary`s that look like: ``` INFO: Found 1 target... ERROR: .../BUILD.bazel:127:10: Linking <some_target> [for host] failed: (Exit 1): clang failed: error executing command /nix/store/q52j3nyvc8947za806109xrxaz4dqdzf-clang-wrapper-13.0.0/bin/clang @bazel-out/host/bin/external/some/target.params Use --sandbox_debug to see verbose messages from the sandbox clang-13: error: invalid linker name in argument '-fuse-ld=/nix/store/izgkyvmb4m35pary1blnpypa9l9j059y-lldb-13.0.0-dev/include' ``` ## To Reproduce ### Using a `nix-shell` In `nix-shell -p bazelisk -p lldb -p llvmPackages_13.bintools -p llvmPackages_13.clang`: ```bash cd $(mktemp -d) touch WORKSPACE echo "int main() { return 3; }" >> main.cc echo "5.0.0" >> .bazelversion cat <<EOF > BUILD.bazel cc_binary( name = "test", srcs = ["main.cc"], ) EOF export BAZEL_USE_CPP_ONLY_TOOLCHAIN=1 bazelisk build --action_env=CC=clang --action_env=CXX=clang++ //:test -s ```` Note the use of `bazelisk` to get Bazel 5.0+; [this commit](https://github.com/bazelbuild/bazel/commit/00e30ca5968d42b4a1e42327fa683debc1063b89) introduces the use of `lld` in autoconfiguration and it landed in Bazel 5.0. Also note that this is currently not reproducible on macOS on arm64, [even with `BAZEL_USE_CPP_ONLY_TOOLCHAIN=1`](https://github.com/bazelbuild/bazel/blob/be21f194ae00d1a21e8e36a6fbb30be9e449086c/tools/cpp/cc_configure.bzl#L121-L125), because of a [bug in `nix`'s cc-wrapper](https://github.com/NixOS/nixpkgs/issues/154203) that causes it to error when attempting to sign the output "file" (`/dev/null`). Thus, `lld` and `gold` are just not used by Bazel in these cases since the compiler invocation fails. ### The Verbose Output Alternatively, here is the output of `clang++ -xc++ - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=lld -v` in such a shell: <details> <summary>Click to expand (grep for "lld")</summary> ``` clang version 13.0.0 Target: x86_64-unknown-linux-gnu Thread model: posix InstalledDir: /nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin Found candidate GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0 Found candidate GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0 Selected GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0 Candidate multilib: .;@m64 Selected multilib: .;@m64 "/nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin/clang-13" -cc1 -triple x86_64-unknown-linux-gnu -emit-obj --mrelax-relocations -disable-free -disable-llvm-verifier -discard-value-names -main-file-name - -mrelocation-model pic -pic-level 2 -fhalf-no-semantic-interposition -mframe-pointer=none -fmath-errno -fno-rounding-math -mconstructor-aliases -munwind-tables -target-cpu x86-64 -tune-cpu generic -debugger-tuning=gdb -v -fcoverage-compilation-dir=/run/user/12064113/tmp.sqsgMmvY2k -nostdsysteminc -resource-dir /nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root -idirafter /nix/store/sr2bm5pp15dhb80w16vrwyhgh40kd4yf-glibc-2.33-59-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0 -isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu -D _FORTIFY_SOURCE=2 -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0 -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/x86_64-unknown-linux-gnu -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/backward -internal-isystem /nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root/include -O2 -Wformat -Wformat-security -Werror=format-security -fdeprecated-macro -fdebug-compilation-dir=/run/user/12064113/tmp.sqsgMmvY2k -ferror-limit 19 -fwrapv -stack-protector 2 -stack-protector-buffer-size 4 -fgnuc-version=4.2.1 -fcxx-exceptions -fexceptions -fcolor-diagnostics -vectorize-loops -vectorize-slp -faddrsig -D__GCC_HAVE_DWARF2_CFI_ASM=1 -o /run/user/12064113/--158e23.o -x c++ - clang -cc1 version 13.0.0 based upon LLVM 13.0.0 default target x86_64-unknown-linux-gnu ignoring duplicate directory "/nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include" ignoring duplicate directory "/nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include" ignoring duplicate directory "/nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include" ignoring duplicate directory "/nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include" ignoring duplicate directory "/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0" ignoring duplicate directory "/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu" #include "..." search starts here: #include <...> search starts here: /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0 /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/backward /nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root/include /nix/store/sr2bm5pp15dhb80w16vrwyhgh40kd4yf-glibc-2.33-59-dev/include End of search list. "/nix/store/kcg1vn8a2xf0fzd789fq16mww4wqvg0n-llvm-binutils-wrapper-13.0.0/bin/ld.lld" -z relro --hash-style=gnu --hash-style=both --enable-new-dtags --eh-frame-hdr -m elf_x86_64 -o /dev/null /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crt1.o /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crti.o /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0/crtbegin.o -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0 -L/nix/store/lwkpcfphv5nwymi0bvnvfga6q4p72v8i-gcc-10.3.0-lib/x86_64-unknown-linux-gnu/lib -L/nix/store/svzwv5yhcp03rysmc0lh5922a6f7qgp8-clang-13.0.0-lib/lib -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0 -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../lib64 -L/lib/../lib64 -L/usr/lib/../lib64 -L/nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin/../lib -L/lib -L/usr/lib -dynamic-linker=/nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/ld-linux-x86-64.so.2 /run/user/12064113/--158e23.o --start-lib --end-lib -rpath /nix/store/4ibh63jd2lzx03mrhjn9km8ihha9wq55-shell/lib64 -rpath /nix/store/4ibh63jd2lzx03mrhjn9km8ihha9wq55-shell/lib -lstdc++ -lm -lgcc_s -lgcc -lc -lgcc_s -lgcc /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0/crtend.o /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crtn.o ``` </details> ## Potential Solutions A trivial fix would be to search for the linker name plus a space or end quote (i.e. `lld `/`lld"`/`lld'`, `gold `/`gold"`/`gold'`) or, for `lld` specifically, to search for `.lld` (afaik all the flavors of `lld` *for unix platforms* are `<something>.lld`: `ld64.lld`, `ld.lld`). This _seems_ like it'd fix this issue without relying on compiler-specific output. Another potential solution is to only use the _last_ line of the verbose output. However, I am not sure we can rely on this always being the linker invocation, especially for compilers that aren't `clang`. Ideally we'd just ask the compiler for the path to the linker but I don't know of any way to do so that's well supported across different compilers. If any of these solutions are deemed a reasonable fix (or if a better path forward is proposed), I'm happy to make a PR. ## Workarounds To anyone facing a similar issue, my workaround for now was to simply wrap the `lldb` derivation and have it use another name: ```nix with nixpkgs; stdenvNoCC.mkDerivation { pname = "llvm-debugger"; inherit (lldb) version outputs; lldb_outputs = lldb.all; src = "${lldb}"; disallowedReferences = [lldb]; buildPhase = '' i=0 lldb_outs=($lldb_outputs) outs=($outputs) for o in "''${outs[@]}"; do echo "[$o]" "''${lldb_outs[$i]}" to "''${!o}" # can't symlink because then the paths used for `LDFLAGS` still # contain "lldb": cp -R "''${lldb_outs[$i]}" "''${!o}" : $((++i)) done # Remove some lingering references to the actual lldb package so that it's # not registered as a runtime dep of this package. scrub() { local i=0 for o in "''${outs[@]}"; do substituteInPlace "$1" \ --replace "''${lldb_outs[$i]}" "''${!o}" : $((++i)) done } scrub "$out/bin/lldb" scrub "$dev/nix-support/propagated-build-inputs" ''; installPhase = "true"; } # Alternatively the below will work too but will cause you to rebuild `lldb` from source: # nixpkgs.lldb.overrideDerivation (o: { name = "llvm-debugger"; }) ```
process
bazel tools s unix cc configure bzl breaks when other link opts contain lld problem when creating unix cc configure bzl s is used to parse to figure out if and what their paths are the parsing essentially looks for the in the in the verbose output and then the is essentially cxx xc int main o dev null wl start lib wl end lib fuse ld linker all together that s essentially bash cxx xc int main o dev null wl start lib wl end lib fuse ld linker v grep linker xargs n grep linker tr d head the problem is that this can potentially isolate things other than the path to the linker i ran into this issue when using a with lldb present nix by default its compilers including the compiler it exposes as cc and cxx which bazel s cc toolchain autoconfiguration picks up by default to make local config cc in a shell script that adds the contents of nix ldflags and other env vars to the actual compiler invocation in its verbose output clang prints out the actual compiler invocation find linker path interprets the l include that nix s wrapper adds to the invocation for lldb s library files to be the linker path this manifests in compile errors on cc binary s that look like info found target error build bazel linking failed exit clang failed error executing command nix store clang wrapper bin clang bazel out host bin external some target params use sandbox debug to see verbose messages from the sandbox clang error invalid linker name in argument fuse ld nix store lldb dev include to reproduce using a nix shell in nix shell p bazelisk p lldb p llvmpackages bintools p llvmpackages clang bash cd mktemp d touch workspace echo int main return main cc echo bazelversion cat build bazel cc binary name test srcs eof export bazel use cpp only toolchain bazelisk build action env cc clang action env cxx clang test s note the use of bazelisk to get bazel introduces the use of lld in autoconfiguration and it landed in bazel also note that this is currently not reproducible on macos on because of a that causes it to error when attempting to sign the output file dev null thus lld and gold are just not used by bazel in these cases since the compiler invocation fails the verbose output alternatively here is the output of clang xc int main o dev null wl start lib wl end lib fuse ld lld v in such a shell click to expand grep for lld clang version target unknown linux gnu thread model posix installeddir nix store clang bin found candidate gcc installation nix store gcc lib gcc unknown linux gnu found candidate gcc installation nix store gcc gcc unknown linux gnu selected gcc installation nix store gcc gcc unknown linux gnu candidate multilib selected multilib nix store clang bin clang triple unknown linux gnu emit obj mrelax relocations disable free disable llvm verifier discard value names main file name mrelocation model pic pic level fhalf no semantic interposition mframe pointer none fmath errno fno rounding math mconstructor aliases munwind tables target cpu tune cpu generic debugger tuning gdb v fcoverage compilation dir run user tmp nostdsysteminc resource dir nix store clang wrapper resource root idirafter nix store glibc dev include isystem nix store lldb dev include isystem nix store compiler rt libc dev include isystem nix store lldb dev include isystem nix store compiler rt libc dev include isystem nix store lldb dev include isystem nix store compiler rt libc dev include isystem nix store gcc include c isystem nix store gcc include c unknown linux gnu d fortify source internal isystem nix store gcc gcc unknown linux gnu include c internal isystem nix store gcc gcc unknown linux gnu include c unknown linux gnu internal isystem nix store gcc gcc unknown linux gnu include c backward internal isystem nix store clang wrapper resource root include wformat wformat security werror format security fdeprecated macro fdebug compilation dir run user tmp ferror limit fwrapv stack protector stack protector buffer size fgnuc version fcxx exceptions fexceptions fcolor diagnostics vectorize loops vectorize slp faddrsig d gcc have cfi asm o run user o x c clang version based upon llvm default target unknown linux gnu ignoring duplicate directory nix store lldb dev include ignoring duplicate directory nix store compiler rt libc dev include ignoring duplicate directory nix store lldb dev include ignoring duplicate directory nix store compiler rt libc dev include ignoring duplicate directory nix store gcc include c ignoring duplicate directory nix store gcc include c unknown linux gnu include search starts here include search starts here nix store lldb dev include nix store compiler rt libc dev include nix store gcc include c nix store gcc include c unknown linux gnu nix store gcc gcc unknown linux gnu include c backward nix store clang wrapper resource root include nix store glibc dev include end of search list nix store llvm binutils wrapper bin ld lld z relro hash style gnu hash style both enable new dtags eh frame hdr m elf o dev null nix store glibc lib o nix store glibc lib crti o nix store gcc lib gcc unknown linux gnu crtbegin o l nix store lldb lib lib l nix store lldb lib lib l nix store lldb lib lib l nix store lldb lib lib l nix store glibc lib l nix store gcc lib gcc unknown linux gnu l nix store gcc lib unknown linux gnu lib l nix store clang lib lib l nix store gcc gcc unknown linux gnu l nix store gcc gcc unknown linux gnu l lib l usr lib l nix store clang bin lib l lib l usr lib dynamic linker nix store glibc lib ld linux so run user o start lib end lib rpath nix store shell rpath nix store shell lib lstdc lm lgcc s lgcc lc lgcc s lgcc nix store gcc lib gcc unknown linux gnu crtend o nix store glibc lib crtn o potential solutions a trivial fix would be to search for the linker name plus a space or end quote i e lld lld lld gold gold gold or for lld specifically to search for lld afaik all the flavors of lld for unix platforms are lld lld ld lld this seems like it d fix this issue without relying on compiler specific output another potential solution is to only use the last line of the verbose output however i am not sure we can rely on this always being the linker invocation especially for compilers that aren t clang ideally we d just ask the compiler for the path to the linker but i don t know of any way to do so that s well supported across different compilers if any of these solutions are deemed a reasonable fix or if a better path forward is proposed i m happy to make a pr workarounds to anyone facing a similar issue my workaround for now was to simply wrap the lldb derivation and have it use another name nix with nixpkgs stdenvnocc mkderivation pname llvm debugger inherit lldb version outputs lldb outputs lldb all src lldb disallowedreferences buildphase i lldb outs lldb outputs outs outputs for o in outs do echo lldb outs to o can t symlink because then the paths used for ldflags still contain lldb cp r lldb outs o i done remove some lingering references to the actual lldb package so that it s not registered as a runtime dep of this package scrub local i for o in outs do substituteinplace replace lldb outs o i done scrub out bin lldb scrub dev nix support propagated build inputs installphase true alternatively the below will work too but will cause you to rebuild lldb from source nixpkgs lldb overridederivation o name llvm debugger
1
11,181
13,957,695,473
IssuesEvent
2020-10-24 08:11:33
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
RO: new harvesting
Geoportal Harvesting process RO - Romania
Dear Angelo, Can we have a new harvesting of our national geoportal, please? We&#39;ve made some changes in order to link the resources. Best regards, Simona Bunea
1.0
RO: new harvesting - Dear Angelo, Can we have a new harvesting of our national geoportal, please? We&#39;ve made some changes in order to link the resources. Best regards, Simona Bunea
process
ro new harvesting dear angelo can we have a new harvesting of our national geoportal please we ve made some changes in order to link the resources best regards simona bunea
1
19,647
26,006,252,622
IssuesEvent
2022-12-20 19:41:13
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Release automation set wrong VERSION
bug regression process
### Description The recent changes to release automation github workflow causes wrong version set and the subsequent issues - other workflows triggered by the new tag fails because the version is `$VERSION` - the create SNAPSHOT PR job fails for the same reason ### Steps to reproduce Check the release automation workflow run for 0.71.0-beta1 and the corresponding triggered workflow runs ### Additional context _No response_ ### Hedera network other ### Version 0.71.0-SNAPSHOT ### Operating system None
1.0
Release automation set wrong VERSION - ### Description The recent changes to release automation github workflow causes wrong version set and the subsequent issues - other workflows triggered by the new tag fails because the version is `$VERSION` - the create SNAPSHOT PR job fails for the same reason ### Steps to reproduce Check the release automation workflow run for 0.71.0-beta1 and the corresponding triggered workflow runs ### Additional context _No response_ ### Hedera network other ### Version 0.71.0-SNAPSHOT ### Operating system None
process
release automation set wrong version description the recent changes to release automation github workflow causes wrong version set and the subsequent issues other workflows triggered by the new tag fails because the version is version the create snapshot pr job fails for the same reason steps to reproduce check the release automation workflow run for and the corresponding triggered workflow runs additional context no response hedera network other version snapshot operating system none
1
14,582
17,703,494,332
IssuesEvent
2021-08-25 03:08:39
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
New Term - MaterialCitation
Term - add Class - new normative Process - complete
## New term * Submitter: Plazi (Donat Agosti, @myrmoteras) * Justification (why is this term necessary?): Specimens are the basis for taxonomic research and are cited in taxonomic treatments and other works. Increasingly these material citations are extracted from publications and submitted as part of data sets to GBIF and reused in studies. Currently GBIF includes 33,199 datasets derived from taxonomic publications and 408,021 material citations which are labeled as occurrences. An estimate of 45,000 species are only represented in GBIF through material citations from publications, mainly covering new species that are extracted from the literature just after the publishing time and thus cut short the time to enter taxonomic names and re-submit datasets to GBIF. This leads to confusion and discussions (https://discourse.gbif.org/t/basisofrecord-for-plazi-datasets/2238/16 ) which need to be resolved, not least because material citations are not specimens per se, but the citation of a specimen. Furthermore, material citations can be part of a specimen, a specimen, or groups of specimens. They can be very verbose or very cursory. * Proponents (at least two independent parties who need this term). Demand requirement. There are four main groups: the publishers of taxonomic and other scholarly publications; the data converters; the re-users of the data, e.g. data aggregators and scientists analysing data. Three publishers already make use of material citations: Pensoft GmbH, the Science Press of the Muséum d’histoire naturelle Paris, and the CETAF publishing group. As a data converter, Plazi so far produces [400,000 material citations](https://www.gbif.org/publisher/7ce8aef0-9e92-11dc-8738-b8a03c50a862) from 33,000 publications and submits them to GBIF. This includes data from xx scientific journals. GBIF is importing material citations as occurrence data and would like to separate these data from other occurrence data. Scientists used data originating from taxonomic publications in over 320 scientific articles. Furthermore there is a demand raised by GBIF users to separate material citations from specimen occurrence data (see above). There is clearly a demand to introduce this new class. Efficacy requirement: Within this community there is a consensus that this new class will accomplish the desired outcome. The equivalent term material-citation (https://terms.tdwg.org/wiki/tp:material-citation) in the TaxPub Journal Article Tag Suit is already used in the production of over 30 scholarly journals (eg. http://plazi.org/resources/schemas-and-ontologies/taxpub/ ) . Stability requirement: Since this is a new class, this will not interfere with existing implementations but rather contribute to resolve a well known issue. Proposed attributes of the new term: * Term name (in lowerCamelCase): MaterialCitation (modified by @tucotuco) * Class (e.g. Location, Taxon): None (modified by @tucotuco - the proposal is for a Class, not a property) * Definition of the term: A reference to or citation of one, a part of, or multiple specimens in scholarly publications. (modified by @tucotuco) * Usage comments (recommendations regarding content, etc.): This class constitutes a new value for the controlled vocabulary in the recommendations for basisOfRecord. When importing Darwin Core Archives of literature-based datasets to GBIF, the basisOfRecord should be changed from “Occurrence”, "PreservedSpecimen" or "Literature" to “MaterialCitation”. * Examples: * the citation of a physical specimen from a scientific collection in a taxonomic treatment in a scientific publication. * the citation of a group of physical specimens, such as paratypes in a taxonomic treatment in a scientific publication. * the occurrence mentioned in a field note book * Refines (identifier of the broader term this term refines, if applicable): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): None * ABCD 2.06 (XPATH of the equivalent term in ABCD, if applicable): Not in ABCD. BioCASe/ABCD provides for a slightly different set of values: ``` <xs:enumeration value="PreservedSpecimen"/> <xs:enumeration value="LivingSpecimen"/> <xs:enumeration value="FossileSpecimen"/> <xs:enumeration value="OtherSpecimen"/> <xs:enumeration value="HumanObservation"/> <xs:enumeration value="MachineObservation"/> <xs:enumeration value="DrawingOrPhotograph"/> <xs:enumeration value="MultimediaObject"/> <xs:enumeration value="AbsenceObservation"/> ```
1.0
New Term - MaterialCitation - ## New term * Submitter: Plazi (Donat Agosti, @myrmoteras) * Justification (why is this term necessary?): Specimens are the basis for taxonomic research and are cited in taxonomic treatments and other works. Increasingly these material citations are extracted from publications and submitted as part of data sets to GBIF and reused in studies. Currently GBIF includes 33,199 datasets derived from taxonomic publications and 408,021 material citations which are labeled as occurrences. An estimate of 45,000 species are only represented in GBIF through material citations from publications, mainly covering new species that are extracted from the literature just after the publishing time and thus cut short the time to enter taxonomic names and re-submit datasets to GBIF. This leads to confusion and discussions (https://discourse.gbif.org/t/basisofrecord-for-plazi-datasets/2238/16 ) which need to be resolved, not least because material citations are not specimens per se, but the citation of a specimen. Furthermore, material citations can be part of a specimen, a specimen, or groups of specimens. They can be very verbose or very cursory. * Proponents (at least two independent parties who need this term). Demand requirement. There are four main groups: the publishers of taxonomic and other scholarly publications; the data converters; the re-users of the data, e.g. data aggregators and scientists analysing data. Three publishers already make use of material citations: Pensoft GmbH, the Science Press of the Muséum d’histoire naturelle Paris, and the CETAF publishing group. As a data converter, Plazi so far produces [400,000 material citations](https://www.gbif.org/publisher/7ce8aef0-9e92-11dc-8738-b8a03c50a862) from 33,000 publications and submits them to GBIF. This includes data from xx scientific journals. GBIF is importing material citations as occurrence data and would like to separate these data from other occurrence data. Scientists used data originating from taxonomic publications in over 320 scientific articles. Furthermore there is a demand raised by GBIF users to separate material citations from specimen occurrence data (see above). There is clearly a demand to introduce this new class. Efficacy requirement: Within this community there is a consensus that this new class will accomplish the desired outcome. The equivalent term material-citation (https://terms.tdwg.org/wiki/tp:material-citation) in the TaxPub Journal Article Tag Suit is already used in the production of over 30 scholarly journals (eg. http://plazi.org/resources/schemas-and-ontologies/taxpub/ ) . Stability requirement: Since this is a new class, this will not interfere with existing implementations but rather contribute to resolve a well known issue. Proposed attributes of the new term: * Term name (in lowerCamelCase): MaterialCitation (modified by @tucotuco) * Class (e.g. Location, Taxon): None (modified by @tucotuco - the proposal is for a Class, not a property) * Definition of the term: A reference to or citation of one, a part of, or multiple specimens in scholarly publications. (modified by @tucotuco) * Usage comments (recommendations regarding content, etc.): This class constitutes a new value for the controlled vocabulary in the recommendations for basisOfRecord. When importing Darwin Core Archives of literature-based datasets to GBIF, the basisOfRecord should be changed from “Occurrence”, "PreservedSpecimen" or "Literature" to “MaterialCitation”. * Examples: * the citation of a physical specimen from a scientific collection in a taxonomic treatment in a scientific publication. * the citation of a group of physical specimens, such as paratypes in a taxonomic treatment in a scientific publication. * the occurrence mentioned in a field note book * Refines (identifier of the broader term this term refines, if applicable): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): None * ABCD 2.06 (XPATH of the equivalent term in ABCD, if applicable): Not in ABCD. BioCASe/ABCD provides for a slightly different set of values: ``` <xs:enumeration value="PreservedSpecimen"/> <xs:enumeration value="LivingSpecimen"/> <xs:enumeration value="FossileSpecimen"/> <xs:enumeration value="OtherSpecimen"/> <xs:enumeration value="HumanObservation"/> <xs:enumeration value="MachineObservation"/> <xs:enumeration value="DrawingOrPhotograph"/> <xs:enumeration value="MultimediaObject"/> <xs:enumeration value="AbsenceObservation"/> ```
process
new term materialcitation new term submitter plazi donat agosti myrmoteras justification why is this term necessary specimens are the basis for taxonomic research and are cited in taxonomic treatments and other works increasingly these material citations are extracted from publications and submitted as part of data sets to gbif and reused in studies currently gbif includes datasets derived from taxonomic publications and material citations which are labeled as occurrences an estimate of species are only represented in gbif through material citations from publications mainly covering new species that are extracted from the literature just after the publishing time and thus cut short the time to enter taxonomic names and re submit datasets to gbif this leads to confusion and discussions which need to be resolved not least because material citations are not specimens per se but the citation of a specimen furthermore material citations can be part of a specimen a specimen or groups of specimens they can be very verbose or very cursory proponents at least two independent parties who need this term demand requirement there are four main groups the publishers of taxonomic and other scholarly publications the data converters the re users of the data e g data aggregators and scientists analysing data three publishers already make use of material citations pensoft gmbh the science press of the muséum d’histoire naturelle paris and the cetaf publishing group as a data converter plazi so far produces from publications and submits them to gbif this includes data from xx scientific journals gbif is importing material citations as occurrence data and would like to separate these data from other occurrence data scientists used data originating from taxonomic publications in over scientific articles furthermore there is a demand raised by gbif users to separate material citations from specimen occurrence data see above there is clearly a demand to introduce this new class efficacy requirement within this community there is a consensus that this new class will accomplish the desired outcome the equivalent term material citation in the taxpub journal article tag suit is already used in the production of over scholarly journals eg stability requirement since this is a new class this will not interfere with existing implementations but rather contribute to resolve a well known issue proposed attributes of the new term term name in lowercamelcase materialcitation modified by tucotuco class e g location taxon none modified by tucotuco the proposal is for a class not a property definition of the term a reference to or citation of one a part of or multiple specimens in scholarly publications modified by tucotuco usage comments recommendations regarding content etc this class constitutes a new value for the controlled vocabulary in the recommendations for basisofrecord when importing darwin core archives of literature based datasets to gbif the basisofrecord should be changed from “occurrence” preservedspecimen or literature to “materialcitation” examples the citation of a physical specimen from a scientific collection in a taxonomic treatment in a scientific publication the citation of a group of physical specimens such as paratypes in a taxonomic treatment in a scientific publication the occurrence mentioned in a field note book refines identifier of the broader term this term refines if applicable none replaces identifier of the existing term that would be deprecated and replaced by this term if applicable none abcd xpath of the equivalent term in abcd if applicable not in abcd biocase abcd provides for a slightly different set of values
1
556,206
16,477,594,493
IssuesEvent
2021-05-24 07:46:28
edwisely-ai/Tech-Bridge
https://api.github.com/repos/edwisely-ai/Tech-Bridge
opened
Dark Mode Issue the App (Student)
Criticality High Priority High
whenever dark mode is activated the students cannot see the Questions. This is a serious issue in the Front End while Writing exams, This should be a Priority and must be resolved immediately Note: The Thing Should be Checked with Faculty App
1.0
Dark Mode Issue the App (Student) - whenever dark mode is activated the students cannot see the Questions. This is a serious issue in the Front End while Writing exams, This should be a Priority and must be resolved immediately Note: The Thing Should be Checked with Faculty App
non_process
dark mode issue the app student whenever dark mode is activated the students cannot see the questions this is a serious issue in the front end while writing exams this should be a priority and must be resolved immediately note the thing should be checked with faculty app
0
65,464
8,815,302,082
IssuesEvent
2018-12-29 16:50:59
KoolTheba/arcade-game-clone
https://api.github.com/repos/KoolTheba/arcade-game-clone
opened
Basic functionality
:book: documentation
In this game you have a Player and Enemies (bugs). The goal of the player is to reach the water, without colliding into any one of the enemies. - [ ] The player can move left, right, up and down - [ ] The enemies move at varying speeds on the paved block portion of the game board - [ ] Once a the player collides with an enemy, the game is reset and the player moves back to the starting square - [ ] Once the player reaches the water (i.e., the top of the game board), the game is won
1.0
Basic functionality - In this game you have a Player and Enemies (bugs). The goal of the player is to reach the water, without colliding into any one of the enemies. - [ ] The player can move left, right, up and down - [ ] The enemies move at varying speeds on the paved block portion of the game board - [ ] Once a the player collides with an enemy, the game is reset and the player moves back to the starting square - [ ] Once the player reaches the water (i.e., the top of the game board), the game is won
non_process
basic functionality in this game you have a player and enemies bugs the goal of the player is to reach the water without colliding into any one of the enemies the player can move left right up and down the enemies move at varying speeds on the paved block portion of the game board once a the player collides with an enemy the game is reset and the player moves back to the starting square once the player reaches the water i e the top of the game board the game is won
0
15,835
20,022,557,495
IssuesEvent
2022-02-01 17:43:55
EKGF/ekg-mm
https://api.github.com/repos/EKGF/ekg-mm
closed
Github Kanban is a linear process?
ekg-mm-process
@jgeluk You are interpreting Kanban your own unique way and with this you are reinventing the OODA loop. Look it up. I have been using and teaching this stuff for 40 years all around the world and I know how it works. GitHub is just another implementation of Kanban. And now you want still another implementation. Kanban is a linear process. OODA is a linear process. Both have decades of research and experienced practitioners. Why would we not want to ride on their coat tails? Sincerely, _Originally posted by @DennisWisnosky in https://github.com/EKGF/ekg-mm/issues/16#issuecomment-754864156_
1.0
Github Kanban is a linear process? - @jgeluk You are interpreting Kanban your own unique way and with this you are reinventing the OODA loop. Look it up. I have been using and teaching this stuff for 40 years all around the world and I know how it works. GitHub is just another implementation of Kanban. And now you want still another implementation. Kanban is a linear process. OODA is a linear process. Both have decades of research and experienced practitioners. Why would we not want to ride on their coat tails? Sincerely, _Originally posted by @DennisWisnosky in https://github.com/EKGF/ekg-mm/issues/16#issuecomment-754864156_
process
github kanban is a linear process jgeluk you are interpreting kanban your own unique way and with this you are reinventing the ooda loop look it up i have been using and teaching this stuff for years all around the world and i know how it works github is just another implementation of kanban and now you want still another implementation kanban is a linear process ooda is a linear process both have decades of research and experienced practitioners why would we not want to ride on their coat tails sincerely originally posted by denniswisnosky in
1
11,837
14,656,376,218
IssuesEvent
2020-12-28 13:17:17
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
window.location constructor properties shadowed incorrectly
AREA: client FREQUENCY: level 1 SYSTEM: client side processing TYPE: bug
Sample code: ``` console.log('-1->', window.location.constructor.name); console.log('-2->', window.location.constructor.toString()); console.log('-3->', Function.prototype.toString.apply(window.location.constructor)); console.log('-4->', window.location instanceof Location); ``` When run in a real browser: -1-> Location -2-> function Location() { [native code] } -3-> function Location() { [native code] } -4-> true When run under Testcafe: -1-> g -2-> function(e,t,r){var n=this ... -3-> function(e,t,r){var n=this ... -4-> false Issue is that many websites (e.g. Google login) use a check like this as part of botguard and change their behaviour as a result. Can testcafe-hammerhead be improved to "stealth" these changes?
1.0
window.location constructor properties shadowed incorrectly - Sample code: ``` console.log('-1->', window.location.constructor.name); console.log('-2->', window.location.constructor.toString()); console.log('-3->', Function.prototype.toString.apply(window.location.constructor)); console.log('-4->', window.location instanceof Location); ``` When run in a real browser: -1-> Location -2-> function Location() { [native code] } -3-> function Location() { [native code] } -4-> true When run under Testcafe: -1-> g -2-> function(e,t,r){var n=this ... -3-> function(e,t,r){var n=this ... -4-> false Issue is that many websites (e.g. Google login) use a check like this as part of botguard and change their behaviour as a result. Can testcafe-hammerhead be improved to "stealth" these changes?
process
window location constructor properties shadowed incorrectly sample code console log window location constructor name console log window location constructor tostring console log function prototype tostring apply window location constructor console log window location instanceof location when run in a real browser location function location function location true when run under testcafe g function e t r var n this function e t r var n this false issue is that many websites e g google login use a check like this as part of botguard and change their behaviour as a result can testcafe hammerhead be improved to stealth these changes
1
41,899
10,821,362,696
IssuesEvent
2019-11-08 18:29:40
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
Tensorflow
subtype:centos type:build/install
<em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em> **System information** - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Scientific Linux 7.6, but also tested on Centos 7.6 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: NA - TensorFlow installed from (source or binary): source - TensorFlow version: 1.12.0 - Python version: 2.6 - Installed using virtualenv? pip? conda?: manually compiling using bazel - Bazel version (if compiling from source): 1.16.1, but also tried 1.15.0 and 1.18.0 - GCC/Compiler version (if compiling from source): 4.8.5 - CUDA/cuDNN version: Cuda 9 CuDNN 7 - GPU model and memory: Compute Cluster: multpile Tesla K80 (24GB) or Tesla P100 (16GB). **Describe the problem** During compilation I'm encountering the following ERROR messages: [205 / 208] Compiling tensorflow/contrib/image/kernels/adjust_hsv_in_yiq_op_gpu.cu.cc; 10s local ERROR: /tmp/Tensorflow/PACKAGES/tensorflow/tensorflow/contrib/image/BUILD:115:1: undeclared inclusion(s) in rule '//tensorflow/contrib/image:python/ops/_distort_image_ops_gpu': this rule is missing dependency declarations for the following files included by 'tensorflow/contrib/image/kernels/adjust_hsv_in_yiq_op_gpu.cu.cc': '/tmp/bazel/userid_output/external/eigen_archive/Eigen/Core' '/tmp/bazel/userid_output/external/eigen_archive/Eigen/src/Core/util/DisableStupidWarnings.h' '/tmp/bazel/userid_output/external/eigen_archive/Eigen/src/Core/util/ReenableStupidWarnings.h However, when I add ---verbose_failures to bazel and execute the failing compilation job manually to debug the issue, the compilation is executed correctly. So I'm assuming only some check of bazel is failing, although the compilation itself could be executed correctly. Restarting the build (after my manually compilation) will also continue, until the next job where Eigen is used again. The build is failing with a similar error. Previous versions, including 1.11.0, are not encountering this issue. I could not find any noticeable changes in the Eigen build files for bazel or related to Eigen. Therefore I'm currently out of ideas..... **Provide the exact sequence of commands / steps that you executed before running into the problem** bazel --output_base=/tmp/bazel/userid_output build --jobs 12 -c opt --copt=-mavx2 --copt=-O --copt=-msse4.2 --copt=-mfma -config=cuda //tensorflow/tools/pip_package:build_pip_package (I'm using bazel output_base because my home directory is on NFS) **Any other info / logs** Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
1.0
Tensorflow - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em> **System information** - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Scientific Linux 7.6, but also tested on Centos 7.6 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: NA - TensorFlow installed from (source or binary): source - TensorFlow version: 1.12.0 - Python version: 2.6 - Installed using virtualenv? pip? conda?: manually compiling using bazel - Bazel version (if compiling from source): 1.16.1, but also tried 1.15.0 and 1.18.0 - GCC/Compiler version (if compiling from source): 4.8.5 - CUDA/cuDNN version: Cuda 9 CuDNN 7 - GPU model and memory: Compute Cluster: multpile Tesla K80 (24GB) or Tesla P100 (16GB). **Describe the problem** During compilation I'm encountering the following ERROR messages: [205 / 208] Compiling tensorflow/contrib/image/kernels/adjust_hsv_in_yiq_op_gpu.cu.cc; 10s local ERROR: /tmp/Tensorflow/PACKAGES/tensorflow/tensorflow/contrib/image/BUILD:115:1: undeclared inclusion(s) in rule '//tensorflow/contrib/image:python/ops/_distort_image_ops_gpu': this rule is missing dependency declarations for the following files included by 'tensorflow/contrib/image/kernels/adjust_hsv_in_yiq_op_gpu.cu.cc': '/tmp/bazel/userid_output/external/eigen_archive/Eigen/Core' '/tmp/bazel/userid_output/external/eigen_archive/Eigen/src/Core/util/DisableStupidWarnings.h' '/tmp/bazel/userid_output/external/eigen_archive/Eigen/src/Core/util/ReenableStupidWarnings.h However, when I add ---verbose_failures to bazel and execute the failing compilation job manually to debug the issue, the compilation is executed correctly. So I'm assuming only some check of bazel is failing, although the compilation itself could be executed correctly. Restarting the build (after my manually compilation) will also continue, until the next job where Eigen is used again. The build is failing with a similar error. Previous versions, including 1.11.0, are not encountering this issue. I could not find any noticeable changes in the Eigen build files for bazel or related to Eigen. Therefore I'm currently out of ideas..... **Provide the exact sequence of commands / steps that you executed before running into the problem** bazel --output_base=/tmp/bazel/userid_output build --jobs 12 -c opt --copt=-mavx2 --copt=-O --copt=-msse4.2 --copt=-mfma -config=cuda //tensorflow/tools/pip_package:build_pip_package (I'm using bazel output_base because my home directory is on NFS) **Any other info / logs** Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
non_process
tensorflow please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution e g linux ubuntu scientific linux but also tested on centos mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device na tensorflow installed from source or binary source tensorflow version python version installed using virtualenv pip conda manually compiling using bazel bazel version if compiling from source but also tried and gcc compiler version if compiling from source cuda cudnn version cuda cudnn gpu model and memory compute cluster multpile tesla or tesla describe the problem during compilation i m encountering the following error messages compiling tensorflow contrib image kernels adjust hsv in yiq op gpu cu cc local error tmp tensorflow packages tensorflow tensorflow contrib image build undeclared inclusion s in rule tensorflow contrib image python ops distort image ops gpu this rule is missing dependency declarations for the following files included by tensorflow contrib image kernels adjust hsv in yiq op gpu cu cc tmp bazel userid output external eigen archive eigen core tmp bazel userid output external eigen archive eigen src core util disablestupidwarnings h tmp bazel userid output external eigen archive eigen src core util reenablestupidwarnings h however when i add verbose failures to bazel and execute the failing compilation job manually to debug the issue the compilation is executed correctly so i m assuming only some check of bazel is failing although the compilation itself could be executed correctly restarting the build after my manually compilation will also continue until the next job where eigen is used again the build is failing with a similar error previous versions including are not encountering this issue i could not find any noticeable changes in the eigen build files for bazel or related to eigen therefore i m currently out of ideas provide the exact sequence of commands steps that you executed before running into the problem bazel output base tmp bazel userid output build jobs c opt copt copt o copt copt mfma config cuda tensorflow tools pip package build pip package i m using bazel output base because my home directory is on nfs any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached
0
685,152
23,445,481,238
IssuesEvent
2022-08-15 19:09:28
UniversityDAO/udao
https://api.github.com/repos/UniversityDAO/udao
closed
Refresh Issue
priority
Right now, as of the reading dataflow rework, we only load all the data onto the page ONCE and that's as soon as a user presses "Launch App" on the landing page. Once they do so, it directs them to a "Loading" page where we then fetch all the data. Once the data is fetched and done loading, we automatically redirect to the dashboard page. The problem with this is what if we're on the dashboard/grants/proposals pages, and the user refreshes the page, the page ends up being empty. It's empty because once the page refreshes the entire app has to re-render, and once it re-renders the user will be left on whatever page they refreshed on, meaning they essentially 'bypassed' the launch app button. This is a bit of an issue. One solution would be to have an event listener for page refresh, and upon refresh we redirect them to the loading page. where data will be loaded up. The issue with this solution is that if somebody is on the grants page and refreshes, once the data loads back in they're going to be on the dashboard page. Not an app-breaking problem but still annoying from a user standpoint.
1.0
Refresh Issue - Right now, as of the reading dataflow rework, we only load all the data onto the page ONCE and that's as soon as a user presses "Launch App" on the landing page. Once they do so, it directs them to a "Loading" page where we then fetch all the data. Once the data is fetched and done loading, we automatically redirect to the dashboard page. The problem with this is what if we're on the dashboard/grants/proposals pages, and the user refreshes the page, the page ends up being empty. It's empty because once the page refreshes the entire app has to re-render, and once it re-renders the user will be left on whatever page they refreshed on, meaning they essentially 'bypassed' the launch app button. This is a bit of an issue. One solution would be to have an event listener for page refresh, and upon refresh we redirect them to the loading page. where data will be loaded up. The issue with this solution is that if somebody is on the grants page and refreshes, once the data loads back in they're going to be on the dashboard page. Not an app-breaking problem but still annoying from a user standpoint.
non_process
refresh issue right now as of the reading dataflow rework we only load all the data onto the page once and that s as soon as a user presses launch app on the landing page once they do so it directs them to a loading page where we then fetch all the data once the data is fetched and done loading we automatically redirect to the dashboard page the problem with this is what if we re on the dashboard grants proposals pages and the user refreshes the page the page ends up being empty it s empty because once the page refreshes the entire app has to re render and once it re renders the user will be left on whatever page they refreshed on meaning they essentially bypassed the launch app button this is a bit of an issue one solution would be to have an event listener for page refresh and upon refresh we redirect them to the loading page where data will be loaded up the issue with this solution is that if somebody is on the grants page and refreshes once the data loads back in they re going to be on the dashboard page not an app breaking problem but still annoying from a user standpoint
0
9,970
13,016,342,927
IssuesEvent
2020-07-26 05:57:39
DrFrankenstein/avpipe
https://api.github.com/repos/DrFrankenstein/avpipe
opened
Source: In/Out Markers
Source processing
As a user, I want to set "in" and "out" markers on a source so that I can cut out the parts I don't want. The "out" marker should only be available for sources that have a set duration. FFmpeg options: `-t`, `-to`, `-ss`, `-sseof`.
1.0
Source: In/Out Markers - As a user, I want to set "in" and "out" markers on a source so that I can cut out the parts I don't want. The "out" marker should only be available for sources that have a set duration. FFmpeg options: `-t`, `-to`, `-ss`, `-sseof`.
process
source in out markers as a user i want to set in and out markers on a source so that i can cut out the parts i don t want the out marker should only be available for sources that have a set duration ffmpeg options t to ss sseof
1
21,156
28,132,167,129
IssuesEvent
2023-04-01 01:31:44
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] You should be able to add `:expression`s as order-bys
Type:Bug .Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
`lib/order-by` currently errors if you try to use it with an expression reference.
1.0
[MLv2] You should be able to add `:expression`s as order-bys - `lib/order-by` currently errors if you try to use it with an expression reference.
process
you should be able to add expression s as order bys lib order by currently errors if you try to use it with an expression reference
1
13,879
16,654,715,446
IssuesEvent
2021-06-05 10:01:41
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Responsive issue in Change password screen > UI issues
Bug P2 Participant manager Process: Fixed Process: Tested dev
Responsive issue in the Change password screen > UI issues ![mbc](https://user-images.githubusercontent.com/71445210/115671667-5f8e0000-a368-11eb-8573-eeda2bfc11b4.png)
2.0
[PM] Responsive issue in Change password screen > UI issues - Responsive issue in the Change password screen > UI issues ![mbc](https://user-images.githubusercontent.com/71445210/115671667-5f8e0000-a368-11eb-8573-eeda2bfc11b4.png)
process
responsive issue in change password screen ui issues responsive issue in the change password screen ui issues
1
356
2,794,528,579
IssuesEvent
2015-05-11 17:10:29
linguisticteam/resource-central
https://api.github.com/repos/linguisticteam/resource-central
closed
US1: Initial sketch of UI for adding resources
Work in Process
Initial very rough sketch of the UI for adding new resources. ![ui_sketch](https://cloud.githubusercontent.com/assets/3427319/6094120/50a8cac2-af25-11e4-82b9-a288dc5fdf9f.jpg)
1.0
US1: Initial sketch of UI for adding resources - Initial very rough sketch of the UI for adding new resources. ![ui_sketch](https://cloud.githubusercontent.com/assets/3427319/6094120/50a8cac2-af25-11e4-82b9-a288dc5fdf9f.jpg)
process
initial sketch of ui for adding resources initial very rough sketch of the ui for adding new resources
1
51,120
13,618,142,278
IssuesEvent
2020-09-23 18:05:15
znanstvenikumeni/competitionmanager
https://api.github.com/repos/znanstvenikumeni/competitionmanager
closed
User id isn't being logged for AccountSecurity/LoginFail/WrongPassword errors
affects competitors affects jury members affects mentors affects organisers bug security
As stated in the title, user ID isn't being logged for AccountSecurity/LoginFail/WrongPassword errors, which should be fixed as it disables account-level login attempt blocking and makes debugging and security harder. ![image](https://user-images.githubusercontent.com/1742806/68335535-ec91ba80-00dc-11ea-8afe-308ce08f4657.png)
True
User id isn't being logged for AccountSecurity/LoginFail/WrongPassword errors - As stated in the title, user ID isn't being logged for AccountSecurity/LoginFail/WrongPassword errors, which should be fixed as it disables account-level login attempt blocking and makes debugging and security harder. ![image](https://user-images.githubusercontent.com/1742806/68335535-ec91ba80-00dc-11ea-8afe-308ce08f4657.png)
non_process
user id isn t being logged for accountsecurity loginfail wrongpassword errors as stated in the title user id isn t being logged for accountsecurity loginfail wrongpassword errors which should be fixed as it disables account level login attempt blocking and makes debugging and security harder
0
15,405
5,956,115,111
IssuesEvent
2017-05-28 13:52:00
Framstag/libosmscout
https://api.github.com/repos/Framstag/libosmscout
closed
Compiler warnings in GenCoordDat.cpp
build enhancement help wanted
See https://ci.appveyor.com/project/Framstag/libosmscout/build/job/644tpk6kyk8u0q3h ``` C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::FindDuplicateCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const': C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:108:16: warning: unused variable 'oldUpperLimit' [-Wunused-variable] Id oldUpperLimit=currentUpperLimit; ^~~~~~~~~~~~~ C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::StoreCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const': C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:283:19: warning: unused variable 'oldUpperLimit' [-Wunused-variable] OSMId oldUpperLimit=currentUpperLimit; ^~~~~~~~~~~~~ ``` <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/44932709-compiler-warnings-in-gencoorddat-cpp?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
Compiler warnings in GenCoordDat.cpp - See https://ci.appveyor.com/project/Framstag/libosmscout/build/job/644tpk6kyk8u0q3h ``` C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::FindDuplicateCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const': C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:108:16: warning: unused variable 'oldUpperLimit' [-Wunused-variable] Id oldUpperLimit=currentUpperLimit; ^~~~~~~~~~~~~ C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::StoreCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const': C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:283:19: warning: unused variable 'oldUpperLimit' [-Wunused-variable] OSMId oldUpperLimit=currentUpperLimit; ^~~~~~~~~~~~~ ``` <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/44932709-compiler-warnings-in-gencoorddat-cpp?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_process
compiler warnings in gencoorddat cpp see c projects libosmscout libosmscout import src osmscout import gencoorddat cpp in member function bool osmscout coorddatagenerator findduplicatecoordinates const osmscout typeconfig const osmscout importparameter osmscout progress std unordered map const c projects libosmscout libosmscout import src osmscout import gencoorddat cpp warning unused variable oldupperlimit id oldupperlimit currentupperlimit c projects libosmscout libosmscout import src osmscout import gencoorddat cpp in member function bool osmscout coorddatagenerator storecoordinates const osmscout typeconfig const osmscout importparameter osmscout progress std unordered map const c projects libosmscout libosmscout import src osmscout import gencoorddat cpp warning unused variable oldupperlimit osmid oldupperlimit currentupperlimit want to back this issue we accept bounties via
0
21,972
30,466,569,085
IssuesEvent
2023-07-17 10:47:35
mikerae/sarah-rae-illustrations
https://api.github.com/repos/mikerae/sarah-rae-illustrations
opened
USER STORY:Stripe: Migrate to using fully integrating Stripe
Could Have Epic Admin and Store Management Site Owner Order Processing & Delivery
As **site owner** I can **fully integrate with Stripe** so that **use Stripe's powerful invoicing, reporting and analytics**. ### Acceptance Criteria - All CRUD functionality on products is integrated with Stripe - All CRUD functionality on Users is integrated with Stripe - Stripe Analytics are accessable by Site owner ### Tasks - [ ] Migrate from cardElement to paymentElement - [ ] Integrate Product Id: - [ ] Integrate Price ID: - [ ] Integrate Users with Customers: - [ ] Integrate Users home and billing address on payments - [ ] set up Analytics - [ ] display analytics
1.0
USER STORY:Stripe: Migrate to using fully integrating Stripe - As **site owner** I can **fully integrate with Stripe** so that **use Stripe's powerful invoicing, reporting and analytics**. ### Acceptance Criteria - All CRUD functionality on products is integrated with Stripe - All CRUD functionality on Users is integrated with Stripe - Stripe Analytics are accessable by Site owner ### Tasks - [ ] Migrate from cardElement to paymentElement - [ ] Integrate Product Id: - [ ] Integrate Price ID: - [ ] Integrate Users with Customers: - [ ] Integrate Users home and billing address on payments - [ ] set up Analytics - [ ] display analytics
process
user story stripe migrate to using fully integrating stripe as site owner i can fully integrate with stripe so that use stripe s powerful invoicing reporting and analytics acceptance criteria all crud functionality on products is integrated with stripe all crud functionality on users is integrated with stripe stripe analytics are accessable by site owner tasks migrate from cardelement to paymentelement integrate product id integrate price id integrate users with customers integrate users home and billing address on payments set up analytics display analytics
1
2,907
5,889,722,895
IssuesEvent
2017-05-17 13:33:28
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
Changing Color on selected filter
enhancement inprocess
I want to change the color of selected filter to a different color. For example see below. I selected filter for Run to Trial-Run-I . I would like to change color in text insider selected filter to something more visible color. ![image](https://cloud.githubusercontent.com/assets/12001907/26015270/ffe1e0f2-3725-11e7-8b5b-b4cf50b4304a.png) Is there any way to pass color for selected?
1.0
Changing Color on selected filter - I want to change the color of selected filter to a different color. For example see below. I selected filter for Run to Trial-Run-I . I would like to change color in text insider selected filter to something more visible color. ![image](https://cloud.githubusercontent.com/assets/12001907/26015270/ffe1e0f2-3725-11e7-8b5b-b4cf50b4304a.png) Is there any way to pass color for selected?
process
changing color on selected filter i want to change the color of selected filter to a different color for example see below i selected filter for run to trial run i i would like to change color in text insider selected filter to something more visible color is there any way to pass color for selected
1
2,918
5,914,478,679
IssuesEvent
2017-05-22 03:04:28
nodejs/node
https://api.github.com/repos/nodejs/node
closed
investigate flaky sequential/test-benchmark-child-process on Windows
benchmark child_process test windows
* **Version**: v8.0.0-pre * **Platform**: win2008r2 * **Subsystem**: test <!-- Enter your issue details below this comment. --> `sequential/test-benchmark-child-process ` is still failing sometimes flaky on Windows in CI. I'll open a PR to mark it as flaky. This issue is for trying to locate the problem and a solution. When the test succeeds, it seems to take just a few seconds. https://ci.nodejs.org/job/node-test-binary-windows/RUN_SUBSET=3,VS_VERSION=vs2015,label=win2008r2/7865/console ```console ok 356 sequential/test-benchmark-child-process --- duration_ms: 1.764 ``` When it fails, it's a timeout. https://ci.nodejs.org/job/node-test-binary-windows/7867/RUN_SUBSET=3,VS_VERSION=vs2015,label=win2008r2/console ```console not ok 356 sequential/test-benchmark-child-process --- duration_ms: 60.70 severity: fail stack: |- timeout ``` This would suggest a race condition or something else causing a child process to hang or something. And that might be the cause. But... Interestingly, a stress test where the five benchmarks that this test calls were all split out into individual tests, succeeded but each test took around 30 seconds to run. Wha??!! I know! (Only other change in those tests is the `dur` option for the benchmarks was increased from 0 to 0.1. Well, that, and that this test was run on win2016 so maybe the results are completely irrelevant? I don't know.) https://ci.nodejs.org/job/node-stress-single-test/1161/nodes=win2016/console: ```console ok 1 sequential/test-benchmark-child-process-exec-stdout --- duration_ms: 3.158 ... ok 2 sequential/test-benchmark-child-process-params --- duration_ms: 36.735 ... ok 3 sequential/test-benchmark-child-process-read-ipc --- duration_ms: 30.116 ... ok 4 sequential/test-benchmark-child-process-read --- duration_ms: 31.844 ... ok 5 sequential/test-benchmark-child-process-spawn-echo --- duration_ms: 31.661 ``` So I'm not sure what's going on here. Maybe it can be worked out by someone more comfortable testing and debugging on Windows or someone more deeply familiar with child_process and/or our benchmarking code. @nodejs/platform-windows @nodejs/benchmarking @mscdex @cjihrig @bnoordhuis @nodejs/testing
1.0
investigate flaky sequential/test-benchmark-child-process on Windows - * **Version**: v8.0.0-pre * **Platform**: win2008r2 * **Subsystem**: test <!-- Enter your issue details below this comment. --> `sequential/test-benchmark-child-process ` is still failing sometimes flaky on Windows in CI. I'll open a PR to mark it as flaky. This issue is for trying to locate the problem and a solution. When the test succeeds, it seems to take just a few seconds. https://ci.nodejs.org/job/node-test-binary-windows/RUN_SUBSET=3,VS_VERSION=vs2015,label=win2008r2/7865/console ```console ok 356 sequential/test-benchmark-child-process --- duration_ms: 1.764 ``` When it fails, it's a timeout. https://ci.nodejs.org/job/node-test-binary-windows/7867/RUN_SUBSET=3,VS_VERSION=vs2015,label=win2008r2/console ```console not ok 356 sequential/test-benchmark-child-process --- duration_ms: 60.70 severity: fail stack: |- timeout ``` This would suggest a race condition or something else causing a child process to hang or something. And that might be the cause. But... Interestingly, a stress test where the five benchmarks that this test calls were all split out into individual tests, succeeded but each test took around 30 seconds to run. Wha??!! I know! (Only other change in those tests is the `dur` option for the benchmarks was increased from 0 to 0.1. Well, that, and that this test was run on win2016 so maybe the results are completely irrelevant? I don't know.) https://ci.nodejs.org/job/node-stress-single-test/1161/nodes=win2016/console: ```console ok 1 sequential/test-benchmark-child-process-exec-stdout --- duration_ms: 3.158 ... ok 2 sequential/test-benchmark-child-process-params --- duration_ms: 36.735 ... ok 3 sequential/test-benchmark-child-process-read-ipc --- duration_ms: 30.116 ... ok 4 sequential/test-benchmark-child-process-read --- duration_ms: 31.844 ... ok 5 sequential/test-benchmark-child-process-spawn-echo --- duration_ms: 31.661 ``` So I'm not sure what's going on here. Maybe it can be worked out by someone more comfortable testing and debugging on Windows or someone more deeply familiar with child_process and/or our benchmarking code. @nodejs/platform-windows @nodejs/benchmarking @mscdex @cjihrig @bnoordhuis @nodejs/testing
process
investigate flaky sequential test benchmark child process on windows version pre platform subsystem test sequential test benchmark child process is still failing sometimes flaky on windows in ci i ll open a pr to mark it as flaky this issue is for trying to locate the problem and a solution when the test succeeds it seems to take just a few seconds console ok sequential test benchmark child process duration ms when it fails it s a timeout console not ok sequential test benchmark child process duration ms severity fail stack timeout this would suggest a race condition or something else causing a child process to hang or something and that might be the cause but interestingly a stress test where the five benchmarks that this test calls were all split out into individual tests succeeded but each test took around seconds to run wha i know only other change in those tests is the dur option for the benchmarks was increased from to well that and that this test was run on so maybe the results are completely irrelevant i don t know console ok sequential test benchmark child process exec stdout duration ms ok sequential test benchmark child process params duration ms ok sequential test benchmark child process read ipc duration ms ok sequential test benchmark child process read duration ms ok sequential test benchmark child process spawn echo duration ms so i m not sure what s going on here maybe it can be worked out by someone more comfortable testing and debugging on windows or someone more deeply familiar with child process and or our benchmarking code nodejs platform windows nodejs benchmarking mscdex cjihrig bnoordhuis nodejs testing
1
66,887
8,973,689,742
IssuesEvent
2019-01-29 21:44:24
PegaSysEng/pantheon
https://api.github.com/repos/PegaSysEng/pantheon
closed
Update installation to include libsodium dependency
doc next release documentation
### Acceptance Criteria Document Pantheon dependency on Orion which depends on libsodium in installation docs.
1.0
Update installation to include libsodium dependency - ### Acceptance Criteria Document Pantheon dependency on Orion which depends on libsodium in installation docs.
non_process
update installation to include libsodium dependency acceptance criteria document pantheon dependency on orion which depends on libsodium in installation docs
0
4,154
7,103,700,373
IssuesEvent
2018-01-16 06:48:15
log2timeline/plaso
https://api.github.com/repos/log2timeline/plaso
closed
add preprocess plugin to detect Linux operating system information
enhancement preprocessing
- [x] add preprocess pluging to detect Linux operating system information (also see #928 and #929) - https://github.com/ForensicArtifacts/artifacts/pull/230 - https://codereview.appspot.com/322490043 - `/etc/system-release` ``` Fedora release 26 (Twenty Six) ``` - [x] redefine processor / knowledge base to continue and use most appropriate match - https://codereview.appspot.com/339180043 - [x] alternative `/etc/os-release` - https://codereview.appspot.com/339180043 ``` NAME=Fedora VERSION="26 (Workstation Edition)" ID=fedora VERSION_ID=26 PRETTY_NAME="Fedora 26 (Workstation Edition)" ANSI_COLOR="0;34" CPE_NAME="cpe:/o:fedoraproject:fedora:26" HOME_URL="https://fedoraproject.org/" BUG_REPORT_URL="https://bugzilla.redhat.com/" REDHAT_BUGZILLA_PRODUCT="Fedora" REDHAT_BUGZILLA_PRODUCT_VERSION=26 REDHAT_SUPPORT_PRODUCT="Fedora" REDHAT_SUPPORT_PRODUCT_VERSION=26 PRIVACY_POLICY_URL=https://fedoraproject.org/wiki/Legal:PrivacyPolicy VARIANT="Workstation Edition" VARIANT_ID=workstation ``` - [x] alternative `/etc/lsb-release` - https://codereview.appspot.com/339180043 - [x] check if operating system information is stored - https://codereview.appspot.com/338350043 ``` ***************************** System configuration ***************************** Hostname : TEST Operating system : N/A Operating system product : N/A Operating system version : N/A Code page : cp1252 Keyboard layout : N/A Time zone : UTC -------------------------------------------------------------------------------- ``` * [x] fall back to `/etc/issue` or `/etc/issue.net` for older releases, for example * https://github.com/ForensicArtifacts/artifacts/pull/244 * https://codereview.appspot.com/337430043/ ``` Debian GNU/Linux 5.0 \n \l ``` This could be unreliable e.g. Fedora: ``` \S Kernel \r on an \m (\l) ```
1.0
add preprocess plugin to detect Linux operating system information - - [x] add preprocess pluging to detect Linux operating system information (also see #928 and #929) - https://github.com/ForensicArtifacts/artifacts/pull/230 - https://codereview.appspot.com/322490043 - `/etc/system-release` ``` Fedora release 26 (Twenty Six) ``` - [x] redefine processor / knowledge base to continue and use most appropriate match - https://codereview.appspot.com/339180043 - [x] alternative `/etc/os-release` - https://codereview.appspot.com/339180043 ``` NAME=Fedora VERSION="26 (Workstation Edition)" ID=fedora VERSION_ID=26 PRETTY_NAME="Fedora 26 (Workstation Edition)" ANSI_COLOR="0;34" CPE_NAME="cpe:/o:fedoraproject:fedora:26" HOME_URL="https://fedoraproject.org/" BUG_REPORT_URL="https://bugzilla.redhat.com/" REDHAT_BUGZILLA_PRODUCT="Fedora" REDHAT_BUGZILLA_PRODUCT_VERSION=26 REDHAT_SUPPORT_PRODUCT="Fedora" REDHAT_SUPPORT_PRODUCT_VERSION=26 PRIVACY_POLICY_URL=https://fedoraproject.org/wiki/Legal:PrivacyPolicy VARIANT="Workstation Edition" VARIANT_ID=workstation ``` - [x] alternative `/etc/lsb-release` - https://codereview.appspot.com/339180043 - [x] check if operating system information is stored - https://codereview.appspot.com/338350043 ``` ***************************** System configuration ***************************** Hostname : TEST Operating system : N/A Operating system product : N/A Operating system version : N/A Code page : cp1252 Keyboard layout : N/A Time zone : UTC -------------------------------------------------------------------------------- ``` * [x] fall back to `/etc/issue` or `/etc/issue.net` for older releases, for example * https://github.com/ForensicArtifacts/artifacts/pull/244 * https://codereview.appspot.com/337430043/ ``` Debian GNU/Linux 5.0 \n \l ``` This could be unreliable e.g. Fedora: ``` \S Kernel \r on an \m (\l) ```
process
add preprocess plugin to detect linux operating system information add preprocess pluging to detect linux operating system information also see and etc system release fedora release twenty six redefine processor knowledge base to continue and use most appropriate match alternative etc os release name fedora version workstation edition id fedora version id pretty name fedora workstation edition ansi color cpe name cpe o fedoraproject fedora home url bug report url redhat bugzilla product fedora redhat bugzilla product version redhat support product fedora redhat support product version privacy policy url variant workstation edition variant id workstation alternative etc lsb release check if operating system information is stored system configuration hostname test operating system n a operating system product n a operating system version n a code page keyboard layout n a time zone utc fall back to etc issue or etc issue net for older releases for example debian gnu linux n l this could be unreliable e g fedora s kernel r on an m l
1
16,774
3,560,795,584
IssuesEvent
2016-01-23 09:52:18
kefirfromperm/kefirbb
https://api.github.com/repos/kefirfromperm/kefirbb
opened
Do tests refactoring
test
It's needed to do refactoring of tests. There are to many double codes and it is not clear which text was parsed on failure.
1.0
Do tests refactoring - It's needed to do refactoring of tests. There are to many double codes and it is not clear which text was parsed on failure.
non_process
do tests refactoring it s needed to do refactoring of tests there are to many double codes and it is not clear which text was parsed on failure
0
12,111
14,740,473,049
IssuesEvent
2021-01-07 09:08:43
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
SAB Error - CC Processing
anc-process anp-0.5 ant-bug has attachment
In GitLab by @kdjstudios on Nov 9, 2018, 11:56 **Submitted by:** "Arianna Screen" <arianna.screen@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-11-09-55722/conversation **Server:** Internal **Client/Site:** Santa Rosa **Account:** NA **Issue:** We are experiencing an error with processing credit card payments in SA Billing. Please see attached. ![SAB+CC+Payment+error](/uploads/0b01cb3914c0ee42d6fe0d31f0c94d0c/SAB+CC+Payment+error.png)
1.0
SAB Error - CC Processing - In GitLab by @kdjstudios on Nov 9, 2018, 11:56 **Submitted by:** "Arianna Screen" <arianna.screen@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-11-09-55722/conversation **Server:** Internal **Client/Site:** Santa Rosa **Account:** NA **Issue:** We are experiencing an error with processing credit card payments in SA Billing. Please see attached. ![SAB+CC+Payment+error](/uploads/0b01cb3914c0ee42d6fe0d31f0c94d0c/SAB+CC+Payment+error.png)
process
sab error cc processing in gitlab by kdjstudios on nov submitted by arianna screen helpdesk server internal client site santa rosa account na issue we are experiencing an error with processing credit card payments in sa billing please see attached uploads sab cc payment error png
1
188,486
14,445,999,233
IssuesEvent
2020-12-08 00:13:54
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
paul-lee-attorney/fabric-ca-1.4.7-gm: cmd/fabric-ca-client/command/main_test.go; 3 LoC
fresh test tiny
Found a possible issue in [paul-lee-attorney/fabric-ca-1.4.7-gm](https://www.github.com/paul-lee-attorney/fabric-ca-1.4.7-gm) at [cmd/fabric-ca-client/command/main_test.go](https://github.com/paul-lee-attorney/fabric-ca-1.4.7-gm/blob/c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578/cmd/fabric-ca-client/command/main_test.go#L152-L154) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to e at line 153 may start a goroutine [Click here to see the code in its original context.](https://github.com/paul-lee-attorney/fabric-ca-1.4.7-gm/blob/c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578/cmd/fabric-ca-client/command/main_test.go#L152-L154) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, e := range errCases { extraArgErrorTest(&e, t) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578
1.0
paul-lee-attorney/fabric-ca-1.4.7-gm: cmd/fabric-ca-client/command/main_test.go; 3 LoC - Found a possible issue in [paul-lee-attorney/fabric-ca-1.4.7-gm](https://www.github.com/paul-lee-attorney/fabric-ca-1.4.7-gm) at [cmd/fabric-ca-client/command/main_test.go](https://github.com/paul-lee-attorney/fabric-ca-1.4.7-gm/blob/c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578/cmd/fabric-ca-client/command/main_test.go#L152-L154) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to e at line 153 may start a goroutine [Click here to see the code in its original context.](https://github.com/paul-lee-attorney/fabric-ca-1.4.7-gm/blob/c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578/cmd/fabric-ca-client/command/main_test.go#L152-L154) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, e := range errCases { extraArgErrorTest(&e, t) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578
non_process
paul lee attorney fabric ca gm cmd fabric ca client command main test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to e at line may start a goroutine click here to show the line s of go which triggered the analyzer go for e range errcases extraargerrortest e t leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
13,887
16,654,864,392
IssuesEvent
2021-06-05 10:36:08
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Responsive issue > UI is broken for toaster messages
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
UI is broken for all the toaster messages [Note: Issue should be fixed for all the toaster messages] ![mbt1](https://user-images.githubusercontent.com/71445210/115513770-14f58080-a2a1-11eb-85a6-b45ef7c92fd2.png)
3.0
[PM] Responsive issue > UI is broken for toaster messages - UI is broken for all the toaster messages [Note: Issue should be fixed for all the toaster messages] ![mbt1](https://user-images.githubusercontent.com/71445210/115513770-14f58080-a2a1-11eb-85a6-b45ef7c92fd2.png)
process
responsive issue ui is broken for toaster messages ui is broken for all the toaster messages
1
344,820
10,349,642,356
IssuesEvent
2019-09-04 23:18:43
oslc-op/jira-migration-landfill
https://api.github.com/repos/oslc-op/jira-migration-landfill
closed
Example 3 in oslc-core-overview is incorrect
Core: Main Spec Jira: formatting Priority: High Xtra: Jira
Example 3 in the OSLC Core Overview is wrong: Example 3: Example URL with oslc.prefix [http://example.com/bugs/4242?oslc.prefix=foaf=&oslc.properties=foaf:lastName](http://example.com/bugs/4242?oslc.prefix=foaf=&oslc.properties=foaf:lastName),... It omits the required Uri\\\_ref\\\_esc for foaf. It should be: Example 3: Example URL with oslc.prefix [http://example.com/bugs/4242?oslc.prefix=foaf=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2F&oslc.properties=foaf:lastName](http://example.com/bugs/4242?oslc.prefix=foaf=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2F&oslc.properties=foaf:lastName),... --- _Migrated from https://issues.oasis-open.org/browse/OSLCCORE-152 (opened by @oslc-bot; previously assigned to @jamsden)_
1.0
Example 3 in oslc-core-overview is incorrect - Example 3 in the OSLC Core Overview is wrong: Example 3: Example URL with oslc.prefix [http://example.com/bugs/4242?oslc.prefix=foaf=&oslc.properties=foaf:lastName](http://example.com/bugs/4242?oslc.prefix=foaf=&oslc.properties=foaf:lastName),... It omits the required Uri\\\_ref\\\_esc for foaf. It should be: Example 3: Example URL with oslc.prefix [http://example.com/bugs/4242?oslc.prefix=foaf=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2F&oslc.properties=foaf:lastName](http://example.com/bugs/4242?oslc.prefix=foaf=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2F&oslc.properties=foaf:lastName),... --- _Migrated from https://issues.oasis-open.org/browse/OSLCCORE-152 (opened by @oslc-bot; previously assigned to @jamsden)_
non_process
example in oslc core overview is incorrect example in the oslc core overview is wrong example example url with oslc prefix it omits the required uri ref esc for foaf it should be example example url with oslc prefix migrated from opened by oslc bot previously assigned to jamsden
0
9,957
12,990,502,135
IssuesEvent
2020-07-23 00:13:56
parcel-bundler/parcel
https://api.github.com/repos/parcel-bundler/parcel
closed
using posthtml without options kills php tags
:bug: Bug HTML Preprocessing Stale
in src/packagers/HTMLPackager.js, this code: ```js async addAsset(asset) { let html = asset.generated.html || ''; // Find child bundles that have JS or CSS sibling bundles, // add them to the head so they are loaded immediately. let siblingBundles = Array.from(this.bundle.childBundles) .reduce((p, b) => p.concat([...b.siblingBundles.values()]), []) .filter(b => b.type === 'css' || b.type === 'js'); if (siblingBundles.length > 0) { html = posthtml( this.insertSiblingBundles.bind(this, siblingBundles) ).process(html, {sync: true}).html; } await this.write(html); } ``` calls posthtml().process with a options `{sync: true}` and that causes it to strip out unknown directives. do you mind if i load up the .posthtmlrc options and add sync: true to them and use that here? or any other suggestion? My goal is to get php to not be clobbered by parcel's html asset handler
1.0
using posthtml without options kills php tags - in src/packagers/HTMLPackager.js, this code: ```js async addAsset(asset) { let html = asset.generated.html || ''; // Find child bundles that have JS or CSS sibling bundles, // add them to the head so they are loaded immediately. let siblingBundles = Array.from(this.bundle.childBundles) .reduce((p, b) => p.concat([...b.siblingBundles.values()]), []) .filter(b => b.type === 'css' || b.type === 'js'); if (siblingBundles.length > 0) { html = posthtml( this.insertSiblingBundles.bind(this, siblingBundles) ).process(html, {sync: true}).html; } await this.write(html); } ``` calls posthtml().process with a options `{sync: true}` and that causes it to strip out unknown directives. do you mind if i load up the .posthtmlrc options and add sync: true to them and use that here? or any other suggestion? My goal is to get php to not be clobbered by parcel's html asset handler
process
using posthtml without options kills php tags in src packagers htmlpackager js this code js async addasset asset let html asset generated html find child bundles that have js or css sibling bundles add them to the head so they are loaded immediately let siblingbundles array from this bundle childbundles reduce p b p concat filter b b type css b type js if siblingbundles length html posthtml this insertsiblingbundles bind this siblingbundles process html sync true html await this write html calls posthtml process with a options sync true and that causes it to strip out unknown directives do you mind if i load up the posthtmlrc options and add sync true to them and use that here or any other suggestion my goal is to get php to not be clobbered by parcel s html asset handler
1
2,154
5,005,712,689
IssuesEvent
2016-12-12 11:39:14
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
closed
Cannot scroll large forms
browser: firefox browser: safari bug comp: activiti-processList
If a large form is attached to start event/user task can only scroll form using up and down arrows on keyboard, scrollbar within browser is disabled/not accessible. Issue only in Firefox and Safari, this maybe browser limitation
1.0
Cannot scroll large forms - If a large form is attached to start event/user task can only scroll form using up and down arrows on keyboard, scrollbar within browser is disabled/not accessible. Issue only in Firefox and Safari, this maybe browser limitation
process
cannot scroll large forms if a large form is attached to start event user task can only scroll form using up and down arrows on keyboard scrollbar within browser is disabled not accessible issue only in firefox and safari this maybe browser limitation
1
9,058
4,389,957,380
IssuesEvent
2016-08-09 00:29:26
jeff1evesque/machine-learning
https://api.github.com/repos/jeff1evesque/machine-learning
opened
Remove 'vagrant_implement' instances
build enhancement
We need to remove all instances of `$vagrant_mounted = $hiera_general['vagrant_implement']`, from any puppet manifests, and make necessary adjustments within `setup_tables.py`. The former puppet variable was created when only a singled puppet environment existed, to contain the following (current) puppet environments: - docker - vagrant
1.0
Remove 'vagrant_implement' instances - We need to remove all instances of `$vagrant_mounted = $hiera_general['vagrant_implement']`, from any puppet manifests, and make necessary adjustments within `setup_tables.py`. The former puppet variable was created when only a singled puppet environment existed, to contain the following (current) puppet environments: - docker - vagrant
non_process
remove vagrant implement instances we need to remove all instances of vagrant mounted hiera general from any puppet manifests and make necessary adjustments within setup tables py the former puppet variable was created when only a singled puppet environment existed to contain the following current puppet environments docker vagrant
0
12,293
14,850,813,958
IssuesEvent
2021-01-18 05:32:43
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Sites tab > Contents are not displaying in the sites tab : getting error message
Blocker Bug P0 Participant manager Process: Fixed Process: Tested dev
Sites tab > Contents are not displaying in the sites tab : getting an error message ![sites tab1](https://user-images.githubusercontent.com/71445210/104813803-7be24400-5831-11eb-8e53-a7ea7551f491.png)
2.0
[PM] Sites tab > Contents are not displaying in the sites tab : getting error message - Sites tab > Contents are not displaying in the sites tab : getting an error message ![sites tab1](https://user-images.githubusercontent.com/71445210/104813803-7be24400-5831-11eb-8e53-a7ea7551f491.png)
process
sites tab contents are not displaying in the sites tab getting error message sites tab contents are not displaying in the sites tab getting an error message
1
18,817
24,718,365,315
IssuesEvent
2022-10-20 08:48:17
hermes-hmc/workflow
https://api.github.com/repos/hermes-hmc/workflow
opened
Implement author deduplication based on config
enhancement 2️ processing
- **Requires:** #39 Author deduplication (including email mapping) should be part of the processing step.
1.0
Implement author deduplication based on config - - **Requires:** #39 Author deduplication (including email mapping) should be part of the processing step.
process
implement author deduplication based on config requires author deduplication including email mapping should be part of the processing step
1
15,086
18,795,513,924
IssuesEvent
2021-11-08 21:47:38
googleapis/python-bigquery
https://api.github.com/repos/googleapis/python-bigquery
closed
type check with mypy as well
api: bigquery type: process
https://issues.apache.org/jira/browse/BEAM-12975 -- https://github.com/googleapis/python-bigquery/pull/976 appears to have broken Apache BEAM, which uses mypy for type checking. This would also help with Mac developers, as I've been unable to run pytype locally due to some missing wheels.
1.0
type check with mypy as well - https://issues.apache.org/jira/browse/BEAM-12975 -- https://github.com/googleapis/python-bigquery/pull/976 appears to have broken Apache BEAM, which uses mypy for type checking. This would also help with Mac developers, as I've been unable to run pytype locally due to some missing wheels.
process
type check with mypy as well appears to have broken apache beam which uses mypy for type checking this would also help with mac developers as i ve been unable to run pytype locally due to some missing wheels
1
375,195
26,151,579,617
IssuesEvent
2022-12-30 14:28:31
FilledStacks/stacked
https://api.github.com/repos/FilledStacks/stacked
closed
Update Get Started to explain the generated Code
documentation
At the moment the Get Started section of the documentation lacks an explanation of the output of the cli. Since we have some counter code in the generated code we should go over explaining how the view and viewmodel works by using the generated code as an example.
1.0
Update Get Started to explain the generated Code - At the moment the Get Started section of the documentation lacks an explanation of the output of the cli. Since we have some counter code in the generated code we should go over explaining how the view and viewmodel works by using the generated code as an example.
non_process
update get started to explain the generated code at the moment the get started section of the documentation lacks an explanation of the output of the cli since we have some counter code in the generated code we should go over explaining how the view and viewmodel works by using the generated code as an example
0
4,842
7,736,175,388
IssuesEvent
2018-05-27 23:29:37
vital-software/scala-redox
https://api.github.com/repos/vital-software/scala-redox
opened
Set up CI to perform releases
process: ci stage: backlog type: devops
Similar to what we've done with our other Scala libs, this library should support versioning the project during the CI build of master commits: https://github.com/vital-software/doc-dna/blob/deb774269b7c56c44fe72aab68e7532dcdcb8fc9/.buildkite/pipeline.yml#L27-L34 The CI system should have Sonatype access, and use it to publish artifacts.
1.0
Set up CI to perform releases - Similar to what we've done with our other Scala libs, this library should support versioning the project during the CI build of master commits: https://github.com/vital-software/doc-dna/blob/deb774269b7c56c44fe72aab68e7532dcdcb8fc9/.buildkite/pipeline.yml#L27-L34 The CI system should have Sonatype access, and use it to publish artifacts.
process
set up ci to perform releases similar to what we ve done with our other scala libs this library should support versioning the project during the ci build of master commits the ci system should have sonatype access and use it to publish artifacts
1
87,744
17,368,201,987
IssuesEvent
2021-07-30 10:15:08
Serrin/Celestra
https://api.github.com/repos/Serrin/Celestra
closed
Changes in v4.4.3
closed - done or fixed code code - CUT code - ESM documentation maintenance type - enhancement
1. Documentation and pdf fixes. 2. Add a new function: `findLast(<collection>,<callback>);` 3. Add an alias: `contains(<collection>,<value>);` -> `includes(<collection>,<value>);` 4. Fix the description of these functions: ````javascript arrayRemove(<array>,<value>[,all=false]); arrayCycle(<collection>[,n=100]); arrayRepeat(<value>[,n=100]); iterRange([start=0[,step=1[,end=Infinity]]]); iterCycle(<iter>[,n=Infinity]); iterRepeat(<value>[,n=Infinity]); take(<collection>[,n=1]); takeRight(<collection>[,n=1]); drop(<collection>[,n=1]); dropRight(<collection>[,n=1]); setUnion(<collection1>[,collectionN]); randomString([length[,specChar=false]]); javaHash(<data>[,hexa=false]); ```` 5. Remove the description of these removed function in __celestra.html__ and __readme.md__ and collect these function names in a new line of the __Collections__ table: ````javascript forOf(<collection>,<callback>); mapOf(<collection>,<callback>); sizeOf(<collection>); filterOf(<collection>,<callback>); hasOf(<collection>,<value>); findOf(<collection>,<callback>); everyOf(<collection>,<callback>); someOf(<collection>,<callback>); noneOf(<collection>,<callback>); firstOf(<collection>); lastOf(<collection>); sliceOf(<collection>[,begin[,end]]); reverseOf(<collection>); sortOf(<collection>); reduceOf(<collection>,<callback>[,initialvalue]); concatOf(<collection1>[,collectionN]); flatOf(<collection>); enumerateOf(<collection>); joinOf(<collection>[,separator=","]); takeOf(<collection>[,n]); dropOf(<collection>[,n]); ```` 6. __CUT v0.8.19__ - Replace the __unittest.dev.html__, __unittest.min.html__ and __unittest.esm.html__ files with the __unittest.html__
3.0
Changes in v4.4.3 - 1. Documentation and pdf fixes. 2. Add a new function: `findLast(<collection>,<callback>);` 3. Add an alias: `contains(<collection>,<value>);` -> `includes(<collection>,<value>);` 4. Fix the description of these functions: ````javascript arrayRemove(<array>,<value>[,all=false]); arrayCycle(<collection>[,n=100]); arrayRepeat(<value>[,n=100]); iterRange([start=0[,step=1[,end=Infinity]]]); iterCycle(<iter>[,n=Infinity]); iterRepeat(<value>[,n=Infinity]); take(<collection>[,n=1]); takeRight(<collection>[,n=1]); drop(<collection>[,n=1]); dropRight(<collection>[,n=1]); setUnion(<collection1>[,collectionN]); randomString([length[,specChar=false]]); javaHash(<data>[,hexa=false]); ```` 5. Remove the description of these removed function in __celestra.html__ and __readme.md__ and collect these function names in a new line of the __Collections__ table: ````javascript forOf(<collection>,<callback>); mapOf(<collection>,<callback>); sizeOf(<collection>); filterOf(<collection>,<callback>); hasOf(<collection>,<value>); findOf(<collection>,<callback>); everyOf(<collection>,<callback>); someOf(<collection>,<callback>); noneOf(<collection>,<callback>); firstOf(<collection>); lastOf(<collection>); sliceOf(<collection>[,begin[,end]]); reverseOf(<collection>); sortOf(<collection>); reduceOf(<collection>,<callback>[,initialvalue]); concatOf(<collection1>[,collectionN]); flatOf(<collection>); enumerateOf(<collection>); joinOf(<collection>[,separator=","]); takeOf(<collection>[,n]); dropOf(<collection>[,n]); ```` 6. __CUT v0.8.19__ - Replace the __unittest.dev.html__, __unittest.min.html__ and __unittest.esm.html__ files with the __unittest.html__
non_process
changes in documentation and pdf fixes add a new function findlast add an alias contains includes fix the description of these functions javascript arrayremove arraycycle arrayrepeat iterrange itercycle iterrepeat take takeright drop dropright setunion randomstring javahash remove the description of these removed function in celestra html and readme md and collect these function names in a new line of the collections table javascript forof mapof sizeof filterof hasof findof everyof someof noneof firstof lastof sliceof reverseof sortof reduceof concatof flatof enumerateof joinof takeof dropof cut replace the unittest dev html unittest min html and unittest esm html files with the unittest html
0
15,013
18,724,271,103
IssuesEvent
2021-11-03 14:51:39
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
closed
Autogen: fix coverage landmines.
api: datastore api: pubsub api: bigtable api: logging api: vision api: monitoring api: translation api: speech api: spanner api: clouderrorreporting api: language api: cloudtrace api: videointelligence api: firestore type: process api: dataproc api: container api: bigquerydatatransfer api: dlp api: texttospeech api: redis api: cloudiot api: cloudtasks api: automl api: cloudkms api: cloudasset api: oslogin api: websecurityscanner api: iam api: irm api: talent api: webrisk api: datacatalog api: phishingprotection api: datalabeling api: cloudscheduler api: securitycenter
The `unit_cov_level` numbers in the various `noxfile.py` files are landmines (e.g., PR #8147 adds a `# pragma: NO COVER` and actually trips the check, because the total number of lines checked in BQDT drops!) The numbers don't represent any **real** goals (why 79% in one library or 95% in another?), but are merely "high water marks." The real goal is 100% coverage across all unit test runs, which we **don't** hit for many of the autogen-only libraries. To address the first issue, I would argue that we should just drop the coverage checks inside the `unit` / `default` sessions. To address the second issue, we need to apply a change like the one in #7413 which actually tests the "shim" modules for the autogen-only libraries (and excludes the namespace package files). Both of these changes need to happen in the generator / templates (dropping the `--cov-fail-under` bit in the `unit` / `default` session, tweaking the `.coveragerc`, adding a testcase module for each generated shim module).
1.0
Autogen: fix coverage landmines. - The `unit_cov_level` numbers in the various `noxfile.py` files are landmines (e.g., PR #8147 adds a `# pragma: NO COVER` and actually trips the check, because the total number of lines checked in BQDT drops!) The numbers don't represent any **real** goals (why 79% in one library or 95% in another?), but are merely "high water marks." The real goal is 100% coverage across all unit test runs, which we **don't** hit for many of the autogen-only libraries. To address the first issue, I would argue that we should just drop the coverage checks inside the `unit` / `default` sessions. To address the second issue, we need to apply a change like the one in #7413 which actually tests the "shim" modules for the autogen-only libraries (and excludes the namespace package files). Both of these changes need to happen in the generator / templates (dropping the `--cov-fail-under` bit in the `unit` / `default` session, tweaking the `.coveragerc`, adding a testcase module for each generated shim module).
process
autogen fix coverage landmines the unit cov level numbers in the various noxfile py files are landmines e g pr adds a pragma no cover and actually trips the check because the total number of lines checked in bqdt drops the numbers don t represent any real goals why in one library or in another but are merely high water marks the real goal is coverage across all unit test runs which we don t hit for many of the autogen only libraries to address the first issue i would argue that we should just drop the coverage checks inside the unit default sessions to address the second issue we need to apply a change like the one in which actually tests the shim modules for the autogen only libraries and excludes the namespace package files both of these changes need to happen in the generator templates dropping the cov fail under bit in the unit default session tweaking the coveragerc adding a testcase module for each generated shim module
1
134,673
19,299,530,891
IssuesEvent
2021-12-13 02:26:17
bookey-dev/bookey.bug
https://api.github.com/repos/bookey-dev/bookey.bug
closed
输入框为空状态下按钮高亮显示
P5 platform: ios easy fix v1.9 design
机型:iphone SE 步骤:1.启动Bookey,登陆首页 ![输入框为空状态下按钮高亮显示](https://user-images.githubusercontent.com/80370536/117533834-72b7f580-b021-11eb-81ff-ea0f0cdb49c5.jpg) 2.sign up 结果:输入框为空状态下next按钮高亮显示 期望:建议为空状态下,操作按钮暗淡色,输入后高亮显示,与android保持一致
1.0
输入框为空状态下按钮高亮显示 - 机型:iphone SE 步骤:1.启动Bookey,登陆首页 ![输入框为空状态下按钮高亮显示](https://user-images.githubusercontent.com/80370536/117533834-72b7f580-b021-11eb-81ff-ea0f0cdb49c5.jpg) 2.sign up 结果:输入框为空状态下next按钮高亮显示 期望:建议为空状态下,操作按钮暗淡色,输入后高亮显示,与android保持一致
non_process
输入框为空状态下按钮高亮显示 机型:iphone se 步骤: 启动bookey,登陆首页 sign up 结果:输入框为空状态下next按钮高亮显示 期望:建议为空状态下,操作按钮暗淡色,输入后高亮显示,与android保持一致
0
745
3,214,363,065
IssuesEvent
2015-10-07 01:11:32
grafeo/grafeo
https://api.github.com/repos/grafeo/grafeo
closed
Read and Write PPM
Component: Image Processing feature request priority: high
- `image_read_ppm` - `image_write_ppm` - link `image_read` and `image_write` with PPM versions
1.0
Read and Write PPM - - `image_read_ppm` - `image_write_ppm` - link `image_read` and `image_write` with PPM versions
process
read and write ppm image read ppm image write ppm link image read and image write with ppm versions
1
122,418
26,126,189,129
IssuesEvent
2022-12-28 18:56:17
shelcia/mocker
https://api.github.com/repos/shelcia/mocker
closed
Fix code scanning alert - Database query built from user-controlled sources
medium codepeak22
<!-- Warning: The suggested title contains the alert rule name. This can expose security information. --> Tracking issue for: - [ ] https://github.com/shelcia/mocker/security/code-scanning/1
1.0
Fix code scanning alert - Database query built from user-controlled sources - <!-- Warning: The suggested title contains the alert rule name. This can expose security information. --> Tracking issue for: - [ ] https://github.com/shelcia/mocker/security/code-scanning/1
non_process
fix code scanning alert database query built from user controlled sources tracking issue for
0
18,155
24,192,984,880
IssuesEvent
2022-09-23 19:41:56
google/android-fhir
https://api.github.com/repos/google/android-fhir
reopened
HAPI Structures, Java11, and Android API levels
process
This is an issue @vitorpamplona raised in the process of implementing #1403. PR #1603 introduces new dependencies such as CQL evalutor, CQL engine, and CQL translator. They use HAPI version 6.0.1 which is compiled using Java11. There are certain Java11 APIs that are not available on older Android versions, especially pre Android API 26. For example, `java.lang.reflect.Method.getParameterCount()`. As a result, with this PR, the workflow library will crash on Android API level 24. Please note, however, that although the API `java.lang.reflect.Method.getParameterCount()` is used in HAPI, we do not invoke code paths in HAPI using this API through our usage in the FHIR Engine and SDC library. Only in the workflow library in this PR. In other words, when I tested the FHIR Engine and SDC library with HAPI 6 on Android API level 24, there's no crash caused by the missing of this API. Android API levels 26+ support Java11 APIs. So the above PR works fine in Android API 26+. This means that we will need to update the min api level for the workflow library to API 26. At the moment, our gradle files are written so that all libraries share the same min api level. This will need to be changed so we do not raise API level above 24 for users of FHIR engine and SDC library. cc: @joiskash
1.0
HAPI Structures, Java11, and Android API levels - This is an issue @vitorpamplona raised in the process of implementing #1403. PR #1603 introduces new dependencies such as CQL evalutor, CQL engine, and CQL translator. They use HAPI version 6.0.1 which is compiled using Java11. There are certain Java11 APIs that are not available on older Android versions, especially pre Android API 26. For example, `java.lang.reflect.Method.getParameterCount()`. As a result, with this PR, the workflow library will crash on Android API level 24. Please note, however, that although the API `java.lang.reflect.Method.getParameterCount()` is used in HAPI, we do not invoke code paths in HAPI using this API through our usage in the FHIR Engine and SDC library. Only in the workflow library in this PR. In other words, when I tested the FHIR Engine and SDC library with HAPI 6 on Android API level 24, there's no crash caused by the missing of this API. Android API levels 26+ support Java11 APIs. So the above PR works fine in Android API 26+. This means that we will need to update the min api level for the workflow library to API 26. At the moment, our gradle files are written so that all libraries share the same min api level. This will need to be changed so we do not raise API level above 24 for users of FHIR engine and SDC library. cc: @joiskash
process
hapi structures and android api levels this is an issue vitorpamplona raised in the process of implementing pr introduces new dependencies such as cql evalutor cql engine and cql translator they use hapi version which is compiled using there are certain apis that are not available on older android versions especially pre android api for example java lang reflect method getparametercount as a result with this pr the workflow library will crash on android api level please note however that although the api java lang reflect method getparametercount is used in hapi we do not invoke code paths in hapi using this api through our usage in the fhir engine and sdc library only in the workflow library in this pr in other words when i tested the fhir engine and sdc library with hapi on android api level there s no crash caused by the missing of this api android api levels support apis so the above pr works fine in android api this means that we will need to update the min api level for the workflow library to api at the moment our gradle files are written so that all libraries share the same min api level this will need to be changed so we do not raise api level above for users of fhir engine and sdc library cc joiskash
1
10,088
13,044,162,001
IssuesEvent
2020-07-29 03:47:28
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `SubTimeDateTimeNull` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `SubTimeDateTimeNull` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `SubTimeDateTimeNull` from TiDB - ## Description Port the scalar function `SubTimeDateTimeNull` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function subtimedatetimenull from tidb description port the scalar function subtimedatetimenull from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
1
10,203
13,066,579,076
IssuesEvent
2020-07-30 22:01:04
googleapis/google-auth-library-ruby
https://api.github.com/repos/googleapis/google-auth-library-ruby
closed
New release enquiry
type: process
Hi there, I am looking to use the recent changes that have been merged to master: https://github.com/googleapis/google-auth-library-ruby/commit/48c689aa93bfe81c5a6ae23362d86fc25ba098cf What is the typical process of a change being merged and then a release being made? Currently, we could track master branch, but this will not work with projects that use gemspec.
1.0
New release enquiry - Hi there, I am looking to use the recent changes that have been merged to master: https://github.com/googleapis/google-auth-library-ruby/commit/48c689aa93bfe81c5a6ae23362d86fc25ba098cf What is the typical process of a change being merged and then a release being made? Currently, we could track master branch, but this will not work with projects that use gemspec.
process
new release enquiry hi there i am looking to use the recent changes that have been merged to master what is the typical process of a change being merged and then a release being made currently we could track master branch but this will not work with projects that use gemspec
1
162,706
20,241,557,365
IssuesEvent
2022-02-14 09:46:27
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Security Solution] Not able to create an EQL rule due to validation error
bug impact:critical fixed Team: SecuritySolution Team:Detection Rules v7.17.1
**Describe the bug:** - Not able to create an EQL rule due to validation error **Kibana/Elasticsearch Stack version:** - 7.17 latest branch (45be56082f7338428b62ebd1c67b222f470be0ce) - 7.17.1 latest snapshot **Steps to reproduce:** 1. Navigate to the rules page 2. Click on `create new rule` 2. Select `Event Correlation type` 3. Enter a valid EQL query **Current behavior:** - A validation error is returned - You cannot proceed with the rule creation - The rule cannot be created <img width="1034" alt="Screenshot 2022-02-09 at 12 19 39" src="https://user-images.githubusercontent.com/17427073/153189817-8109045d-0515-4794-8a1c-3ec65f33bbd3.png"> <img width="769" alt="Screenshot 2022-02-09 at 12 19 55" src="https://user-images.githubusercontent.com/17427073/153189858-ba2a4bd0-f78e-44b5-a752-5d6edcf61414.png"> **Expected behavior:** - No validation error is displayed - The rule can be correctly created **Additional information:** - The displayed error: ```` {"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]"}],"type":"illegal_argument_exception","reason":"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]"},"status":400} ```` ```` { "name": "Error", "message": "{\"error\":{\"root_cause\":[{\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"}],\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"},\"status\":400}", "stack": "Error: {\"error\":{\"root_cause\":[{\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"}],\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"},\"status\":400}\n at u (https://upgrades.kb.us-central1.gcp.qa.cld.elstc.co:9243/46574/bundles/plugin/securitySolution/8.0.0/securitySolution.chunk.12.js:3:16800)\n at async f (https://upgrades.kb.us-central1.gcp.qa.cld.elstc.co:9243/46574/bundles/plugin/securitySolution/8.0.0/securitySolution.chunk.12.js:3:17361)" } ```` - After the upgrade of a 7.17 EQL rule to 7.17.1 version, the rule still works and generate alerts - After the upgrade of a 7.17 EQL rule to 7.17.1 version, the same error validation error is displayed when trying to edit the rule. So the rule cannot be edited.
True
[Security Solution] Not able to create an EQL rule due to validation error - **Describe the bug:** - Not able to create an EQL rule due to validation error **Kibana/Elasticsearch Stack version:** - 7.17 latest branch (45be56082f7338428b62ebd1c67b222f470be0ce) - 7.17.1 latest snapshot **Steps to reproduce:** 1. Navigate to the rules page 2. Click on `create new rule` 2. Select `Event Correlation type` 3. Enter a valid EQL query **Current behavior:** - A validation error is returned - You cannot proceed with the rule creation - The rule cannot be created <img width="1034" alt="Screenshot 2022-02-09 at 12 19 39" src="https://user-images.githubusercontent.com/17427073/153189817-8109045d-0515-4794-8a1c-3ec65f33bbd3.png"> <img width="769" alt="Screenshot 2022-02-09 at 12 19 55" src="https://user-images.githubusercontent.com/17427073/153189858-ba2a4bd0-f78e-44b5-a752-5d6edcf61414.png"> **Expected behavior:** - No validation error is displayed - The rule can be correctly created **Additional information:** - The displayed error: ```` {"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]"}],"type":"illegal_argument_exception","reason":"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]"},"status":400} ```` ```` { "name": "Error", "message": "{\"error\":{\"root_cause\":[{\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"}],\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"},\"status\":400}", "stack": "Error: {\"error\":{\"root_cause\":[{\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"}],\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"},\"status\":400}\n at u (https://upgrades.kb.us-central1.gcp.qa.cld.elstc.co:9243/46574/bundles/plugin/securitySolution/8.0.0/securitySolution.chunk.12.js:3:16800)\n at async f (https://upgrades.kb.us-central1.gcp.qa.cld.elstc.co:9243/46574/bundles/plugin/securitySolution/8.0.0/securitySolution.chunk.12.js:3:17361)" } ```` - After the upgrade of a 7.17 EQL rule to 7.17.1 version, the rule still works and generate alerts - After the upgrade of a 7.17 EQL rule to 7.17.1 version, the same error validation error is displayed when trying to edit the rule. So the rule cannot be edited.
non_process
not able to create an eql rule due to validation error describe the bug not able to create an eql rule due to validation error kibana elasticsearch stack version latest branch latest snapshot steps to reproduce navigate to the rules page click on create new rule select event correlation type enter a valid eql query current behavior a validation error is returned you cannot proceed with the rule creation the rule cannot be created img width alt screenshot at src img width alt screenshot at src expected behavior no validation error is displayed the rule can be correctly created additional information the displayed error error root cause contains unrecognized parameter type illegal argument exception reason request contains unrecognized parameter status name error message error root cause contains unrecognized parameter type illegal argument exception reason request contains unrecognized parameter status stack error error root cause contains unrecognized parameter type illegal argument exception reason request contains unrecognized parameter status n at u at async f after the upgrade of a eql rule to version the rule still works and generate alerts after the upgrade of a eql rule to version the same error validation error is displayed when trying to edit the rule so the rule cannot be edited
0
284,484
8,742,809,063
IssuesEvent
2018-12-12 17:20:25
K-Ho/code
https://api.github.com/repos/K-Ho/code
opened
Fixed Initial sizing/ position of Graph Network
UI/UX high priority
Make the graph network render larger by default and always in an open space (center of screen)
1.0
Fixed Initial sizing/ position of Graph Network - Make the graph network render larger by default and always in an open space (center of screen)
non_process
fixed initial sizing position of graph network make the graph network render larger by default and always in an open space center of screen
0
189,344
22,047,021,634
IssuesEvent
2022-05-30 03:43:46
dpteam/RK3188_TABLET
https://api.github.com/repos/dpteam/RK3188_TABLET
closed
WS-2021-0279 (Medium) detected in randomv3.0.66 - autoclosed
security vulnerability
## WS-2021-0279 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>randomv3.0.66</b></p></summary> <p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git>https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/tree-log.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Linux Kernel in versions v2.6.29-rc1 to v5.12.9 there is an error handling in fixup_inode_link_counts which could lead to memory leak. <p>Publish Date: 2021-06-25 <p>URL: <a href=https://github.com/gregkh/linux/commit/4cd303735bdfacd115ee20a6f3235b0084924174>WS-2021-0279</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1000819">https://osv.dev/vulnerability/UVI-2021-1000819</a></p> <p>Release Date: 2021-06-25</p> <p>Fix Resolution: v5.12.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0279 (Medium) detected in randomv3.0.66 - autoclosed - ## WS-2021-0279 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>randomv3.0.66</b></p></summary> <p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git>https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/tree-log.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Linux Kernel in versions v2.6.29-rc1 to v5.12.9 there is an error handling in fixup_inode_link_counts which could lead to memory leak. <p>Publish Date: 2021-06-25 <p>URL: <a href=https://github.com/gregkh/linux/commit/4cd303735bdfacd115ee20a6f3235b0084924174>WS-2021-0279</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1000819">https://osv.dev/vulnerability/UVI-2021-1000819</a></p> <p>Release Date: 2021-06-25</p> <p>Fix Resolution: v5.12.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws medium detected in autoclosed ws medium severity vulnerability vulnerable library library home page a href found in head commit a href found in base branch master vulnerable source files fs btrfs tree log c vulnerability details linux kernel in versions to there is an error handling in fixup inode link counts which could lead to memory leak publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
36,979
15,110,609,649
IssuesEvent
2021-02-08 19:27:52
microsoft/BotFramework-Composer
https://api.github.com/repos/microsoft/BotFramework-Composer
closed
End conversation by the Bot
Bot Services Support Type: Question customer-replied-to customer-reported
I have created a Bot which I have attached to Omnichannel for customer services (D365 extension). When a user starts a conversation with he Bot through any channel (e.g. Teams) the conversation goes for ever with the Bot. I am not able to end up the conversation in Omnichannel for customer service as there is no standard Action in the Bot Framework Composer to end the conversation. Could you advise what is the easiest way to end up the conversation by the Bot so that it ends in Omnichannel for customer services (D365) as well?
1.0
End conversation by the Bot - I have created a Bot which I have attached to Omnichannel for customer services (D365 extension). When a user starts a conversation with he Bot through any channel (e.g. Teams) the conversation goes for ever with the Bot. I am not able to end up the conversation in Omnichannel for customer service as there is no standard Action in the Bot Framework Composer to end the conversation. Could you advise what is the easiest way to end up the conversation by the Bot so that it ends in Omnichannel for customer services (D365) as well?
non_process
end conversation by the bot i have created a bot which i have attached to omnichannel for customer services extension when a user starts a conversation with he bot through any channel e g teams the conversation goes for ever with the bot i am not able to end up the conversation in omnichannel for customer service as there is no standard action in the bot framework composer to end the conversation could you advise what is the easiest way to end up the conversation by the bot so that it ends in omnichannel for customer services as well
0
51,938
13,690,550,323
IssuesEvent
2020-09-30 14:30:43
mrgadgil/test
https://api.github.com/repos/mrgadgil/test
opened
HIGH severity finding reported by IBM Security Advisor
IBM Security Advisor
**Source**: Security Advisor **Finding**: 15c57fd8bfbff988b328b55b167a1689/providers/security-advisor/occurrences/ata-1601476234889 **Severity**: HIGH [View in Security Advisor Dashboard](https://cloud.ibm.com/security-advisor#/findings?id=15c57fd8bfbff988b328b55b167a1689/providers/security-advisor/occurrences/ata-1601476234889&region=us-south)
True
HIGH severity finding reported by IBM Security Advisor - **Source**: Security Advisor **Finding**: 15c57fd8bfbff988b328b55b167a1689/providers/security-advisor/occurrences/ata-1601476234889 **Severity**: HIGH [View in Security Advisor Dashboard](https://cloud.ibm.com/security-advisor#/findings?id=15c57fd8bfbff988b328b55b167a1689/providers/security-advisor/occurrences/ata-1601476234889&region=us-south)
non_process
high severity finding reported by ibm security advisor source security advisor finding providers security advisor occurrences ata severity high
0
13,394
15,866,825,496
IssuesEvent
2021-04-08 16:10:14
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Improve error for invalid permissions on Linux
bug/2-confirmed kind/bug process/candidate team/client
I'm running into this on ubuntu 20.04 while having `prisma@2.20.1` installed globally. `prisma db push --preview-feature` ```sh Environment variables loaded from prisma/.env Prisma schema loaded from prisma/schema.prisma Datasource "db": PostgreSQL database "debug-migrate", schema "public" at "playground-database-postgres.cluster-clfeqqifnebj.eu-west-1.rds.amazonaws.com:5432" PostgreSQL database debug-migrate created at playground-database-postgres.cluster-clfeqqifnebj.eu-west-1.rds.amazonaws.com:5432 🚀 Your database is now in sync with your schema. Done in 2.54s Running generate... (Use --skip-generate to skip the generators) Error: EACCES: permission denied, copyfile '/home/millsp/.cache/prisma/master/60ba6551f29b17d7d6ce479e5733c70d9c00860e/debian-openssl-1.1.x/query-engine' -> '/usr/local/lib/node_modules/prisma/query-engine-debian-openssl-1.1.x' ```
1.0
Improve error for invalid permissions on Linux - I'm running into this on ubuntu 20.04 while having `prisma@2.20.1` installed globally. `prisma db push --preview-feature` ```sh Environment variables loaded from prisma/.env Prisma schema loaded from prisma/schema.prisma Datasource "db": PostgreSQL database "debug-migrate", schema "public" at "playground-database-postgres.cluster-clfeqqifnebj.eu-west-1.rds.amazonaws.com:5432" PostgreSQL database debug-migrate created at playground-database-postgres.cluster-clfeqqifnebj.eu-west-1.rds.amazonaws.com:5432 🚀 Your database is now in sync with your schema. Done in 2.54s Running generate... (Use --skip-generate to skip the generators) Error: EACCES: permission denied, copyfile '/home/millsp/.cache/prisma/master/60ba6551f29b17d7d6ce479e5733c70d9c00860e/debian-openssl-1.1.x/query-engine' -> '/usr/local/lib/node_modules/prisma/query-engine-debian-openssl-1.1.x' ```
process
improve error for invalid permissions on linux i m running into this on ubuntu while having prisma installed globally prisma db push preview feature sh environment variables loaded from prisma env prisma schema loaded from prisma schema prisma datasource db postgresql database debug migrate schema public at playground database postgres cluster clfeqqifnebj eu west rds amazonaws com postgresql database debug migrate created at playground database postgres cluster clfeqqifnebj eu west rds amazonaws com 🚀 your database is now in sync with your schema done in running generate use skip generate to skip the generators error eacces permission denied copyfile home millsp cache prisma master debian openssl x query engine usr local lib node modules prisma query engine debian openssl x
1
791,756
27,874,745,612
IssuesEvent
2023-03-21 15:25:39
pastas/pastas
https://api.github.com/repos/pastas/pastas
closed
[ENHANCEMENT] allow ffill and bfill as fill_before and fill_after options for time series
enhancement priority 2
**Describe the proposed enhancement** Allow the following settings dictionary: ```python settings = { "sample_up": "interpolate", "sample_down": "mean", "fill_before": "bfill", # <- new option "fill_after": "ffill", # <- new option "fill_nan": "interpolate", } ``` I know this is the same as getting the first/last value from the timeseries and entering that into the settings dictionary for fill before/after, but I like not having to think about it :). When would you ever use this? I used it for a pumping test, where I used a reference head time series (with no pumping influence) as input for my time series models, that in some cases needed to be extended slightly into the past/future. In my case, extending the first/last values as fill values worked pretty well.
1.0
[ENHANCEMENT] allow ffill and bfill as fill_before and fill_after options for time series - **Describe the proposed enhancement** Allow the following settings dictionary: ```python settings = { "sample_up": "interpolate", "sample_down": "mean", "fill_before": "bfill", # <- new option "fill_after": "ffill", # <- new option "fill_nan": "interpolate", } ``` I know this is the same as getting the first/last value from the timeseries and entering that into the settings dictionary for fill before/after, but I like not having to think about it :). When would you ever use this? I used it for a pumping test, where I used a reference head time series (with no pumping influence) as input for my time series models, that in some cases needed to be extended slightly into the past/future. In my case, extending the first/last values as fill values worked pretty well.
non_process
allow ffill and bfill as fill before and fill after options for time series describe the proposed enhancement allow the following settings dictionary python settings sample up interpolate sample down mean fill before bfill new option fill after ffill new option fill nan interpolate i know this is the same as getting the first last value from the timeseries and entering that into the settings dictionary for fill before after but i like not having to think about it when would you ever use this i used it for a pumping test where i used a reference head time series with no pumping influence as input for my time series models that in some cases needed to be extended slightly into the past future in my case extending the first last values as fill values worked pretty well
0
10,197
13,064,284,901
IssuesEvent
2020-07-30 17:50:32
GetTerminus/terminus-oss
https://api.github.com/repos/GetTerminus/terminus-oss
closed
Split NGX-Tools repo and move parts into monorepo
Goal: Process Improvement Type: chore
- [ ] Follow existing endpoints for package separation? - [ ] Publish to existing NPM package - [ ] Move any applicable issues
1.0
Split NGX-Tools repo and move parts into monorepo - - [ ] Follow existing endpoints for package separation? - [ ] Publish to existing NPM package - [ ] Move any applicable issues
process
split ngx tools repo and move parts into monorepo follow existing endpoints for package separation publish to existing npm package move any applicable issues
1
9,176
12,226,503,509
IssuesEvent
2020-05-03 11:14:24
gfleetwood/asteres
https://api.github.com/repos/gfleetwood/asteres
opened
AdvancerTechnologies/MyoWare_MuscleSensor (35062662)
Processing engineering
https://github.com/AdvancerTechnologies/MyoWare_MuscleSensor Example code and documentation for the MyoWare™ Muscle Sensor
1.0
AdvancerTechnologies/MyoWare_MuscleSensor (35062662) - https://github.com/AdvancerTechnologies/MyoWare_MuscleSensor Example code and documentation for the MyoWare™ Muscle Sensor
process
advancertechnologies myoware musclesensor example code and documentation for the myoware™ muscle sensor
1
20,628
27,300,060,398
IssuesEvent
2023-02-24 00:39:09
googleapis/google-cloudevents-python
https://api.github.com/repos/googleapis/google-cloudevents-python
closed
Warning: a recent release failed
type: process status: will not fix api: eventarc
The following release PRs may have failed: * #190 - The release job was triggered, but has not reported back success. * #187 - The release job was triggered, but has not reported back success. * #185 - The release job was triggered, but has not reported back success. * #183 - The release job was triggered, but has not reported back success. * #181 - The release job was triggered, but has not reported back success. * #136 - The release job was triggered, but has not reported back success.
1.0
Warning: a recent release failed - The following release PRs may have failed: * #190 - The release job was triggered, but has not reported back success. * #187 - The release job was triggered, but has not reported back success. * #185 - The release job was triggered, but has not reported back success. * #183 - The release job was triggered, but has not reported back success. * #181 - The release job was triggered, but has not reported back success. * #136 - The release job was triggered, but has not reported back success.
process
warning a recent release failed the following release prs may have failed the release job was triggered but has not reported back success the release job was triggered but has not reported back success the release job was triggered but has not reported back success the release job was triggered but has not reported back success the release job was triggered but has not reported back success the release job was triggered but has not reported back success
1
2,996
5,970,835,671
IssuesEvent
2017-05-31 00:03:08
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Desktop: System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper failed with "Xunit.Sdk.EqualException"
area-System.ServiceProcess test-run-desktop
Failed test: System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper Detail: https://ci.dot.net/job/dotnet_corefx/job/master/job/outerloop_netfx_windows_nt_debug/66/testReport/System.ServiceProcess.Tests/ServiceControllerTests/ConstructWithServiceName_ToUpper/ Configuration: outerloop_netfx_windows_nt_debug MESSAGE: ~~~ Assert.Equal() Failure ↓ (pos 3) Expected: 749b386a-fd41-4bed-951f-75d4a705f7d9 Actual: 749B386A-FD41-4BED-951F-75D4A705F7D9 ↑ (pos 3) ~~~ STACK TRACE: ~~~ at System.ServiceProcess.Tests.ServiceControllerTests.AssertExpectedProperties(ServiceController testServiceController) in D:\j\workspace\outerloop_net---903ddde6\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.Tests\ServiceControllerTests.cs:line 92 at System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper() in D:\j\workspace\outerloop_net---903ddde6\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.Tests\ServiceControllerTests.cs:line 109 ~~~
1.0
Desktop: System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper failed with "Xunit.Sdk.EqualException" - Failed test: System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper Detail: https://ci.dot.net/job/dotnet_corefx/job/master/job/outerloop_netfx_windows_nt_debug/66/testReport/System.ServiceProcess.Tests/ServiceControllerTests/ConstructWithServiceName_ToUpper/ Configuration: outerloop_netfx_windows_nt_debug MESSAGE: ~~~ Assert.Equal() Failure ↓ (pos 3) Expected: 749b386a-fd41-4bed-951f-75d4a705f7d9 Actual: 749B386A-FD41-4BED-951F-75D4A705F7D9 ↑ (pos 3) ~~~ STACK TRACE: ~~~ at System.ServiceProcess.Tests.ServiceControllerTests.AssertExpectedProperties(ServiceController testServiceController) in D:\j\workspace\outerloop_net---903ddde6\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.Tests\ServiceControllerTests.cs:line 92 at System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper() in D:\j\workspace\outerloop_net---903ddde6\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.Tests\ServiceControllerTests.cs:line 109 ~~~
process
desktop system serviceprocess tests servicecontrollertests constructwithservicename toupper failed with xunit sdk equalexception failed test system serviceprocess tests servicecontrollertests constructwithservicename toupper detail configuration outerloop netfx windows nt debug message assert equal failure ↓ pos expected actual ↑ pos stack trace at system serviceprocess tests servicecontrollertests assertexpectedproperties servicecontroller testservicecontroller in d j workspace outerloop net src system serviceprocess servicecontroller tests system serviceprocess servicecontroller tests servicecontrollertests cs line at system serviceprocess tests servicecontrollertests constructwithservicename toupper in d j workspace outerloop net src system serviceprocess servicecontroller tests system serviceprocess servicecontroller tests servicecontrollertests cs line
1
174,427
27,637,871,856
IssuesEvent
2023-03-10 15:44:41
coder/coder
https://api.github.com/repos/coder/coder
closed
Discussion: How should we handle joins?
chore design
# Problem Our current use of SQLc uses almost 0 joins throughout all of our queries. This causes excessive db round trips. The function [`workspaceData`][workspaceData] does 9 database calls to completely populate the returned workspace from the api. This can be reduced with [joins][workspace-joined]. When `dbauthz` is enabled by default, this compounds the issue as many objects require fetching related objects to run authorization. Examples: [workspace builds][build-rbac], [template versions][version-rbac], jobs, build parameters, ... . [build-rbac]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/database/dbauthz/querier.go#L1241-L1251 [version-rbac]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/database/dbauthz/querier.go#L628-L643 [workspaceData]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/workspaces.go#L1047-L1047 [workspace-joined]: https://github.com/coder/coder/blob/6adacd74fa79ae2c424e2144afc65c6e5994510c/coderd/database/sqlxqueries/workspace.gosql#L2-L75 # Solutions [(this PR talks about both)](https://github.com/coder/coder/pull/6371) Allowing **queries to leverage joins** can reduce db round trips. The issue is how to handle these. We currently use SQLc, which would create a new model for each query and it becomes cumbersome. Additionally, we often have multiple `SELECT` queries for each datatype. I have 2 proposals. ## Keeping SQLc [(originally investigated here)](https://github.com/coder/coder/issues/2201) If we want to keep SQLc, the best way we can do this is with **views**. I am suggesting non-materialized views. So essentially these are saved queries in our postgres database that we can reference from SQLc. Because views are saved in postgres, they require the same migration maintenance a table would. https://github.com/coder/coder/blob/c3218f71cab8d571c2a95bca08efe2816a41fa86/coderd/database/migrations/000102_workspace_build_view.up.sql#L1-L14 A view is a "table" for SQLc, so a model is generated for it. In our `.sql` files, we just reference the view. https://github.com/coder/coder/blob/c3218f71cab8d571c2a95bca08efe2816a41fa86/coderd/database/queries/workspacebuilds.sql#L1-L9 ## Abandon SQLc and use SQLx + Go Templates [(an implementation here)](https://github.com/coder/coder/pull/6356) We use SQLc to generate Golang code from our sql. But this code isn't actually that complex in Golang if we use SQLx. And then we can use Go templates to build dynamic queries. A very basic implementation of this makes our `*.sql` files look like this. Note the highlighting will be a bit off as we are mixing SQL and Go templates. ```sql {{ define "workspace_builds_rbac" }} ( SELECT workspace_builds.*, workspaces.organization_id AS organization_id, workspaces.owner_id AS workspace_owner_id FROM workspace_builds INNER JOIN workspaces ON workspace_builds.workspace_id = workspaces.id ) {{ end }} -- To use the template above {{ define "GetWorkspaceBuildByID" }} SELECT * FROM {{ template "workspace_builds_rbac" }} WHERE id = @build_id {{ end }} ``` I investigated [IDE Highlighting][IDE-High], but the tl;dr is that just using `sql` highlighting is likely the best. It will never be perfect with Go templates :cry:. ### How to use template query in golang? I made a package called `sqxqueriers` that handles templates and keeping our `@param` named parameters for easier to read sqlc. You can see that [here](sqlxqueries). To call a query you made, you can use a generic function called sqlxGet for fetching one row: https://github.com/coder/coder/blob/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/sqlx.go#L10 The type embeds the sqlc `WorkspaceBuild` type. We would keep SQLc for generating models. We'd just move queries to this SQLx. The `db` struct tags are used for matching to columns. Obviously the query can be handled more manually if the tags do not match. https://github.com/coder/coder/blob/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/modelqueries.go#L183-L191 [IDE-High]: https://github.com/coder/coder/blob/6adacd74fa79ae2c424e2144afc65c6e5994510c/coderd/database/sqlxqueries/README.md [sqlxqueries]: https://github.com/coder/coder/tree/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/sqlxqueries # Comparison | | SQLc Views | SQLx + Templates | |--------------------------|--------------------------------|------------------| | SQL highlighting | ✅ | ⚠️ (kinda) | | Golang Code | ✅ | ✅ | | Golang Simple Types | ✅ | ✅ | | Golang Joined Types | No embeds, duplicated | Uses Embeds | | Supports Dynamic Queries | ❌ (gross strings replace) | ✅ | | No Migrations | ❌ (migration to maintain view) | ✅ | ## Highligting SQL template highlighting in vscode isn't that bad. In Goland it's pretty terrible. ## Golang Code SQLx + Templates requires more infrastructure code to support, but this also gives us opportunity to add in features. SQLc is slow to update and add features we need. ## Joined Types SQLc joined types are duplicated structs with identical fields. We can do anonymous embeds for template joined types. I would really like to see `dbauthz` not adhere to `db.Store` and then we can do some better type handling at this layer to make types more consistent. You cannot insert into a view, so when updating or inserting data, you cannot return the joined data. Meaning 2 types will exist. A `thin` and a `joined`. ## Dynamic Queries We currently use `CASE WHEN` statements for dynamic where clauses. This works fine, but is only supported in `WHERE`. Things not supported currently: - Conditional sort order (ASC vs DESC) - Conditional sort column - Conditional updates, only update fields provided. [Caused bug in v1](https://github.com/coder/v1/issues/13767)
1.0
Discussion: How should we handle joins? - # Problem Our current use of SQLc uses almost 0 joins throughout all of our queries. This causes excessive db round trips. The function [`workspaceData`][workspaceData] does 9 database calls to completely populate the returned workspace from the api. This can be reduced with [joins][workspace-joined]. When `dbauthz` is enabled by default, this compounds the issue as many objects require fetching related objects to run authorization. Examples: [workspace builds][build-rbac], [template versions][version-rbac], jobs, build parameters, ... . [build-rbac]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/database/dbauthz/querier.go#L1241-L1251 [version-rbac]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/database/dbauthz/querier.go#L628-L643 [workspaceData]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/workspaces.go#L1047-L1047 [workspace-joined]: https://github.com/coder/coder/blob/6adacd74fa79ae2c424e2144afc65c6e5994510c/coderd/database/sqlxqueries/workspace.gosql#L2-L75 # Solutions [(this PR talks about both)](https://github.com/coder/coder/pull/6371) Allowing **queries to leverage joins** can reduce db round trips. The issue is how to handle these. We currently use SQLc, which would create a new model for each query and it becomes cumbersome. Additionally, we often have multiple `SELECT` queries for each datatype. I have 2 proposals. ## Keeping SQLc [(originally investigated here)](https://github.com/coder/coder/issues/2201) If we want to keep SQLc, the best way we can do this is with **views**. I am suggesting non-materialized views. So essentially these are saved queries in our postgres database that we can reference from SQLc. Because views are saved in postgres, they require the same migration maintenance a table would. https://github.com/coder/coder/blob/c3218f71cab8d571c2a95bca08efe2816a41fa86/coderd/database/migrations/000102_workspace_build_view.up.sql#L1-L14 A view is a "table" for SQLc, so a model is generated for it. In our `.sql` files, we just reference the view. https://github.com/coder/coder/blob/c3218f71cab8d571c2a95bca08efe2816a41fa86/coderd/database/queries/workspacebuilds.sql#L1-L9 ## Abandon SQLc and use SQLx + Go Templates [(an implementation here)](https://github.com/coder/coder/pull/6356) We use SQLc to generate Golang code from our sql. But this code isn't actually that complex in Golang if we use SQLx. And then we can use Go templates to build dynamic queries. A very basic implementation of this makes our `*.sql` files look like this. Note the highlighting will be a bit off as we are mixing SQL and Go templates. ```sql {{ define "workspace_builds_rbac" }} ( SELECT workspace_builds.*, workspaces.organization_id AS organization_id, workspaces.owner_id AS workspace_owner_id FROM workspace_builds INNER JOIN workspaces ON workspace_builds.workspace_id = workspaces.id ) {{ end }} -- To use the template above {{ define "GetWorkspaceBuildByID" }} SELECT * FROM {{ template "workspace_builds_rbac" }} WHERE id = @build_id {{ end }} ``` I investigated [IDE Highlighting][IDE-High], but the tl;dr is that just using `sql` highlighting is likely the best. It will never be perfect with Go templates :cry:. ### How to use template query in golang? I made a package called `sqxqueriers` that handles templates and keeping our `@param` named parameters for easier to read sqlc. You can see that [here](sqlxqueries). To call a query you made, you can use a generic function called sqlxGet for fetching one row: https://github.com/coder/coder/blob/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/sqlx.go#L10 The type embeds the sqlc `WorkspaceBuild` type. We would keep SQLc for generating models. We'd just move queries to this SQLx. The `db` struct tags are used for matching to columns. Obviously the query can be handled more manually if the tags do not match. https://github.com/coder/coder/blob/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/modelqueries.go#L183-L191 [IDE-High]: https://github.com/coder/coder/blob/6adacd74fa79ae2c424e2144afc65c6e5994510c/coderd/database/sqlxqueries/README.md [sqlxqueries]: https://github.com/coder/coder/tree/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/sqlxqueries # Comparison | | SQLc Views | SQLx + Templates | |--------------------------|--------------------------------|------------------| | SQL highlighting | ✅ | ⚠️ (kinda) | | Golang Code | ✅ | ✅ | | Golang Simple Types | ✅ | ✅ | | Golang Joined Types | No embeds, duplicated | Uses Embeds | | Supports Dynamic Queries | ❌ (gross strings replace) | ✅ | | No Migrations | ❌ (migration to maintain view) | ✅ | ## Highligting SQL template highlighting in vscode isn't that bad. In Goland it's pretty terrible. ## Golang Code SQLx + Templates requires more infrastructure code to support, but this also gives us opportunity to add in features. SQLc is slow to update and add features we need. ## Joined Types SQLc joined types are duplicated structs with identical fields. We can do anonymous embeds for template joined types. I would really like to see `dbauthz` not adhere to `db.Store` and then we can do some better type handling at this layer to make types more consistent. You cannot insert into a view, so when updating or inserting data, you cannot return the joined data. Meaning 2 types will exist. A `thin` and a `joined`. ## Dynamic Queries We currently use `CASE WHEN` statements for dynamic where clauses. This works fine, but is only supported in `WHERE`. Things not supported currently: - Conditional sort order (ASC vs DESC) - Conditional sort column - Conditional updates, only update fields provided. [Caused bug in v1](https://github.com/coder/v1/issues/13767)
non_process
discussion how should we handle joins problem our current use of sqlc uses almost joins throughout all of our queries this causes excessive db round trips the function does database calls to completely populate the returned workspace from the api this can be reduced with when dbauthz is enabled by default this compounds the issue as many objects require fetching related objects to run authorization examples jobs build parameters solutions allowing queries to leverage joins can reduce db round trips the issue is how to handle these we currently use sqlc which would create a new model for each query and it becomes cumbersome additionally we often have multiple select queries for each datatype i have proposals keeping sqlc if we want to keep sqlc the best way we can do this is with views i am suggesting non materialized views so essentially these are saved queries in our postgres database that we can reference from sqlc because views are saved in postgres they require the same migration maintenance a table would a view is a table for sqlc so a model is generated for it in our sql files we just reference the view abandon sqlc and use sqlx go templates we use sqlc to generate golang code from our sql but this code isn t actually that complex in golang if we use sqlx and then we can use go templates to build dynamic queries a very basic implementation of this makes our sql files look like this note the highlighting will be a bit off as we are mixing sql and go templates sql define workspace builds rbac select workspace builds workspaces organization id as organization id workspaces owner id as workspace owner id from workspace builds inner join workspaces on workspace builds workspace id workspaces id end to use the template above define getworkspacebuildbyid select from template workspace builds rbac where id build id end i investigated but the tl dr is that just using sql highlighting is likely the best it will never be perfect with go templates cry how to use template query in golang i made a package called sqxqueriers that handles templates and keeping our param named parameters for easier to read sqlc you can see that sqlxqueries to call a query you made you can use a generic function called sqlxget for fetching one row the type embeds the sqlc workspacebuild type we would keep sqlc for generating models we d just move queries to this sqlx the db struct tags are used for matching to columns obviously the query can be handled more manually if the tags do not match comparison sqlc views sqlx templates sql highlighting ✅ ⚠️ kinda golang code ✅ ✅ golang simple types ✅ ✅ golang joined types no embeds duplicated uses embeds supports dynamic queries ❌ gross strings replace ✅ no migrations ❌ migration to maintain view ✅ highligting sql template highlighting in vscode isn t that bad in goland it s pretty terrible golang code sqlx templates requires more infrastructure code to support but this also gives us opportunity to add in features sqlc is slow to update and add features we need joined types sqlc joined types are duplicated structs with identical fields we can do anonymous embeds for template joined types i would really like to see dbauthz not adhere to db store and then we can do some better type handling at this layer to make types more consistent you cannot insert into a view so when updating or inserting data you cannot return the joined data meaning types will exist a thin and a joined dynamic queries we currently use case when statements for dynamic where clauses this works fine but is only supported in where things not supported currently conditional sort order asc vs desc conditional sort column conditional updates only update fields provided
0
125,235
4,954,634,826
IssuesEvent
2016-12-01 18:09:09
orcidee/rpgconmanager
https://api.github.com/repos/orcidee/rpgconmanager
closed
Griser le nb de tables
bug Priority : 1
Il faudrait griser/rendre visiblement non-éditable le champs nombre de tables quand la partie a été vérifiée et validée.
1.0
Griser le nb de tables - Il faudrait griser/rendre visiblement non-éditable le champs nombre de tables quand la partie a été vérifiée et validée.
non_process
griser le nb de tables il faudrait griser rendre visiblement non éditable le champs nombre de tables quand la partie a été vérifiée et validée
0
5,761
8,598,978,418
IssuesEvent
2018-11-15 23:47:47
w3c/w3process
https://api.github.com/repos/w3c/w3process
closed
Does a spec need to be Rec-ready to be a CR? Align defn with process
Agenda+ PendingReview Process2019Candidate
https://www.w3.org/2018/Process-20180201/#candidate-rec sets out requirements for advancement to CR. It does not require any statement of reasons why a CR may [update: instating omitted text from original issue filing, with apologies] _not_ advance to Rec. [/update] However https://www.w3.org/2018/Process-20180201/#RecsCR defines what a CR is and adds informatively: > Note: Candidate Recommendations are expected to be acceptable as Recommendations. Announcement of a different next step should include the reasons why the change in expectations comes at so late a stage. This (first part) is not tested in the requirements for advancing to CR, and in fact it is not unusual for substantive changes to be made after first CR, on the path to Rec. The second part is very reasonable, however it omits an important real world case, which is non-advancement for an unreasonably long period. As a minimum, the two sections should be aligned so that the tests for advancement clearly verify the requirements. My preference for achieving the alignment would be to remove or modify the expectation from the note, since I don't think it reflects reality or is particularly helpful. It is obvious enough that a CR is a step along the path to Rec and that there are extra steps needed if making substantive changes between CR and Rec. Conversely, if the substantive content of a CR is acceptable as a Rec, there are good reasons why we do not just publish it as a Rec: the "acceptability" depends not on the document in itself, but rather on the demonstration that implementers and the AC and Director agree that it is acceptable. So the statement of expectation doesn't seem to be clear enough. I would propose: > Note: Candidate Recommendations are expected to be advanced to Recommendation. Announcement of a different next step should include the reasons why the change in expectations comes at so late a stage. Handling the second problem, omission of what should happen if a document stays in CR for an unreasonably long time, is harder to fix and probably needs some discussion. Automatically revert the CR to a Note? I imagine that would be unacceptable to some people, but it might focus attention on tests and implementation reports... If it is going to be picked up for discussion, probably better for someone to raise a new issue for tracking purposes.
1.0
Does a spec need to be Rec-ready to be a CR? Align defn with process - https://www.w3.org/2018/Process-20180201/#candidate-rec sets out requirements for advancement to CR. It does not require any statement of reasons why a CR may [update: instating omitted text from original issue filing, with apologies] _not_ advance to Rec. [/update] However https://www.w3.org/2018/Process-20180201/#RecsCR defines what a CR is and adds informatively: > Note: Candidate Recommendations are expected to be acceptable as Recommendations. Announcement of a different next step should include the reasons why the change in expectations comes at so late a stage. This (first part) is not tested in the requirements for advancing to CR, and in fact it is not unusual for substantive changes to be made after first CR, on the path to Rec. The second part is very reasonable, however it omits an important real world case, which is non-advancement for an unreasonably long period. As a minimum, the two sections should be aligned so that the tests for advancement clearly verify the requirements. My preference for achieving the alignment would be to remove or modify the expectation from the note, since I don't think it reflects reality or is particularly helpful. It is obvious enough that a CR is a step along the path to Rec and that there are extra steps needed if making substantive changes between CR and Rec. Conversely, if the substantive content of a CR is acceptable as a Rec, there are good reasons why we do not just publish it as a Rec: the "acceptability" depends not on the document in itself, but rather on the demonstration that implementers and the AC and Director agree that it is acceptable. So the statement of expectation doesn't seem to be clear enough. I would propose: > Note: Candidate Recommendations are expected to be advanced to Recommendation. Announcement of a different next step should include the reasons why the change in expectations comes at so late a stage. Handling the second problem, omission of what should happen if a document stays in CR for an unreasonably long time, is harder to fix and probably needs some discussion. Automatically revert the CR to a Note? I imagine that would be unacceptable to some people, but it might focus attention on tests and implementation reports... If it is going to be picked up for discussion, probably better for someone to raise a new issue for tracking purposes.
process
does a spec need to be rec ready to be a cr align defn with process sets out requirements for advancement to cr it does not require any statement of reasons why a cr may not advance to rec however defines what a cr is and adds informatively note candidate recommendations are expected to be acceptable as recommendations announcement of a different next step should include the reasons why the change in expectations comes at so late a stage this first part is not tested in the requirements for advancing to cr and in fact it is not unusual for substantive changes to be made after first cr on the path to rec the second part is very reasonable however it omits an important real world case which is non advancement for an unreasonably long period as a minimum the two sections should be aligned so that the tests for advancement clearly verify the requirements my preference for achieving the alignment would be to remove or modify the expectation from the note since i don t think it reflects reality or is particularly helpful it is obvious enough that a cr is a step along the path to rec and that there are extra steps needed if making substantive changes between cr and rec conversely if the substantive content of a cr is acceptable as a rec there are good reasons why we do not just publish it as a rec the acceptability depends not on the document in itself but rather on the demonstration that implementers and the ac and director agree that it is acceptable so the statement of expectation doesn t seem to be clear enough i would propose note candidate recommendations are expected to be advanced to recommendation announcement of a different next step should include the reasons why the change in expectations comes at so late a stage handling the second problem omission of what should happen if a document stays in cr for an unreasonably long time is harder to fix and probably needs some discussion automatically revert the cr to a note i imagine that would be unacceptable to some people but it might focus attention on tests and implementation reports if it is going to be picked up for discussion probably better for someone to raise a new issue for tracking purposes
1
17,230
22,917,655,768
IssuesEvent
2022-07-17 07:34:08
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[SQL Connector] create default database when PulsarCatalog is created if default database is not present.
compute/data-processing
Currently, the default database might not be present if users didn't create the database explicitly. We should create the database upon PulsarCatalog creation
1.0
[SQL Connector] create default database when PulsarCatalog is created if default database is not present. - Currently, the default database might not be present if users didn't create the database explicitly. We should create the database upon PulsarCatalog creation
process
create default database when pulsarcatalog is created if default database is not present currently the default database might not be present if users didn t create the database explicitly we should create the database upon pulsarcatalog creation
1
2,208
5,049,113,886
IssuesEvent
2016-12-20 15:02:43
CERNDocumentServer/cds
https://api.github.com/repos/CERNDocumentServer/cds
closed
webhooks: cancel tasks
avc_processing review
When canceling an event that **is still running** we need to stop all the celery tasks associated with it and also delete any file (on disk and/or db) that was created. - Downloader: - [x] Stop download task and delete the `ObjectVersion` and the file on disk - AVCWorflow: - [x] If the master file was (is being) downloaded by the event, stop the task if still running and delete the `ObjectVersion` and the file on disk - [x] Delete any extracted metadata that might have being added to the deposit and also any `ObjectVersionTag` that might have being created (specially if the file was upload via http) - [x] Stop all the Sorenson jobs - [x] Delete any `ObjectVersion` that the transcode task might have created, if any of the jobs have finished we also need to delete the file from disk
1.0
webhooks: cancel tasks - When canceling an event that **is still running** we need to stop all the celery tasks associated with it and also delete any file (on disk and/or db) that was created. - Downloader: - [x] Stop download task and delete the `ObjectVersion` and the file on disk - AVCWorflow: - [x] If the master file was (is being) downloaded by the event, stop the task if still running and delete the `ObjectVersion` and the file on disk - [x] Delete any extracted metadata that might have being added to the deposit and also any `ObjectVersionTag` that might have being created (specially if the file was upload via http) - [x] Stop all the Sorenson jobs - [x] Delete any `ObjectVersion` that the transcode task might have created, if any of the jobs have finished we also need to delete the file from disk
process
webhooks cancel tasks when canceling an event that is still running we need to stop all the celery tasks associated with it and also delete any file on disk and or db that was created downloader stop download task and delete the objectversion and the file on disk avcworflow if the master file was is being downloaded by the event stop the task if still running and delete the objectversion and the file on disk delete any extracted metadata that might have being added to the deposit and also any objectversiontag that might have being created specially if the file was upload via http stop all the sorenson jobs delete any objectversion that the transcode task might have created if any of the jobs have finished we also need to delete the file from disk
1
142,301
19,089,396,360
IssuesEvent
2021-11-29 10:22:03
tharun453/samples
https://api.github.com/repos/tharun453/samples
opened
CVE-2021-23358 (High) detected in underscore-1.7.0.tgz
security vulnerability
## CVE-2021-23358 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore-1.7.0.tgz</b></p></summary> <p>JavaScript's functional programming helper library.</p> <p>Library home page: <a href="https://registry.npmjs.org/underscore/-/underscore-1.7.0.tgz">https://registry.npmjs.org/underscore/-/underscore-1.7.0.tgz</a></p> <p>Path to dependency file: samples/core/tutorials/buggyamb/BuggyAmb/wwwroot/scripts/jquery-ui-1.12.1/package.json</p> <p>Path to vulnerable library: samples/core/tutorials/buggyamb/BuggyAmb/wwwroot/scripts/jquery-ui-1.12.1/node_modules/underscore/package.json</p> <p> Dependency Hierarchy: - grunt-0.4.5.tgz (Root Library) - js-yaml-2.0.5.tgz - argparse-0.1.16.tgz - :x: **underscore-1.7.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tharun453/samples/commit/0d7f1931b9759c22f0469b959114a5d94f8f92e4">0d7f1931b9759c22f0469b959114a5d94f8f92e4</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package underscore from 1.13.0-0 and before 1.13.0-2, from 1.3.2 and before 1.12.1 are vulnerable to Arbitrary Code Injection via the template function, particularly when a variable property is passed as an argument as it is not sanitized. <p>Publish Date: 2021-03-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23358>CVE-2021-23358</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358</a></p> <p>Release Date: 2021-03-29</p> <p>Fix Resolution: underscore - 1.12.1,1.13.0-2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23358 (High) detected in underscore-1.7.0.tgz - ## CVE-2021-23358 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore-1.7.0.tgz</b></p></summary> <p>JavaScript's functional programming helper library.</p> <p>Library home page: <a href="https://registry.npmjs.org/underscore/-/underscore-1.7.0.tgz">https://registry.npmjs.org/underscore/-/underscore-1.7.0.tgz</a></p> <p>Path to dependency file: samples/core/tutorials/buggyamb/BuggyAmb/wwwroot/scripts/jquery-ui-1.12.1/package.json</p> <p>Path to vulnerable library: samples/core/tutorials/buggyamb/BuggyAmb/wwwroot/scripts/jquery-ui-1.12.1/node_modules/underscore/package.json</p> <p> Dependency Hierarchy: - grunt-0.4.5.tgz (Root Library) - js-yaml-2.0.5.tgz - argparse-0.1.16.tgz - :x: **underscore-1.7.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tharun453/samples/commit/0d7f1931b9759c22f0469b959114a5d94f8f92e4">0d7f1931b9759c22f0469b959114a5d94f8f92e4</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package underscore from 1.13.0-0 and before 1.13.0-2, from 1.3.2 and before 1.12.1 are vulnerable to Arbitrary Code Injection via the template function, particularly when a variable property is passed as an argument as it is not sanitized. <p>Publish Date: 2021-03-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23358>CVE-2021-23358</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358</a></p> <p>Release Date: 2021-03-29</p> <p>Fix Resolution: underscore - 1.12.1,1.13.0-2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in underscore tgz cve high severity vulnerability vulnerable library underscore tgz javascript s functional programming helper library library home page a href path to dependency file samples core tutorials buggyamb buggyamb wwwroot scripts jquery ui package json path to vulnerable library samples core tutorials buggyamb buggyamb wwwroot scripts jquery ui node modules underscore package json dependency hierarchy grunt tgz root library js yaml tgz argparse tgz x underscore tgz vulnerable library found in head commit a href found in base branch main vulnerability details the package underscore from and before from and before are vulnerable to arbitrary code injection via the template function particularly when a variable property is passed as an argument as it is not sanitized publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution underscore step up your open source security game with whitesource
0
3,903
6,823,007,495
IssuesEvent
2017-11-07 22:06:15
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
whenBlock search by date seems broken
status-inprocess tools-whenBlock type-bug
When I run this (following the command help): ./whenBlock 2017-03-02 We obtain this strange error. Notice that I am using the date returned for block 0. I could not make it work with any date, most of the times it returns this error ------------------------------------------------------------------------------------------------------------------- whenBlock argc: 2 [1:2017-03-02] whenBlock 2017-03-02 Could not open the mini-block database: /home/carlos/.quickBlocks/cache/miniBlocks.bin. Usage: whenBlock [-a|-l|-v|-h] date / block Purpose: Finds the nearest block before the JSON-formatted date, or the nearest date before the given block. Where: date / block one of the special values listed below or YYYY-MM-DD-[HH[:MM[:SS]]] or a blockNumber (required) -a (--alone) show the found block or found date unadorned (useful for scripting) -l (--list) list special blocks timestamps and dates -v (--verbose) set verbose level. Either -v, --verbose or -v:n where 'n' is level -h (--help) display this help screen Powered by QuickBlocks
1.0
whenBlock search by date seems broken - When I run this (following the command help): ./whenBlock 2017-03-02 We obtain this strange error. Notice that I am using the date returned for block 0. I could not make it work with any date, most of the times it returns this error ------------------------------------------------------------------------------------------------------------------- whenBlock argc: 2 [1:2017-03-02] whenBlock 2017-03-02 Could not open the mini-block database: /home/carlos/.quickBlocks/cache/miniBlocks.bin. Usage: whenBlock [-a|-l|-v|-h] date / block Purpose: Finds the nearest block before the JSON-formatted date, or the nearest date before the given block. Where: date / block one of the special values listed below or YYYY-MM-DD-[HH[:MM[:SS]]] or a blockNumber (required) -a (--alone) show the found block or found date unadorned (useful for scripting) -l (--list) list special blocks timestamps and dates -v (--verbose) set verbose level. Either -v, --verbose or -v:n where 'n' is level -h (--help) display this help screen Powered by QuickBlocks
process
whenblock search by date seems broken when i run this following the command help whenblock we obtain this strange error notice that i am using the date returned for block i could not make it work with any date most of the times it returns this error whenblock argc whenblock could not open the mini block database home carlos quickblocks cache miniblocks bin usage whenblock date block purpose finds the nearest block before the json formatted date or the nearest date before the given block where date block one of the special values listed below or yyyy mm dd or a blocknumber required a alone show the found block or found date unadorned useful for scripting l list list special blocks timestamps and dates v verbose set verbose level either v verbose or v n where n is level h help display this help screen powered by quickblocks
1
273,128
20,772,510,979
IssuesEvent
2022-03-16 07:05:01
DLR-RM/stable-baselines3
https://api.github.com/repos/DLR-RM/stable-baselines3
closed
The code problem of Custom Feature Extractor
documentation
**Important Note: We do not do technical support, nor consulting** and don't answer personal questions per email. Please post your question on the [RL Discord](https://discord.com/invite/xhfNqQv), [Reddit](https://www.reddit.com/r/reinforcementlearning/) or [Stack Overflow](https://stackoverflow.com/) in that case. ### 📚 Documentation ####problem In [this code](https://github.com/DLR-RM/stable-baselines3/blob/master/docs/guide/custom_policy.rst#custom-feature-extractor) 1:what's the function of PART"with torch.no_grad...." ? why use torch.no_grad to compute "n_flatten" 2: what's the meaning of n_flatten ? ### Checklist - [x] I have read the [documentation](https://stable-baselines3.readthedocs.io/en/master/) (**required**) - [x] I have checked that there is no similar [issue](https://github.com/DLR-RM/stable-baselines3/issues) in the repo (**required**) ### <!--- This Template is an edited version of the one from https://github.com/pytorch/pytorch -->
1.0
The code problem of Custom Feature Extractor - **Important Note: We do not do technical support, nor consulting** and don't answer personal questions per email. Please post your question on the [RL Discord](https://discord.com/invite/xhfNqQv), [Reddit](https://www.reddit.com/r/reinforcementlearning/) or [Stack Overflow](https://stackoverflow.com/) in that case. ### 📚 Documentation ####problem In [this code](https://github.com/DLR-RM/stable-baselines3/blob/master/docs/guide/custom_policy.rst#custom-feature-extractor) 1:what's the function of PART"with torch.no_grad...." ? why use torch.no_grad to compute "n_flatten" 2: what's the meaning of n_flatten ? ### Checklist - [x] I have read the [documentation](https://stable-baselines3.readthedocs.io/en/master/) (**required**) - [x] I have checked that there is no similar [issue](https://github.com/DLR-RM/stable-baselines3/issues) in the repo (**required**) ### <!--- This Template is an edited version of the one from https://github.com/pytorch/pytorch -->
non_process
the code problem of custom feature extractor important note we do not do technical support nor consulting and don t answer personal questions per email please post your question on the or in that case 📚 documentation problem in what s the function of part with torch no grad why use torch no grad to compute n flatten what s the meaning of n flatten  checklist i have read the required i have checked that there is no similar in the repo required
0
5,264
8,057,651,126
IssuesEvent
2018-08-02 15:57:34
GoogleCloudPlatform/google-cloud-python
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
closed
BigQuery: 'Client.update_dataset' systest error, 412 PreconditionFailed
api: bigquery backend flaky testing type: process
First error in [this CI run](https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/7126): ```python _______________________ TestBigQuery.test_update_dataset _______________________ self = <tests.system.TestBigQuery testMethod=test_update_dataset> def test_update_dataset(self): dataset = self.temp_dataset(_make_dataset_id('update_dataset')) self.assertTrue(_dataset_exists(dataset)) self.assertIsNone(dataset.friendly_name) self.assertIsNone(dataset.description) self.assertEquals(dataset.labels, {}) dataset.friendly_name = 'Friendly' dataset.description = 'Description' dataset.labels = {'priority': 'high', 'color': 'blue'} ds2 = Config.CLIENT.update_dataset( dataset, > ('friendly_name', 'description', 'labels')) tests/system.py:189: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/bigquery/client.py:387: in update_dataset headers=headers) ... if not 200 <= response.status_code < 300: > raise exceptions.from_http_response(response) E PreconditionFailed: 412 PATCH https://www.googleapis.com/bigquery/v2/projects/precise-truck-742/datasets/update_dataset_7126_1531263324: Precondition Failed ../.nox/sys-2-7/lib/python2.7/site-packages/google/cloud/_http.py:293: PreconditionFailed ```
1.0
BigQuery: 'Client.update_dataset' systest error, 412 PreconditionFailed - First error in [this CI run](https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/7126): ```python _______________________ TestBigQuery.test_update_dataset _______________________ self = <tests.system.TestBigQuery testMethod=test_update_dataset> def test_update_dataset(self): dataset = self.temp_dataset(_make_dataset_id('update_dataset')) self.assertTrue(_dataset_exists(dataset)) self.assertIsNone(dataset.friendly_name) self.assertIsNone(dataset.description) self.assertEquals(dataset.labels, {}) dataset.friendly_name = 'Friendly' dataset.description = 'Description' dataset.labels = {'priority': 'high', 'color': 'blue'} ds2 = Config.CLIENT.update_dataset( dataset, > ('friendly_name', 'description', 'labels')) tests/system.py:189: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/bigquery/client.py:387: in update_dataset headers=headers) ... if not 200 <= response.status_code < 300: > raise exceptions.from_http_response(response) E PreconditionFailed: 412 PATCH https://www.googleapis.com/bigquery/v2/projects/precise-truck-742/datasets/update_dataset_7126_1531263324: Precondition Failed ../.nox/sys-2-7/lib/python2.7/site-packages/google/cloud/_http.py:293: PreconditionFailed ```
process
bigquery client update dataset systest error preconditionfailed first error in python testbigquery test update dataset self def test update dataset self dataset self temp dataset make dataset id update dataset self asserttrue dataset exists dataset self assertisnone dataset friendly name self assertisnone dataset description self assertequals dataset labels dataset friendly name friendly dataset description description dataset labels priority high color blue config client update dataset dataset friendly name description labels tests system py google cloud bigquery client py in update dataset headers headers if not response status code raise exceptions from http response response e preconditionfailed patch precondition failed nox sys lib site packages google cloud http py preconditionfailed
1
14,793
18,066,914,025
IssuesEvent
2021-09-20 20:21:34
Jeffail/benthos
https://api.github.com/repos/Jeffail/benthos
closed
Support dynamic queries in sql processors
enhancement processors effort: lower
Hello, could you support dynamic queries also for SQL processors like [AWS dynamo db] (https://www.benthos.dev/docs/components/processors/aws_dynamodb_partiql/) does? Otherwise, the user has to write multiple configs for different queries.
1.0
Support dynamic queries in sql processors - Hello, could you support dynamic queries also for SQL processors like [AWS dynamo db] (https://www.benthos.dev/docs/components/processors/aws_dynamodb_partiql/) does? Otherwise, the user has to write multiple configs for different queries.
process
support dynamic queries in sql processors hello could you support dynamic queries also for sql processors like does otherwise the user has to write multiple configs for different queries
1
4,395
7,286,431,786
IssuesEvent
2018-02-23 09:43:36
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[needs-docs]Rename Node Editor and Node Tool (fixes #17807)
Automatic new feature Easy Processing Screenshots
Original commit: https://github.com/qgis/QGIS/commit/5461d3c52ace8f7afe72b9215a0b0532574294e2 by DelazJ Unfortunately this naughty coder did not write a description... :-(
1.0
[needs-docs]Rename Node Editor and Node Tool (fixes #17807) - Original commit: https://github.com/qgis/QGIS/commit/5461d3c52ace8f7afe72b9215a0b0532574294e2 by DelazJ Unfortunately this naughty coder did not write a description... :-(
process
rename node editor and node tool fixes original commit by delazj unfortunately this naughty coder did not write a description
1
204,336
7,086,994,848
IssuesEvent
2018-01-11 16:25:31
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Do stuff with cloud provider's DNS
lifecycle/stale priority/backlog sig/network
In #9682, @pnovotnak wrote: _I'd like to create resources that are accessible within my private network at my cloud provider (though outside my Kubernetes cluster), but not to the outside internet, that are resolvable via DNS within the cluster. Ironically it seems like endpoint IPs fit the bill (although they change and are not resolvable) whereas service IPs, which are stable and what, are not accessible from outside the cluster._ to which @thockin responded: _Reaching up into the cloud-provider's DNS is in the cards for sometime after 1.0_
1.0
Do stuff with cloud provider's DNS - In #9682, @pnovotnak wrote: _I'd like to create resources that are accessible within my private network at my cloud provider (though outside my Kubernetes cluster), but not to the outside internet, that are resolvable via DNS within the cluster. Ironically it seems like endpoint IPs fit the bill (although they change and are not resolvable) whereas service IPs, which are stable and what, are not accessible from outside the cluster._ to which @thockin responded: _Reaching up into the cloud-provider's DNS is in the cards for sometime after 1.0_
non_process
do stuff with cloud provider s dns in pnovotnak wrote i d like to create resources that are accessible within my private network at my cloud provider though outside my kubernetes cluster but not to the outside internet that are resolvable via dns within the cluster ironically it seems like endpoint ips fit the bill although they change and are not resolvable whereas service ips which are stable and what are not accessible from outside the cluster to which thockin responded reaching up into the cloud provider s dns is in the cards for sometime after
0
11,169
13,957,694,554
IssuesEvent
2020-10-24 08:11:17
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
SE: Harvesting Request
Geoportal Harvesting process SE - Sweden
Hi! After some updates by our contributers we would neeed a new harvesting of the Swedish node. Have a great weekend! Bj&ouml;rn Olofsson, The Swedish Geoportal
1.0
SE: Harvesting Request - Hi! After some updates by our contributers we would neeed a new harvesting of the Swedish node. Have a great weekend! Bj&ouml;rn Olofsson, The Swedish Geoportal
process
se harvesting request hi after some updates by our contributers we would neeed a new harvesting of the swedish node have a great weekend bj ouml rn olofsson the swedish geoportal
1
11,888
14,681,300,946
IssuesEvent
2020-12-31 12:49:28
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
Studies tab > Loader icon is missing at the bottom of the page
Bug P1 Participant manager Process: Dev Process: Fixed Process: Tested QA
AR : Studies tab > Loader icon is missing at the bottom of the page ER : Loader icon should be present (Note : Only 10 sets of data should load at a time) ![apps tab](https://user-images.githubusercontent.com/71445210/98634091-b66ac480-2348-11eb-97e0-1c76b3c8a25b.png)
3.0
Studies tab > Loader icon is missing at the bottom of the page - AR : Studies tab > Loader icon is missing at the bottom of the page ER : Loader icon should be present (Note : Only 10 sets of data should load at a time) ![apps tab](https://user-images.githubusercontent.com/71445210/98634091-b66ac480-2348-11eb-97e0-1c76b3c8a25b.png)
process
studies tab loader icon is missing at the bottom of the page ar studies tab loader icon is missing at the bottom of the page er loader icon should be present note only sets of data should load at a time
1
821,601
30,828,067,633
IssuesEvent
2023-08-01 21:57:45
dotCMS/core
https://api.github.com/repos/dotCMS/core
closed
Design System: Implement `button` component
Team : Lunik Type : New Functionality Triage OKR : Core Features Priority : 2 High
### Parent Issue #25355 ### User Story As a stakeholder, I want you to implement the button component new design for primeng and dojo widgets. ### Acceptance Criteria 1. Match the design provided with all the types and severities in primeng 2. Add Storybook stories for the buttons with the same structure of primeng 3. Implement the buttons in dojo 4. User the CSS variables accordingly ### Proposed Objective Core Features ### Proposed Priority Priority 3 - Average ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. [Figma](https://www.figma.com/file/7CzbMcGSbTIJjerWtcapAm/Design-System-2023?type=design&node-id=42%3A689&mode=dev) ### Assumptions & Initiation Needs [_No response_](https://github.com/dotCMS/core/assets/751424/52512905-2ff6-46c3-8244-cb472ab1a83e) ### Quality Assurance Notes & Workarounds _No response_ ### Sub-Tasks & Estimates _No response_
1.0
Design System: Implement `button` component - ### Parent Issue #25355 ### User Story As a stakeholder, I want you to implement the button component new design for primeng and dojo widgets. ### Acceptance Criteria 1. Match the design provided with all the types and severities in primeng 2. Add Storybook stories for the buttons with the same structure of primeng 3. Implement the buttons in dojo 4. User the CSS variables accordingly ### Proposed Objective Core Features ### Proposed Priority Priority 3 - Average ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. [Figma](https://www.figma.com/file/7CzbMcGSbTIJjerWtcapAm/Design-System-2023?type=design&node-id=42%3A689&mode=dev) ### Assumptions & Initiation Needs [_No response_](https://github.com/dotCMS/core/assets/751424/52512905-2ff6-46c3-8244-cb472ab1a83e) ### Quality Assurance Notes & Workarounds _No response_ ### Sub-Tasks & Estimates _No response_
non_process
design system implement button component parent issue user story as a stakeholder i want you to implement the button component new design for primeng and dojo widgets acceptance criteria match the design provided with all the types and severities in primeng add storybook stories for the buttons with the same structure of primeng implement the buttons in dojo user the css variables accordingly proposed objective core features proposed priority priority average external links slack conversations support tickets figma designs etc assumptions initiation needs quality assurance notes workarounds no response sub tasks estimates no response
0
12,057
14,231,685,221
IssuesEvent
2020-11-18 09:54:27
ValveSoftware/Proton
https://api.github.com/repos/ValveSoftware/Proton
reopened
428: Shibuya Scramble (648580)
Game compatibility - Unofficial
# Compatibility Report - Name of the game with compatibility issues: 428: Shibuya Scramble - Steam AppID of the game: 648580 ## System Information - GPU: R9 390 - Driver/LLVM version: Mesa 20.2.2-2 - Kernel version: 5.9.8.zen1-1 (identical behavior with 5.9.8.arch1-1) - Link to full system information report as [Gist](https://gist.github.com/): https://gist.github.com/hajnal-endot/8028dacae9ca79bac10d82aa842ccb93 - Proton version: 5.13-2 ## I confirm: - [yes ] that I haven't found an existing compatibility report for this game. - [yes ] that I have checked whether there are updates for my system available. <!-- Please add `PROTON_LOG=1 %command%` to the game's launch options and drag and drop the generated `$HOME/steam-$APPID.log` into this issue report --> [steam-648580.log](https://github.com/ValveSoftware/Proton/files/5557001/steam-648580.log) ## Symptoms <!-- What's the problem? --> Game attempts to open in fullscreen, black screen appears in fullscreen, game immediately crashes to desktop. ## Reproduction 1. Open game through Steam. 2. Observe issue. COMMENT: There's a good chance the answer is staring me in the face and I'm just really stupid. Others seem to have very little trouble with this game per ProtonDB, and this is a newly-configured system that I haven't ironed out all the issues with yet.
True
428: Shibuya Scramble (648580) - # Compatibility Report - Name of the game with compatibility issues: 428: Shibuya Scramble - Steam AppID of the game: 648580 ## System Information - GPU: R9 390 - Driver/LLVM version: Mesa 20.2.2-2 - Kernel version: 5.9.8.zen1-1 (identical behavior with 5.9.8.arch1-1) - Link to full system information report as [Gist](https://gist.github.com/): https://gist.github.com/hajnal-endot/8028dacae9ca79bac10d82aa842ccb93 - Proton version: 5.13-2 ## I confirm: - [yes ] that I haven't found an existing compatibility report for this game. - [yes ] that I have checked whether there are updates for my system available. <!-- Please add `PROTON_LOG=1 %command%` to the game's launch options and drag and drop the generated `$HOME/steam-$APPID.log` into this issue report --> [steam-648580.log](https://github.com/ValveSoftware/Proton/files/5557001/steam-648580.log) ## Symptoms <!-- What's the problem? --> Game attempts to open in fullscreen, black screen appears in fullscreen, game immediately crashes to desktop. ## Reproduction 1. Open game through Steam. 2. Observe issue. COMMENT: There's a good chance the answer is staring me in the face and I'm just really stupid. Others seem to have very little trouble with this game per ProtonDB, and this is a newly-configured system that I haven't ironed out all the issues with yet.
non_process
shibuya scramble compatibility report name of the game with compatibility issues shibuya scramble steam appid of the game system information gpu driver llvm version mesa kernel version identical behavior with link to full system information report as proton version i confirm that i haven t found an existing compatibility report for this game that i have checked whether there are updates for my system available please add proton log command to the game s launch options and drag and drop the generated home steam appid log into this issue report symptoms game attempts to open in fullscreen black screen appears in fullscreen game immediately crashes to desktop reproduction open game through steam observe issue comment there s a good chance the answer is staring me in the face and i m just really stupid others seem to have very little trouble with this game per protondb and this is a newly configured system that i haven t ironed out all the issues with yet
0
4,990
7,822,175,201
IssuesEvent
2018-06-14 00:49:16
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
Add a function to convert.m to changed a pressure profile into altitude (Trac #4)
Migrated from Trac enhancement fasching@uwm.edu post_processing
Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/4 ```json { "status": "closed", "changetime": "2009-05-16T10:07:24", "description": "Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.", "reporter": "fasching@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1242468444000000", "component": "post_processing", "summary": "Add a function to convert.m to changed a pressure profile into altitude", "priority": "minor", "keywords": "conversion, MATLAB", "time": "2009-05-01T21:20:08", "milestone": "", "owner": "fasching@uwm.edu", "type": "enhancement" } ```
1.0
Add a function to convert.m to changed a pressure profile into altitude (Trac #4) - Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/4 ```json { "status": "closed", "changetime": "2009-05-16T10:07:24", "description": "Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.", "reporter": "fasching@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1242468444000000", "component": "post_processing", "summary": "Add a function to convert.m to changed a pressure profile into altitude", "priority": "minor", "keywords": "conversion, MATLAB", "time": "2009-05-01T21:20:08", "milestone": "", "owner": "fasching@uwm.edu", "type": "enhancement" } ```
process
add a function to convert m to changed a pressure profile into altitude trac add a function to convert m to changed a pressure profile into altitude this would be useful for cases that do not specify things in terms of altitude attachments migrated from json status closed changetime description add a function to convert m to changed a pressure profile into altitude this would be useful for cases that do not specify things in terms of altitude reporter fasching uwm edu cc resolution verified by v larson ts component post processing summary add a function to convert m to changed a pressure profile into altitude priority minor keywords conversion matlab time milestone owner fasching uwm edu type enhancement
1
81,617
15,781,626,709
IssuesEvent
2021-04-01 11:39:45
GTNewHorizons/GT-New-Horizons-Modpack
https://api.github.com/repos/GTNewHorizons/GT-New-Horizons-Modpack
closed
(Enderio) Inventory Panel Sorting Issues
Status: stale Type: Need Code changes Type: suggestion
_Using the EnderIO the inventory panel has options for how to sort its contents. However unlike in other systems such as an ME system it does not keep the sorting preference once left and re-entered, which can be quite annoying. Bart recommended me to make this ticket as its likely not a terribly hard fix._
1.0
(Enderio) Inventory Panel Sorting Issues - _Using the EnderIO the inventory panel has options for how to sort its contents. However unlike in other systems such as an ME system it does not keep the sorting preference once left and re-entered, which can be quite annoying. Bart recommended me to make this ticket as its likely not a terribly hard fix._
non_process
enderio inventory panel sorting issues using the enderio the inventory panel has options for how to sort its contents however unlike in other systems such as an me system it does not keep the sorting preference once left and re entered which can be quite annoying bart recommended me to make this ticket as its likely not a terribly hard fix
0
9,493
12,486,892,589
IssuesEvent
2020-05-31 05:35:16
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
TIN Interpolation - output raster format and size issues
Bug Feedback Processing
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). --> **Describe the bug** 1) Produces an ArcInfo ASCII Grid, but defaults to tif extension. 2) The rows/cols and pixel sizes (x/y) are not respected - ex. with the default 0.10 size, the output raster reports: Origin 283796,5.04147e+06 Pixel Size 0.09977998600000000107,-0.09982621899999999382 3) Break and structure lines are not respected - see https://github.com/qgis/QGIS/issues/27048 **How to Reproduce** <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error --> **QGIS and OS versions** QGIS version 3.8.3-Zanzibar QGIS code revision 685d8b15d2 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 2.4.1 Running against GDAL/OGR 2.4.1 Compiled against GEOS 3.7.2-CAPI-1.11.0 Running against GEOS 3.7.2-CAPI-1.11.0 b55d2125 PostgreSQL Client Version 10.8 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 5.2.0 Running against PROJ Rel. 5.2.0, September 15th, 2018 OS Version Windows 10 (10.0) **Additional context** <!-- Add any other context about the problem here. -->
1.0
TIN Interpolation - output raster format and size issues - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). --> **Describe the bug** 1) Produces an ArcInfo ASCII Grid, but defaults to tif extension. 2) The rows/cols and pixel sizes (x/y) are not respected - ex. with the default 0.10 size, the output raster reports: Origin 283796,5.04147e+06 Pixel Size 0.09977998600000000107,-0.09982621899999999382 3) Break and structure lines are not respected - see https://github.com/qgis/QGIS/issues/27048 **How to Reproduce** <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error --> **QGIS and OS versions** QGIS version 3.8.3-Zanzibar QGIS code revision 685d8b15d2 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 2.4.1 Running against GDAL/OGR 2.4.1 Compiled against GEOS 3.7.2-CAPI-1.11.0 Running against GEOS 3.7.2-CAPI-1.11.0 b55d2125 PostgreSQL Client Version 10.8 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 5.2.0 Running against PROJ Rel. 5.2.0, September 15th, 2018 OS Version Windows 10 (10.0) **Additional context** <!-- Add any other context about the problem here. -->
process
tin interpolation output raster format and size issues bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue if the issue concerns a third party plugin then it cannot be fixed by the qgis team please raise your issue in the dedicated bug tracker for that specific plugin as listed in the plugin s description describe the bug produces an arcinfo ascii grid but defaults to tif extension the rows cols and pixel sizes x y are not respected ex with the default size the output raster reports origin pixel size break and structure lines are not respected see how to reproduce steps sample datasets and qgis project file to reproduce the behavior screencasts or screenshots welcome go to click on scroll down to see error qgis and os versions qgis version zanzibar qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi postgresql client version spatialite version qwt version version compiled against proj running against proj rel september os version windows additional context
1
21,692
30,190,141,149
IssuesEvent
2023-07-04 14:45:50
UnitTestBot/UTBotJava
https://api.github.com/repos/UnitTestBot/UTBotJava
opened
Rerun Spring integration tests after minimizations with full context reset
ctg-enhancement comp-instrumented-process comp-spring
**Description** Right now we don't fully reset Spring context between concrete executions when generating integration tests with fuzzer, because it's too time consuming (can take several seconds per reset). We do our best to reset relevant parts of context (e.g. reset relevant beans and rollback transactions), however that may still not be enough because, for example, database id generators are not rollbacked with the transaction. Partial reset of Spring context may lead to generation of unreproducible tests that rely on some code from earlier concrete executions being executed before them. **Action plan** To cope we that, it is suggested to rerun tests that are left after test case minimization with full context reset and use results obtained from these reruns.
1.0
Rerun Spring integration tests after minimizations with full context reset - **Description** Right now we don't fully reset Spring context between concrete executions when generating integration tests with fuzzer, because it's too time consuming (can take several seconds per reset). We do our best to reset relevant parts of context (e.g. reset relevant beans and rollback transactions), however that may still not be enough because, for example, database id generators are not rollbacked with the transaction. Partial reset of Spring context may lead to generation of unreproducible tests that rely on some code from earlier concrete executions being executed before them. **Action plan** To cope we that, it is suggested to rerun tests that are left after test case minimization with full context reset and use results obtained from these reruns.
process
rerun spring integration tests after minimizations with full context reset description right now we don t fully reset spring context between concrete executions when generating integration tests with fuzzer because it s too time consuming can take several seconds per reset we do our best to reset relevant parts of context e g reset relevant beans and rollback transactions however that may still not be enough because for example database id generators are not rollbacked with the transaction partial reset of spring context may lead to generation of unreproducible tests that rely on some code from earlier concrete executions being executed before them action plan to cope we that it is suggested to rerun tests that are left after test case minimization with full context reset and use results obtained from these reruns
1
115,889
11,889,727,323
IssuesEvent
2020-03-28 15:10:11
samiha-rahman/soen390
https://api.github.com/repos/samiha-rahman/soen390
opened
US-59: Update Defect Tracking and Report Section for sprint 4
developer story documentation epic 10
As a developer, I need to update the defect tracking and report
1.0
US-59: Update Defect Tracking and Report Section for sprint 4 - As a developer, I need to update the defect tracking and report
non_process
us update defect tracking and report section for sprint as a developer i need to update the defect tracking and report
0
143,953
5,533,275,796
IssuesEvent
2017-03-21 12:57:39
sbpp/sourcebans-pp
https://api.github.com/repos/sbpp/sourcebans-pp
closed
Unable to login via steam
Priority: Medium Status: Abandoned Status: Completed Status: Review Needed Type: Bug
I have installed php-curl module and tried to update `steamopenid.php` and `includes/openid.php` to the latest version from github. All that happens is I get taken to steam as I am supposed to, but when I am returned to sourcebans I am not logged in. No errors on the page or in any log files I know of (php and nginx). - SourceBans++ 1.5.4.7 - PHP 7.0.8-0 - Ubuntu 16.04.1 LTS
1.0
Unable to login via steam - I have installed php-curl module and tried to update `steamopenid.php` and `includes/openid.php` to the latest version from github. All that happens is I get taken to steam as I am supposed to, but when I am returned to sourcebans I am not logged in. No errors on the page or in any log files I know of (php and nginx). - SourceBans++ 1.5.4.7 - PHP 7.0.8-0 - Ubuntu 16.04.1 LTS
non_process
unable to login via steam i have installed php curl module and tried to update steamopenid php and includes openid php to the latest version from github all that happens is i get taken to steam as i am supposed to but when i am returned to sourcebans i am not logged in no errors on the page or in any log files i know of php and nginx sourcebans php ubuntu lts
0
341,097
24,682,902,845
IssuesEvent
2022-10-18 23:38:18
apollographql/apollo-server
https://api.github.com/repos/apollographql/apollo-server
opened
Docs: reinstate serverless deployment guides
:memo: documentation
We should reintroduce (and update) our deployment guides for getting Apollo Server up and running in the popular serverless frameworks (lambda, azure, cloud, ...others?) @glasser proposed a one-size-fits-most approach using @vendia/serverless here which we should lean in to: https://github.com/apollo-server-integrations/apollo-server-integration-aws-lambda/issues/38 We can suggest more specific implementations secondarily or as an "advanced usage" (i.e. `@as-integrations/aws-lambda` for users who are concerned about bundle size). <!-- - Some features can be built as plugins. We encourage exploring the plugin API prior to opening a feature request: https://www.apollographql.com/docs/apollo-server/integrations/plugins/ In the event that the plugin API doesn't allow you to build a feature, it may be that expanding the plugin API *itself* is the best place for the feature to be introduced! Consider this flexible solution when opening a new feature request since it also unlocks new opportunities. - Prior to opening a feature request, please search for existing requests. If you find an existing feature that matches your needs, use the 👍 emote to show your support for it. If the specifics of your use case are not covered in the existing feature request but the idea seems similar enough, please take the time to *add new conversation* which helps the feature's design evolve. - If you do not find any other existing requests for the feature you desire, you should open a new feature request. Please take the time to help us understand your use-case as precisely as possible. Be sure to demonstrate that you've evaluated existing features and found them unsuitable and were unable to implement the functionality with the plugin API. Be flexible in your design and consider slight variations which might necessitate a specific API design. We also hope you'll be willing to engage in the on-going design discussion prior to opening a pull-request. -->
1.0
Docs: reinstate serverless deployment guides - We should reintroduce (and update) our deployment guides for getting Apollo Server up and running in the popular serverless frameworks (lambda, azure, cloud, ...others?) @glasser proposed a one-size-fits-most approach using @vendia/serverless here which we should lean in to: https://github.com/apollo-server-integrations/apollo-server-integration-aws-lambda/issues/38 We can suggest more specific implementations secondarily or as an "advanced usage" (i.e. `@as-integrations/aws-lambda` for users who are concerned about bundle size). <!-- - Some features can be built as plugins. We encourage exploring the plugin API prior to opening a feature request: https://www.apollographql.com/docs/apollo-server/integrations/plugins/ In the event that the plugin API doesn't allow you to build a feature, it may be that expanding the plugin API *itself* is the best place for the feature to be introduced! Consider this flexible solution when opening a new feature request since it also unlocks new opportunities. - Prior to opening a feature request, please search for existing requests. If you find an existing feature that matches your needs, use the 👍 emote to show your support for it. If the specifics of your use case are not covered in the existing feature request but the idea seems similar enough, please take the time to *add new conversation* which helps the feature's design evolve. - If you do not find any other existing requests for the feature you desire, you should open a new feature request. Please take the time to help us understand your use-case as precisely as possible. Be sure to demonstrate that you've evaluated existing features and found them unsuitable and were unable to implement the functionality with the plugin API. Be flexible in your design and consider slight variations which might necessitate a specific API design. We also hope you'll be willing to engage in the on-going design discussion prior to opening a pull-request. -->
non_process
docs reinstate serverless deployment guides we should reintroduce and update our deployment guides for getting apollo server up and running in the popular serverless frameworks lambda azure cloud others glasser proposed a one size fits most approach using vendia serverless here which we should lean in to we can suggest more specific implementations secondarily or as an advanced usage i e as integrations aws lambda for users who are concerned about bundle size some features can be built as plugins we encourage exploring the plugin api prior to opening a feature request in the event that the plugin api doesn t allow you to build a feature it may be that expanding the plugin api itself is the best place for the feature to be introduced consider this flexible solution when opening a new feature request since it also unlocks new opportunities prior to opening a feature request please search for existing requests if you find an existing feature that matches your needs use the 👍 emote to show your support for it if the specifics of your use case are not covered in the existing feature request but the idea seems similar enough please take the time to add new conversation which helps the feature s design evolve if you do not find any other existing requests for the feature you desire you should open a new feature request please take the time to help us understand your use case as precisely as possible be sure to demonstrate that you ve evaluated existing features and found them unsuitable and were unable to implement the functionality with the plugin api be flexible in your design and consider slight variations which might necessitate a specific api design we also hope you ll be willing to engage in the on going design discussion prior to opening a pull request
0