Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
39,075
| 5,037,388,063
|
IssuesEvent
|
2016-12-17 16:50:13
|
eloipuertas/ES2016F
|
https://api.github.com/repos/eloipuertas/ES2016F
|
closed
|
[A] Animation model saruman
|
Animations Design Team-A
|
3.21 As a graphic designer, I want a set of animations for Saruman, to give movement to the character. [Low priority]
|
1.0
|
[A] Animation model saruman - 3.21 As a graphic designer, I want a set of animations for Saruman, to give movement to the character. [Low priority]
|
non_process
|
animation model saruman as a graphic designer i want a set of animations for saruman to give movement to the character
| 0
|
4,320
| 7,214,021,250
|
IssuesEvent
|
2018-02-08 00:00:22
|
lvergergsk/BibGallery-FrontEnd
|
https://api.github.com/repos/lvergergsk/BibGallery-FrontEnd
|
closed
|
Using SQL to check free space for a user
|
Data Processing
|
## Oracle
* Possible way 1
```
SELECT
ts.tablespace_name,
TO_CHAR(SUM(NVL(fs.bytes,0))/1024/1024, '99,999,990.99') AS MB_FREE
FROM
user_free_space fs,
user_tablespaces ts,
user_users us
WHERE
fs.tablespace_name(+) = ts.tablespace_name
AND ts.tablespace_name(+) = us.default_tablespace
GROUP BY
ts.tablespace_name;
```
* Possible way 2
```
SELECT table_name as Table_Name, row_cnt as Row_Count, SUM(mb) as Size_MB
FROM
(SELECT in_tbl.table_name, to_number(extractvalue(xmltype(dbms_xmlgen.getxml('select count(*) c from ' ||ut.table_name)),'/ROWSET/ROW/C')) AS row_cnt , mb
FROM
(SELECT CASE WHEN lob_tables IS NULL THEN table_name WHEN lob_tables IS NOT NULL THEN lob_tables END AS table_name , mb
FROM (SELECT ul.table_name AS lob_tables, us.segment_name AS table_name , us.bytes/1024/1024 MB FROM user_segments us
LEFT JOIN user_lobs ul ON us.segment_name = ul.segment_name ) ) in_tbl INNER JOIN user_tables ut ON in_tbl.table_name = ut.table_name ) GROUP BY table_name, row_cnt ORDER BY 3 DESC;
```
* Possible way 3
```
SELECT SUM(bytes)
FROM user_segments
```
## MySQL
```
select table_schema, sum((data_length+index_length)/1024/1024) AS MB from information_schema.tables group by 1;
```
|
1.0
|
Using SQL to check free space for a user - ## Oracle
* Possible way 1
```
SELECT
ts.tablespace_name,
TO_CHAR(SUM(NVL(fs.bytes,0))/1024/1024, '99,999,990.99') AS MB_FREE
FROM
user_free_space fs,
user_tablespaces ts,
user_users us
WHERE
fs.tablespace_name(+) = ts.tablespace_name
AND ts.tablespace_name(+) = us.default_tablespace
GROUP BY
ts.tablespace_name;
```
* Possible way 2
```
SELECT table_name as Table_Name, row_cnt as Row_Count, SUM(mb) as Size_MB
FROM
(SELECT in_tbl.table_name, to_number(extractvalue(xmltype(dbms_xmlgen.getxml('select count(*) c from ' ||ut.table_name)),'/ROWSET/ROW/C')) AS row_cnt , mb
FROM
(SELECT CASE WHEN lob_tables IS NULL THEN table_name WHEN lob_tables IS NOT NULL THEN lob_tables END AS table_name , mb
FROM (SELECT ul.table_name AS lob_tables, us.segment_name AS table_name , us.bytes/1024/1024 MB FROM user_segments us
LEFT JOIN user_lobs ul ON us.segment_name = ul.segment_name ) ) in_tbl INNER JOIN user_tables ut ON in_tbl.table_name = ut.table_name ) GROUP BY table_name, row_cnt ORDER BY 3 DESC;
```
* Possible way 3
```
SELECT SUM(bytes)
FROM user_segments
```
## MySQL
```
select table_schema, sum((data_length+index_length)/1024/1024) AS MB from information_schema.tables group by 1;
```
|
process
|
using sql to check free space for a user oracle possible way select ts tablespace name to char sum nvl fs bytes as mb free from user free space fs user tablespaces ts user users us where fs tablespace name ts tablespace name and ts tablespace name us default tablespace group by ts tablespace name possible way select table name as table name row cnt as row count sum mb as size mb from select in tbl table name to number extractvalue xmltype dbms xmlgen getxml select count c from ut table name rowset row c as row cnt mb from select case when lob tables is null then table name when lob tables is not null then lob tables end as table name mb from select ul table name as lob tables us segment name as table name us bytes mb from user segments us left join user lobs ul on us segment name ul segment name in tbl inner join user tables ut on in tbl table name ut table name group by table name row cnt order by desc possible way select sum bytes from user segments mysql select table schema sum data length index length as mb from information schema tables group by
| 1
|
6,905
| 10,056,518,350
|
IssuesEvent
|
2019-07-22 09:20:31
|
CymChad/BaseRecyclerViewAdapterHelper
|
https://api.github.com/repos/CymChad/BaseRecyclerViewAdapterHelper
|
closed
|
树形列表删除bug
|
processing
|

在树形列表中,如果一级目录下没有子目录时,在调用BaseMultiItemQuickAdapter中remove(int position),删除,会导致程序崩溃。你使用你写的列子也是这样。


|
1.0
|
树形列表删除bug - 
在树形列表中,如果一级目录下没有子目录时,在调用BaseMultiItemQuickAdapter中remove(int position),删除,会导致程序崩溃。你使用你写的列子也是这样。


|
process
|
树形列表删除bug 在树形列表中,如果一级目录下没有子目录时,在调用basemultiitemquickadapter中remove int position 删除,会导致程序崩溃。你使用你写的列子也是这样。
| 1
|
262,022
| 8,249,166,046
|
IssuesEvent
|
2018-09-11 20:42:16
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
opened
|
[studio] Non-site members have access to site and full permissions
|
bug priority: high
|
### Expected behavior
Non-site members shouldn't have access to sites
### Actual behavior
Everyone has full access to everything (seems that all groups are being returned for a user).
### Steps to reproduce the problem
* Create a user and a site, but don't assign the user to any groups
* Login as that user
* See the site
### Log/stack trace (use https://gist.github.com)
### Specs
#### Version
```
Studio Version Number: 3.1.0-SNAPSHOT-0a0946
Build Number: 0a0946d689773516fda8075b4058b38e67f8920c
Build Date/Time: 09-11-2018 13:57:45 -0400
```
#### OS
Any
#### Browser
Any
|
1.0
|
[studio] Non-site members have access to site and full permissions - ### Expected behavior
Non-site members shouldn't have access to sites
### Actual behavior
Everyone has full access to everything (seems that all groups are being returned for a user).
### Steps to reproduce the problem
* Create a user and a site, but don't assign the user to any groups
* Login as that user
* See the site
### Log/stack trace (use https://gist.github.com)
### Specs
#### Version
```
Studio Version Number: 3.1.0-SNAPSHOT-0a0946
Build Number: 0a0946d689773516fda8075b4058b38e67f8920c
Build Date/Time: 09-11-2018 13:57:45 -0400
```
#### OS
Any
#### Browser
Any
|
non_process
|
non site members have access to site and full permissions expected behavior non site members shouldn t have access to sites actual behavior everyone has full access to everything seems that all groups are being returned for a user steps to reproduce the problem create a user and a site but don t assign the user to any groups login as that user see the site log stack trace use specs version studio version number snapshot build number build date time os any browser any
| 0
|
44,091
| 13,048,237,902
|
IssuesEvent
|
2020-07-29 12:11:54
|
jgeraigery/imhotep
|
https://api.github.com/repos/jgeraigery/imhotep
|
opened
|
WS-2018-0125 (Medium) detected in jackson-core-2.2.3.jar, jackson-core-2.6.7.jar
|
security vulnerability
|
## WS-2018-0125 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-core-2.2.3.jar</b>, <b>jackson-core-2.6.7.jar</b></p></summary>
<p>
<details><summary><b>jackson-core-2.2.3.jar</b></p></summary>
<p>Core Jackson abstractions, basic JSON streaming API implementation</p>
<p>Path to dependency file: /tmp/ws-scm/imhotep/imhotep-archive/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-core/2.2.3/jackson-core-2.2.3.jar</p>
<p>
Dependency Hierarchy:
- hadoop-client-2.6.0-cdh5.4.11.jar (Root Library)
- hadoop-aws-2.6.0-cdh5.4.11.jar
- jackson-databind-2.2.3.jar
- :x: **jackson-core-2.2.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-core-2.6.7.jar</b></p></summary>
<p>Core Jackson abstractions, basic JSON streaming API implementation</p>
<p>Library home page: <a href="https://github.com/FasterXML/jackson-core">https://github.com/FasterXML/jackson-core</a></p>
<p>Path to dependency file: /tmp/ws-scm/imhotep/imhotep-server/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-core/2.6.7/jackson-core-2.6.7.jar</p>
<p>
Dependency Hierarchy:
- jackson-databind-2.6.7.1.jar (Root Library)
- :x: **jackson-core-2.6.7.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/imhotep/commit/4432df39a5fc652b4512ad35a6db8f1a3776b771">4432df39a5fc652b4512ad35a6db8f1a3776b771</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
OutOfMemoryError when writing BigDecimal In Jackson Core before version 2.7.7.
When enabled the WRITE_BIGDECIMAL_AS_PLAIN setting, Jackson will attempt to write out the whole number, no matter how large the exponent.
<p>Publish Date: 2016-08-25
<p>URL: <a href=https://github.com/FasterXML/jackson-core/issues/315>WS-2018-0125</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-core/releases/tag/jackson-core-2.7.7">https://github.com/FasterXML/jackson-core/releases/tag/jackson-core-2.7.7</a></p>
<p>Release Date: 2016-08-25</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-core:2.7.7</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-core","packageVersion":"2.2.3","isTransitiveDependency":true,"dependencyTree":"org.apache.hadoop:hadoop-client:2.6.0-cdh5.4.11;org.apache.hadoop:hadoop-aws:2.6.0-cdh5.4.11;com.fasterxml.jackson.core:jackson-databind:2.2.3;com.fasterxml.jackson.core:jackson-core:2.2.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-core:2.7.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-core","packageVersion":"2.6.7","isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.7.1;com.fasterxml.jackson.core:jackson-core:2.6.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-core:2.7.7"}],"vulnerabilityIdentifier":"WS-2018-0125","vulnerabilityDetails":"OutOfMemoryError when writing BigDecimal In Jackson Core before version 2.7.7.\nWhen enabled the WRITE_BIGDECIMAL_AS_PLAIN setting, Jackson will attempt to write out the whole number, no matter how large the exponent.","vulnerabilityUrl":"https://github.com/FasterXML/jackson-core/issues/315","cvss2Severity":"medium","cvss2Score":"5.5","extraData":{}}</REMEDIATE> -->
|
True
|
WS-2018-0125 (Medium) detected in jackson-core-2.2.3.jar, jackson-core-2.6.7.jar - ## WS-2018-0125 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-core-2.2.3.jar</b>, <b>jackson-core-2.6.7.jar</b></p></summary>
<p>
<details><summary><b>jackson-core-2.2.3.jar</b></p></summary>
<p>Core Jackson abstractions, basic JSON streaming API implementation</p>
<p>Path to dependency file: /tmp/ws-scm/imhotep/imhotep-archive/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-core/2.2.3/jackson-core-2.2.3.jar</p>
<p>
Dependency Hierarchy:
- hadoop-client-2.6.0-cdh5.4.11.jar (Root Library)
- hadoop-aws-2.6.0-cdh5.4.11.jar
- jackson-databind-2.2.3.jar
- :x: **jackson-core-2.2.3.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-core-2.6.7.jar</b></p></summary>
<p>Core Jackson abstractions, basic JSON streaming API implementation</p>
<p>Library home page: <a href="https://github.com/FasterXML/jackson-core">https://github.com/FasterXML/jackson-core</a></p>
<p>Path to dependency file: /tmp/ws-scm/imhotep/imhotep-server/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-core/2.6.7/jackson-core-2.6.7.jar</p>
<p>
Dependency Hierarchy:
- jackson-databind-2.6.7.1.jar (Root Library)
- :x: **jackson-core-2.6.7.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/imhotep/commit/4432df39a5fc652b4512ad35a6db8f1a3776b771">4432df39a5fc652b4512ad35a6db8f1a3776b771</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
OutOfMemoryError when writing BigDecimal In Jackson Core before version 2.7.7.
When enabled the WRITE_BIGDECIMAL_AS_PLAIN setting, Jackson will attempt to write out the whole number, no matter how large the exponent.
<p>Publish Date: 2016-08-25
<p>URL: <a href=https://github.com/FasterXML/jackson-core/issues/315>WS-2018-0125</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-core/releases/tag/jackson-core-2.7.7">https://github.com/FasterXML/jackson-core/releases/tag/jackson-core-2.7.7</a></p>
<p>Release Date: 2016-08-25</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-core:2.7.7</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-core","packageVersion":"2.2.3","isTransitiveDependency":true,"dependencyTree":"org.apache.hadoop:hadoop-client:2.6.0-cdh5.4.11;org.apache.hadoop:hadoop-aws:2.6.0-cdh5.4.11;com.fasterxml.jackson.core:jackson-databind:2.2.3;com.fasterxml.jackson.core:jackson-core:2.2.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-core:2.7.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-core","packageVersion":"2.6.7","isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.7.1;com.fasterxml.jackson.core:jackson-core:2.6.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-core:2.7.7"}],"vulnerabilityIdentifier":"WS-2018-0125","vulnerabilityDetails":"OutOfMemoryError when writing BigDecimal In Jackson Core before version 2.7.7.\nWhen enabled the WRITE_BIGDECIMAL_AS_PLAIN setting, Jackson will attempt to write out the whole number, no matter how large the exponent.","vulnerabilityUrl":"https://github.com/FasterXML/jackson-core/issues/315","cvss2Severity":"medium","cvss2Score":"5.5","extraData":{}}</REMEDIATE> -->
|
non_process
|
ws medium detected in jackson core jar jackson core jar ws medium severity vulnerability vulnerable libraries jackson core jar jackson core jar jackson core jar core jackson abstractions basic json streaming api implementation path to dependency file tmp ws scm imhotep imhotep archive pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson core jackson core jar dependency hierarchy hadoop client jar root library hadoop aws jar jackson databind jar x jackson core jar vulnerable library jackson core jar core jackson abstractions basic json streaming api implementation library home page a href path to dependency file tmp ws scm imhotep imhotep server pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson core jackson core jar dependency hierarchy jackson databind jar root library x jackson core jar vulnerable library found in head commit a href vulnerability details outofmemoryerror when writing bigdecimal in jackson core before version when enabled the write bigdecimal as plain setting jackson will attempt to write out the whole number no matter how large the exponent publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson core isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails outofmemoryerror when writing bigdecimal in jackson core before version nwhen enabled the write bigdecimal as plain setting jackson will attempt to write out the whole number no matter how large the exponent vulnerabilityurl
| 0
|
11,872
| 14,673,243,470
|
IssuesEvent
|
2020-12-30 12:38:14
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[iOS] UI Issues in Overview screen
|
Bug P2 Process: Tested dev iOS
|
Steps:
1. Launch the app
2. Observe the below issues in Overview screen
Actual result
1. 'Sign in' text is displayed
2. Pagination button is missing above 'Get Started' button
Expected result
1. 'Sign In' text should be displayed
2. Pagination button should be displayed above 'Get Started' button

|
1.0
|
[iOS] UI Issues in Overview screen - Steps:
1. Launch the app
2. Observe the below issues in Overview screen
Actual result
1. 'Sign in' text is displayed
2. Pagination button is missing above 'Get Started' button
Expected result
1. 'Sign In' text should be displayed
2. Pagination button should be displayed above 'Get Started' button

|
process
|
ui issues in overview screen steps launch the app observe the below issues in overview screen actual result sign in text is displayed pagination button is missing above get started button expected result sign in text should be displayed pagination button should be displayed above get started button
| 1
|
678,661
| 23,205,980,180
|
IssuesEvent
|
2022-08-02 05:22:44
|
phetsims/axon
|
https://api.github.com/repos/phetsims/axon
|
closed
|
Can we get rid of getListenerCount?
|
priority:2-high dev:typescript
|
From https://github.com/phetsims/axon/issues/402, @marlitas and I would like to remove getListenerCount from the Emitter and Property interfaces. Current usages seem to only be in tests. Can we get rid of the tests? If not, perhaps subclass and make that method public?
|
1.0
|
Can we get rid of getListenerCount? - From https://github.com/phetsims/axon/issues/402, @marlitas and I would like to remove getListenerCount from the Emitter and Property interfaces. Current usages seem to only be in tests. Can we get rid of the tests? If not, perhaps subclass and make that method public?
|
non_process
|
can we get rid of getlistenercount from marlitas and i would like to remove getlistenercount from the emitter and property interfaces current usages seem to only be in tests can we get rid of the tests if not perhaps subclass and make that method public
| 0
|
172,550
| 27,297,208,425
|
IssuesEvent
|
2023-02-23 21:28:58
|
AlaskaAirlines/auro-nav
|
https://api.github.com/repos/AlaskaAirlines/auro-nav
|
closed
|
Left navigation design
|
auro-nav Type: Design
|
# Blueprint
[Documentation & research
](https://www.figma.com/file/e5SFMd5WwEB27iG2rcdPcU/Navigation?node-id=1%3A9088)
[Desktop design](https://www.figma.com/file/vgBapdyc1pqZGONvOOvmcJ/Left-Navigation?node-id=1908%3A8814&t=7L24WZtyFeK5ziZf-1)
[Mobile design](https://www.figma.com/file/vgBapdyc1pqZGONvOOvmcJ/Left-Navigation?node-id=1952%3A35184&t=7L24WZtyFeK5ziZf-1)
## Outline tasks
- [x] anatomy
- [x] color
- [x] typography
- [x] layout
- [x] spacing
- [x] animation/behavior
- [x] variants
- [x] states (hover, focus, active, focus-visible)
- [x] a11y
## Optional
- [x] Competitive analysis
- [x] research
- [x] site audit
- [x] usage audit
- [x] inspirational work
|
1.0
|
Left navigation design - # Blueprint
[Documentation & research
](https://www.figma.com/file/e5SFMd5WwEB27iG2rcdPcU/Navigation?node-id=1%3A9088)
[Desktop design](https://www.figma.com/file/vgBapdyc1pqZGONvOOvmcJ/Left-Navigation?node-id=1908%3A8814&t=7L24WZtyFeK5ziZf-1)
[Mobile design](https://www.figma.com/file/vgBapdyc1pqZGONvOOvmcJ/Left-Navigation?node-id=1952%3A35184&t=7L24WZtyFeK5ziZf-1)
## Outline tasks
- [x] anatomy
- [x] color
- [x] typography
- [x] layout
- [x] spacing
- [x] animation/behavior
- [x] variants
- [x] states (hover, focus, active, focus-visible)
- [x] a11y
## Optional
- [x] Competitive analysis
- [x] research
- [x] site audit
- [x] usage audit
- [x] inspirational work
|
non_process
|
left navigation design blueprint documentation research outline tasks anatomy color typography layout spacing animation behavior variants states hover focus active focus visible optional competitive analysis research site audit usage audit inspirational work
| 0
|
77,453
| 3,506,387,130
|
IssuesEvent
|
2016-01-08 06:22:01
|
OregonCore/OregonCore
|
https://api.github.com/repos/OregonCore/OregonCore
|
closed
|
A RAID of the 5 into the team, can enter the 40 game player (BB #498)
|
migrated Priority: Medium Type: Bug
|
This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 24.02.2014 12:42:32 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** invalid
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/498
<hr>
A RAID of the 5 into the team, can enter the 40 game player
I'm from china, and my english is not good.
|
1.0
|
A RAID of the 5 into the team, can enter the 40 game player (BB #498) - This issue was migrated from bitbucket.
**Original Reporter:**
**Original Date:** 24.02.2014 12:42:32 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** invalid
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/498
<hr>
A RAID of the 5 into the team, can enter the 40 game player
I'm from china, and my english is not good.
|
non_process
|
a raid of the into the team can enter the game player bb this issue was migrated from bitbucket original reporter original date gmt original priority major original type bug original state invalid direct link a raid of the into the team can enter the game player i m from china and my english is not good
| 0
|
51,966
| 7,739,946,667
|
IssuesEvent
|
2018-05-28 18:32:09
|
ProjectEvergreen/project-evergreen
|
https://api.github.com/repos/ProjectEvergreen/project-evergreen
|
closed
|
Create a basic Todo App example
|
documentation enhancement todo-mvc website
|
based [TodoMVC](http://todomvc.com/)
<img width="621" alt="screen shot 2018-05-22 at 3 40 20 pm" src="https://user-images.githubusercontent.com/895923/40385989-7bd92968-5dd6-11e8-9cf0-c18a6f4bed5f.png">
1. Basic workflows
1. Basic CRUD functionality
1. Some basic styles / examples of things like Web Components, CSS Grid, etc
|
1.0
|
Create a basic Todo App example - based [TodoMVC](http://todomvc.com/)
<img width="621" alt="screen shot 2018-05-22 at 3 40 20 pm" src="https://user-images.githubusercontent.com/895923/40385989-7bd92968-5dd6-11e8-9cf0-c18a6f4bed5f.png">
1. Basic workflows
1. Basic CRUD functionality
1. Some basic styles / examples of things like Web Components, CSS Grid, etc
|
non_process
|
create a basic todo app example based img width alt screen shot at pm src basic workflows basic crud functionality some basic styles examples of things like web components css grid etc
| 0
|
350,016
| 10,477,244,674
|
IssuesEvent
|
2019-09-23 20:27:36
|
wherebyus/general-tasks
|
https://api.github.com/repos/wherebyus/general-tasks
|
closed
|
events planner beta
|
Added After Sprint Planning Priority: High Product: Events Type: Bug UX: Validated
|
## Feature or problem
can't enter date-says invalid date
## UX Validation
Validated
### Suggested priority
High
### Stakeholders
*Submitted:* cristina349
### Definition of done
How will we know when this feature is complete?
### Subtasks
A detailed list of changes that need to be made or subtasks. One checkbox per.
- [ ] Brew the coffee
## Developer estimate
To help the team accurately estimate the complexity of this task,
take a moment to walk through this list and estimate each item. At the end, you can total
the estimates and round to the nearest prime number.
If any of these are at a `5` or higher, or if the total is above a `5`, consider breaking
this issue into multiple smaller issues.
- [ ] Changes to the database ()
- [ ] Changes to the API ()
- [ ] Testing Changes to the API ()
- [ ] Changes to Application Code ()
- [ ] Adding or updating unit tests ()
- [ ] Local developer testing ()
### Total developer estimate: 0
## Additional estimate
- [ ] Code review ()
- [ ] QA Testing ()
- [ ] Stakeholder Sign-off ()
- [ ] Deploy to Production ()
### Total additional estimate:
## QA Notes
Detailed instructions for testing, one checkbox per test to be completed.
### Contextual tests
- [ ] Accessibility check
- [ ] Cross-browser check (Edge, Chrome, Firefox)
- [ ] Responsive check
|
1.0
|
events planner beta - ## Feature or problem
can't enter date-says invalid date
## UX Validation
Validated
### Suggested priority
High
### Stakeholders
*Submitted:* cristina349
### Definition of done
How will we know when this feature is complete?
### Subtasks
A detailed list of changes that need to be made or subtasks. One checkbox per.
- [ ] Brew the coffee
## Developer estimate
To help the team accurately estimate the complexity of this task,
take a moment to walk through this list and estimate each item. At the end, you can total
the estimates and round to the nearest prime number.
If any of these are at a `5` or higher, or if the total is above a `5`, consider breaking
this issue into multiple smaller issues.
- [ ] Changes to the database ()
- [ ] Changes to the API ()
- [ ] Testing Changes to the API ()
- [ ] Changes to Application Code ()
- [ ] Adding or updating unit tests ()
- [ ] Local developer testing ()
### Total developer estimate: 0
## Additional estimate
- [ ] Code review ()
- [ ] QA Testing ()
- [ ] Stakeholder Sign-off ()
- [ ] Deploy to Production ()
### Total additional estimate:
## QA Notes
Detailed instructions for testing, one checkbox per test to be completed.
### Contextual tests
- [ ] Accessibility check
- [ ] Cross-browser check (Edge, Chrome, Firefox)
- [ ] Responsive check
|
non_process
|
events planner beta feature or problem can t enter date says invalid date ux validation validated suggested priority high stakeholders submitted definition of done how will we know when this feature is complete subtasks a detailed list of changes that need to be made or subtasks one checkbox per brew the coffee developer estimate to help the team accurately estimate the complexity of this task take a moment to walk through this list and estimate each item at the end you can total the estimates and round to the nearest prime number if any of these are at a or higher or if the total is above a consider breaking this issue into multiple smaller issues changes to the database changes to the api testing changes to the api changes to application code adding or updating unit tests local developer testing total developer estimate additional estimate code review qa testing stakeholder sign off deploy to production total additional estimate qa notes detailed instructions for testing one checkbox per test to be completed contextual tests accessibility check cross browser check edge chrome firefox responsive check
| 0
|
12,854
| 15,239,399,279
|
IssuesEvent
|
2021-02-19 04:22:27
|
topcoder-platform/community-app
|
https://api.github.com/repos/topcoder-platform/community-app
|
opened
|
Extract Member Skills History for past challenges
|
ShapeupProcess challenge- recommender-tool enhancement
|
Extract Member Skills History for past challenges.
|
1.0
|
Extract Member Skills History for past challenges - Extract Member Skills History for past challenges.
|
process
|
extract member skills history for past challenges extract member skills history for past challenges
| 1
|
665,014
| 22,296,061,695
|
IssuesEvent
|
2022-06-13 01:48:25
|
TencentBlueKing/bk-iam-saas
|
https://api.github.com/repos/TencentBlueKing/bk-iam-saas
|
closed
|
[RBAC] open api: 校验用户是否某个用户组的成员
|
Type: Enhancement Layer: Backend Priority: High Size: S backlog
|
支持批量用户组,批量数有限制
在目前的接口新增, 但是需要考虑, 大表查询校验关系`存在`
-----------------
需要重新梳理现在的接口列表, 如果蓝盾切换, 去掉依赖, 那么以前的部分接口可以下掉?
|
1.0
|
[RBAC] open api: 校验用户是否某个用户组的成员 - 支持批量用户组,批量数有限制
在目前的接口新增, 但是需要考虑, 大表查询校验关系`存在`
-----------------
需要重新梳理现在的接口列表, 如果蓝盾切换, 去掉依赖, 那么以前的部分接口可以下掉?
|
non_process
|
open api 校验用户是否某个用户组的成员 支持批量用户组,批量数有限制 在目前的接口新增 但是需要考虑 大表查询校验关系 存在 需要重新梳理现在的接口列表 如果蓝盾切换 去掉依赖 那么以前的部分接口可以下掉
| 0
|
217,391
| 16,855,762,320
|
IssuesEvent
|
2021-06-21 06:23:37
|
pingcap/tidb
|
https://api.github.com/repos/pingcap/tidb
|
closed
|
DATA_RACE:runtime.mapassign_fast64() failed
|
component/test
|
DATA_RACE:runtime.mapassign_fast64()
```
[2020-11-12T05:22:56.562Z] WARNING: DATA RACE
[2020-11-12T05:22:56.562Z] runtime.mapassign_fast64()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/map_fast64.go:92 +0x0
[2020-11-12T05:22:56.562Z] github.com/pingcap/parser/terror.ErrClass.NewStdErr()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/parser@v0.0.0-20201109022253-d384bee1451e/terror/terror.go:163 +0x19a
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/hint.checkInsertStmtHintDuplicated()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/dbterror/terror.go:55 +0x4be
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/hint.ExtractTableHintsFromStmtNode()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/hint/hint_processor.go:76 +0x108
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/planner.Optimize()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/planner/optimize.go:108 +0x1aa
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*Compiler).Compile()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/compiler.go:62 +0x2fa
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.(*session).ExecuteStmt()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1207 +0x270
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).Exec()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:170 +0x2f1
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).MustExec()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:216 +0x91
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session_test.(*testSessionSuite2).TestStmtHints()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session_test.go:3097 +0xd1f
[2020-11-12T05:22:56.562Z] runtime.call32()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/asm_amd64.s:539 +0x3a
[2020-11-12T05:22:56.562Z] reflect.Value.Call()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/reflect/value.go:321 +0xd3
[2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkTest.func1()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:850 +0x9aa
[2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkCall.func1()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:739 +0x113
[2020-11-12T05:22:56.562Z]
[2020-11-12T05:22:56.562Z] Previous write at 0x00c0001e0420 by goroutine 456:
[2020-11-12T05:22:56.562Z] runtime.mapassign_fast64()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/map_fast64.go:92 +0x0
[2020-11-12T05:22:56.562Z] github.com/pingcap/parser/terror.ErrClass.NewStdErr()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/parser@v0.0.0-20201109022253-d384bee1451e/terror/terror.go:163 +0x19a
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).handleErr()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:300 +0xa54
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).fastEvalRow()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:373 +0x569
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).fastEvalRow-fm()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:359 +0xaa
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.insertRows()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:233 +0x3b6
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertExec).Next()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert.go:288 +0x117
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.Next()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/executor.go:268 +0x27d
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).handleNoDelayExecutor()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:522 +0x38e
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).handleNoDelay()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:404 +0x254
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).Exec()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:354 +0x3f6
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.runStmt()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1285 +0x2c1
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.(*session).ExecuteStmt()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1229 +0xa57
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).Exec()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:170 +0x2f1
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session_test.(*testSessionSuite).TestPrepareZero()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session_test.go:1021 +0x209
[2020-11-12T05:22:56.562Z] runtime.call32()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/asm_amd64.s:539 +0x3a
[2020-11-12T05:22:56.562Z] reflect.Value.Call()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/reflect/value.go:321 +0xd3
[2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkTest.func1()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:850 +0x9aa
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall.func1()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:739 +0x113
[2020-11-12T05:22:56.563Z]
[2020-11-12T05:22:56.563Z] Goroutine 459 (running) created at:
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:734 +0x4a3
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkTest()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:832 +0x1b9
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).doRun()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:666 +0x13a
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).asyncRun.func1()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:650 +0xf7
[2020-11-12T05:22:56.563Z]
[2020-11-12T05:22:56.563Z] Goroutine 456 (running) created at:
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:734 +0x4a3
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkTest()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:832 +0x1b9
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).doRun()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:666 +0x13a
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).asyncRun.func1()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:650 +0xf7
[2020-11-12T05:22:56.563Z] ==================
```
Latest failed builds:
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/57742/display/redirect
|
1.0
|
DATA_RACE:runtime.mapassign_fast64() failed - DATA_RACE:runtime.mapassign_fast64()
```
[2020-11-12T05:22:56.562Z] WARNING: DATA RACE
[2020-11-12T05:22:56.562Z] runtime.mapassign_fast64()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/map_fast64.go:92 +0x0
[2020-11-12T05:22:56.562Z] github.com/pingcap/parser/terror.ErrClass.NewStdErr()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/parser@v0.0.0-20201109022253-d384bee1451e/terror/terror.go:163 +0x19a
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/hint.checkInsertStmtHintDuplicated()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/dbterror/terror.go:55 +0x4be
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/hint.ExtractTableHintsFromStmtNode()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/hint/hint_processor.go:76 +0x108
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/planner.Optimize()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/planner/optimize.go:108 +0x1aa
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*Compiler).Compile()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/compiler.go:62 +0x2fa
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.(*session).ExecuteStmt()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1207 +0x270
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).Exec()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:170 +0x2f1
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).MustExec()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:216 +0x91
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session_test.(*testSessionSuite2).TestStmtHints()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session_test.go:3097 +0xd1f
[2020-11-12T05:22:56.562Z] runtime.call32()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/asm_amd64.s:539 +0x3a
[2020-11-12T05:22:56.562Z] reflect.Value.Call()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/reflect/value.go:321 +0xd3
[2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkTest.func1()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:850 +0x9aa
[2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkCall.func1()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:739 +0x113
[2020-11-12T05:22:56.562Z]
[2020-11-12T05:22:56.562Z] Previous write at 0x00c0001e0420 by goroutine 456:
[2020-11-12T05:22:56.562Z] runtime.mapassign_fast64()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/map_fast64.go:92 +0x0
[2020-11-12T05:22:56.562Z] github.com/pingcap/parser/terror.ErrClass.NewStdErr()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/parser@v0.0.0-20201109022253-d384bee1451e/terror/terror.go:163 +0x19a
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).handleErr()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:300 +0xa54
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).fastEvalRow()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:373 +0x569
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertValues).fastEvalRow-fm()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:359 +0xaa
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.insertRows()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert_common.go:233 +0x3b6
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*InsertExec).Next()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/insert.go:288 +0x117
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.Next()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/executor.go:268 +0x27d
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).handleNoDelayExecutor()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:522 +0x38e
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).handleNoDelay()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:404 +0x254
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/executor.(*ExecStmt).Exec()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/executor/adapter.go:354 +0x3f6
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.runStmt()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1285 +0x2c1
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session.(*session).ExecuteStmt()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session.go:1229 +0xa57
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/util/testkit.(*TestKit).Exec()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:170 +0x2f1
[2020-11-12T05:22:56.562Z] github.com/pingcap/tidb/session_test.(*testSessionSuite).TestPrepareZero()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/session/session_test.go:1021 +0x209
[2020-11-12T05:22:56.562Z] runtime.call32()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/runtime/asm_amd64.s:539 +0x3a
[2020-11-12T05:22:56.562Z] reflect.Value.Call()
[2020-11-12T05:22:56.562Z] /usr/local/go/src/reflect/value.go:321 +0xd3
[2020-11-12T05:22:56.562Z] github.com/pingcap/check.(*suiteRunner).forkTest.func1()
[2020-11-12T05:22:56.562Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:850 +0x9aa
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall.func1()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:739 +0x113
[2020-11-12T05:22:56.563Z]
[2020-11-12T05:22:56.563Z] Goroutine 459 (running) created at:
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:734 +0x4a3
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkTest()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:832 +0x1b9
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).doRun()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:666 +0x13a
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).asyncRun.func1()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:650 +0xf7
[2020-11-12T05:22:56.563Z]
[2020-11-12T05:22:56.563Z] Goroutine 456 (running) created at:
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkCall()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:734 +0x4a3
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).forkTest()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:832 +0x1b9
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).doRun()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:666 +0x13a
[2020-11-12T05:22:56.563Z] github.com/pingcap/check.(*suiteRunner).asyncRun.func1()
[2020-11-12T05:22:56.563Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/pkg/mod/github.com/pingcap/check@v0.0.0-20200212061837-5e12011dc712/check.go:650 +0xf7
[2020-11-12T05:22:56.563Z] ==================
```
Latest failed builds:
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/57742/display/redirect
|
non_process
|
data race runtime mapassign failed data race runtime mapassign warning data race runtime mapassign usr local go src runtime map go github com pingcap parser terror errclass newstderr home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap parser terror terror go github com pingcap tidb util hint checkinsertstmthintduplicated home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util dbterror terror go github com pingcap tidb util hint extracttablehintsfromstmtnode home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util hint hint processor go github com pingcap tidb planner optimize home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb planner optimize go github com pingcap tidb executor compiler compile home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor compiler go github com pingcap tidb session session executestmt home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb session session go github com pingcap tidb util testkit testkit exec home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util testkit testkit go github com pingcap tidb util testkit testkit mustexec home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util testkit testkit go github com pingcap tidb session test teststmthints home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb session session test go runtime usr local go src runtime asm s reflect value call usr local go src reflect value go github com pingcap check suiterunner forktest home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner forkcall home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go previous write at by goroutine runtime mapassign usr local go src runtime map go github com pingcap parser terror errclass newstderr home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap parser terror terror go github com pingcap tidb executor insertvalues handleerr home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor insert common go github com pingcap tidb executor insertvalues fastevalrow home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor insert common go github com pingcap tidb executor insertvalues fastevalrow fm home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor insert common go github com pingcap tidb executor insertrows home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor insert common go github com pingcap tidb executor insertexec next home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor insert go github com pingcap tidb executor next home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor executor go github com pingcap tidb executor execstmt handlenodelayexecutor home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor adapter go github com pingcap tidb executor execstmt handlenodelay home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor adapter go github com pingcap tidb executor execstmt exec home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb executor adapter go github com pingcap tidb session runstmt home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb session session go github com pingcap tidb session session executestmt home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb session session go github com pingcap tidb util testkit testkit exec home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util testkit testkit go github com pingcap tidb session test testsessionsuite testpreparezero home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb session session test go runtime usr local go src runtime asm s reflect value call usr local go src reflect value go github com pingcap check suiterunner forktest home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner forkcall home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go goroutine running created at github com pingcap check suiterunner forkcall home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner forktest home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner dorun home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner asyncrun home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go goroutine running created at github com pingcap check suiterunner forkcall home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner forktest home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner dorun home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go github com pingcap check suiterunner asyncrun home jenkins agent workspace tidb ghpr unit test go pkg mod github com pingcap check check go latest failed builds
| 0
|
19,847
| 26,247,923,079
|
IssuesEvent
|
2023-01-05 16:40:57
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
closed
|
Insecure MD5 usage in Multiprocessing.connection
|
type-feature type-security expert-multiprocessing
|
# Feature or enhancement
Remove insecure use of md5 in Multiprocessing.connection
# Pitch
We discovered uses off the md5 hash, which has been proven insecure for more than a decade, in the Multiprocessing.connection library in the methods `deliver_challenge` and `answer_challenge`. This usage was apparently added in 2013 since the default implicit hashing mode for `hmac.new` was deprecated at that time. `hmac.new` previously defaulted to MD5 if a hashing algorithm was not specified. The 2013 change brings to code back to consistency with its prior use, but that use is insecure. It should be trivial to change the two uses in this library to a SHA2/3 secure hashing function (e.g., SHA512).
Failure to update the hashing algorithm may require organizations to fully cease use of the Multiprocessing library or components of the library to meet industry security standards with respect to acceptable uses of hashing algorithms.
<!-- gh-linked-prs -->
### Linked PRs
* gh-100772
<!-- /gh-linked-prs -->
|
1.0
|
Insecure MD5 usage in Multiprocessing.connection - # Feature or enhancement
Remove insecure use of md5 in Multiprocessing.connection
# Pitch
We discovered uses off the md5 hash, which has been proven insecure for more than a decade, in the Multiprocessing.connection library in the methods `deliver_challenge` and `answer_challenge`. This usage was apparently added in 2013 since the default implicit hashing mode for `hmac.new` was deprecated at that time. `hmac.new` previously defaulted to MD5 if a hashing algorithm was not specified. The 2013 change brings to code back to consistency with its prior use, but that use is insecure. It should be trivial to change the two uses in this library to a SHA2/3 secure hashing function (e.g., SHA512).
Failure to update the hashing algorithm may require organizations to fully cease use of the Multiprocessing library or components of the library to meet industry security standards with respect to acceptable uses of hashing algorithms.
<!-- gh-linked-prs -->
### Linked PRs
* gh-100772
<!-- /gh-linked-prs -->
|
process
|
insecure usage in multiprocessing connection feature or enhancement remove insecure use of in multiprocessing connection pitch we discovered uses off the hash which has been proven insecure for more than a decade in the multiprocessing connection library in the methods deliver challenge and answer challenge this usage was apparently added in since the default implicit hashing mode for hmac new was deprecated at that time hmac new previously defaulted to if a hashing algorithm was not specified the change brings to code back to consistency with its prior use but that use is insecure it should be trivial to change the two uses in this library to a secure hashing function e g failure to update the hashing algorithm may require organizations to fully cease use of the multiprocessing library or components of the library to meet industry security standards with respect to acceptable uses of hashing algorithms linked prs gh
| 1
|
3,809
| 6,795,273,248
|
IssuesEvent
|
2017-11-01 15:12:12
|
coala/projects
|
https://api.github.com/repos/coala/projects
|
opened
|
Relicense text
|
process/pending_review
|
AGPL is not ideal for large chunks of text.
I suggest that we use CC-BY-SA 4.0
|
1.0
|
Relicense text - AGPL is not ideal for large chunks of text.
I suggest that we use CC-BY-SA 4.0
|
process
|
relicense text agpl is not ideal for large chunks of text i suggest that we use cc by sa
| 1
|
167,307
| 26,484,054,017
|
IssuesEvent
|
2023-01-17 16:37:54
|
influxdata/ui
|
https://api.github.com/repos/influxdata/ui
|
closed
|
Script Editor SQL Adjustments
|
needs/design team/automation release/marty-23.01
|
**Task Description**
<p></p>
We need to do some UI Re-Design to make the script editor SQL experience better:
<p></p>
1 \- Separate the Bucket Selection from the rest of the schema browser so it is more obvious to the user that the bucket selection is required, and that the schema browser is informational only \(see design from Julia\)\.
<p></p>
2 \- Change the text wording to indicate that the user needs to select a database\/bucket first\. See design from Julia\.
<p></p>
3 \- When a user runs a SQL query without a bucket selected, the user will be shown an error message that tells them they must select a database\.
<p></p>
4 \- The raw text that is in the editor should go away once the user clicks into the box and starts typing \(they shouldn't have to delete it manually\)\.
<p></p>
5 \- Remove the time range selection as it will not apply to SQL queries right now\.
### Figma
https://www.figma.com/file/0qAntPk5LVangAHguWT74X/Query-Experience-Project?node-id=2651%3A112467&t=tGT6F3oSYXF2KSZz-1
|
1.0
|
Script Editor SQL Adjustments - **Task Description**
<p></p>
We need to do some UI Re-Design to make the script editor SQL experience better:
<p></p>
1 \- Separate the Bucket Selection from the rest of the schema browser so it is more obvious to the user that the bucket selection is required, and that the schema browser is informational only \(see design from Julia\)\.
<p></p>
2 \- Change the text wording to indicate that the user needs to select a database\/bucket first\. See design from Julia\.
<p></p>
3 \- When a user runs a SQL query without a bucket selected, the user will be shown an error message that tells them they must select a database\.
<p></p>
4 \- The raw text that is in the editor should go away once the user clicks into the box and starts typing \(they shouldn't have to delete it manually\)\.
<p></p>
5 \- Remove the time range selection as it will not apply to SQL queries right now\.
### Figma
https://www.figma.com/file/0qAntPk5LVangAHguWT74X/Query-Experience-Project?node-id=2651%3A112467&t=tGT6F3oSYXF2KSZz-1
|
non_process
|
script editor sql adjustments task description we need to do some ui re design to make the script editor sql experience better separate the bucket selection from the rest of the schema browser so it is more obvious to the user that the bucket selection is required and that the schema browser is informational only see design from julia change the text wording to indicate that the user needs to select a database bucket first see design from julia when a user runs a sql query without a bucket selected the user will be shown an error message that tells them they must select a database the raw text that is in the editor should go away once the user clicks into the box and starts typing they shouldn t have to delete it manually remove the time range selection as it will not apply to sql queries right now figma
| 0
|
139,823
| 11,287,214,652
|
IssuesEvent
|
2020-01-16 03:32:31
|
opentracing-contrib/java-specialagent
|
https://api.github.com/repos/opentracing-contrib/java-specialagent
|
closed
|
`spring-kafka` test failing after refactor
|
.25 bug test
|
@malafeev, would you be able to help me with a test? I have refactored the module structure under `/test`, and for some reason the `spring-kafka` test is failing. Could you take a look at it? The work is in [`circleci` branch](https://github.com/opentracing-contrib/java-specialagent/blob/circleci/test/spring-kafka/spring-kafka-2.3.3/pom.xml), and here's the log of the spring-kafka tests:
https://api.travis-ci.org/v3/job/637261261/log.txt
|
1.0
|
`spring-kafka` test failing after refactor - @malafeev, would you be able to help me with a test? I have refactored the module structure under `/test`, and for some reason the `spring-kafka` test is failing. Could you take a look at it? The work is in [`circleci` branch](https://github.com/opentracing-contrib/java-specialagent/blob/circleci/test/spring-kafka/spring-kafka-2.3.3/pom.xml), and here's the log of the spring-kafka tests:
https://api.travis-ci.org/v3/job/637261261/log.txt
|
non_process
|
spring kafka test failing after refactor malafeev would you be able to help me with a test i have refactored the module structure under test and for some reason the spring kafka test is failing could you take a look at it the work is in and here s the log of the spring kafka tests
| 0
|
130,758
| 12,462,194,340
|
IssuesEvent
|
2020-05-28 08:29:47
|
adriens/covid19-action-plan-nc
|
https://api.github.com/repos/adriens/covid19-action-plan-nc
|
closed
|
Scénario de test du tableau de bord pour crowd testing
|
PRODUCTION Tableau de Bord documentation enhancement
|
# Contexte
La qualité de notre application dépend de:
1 - La justesse des données: ie. sont-t-elles bien alignées avec ce qui est annoncé et la réalité du terrain ?
2 La qualité de l'affichage: typos, texte impossible à lire, taille d'image pas adaptée, graphique gros gros/petit
En Nouvelle-Calédonie a été lancée une plateforme de crowdtesting basée sur le blockchain
# Ressources
- Article [dédié sur le Blog Hightest](https://hightest.nc/blog/posts/le-crowdtesting-met-il-en-danger-les-testeurs-professionnels
)
# Scénarios envisagés
## Dernières données
Si au vu des données du jour, les chiffres de correspondent pas.
## Dates
- Vérifier que le nom du jour et le numéro matchent bien (ie. que ce jour existe bien dans l'année en cours)
## Typo
Toute erreur de typo: orthographe, grammaire, qui nuit à la qualité de la lectre du site
|
1.0
|
Scénario de test du tableau de bord pour crowd testing - # Contexte
La qualité de notre application dépend de:
1 - La justesse des données: ie. sont-t-elles bien alignées avec ce qui est annoncé et la réalité du terrain ?
2 La qualité de l'affichage: typos, texte impossible à lire, taille d'image pas adaptée, graphique gros gros/petit
En Nouvelle-Calédonie a été lancée une plateforme de crowdtesting basée sur le blockchain
# Ressources
- Article [dédié sur le Blog Hightest](https://hightest.nc/blog/posts/le-crowdtesting-met-il-en-danger-les-testeurs-professionnels
)
# Scénarios envisagés
## Dernières données
Si au vu des données du jour, les chiffres de correspondent pas.
## Dates
- Vérifier que le nom du jour et le numéro matchent bien (ie. que ce jour existe bien dans l'année en cours)
## Typo
Toute erreur de typo: orthographe, grammaire, qui nuit à la qualité de la lectre du site
|
non_process
|
scénario de test du tableau de bord pour crowd testing contexte la qualité de notre application dépend de la justesse des données ie sont t elles bien alignées avec ce qui est annoncé et la réalité du terrain la qualité de l affichage typos texte impossible à lire taille d image pas adaptée graphique gros gros petit en nouvelle calédonie a été lancée une plateforme de crowdtesting basée sur le blockchain ressources article scénarios envisagés dernières données si au vu des données du jour les chiffres de correspondent pas dates vérifier que le nom du jour et le numéro matchent bien ie que ce jour existe bien dans l année en cours typo toute erreur de typo orthographe grammaire qui nuit à la qualité de la lectre du site
| 0
|
55,544
| 13,639,169,004
|
IssuesEvent
|
2020-09-25 10:32:28
|
astropy/astropy
|
https://api.github.com/repos/astropy/astropy
|
opened
|
Astropy does not build on MacOSX with xcode 12.0.0
|
build
|
<!-- This comments are hidden when you submit the issue,
so you do not need to remove them! -->
<!-- Please be sure to check out our contributing guidelines,
https://github.com/astropy/astropy/blob/master/CONTRIBUTING.md .
Please be sure to check out our code of conduct,
https://github.com/astropy/astropy/blob/master/CODE_OF_CONDUCT.md . -->
<!-- Please have a search on our GitHub repository to see if a similar
issue has already been posted.
If a similar issue is closed, have a quick look to see if you are satisfied
by the resolution.
If not please go ahead and open an issue! -->
<!-- Please check that the development version still produces the same bug.
You can install development version with
pip install git+https://github.com/astropy/astropy
command. -->
### Description
<!-- Provide a general description of the bug. -->
When I try to build astropy on my Mac (Mojave + Xcode 12.0) it fails with the following error:
```
cextern/cfitsio/lib/group.c:5664:8: error: implicit declaration of function 'getcwd' is invalid in C99 [-Werror,-Wimplicit-function-declaration]
if (!getcwd(buff,FLEN_FILENAME))
^
1 warning and 1 error generated.
error: command 'gcc' failed with exit status 1
(astropy) ➜ astropy-tmp git:(master) gcc --version
Configured with: --prefix=/Applications/Xcode.app/Contents/Developer/usr --with-gxx-include-dir=/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX.sdk/usr/include/c++/4.2.1
Apple clang version 12.0.0 (clang-1200.0.32.2)
Target: x86_64-apple-darwin19.5.0
Thread model: posix
InstalledDir: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin
```
Xcode 12.0 was just released Sep. 16 and I'm guessing that it was automatically updated on my Mac since I did not request an update. Astropy is apparently not alone, e.g. https://gitlab.com/graphviz/graphviz/-/issues/1826.
There is a long slack thread with helpful inputs from @saimn and @manodeep here: https://astropy.slack.com/archives/C067V74GK/p1600984862007700
A workaround is:
```
CFLAGS=-Wno-error=implicit-function-declaration pip install -e .
```
### Expected behavior
<!-- What did you expect to happen. -->
Astropy builds and runs tests.
### Actual behavior
<!-- What actually happened. -->
<!-- Was the output confusing or poorly described? -->
Build failed as shown.
### Steps to Reproduce
<!-- Ideally a code example could be provided so we can run it ourselves. -->
<!-- If you are pasting code, use triple backticks (```) around
your code snippet. -->
<!-- If necessary, sanitize your screen output to be pasted so you do not
reveal secrets like tokens and passwords. -->
1. Checkout astropy at master, `git clean -fxd`
2. `tox -e test` or `pip install -e .`
### System Details
<!-- Even if you do not think this is necessary, it is useful information for the maintainers.
Please run the following snippet and paste the output below:
import platform; print(platform.platform())
import sys; print("Python", sys.version)
import numpy; print("Numpy", numpy.__version__)
import astropy; print("astropy", astropy.__version__)
import scipy; print("Scipy", scipy.__version__)
import matplotlib; print("Matplotlib", matplotlib.__version__)
-->
```
Darwin-19.5.0-x86_64-i386-64bit
Python 3.7.7 (default, Mar 26 2020, 10:32:53)
[Clang 4.0.1 (tags/RELEASE_401/final)]
Numpy 1.18.1
astropy 4.2.dev715+gfaccb8b41
Scipy 1.4.1
Matplotlib 3.1.3
```
|
1.0
|
Astropy does not build on MacOSX with xcode 12.0.0 - <!-- This comments are hidden when you submit the issue,
so you do not need to remove them! -->
<!-- Please be sure to check out our contributing guidelines,
https://github.com/astropy/astropy/blob/master/CONTRIBUTING.md .
Please be sure to check out our code of conduct,
https://github.com/astropy/astropy/blob/master/CODE_OF_CONDUCT.md . -->
<!-- Please have a search on our GitHub repository to see if a similar
issue has already been posted.
If a similar issue is closed, have a quick look to see if you are satisfied
by the resolution.
If not please go ahead and open an issue! -->
<!-- Please check that the development version still produces the same bug.
You can install development version with
pip install git+https://github.com/astropy/astropy
command. -->
### Description
<!-- Provide a general description of the bug. -->
When I try to build astropy on my Mac (Mojave + Xcode 12.0) it fails with the following error:
```
cextern/cfitsio/lib/group.c:5664:8: error: implicit declaration of function 'getcwd' is invalid in C99 [-Werror,-Wimplicit-function-declaration]
if (!getcwd(buff,FLEN_FILENAME))
^
1 warning and 1 error generated.
error: command 'gcc' failed with exit status 1
(astropy) ➜ astropy-tmp git:(master) gcc --version
Configured with: --prefix=/Applications/Xcode.app/Contents/Developer/usr --with-gxx-include-dir=/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs/MacOSX.sdk/usr/include/c++/4.2.1
Apple clang version 12.0.0 (clang-1200.0.32.2)
Target: x86_64-apple-darwin19.5.0
Thread model: posix
InstalledDir: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin
```
Xcode 12.0 was just released Sep. 16 and I'm guessing that it was automatically updated on my Mac since I did not request an update. Astropy is apparently not alone, e.g. https://gitlab.com/graphviz/graphviz/-/issues/1826.
There is a long slack thread with helpful inputs from @saimn and @manodeep here: https://astropy.slack.com/archives/C067V74GK/p1600984862007700
A workaround is:
```
CFLAGS=-Wno-error=implicit-function-declaration pip install -e .
```
### Expected behavior
<!-- What did you expect to happen. -->
Astropy builds and runs tests.
### Actual behavior
<!-- What actually happened. -->
<!-- Was the output confusing or poorly described? -->
Build failed as shown.
### Steps to Reproduce
<!-- Ideally a code example could be provided so we can run it ourselves. -->
<!-- If you are pasting code, use triple backticks (```) around
your code snippet. -->
<!-- If necessary, sanitize your screen output to be pasted so you do not
reveal secrets like tokens and passwords. -->
1. Checkout astropy at master, `git clean -fxd`
2. `tox -e test` or `pip install -e .`
### System Details
<!-- Even if you do not think this is necessary, it is useful information for the maintainers.
Please run the following snippet and paste the output below:
import platform; print(platform.platform())
import sys; print("Python", sys.version)
import numpy; print("Numpy", numpy.__version__)
import astropy; print("astropy", astropy.__version__)
import scipy; print("Scipy", scipy.__version__)
import matplotlib; print("Matplotlib", matplotlib.__version__)
-->
```
Darwin-19.5.0-x86_64-i386-64bit
Python 3.7.7 (default, Mar 26 2020, 10:32:53)
[Clang 4.0.1 (tags/RELEASE_401/final)]
Numpy 1.18.1
astropy 4.2.dev715+gfaccb8b41
Scipy 1.4.1
Matplotlib 3.1.3
```
|
non_process
|
astropy does not build on macosx with xcode this comments are hidden when you submit the issue so you do not need to remove them please be sure to check out our contributing guidelines please be sure to check out our code of conduct please have a search on our github repository to see if a similar issue has already been posted if a similar issue is closed have a quick look to see if you are satisfied by the resolution if not please go ahead and open an issue please check that the development version still produces the same bug you can install development version with pip install git command description when i try to build astropy on my mac mojave xcode it fails with the following error cextern cfitsio lib group c error implicit declaration of function getcwd is invalid in if getcwd buff flen filename warning and error generated error command gcc failed with exit status astropy ➜ astropy tmp git master gcc version configured with prefix applications xcode app contents developer usr with gxx include dir applications xcode app contents developer platforms macosx platform developer sdks macosx sdk usr include c apple clang version clang target apple thread model posix installeddir applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin xcode was just released sep and i m guessing that it was automatically updated on my mac since i did not request an update astropy is apparently not alone e g there is a long slack thread with helpful inputs from saimn and manodeep here a workaround is cflags wno error implicit function declaration pip install e expected behavior astropy builds and runs tests actual behavior build failed as shown steps to reproduce if you are pasting code use triple backticks around your code snippet if necessary sanitize your screen output to be pasted so you do not reveal secrets like tokens and passwords checkout astropy at master git clean fxd tox e test or pip install e system details even if you do not think this is necessary it is useful information for the maintainers please run the following snippet and paste the output below import platform print platform platform import sys print python sys version import numpy print numpy numpy version import astropy print astropy astropy version import scipy print scipy scipy version import matplotlib print matplotlib matplotlib version darwin python default mar numpy astropy scipy matplotlib
| 0
|
11,195
| 13,957,701,492
|
IssuesEvent
|
2020-10-24 08:13:18
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
NO: Question regarding the INSPIRE Geoportal Discovery Service Register
|
Geoportal Harvesting process NO - Norway
|
From: Lars Inge Arnevik
Sent: 05 June 2018 14:04:23 (UTC+01:00) Brussels, Copenhagen, Madrid, Paris
To: inspire-geoportal@jrc.ec.europa.eu
Subject: Question regarding the INSPIRE Geoportal Discovery Service Register
Hi!
I wonder when the next harvesting from Geonorge will take place? I think last harvest was 31. of May and I guess it was interrupted because of our monthly maintenance. We are working on improving the metadata and it is motivating to see the results from the error report J
Best regards
Lars-Inge
|
1.0
|
NO: Question regarding the INSPIRE Geoportal Discovery Service Register - From: Lars Inge Arnevik
Sent: 05 June 2018 14:04:23 (UTC+01:00) Brussels, Copenhagen, Madrid, Paris
To: inspire-geoportal@jrc.ec.europa.eu
Subject: Question regarding the INSPIRE Geoportal Discovery Service Register
Hi!
I wonder when the next harvesting from Geonorge will take place? I think last harvest was 31. of May and I guess it was interrupted because of our monthly maintenance. We are working on improving the metadata and it is motivating to see the results from the error report J
Best regards
Lars-Inge
|
process
|
no question regarding the inspire geoportal discovery service register from lars inge arnevik sent june utc brussels copenhagen madrid paris to inspire geoportal jrc ec europa eu subject question regarding the inspire geoportal discovery service register hi i wonder when the next harvesting from geonorge will take place i think last harvest was of may and i guess it was interrupted because of our monthly maintenance we are working on improving the metadata and it is motivating to see the results from the error report j best regards lars inge
| 1
|
884
| 2,582,912,635
|
IssuesEvent
|
2015-02-15 19:52:34
|
code-cracker/code-cracker
|
https://api.github.com/repos/code-cracker/code-cracker
|
closed
|
Introduce field from constructor
|
3 - Done analyzer C# code-fix enhancement
|
This is a refactoring.
The idea is that on a constructor parameter you CTRL DOT and get a "Introduce field" dialog, that will create a field and assign the value from the parameter to the field.
So this:
````csharp
class Foo
{
public Foo(string bar)
{
}
}
````
Becomes:
````csharp
class Foo
{
private readonly string bar;
public Foo(string bar)
{
this.bar = bar;
}
}
````
If the field is already assigned then no diagnostic is offered.
If the field is already present, it is just assigned to.
If the type does not match (e.g. parameter is `string`, existing field is `int`, a new field with `1` postfixed is added, like that:
````csharp
class Foo
{
private int bar;
public Foo(string bar)
{
}
}
````
Becomes:
````csharp
class Foo
{
private int bar;
private readonly string bar1;
public Foo(string bar)
{
this.bar1 = bar;
}
}
````
Diagnostic Id: `CC0071`
Severity: `Hidden` (refactoring)
Category: `Refactoring`
|
1.0
|
Introduce field from constructor - This is a refactoring.
The idea is that on a constructor parameter you CTRL DOT and get a "Introduce field" dialog, that will create a field and assign the value from the parameter to the field.
So this:
````csharp
class Foo
{
public Foo(string bar)
{
}
}
````
Becomes:
````csharp
class Foo
{
private readonly string bar;
public Foo(string bar)
{
this.bar = bar;
}
}
````
If the field is already assigned then no diagnostic is offered.
If the field is already present, it is just assigned to.
If the type does not match (e.g. parameter is `string`, existing field is `int`, a new field with `1` postfixed is added, like that:
````csharp
class Foo
{
private int bar;
public Foo(string bar)
{
}
}
````
Becomes:
````csharp
class Foo
{
private int bar;
private readonly string bar1;
public Foo(string bar)
{
this.bar1 = bar;
}
}
````
Diagnostic Id: `CC0071`
Severity: `Hidden` (refactoring)
Category: `Refactoring`
|
non_process
|
introduce field from constructor this is a refactoring the idea is that on a constructor parameter you ctrl dot and get a introduce field dialog that will create a field and assign the value from the parameter to the field so this csharp class foo public foo string bar becomes csharp class foo private readonly string bar public foo string bar this bar bar if the field is already assigned then no diagnostic is offered if the field is already present it is just assigned to if the type does not match e g parameter is string existing field is int a new field with postfixed is added like that csharp class foo private int bar public foo string bar becomes csharp class foo private int bar private readonly string public foo string bar this bar diagnostic id severity hidden refactoring category refactoring
| 0
|
58,635
| 6,612,413,518
|
IssuesEvent
|
2017-09-20 03:38:37
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
StorageClass.ReclaimPolicy, PV.MountOptions, and SC.MountOptions need e2e tests
|
kind/cleanup kind/e2e-test-failure milestone-labels-complete priority/important-soon sig/storage
|
https://github.com/kubernetes/kubernetes/pull/47987 for 1.8.
/assign
|
1.0
|
StorageClass.ReclaimPolicy, PV.MountOptions, and SC.MountOptions need e2e tests - https://github.com/kubernetes/kubernetes/pull/47987 for 1.8.
/assign
|
non_process
|
storageclass reclaimpolicy pv mountoptions and sc mountoptions need tests for assign
| 0
|
20,434
| 27,098,832,376
|
IssuesEvent
|
2023-02-15 06:42:23
|
alibaba/MNN
|
https://api.github.com/repos/alibaba/MNN
|
closed
|
How can initialize image and rect?
|
question cv/ImageProcess
|
Hi,
I think if I will use tracker model by mnn, I have to initialize image and ROI rect but I don't know how to this.
Exactly, I'd like to implement like this https://github.com/dongfangduoshou123/DaSiamRPN-Caffe2/blob/master/main.cpp#L46
Please let me know about this.
Thank you.
|
1.0
|
How can initialize image and rect? - Hi,
I think if I will use tracker model by mnn, I have to initialize image and ROI rect but I don't know how to this.
Exactly, I'd like to implement like this https://github.com/dongfangduoshou123/DaSiamRPN-Caffe2/blob/master/main.cpp#L46
Please let me know about this.
Thank you.
|
process
|
how can initialize image and rect hi i think if i will use tracker model by mnn i have to initialize image and roi rect but i don t know how to this exactly i d like to implement like this please let me know about this thank you
| 1
|
18,115
| 24,146,454,150
|
IssuesEvent
|
2022-09-21 19:13:02
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
`@bazel_tools`'s `unix_cc_configure.bzl` breaks when other link opts contain `lld`
|
P4 type: support / not a bug (process) team-Rules-CPP
|
## Problem
When creating [`@local_config_cc` during autoconfiguration](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/cc_configure.bzl#L125), `unix_cc_configure.bzl`'s [`_find_linker_path`](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L174-L218) is used to parse [`-v` command line output from a compiler invocation](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L197) to figure out if [`lld`/`gold` are present](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L421-L422) and what their paths are.
The parsing essentially looks for the [first space separated occurrence of the linker name](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L208-L210) in the [first line containing the linker name](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L205-L207) in the verbose output and then [strips spaces and quotes from it](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L213).
The [compiler invocation](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L186-L198) is essentially: `$CXX -xc - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=$linker`.
All together that's essentially:
```bash
$CXX -xc++ - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=$linker -v \
|& grep $linker \
| xargs -n 1 \
| grep $linker \
| tr -d ' ''"'"'" \
| head -1
```
The problem is that this can potentially isolate things other than the path to the linker.
I ran into this issue when using a [nix-shell](https://nixos.org/manual/nix/stable/command-ref/nix-shell.html) with `lldb` present; `nix` by default [wraps](https://github.com/NixOS/nixpkgs/tree/master/pkgs/build-support/cc-wrapper) its compilers (including the compiler it exposes as `$CC` and `$CXX` which Bazel's cc toolchain autoconfiguration picks up by default to make `@local_config_cc`) in a shell script that adds the contents of `$NIX_LDFLAGS` (and other env vars) to the actual compiler invocation.
In its verbose output, `clang` prints out the actual compiler invocation; `_find_linker_path` interprets the `-L` include that `nix`'s wrapper adds to the invocation for `lldb`'s library files to be the linker path.
This manifests in compile errors on `cc_binary`s that look like:
```
INFO: Found 1 target...
ERROR: .../BUILD.bazel:127:10: Linking <some_target> [for host] failed: (Exit 1): clang failed: error executing command /nix/store/q52j3nyvc8947za806109xrxaz4dqdzf-clang-wrapper-13.0.0/bin/clang @bazel-out/host/bin/external/some/target.params
Use --sandbox_debug to see verbose messages from the sandbox
clang-13: error: invalid linker name in argument '-fuse-ld=/nix/store/izgkyvmb4m35pary1blnpypa9l9j059y-lldb-13.0.0-dev/include'
```
## To Reproduce
### Using a `nix-shell`
In `nix-shell -p bazelisk -p lldb -p llvmPackages_13.bintools -p llvmPackages_13.clang`:
```bash
cd $(mktemp -d)
touch WORKSPACE
echo "int main() { return 3; }" >> main.cc
echo "5.0.0" >> .bazelversion
cat <<EOF > BUILD.bazel
cc_binary(
name = "test",
srcs = ["main.cc"],
)
EOF
export BAZEL_USE_CPP_ONLY_TOOLCHAIN=1
bazelisk build --action_env=CC=clang --action_env=CXX=clang++ //:test -s
````
Note the use of `bazelisk` to get Bazel 5.0+; [this commit](https://github.com/bazelbuild/bazel/commit/00e30ca5968d42b4a1e42327fa683debc1063b89) introduces the use of `lld` in autoconfiguration and it landed in Bazel 5.0.
Also note that this is currently not reproducible on macOS on arm64, [even with `BAZEL_USE_CPP_ONLY_TOOLCHAIN=1`](https://github.com/bazelbuild/bazel/blob/be21f194ae00d1a21e8e36a6fbb30be9e449086c/tools/cpp/cc_configure.bzl#L121-L125), because of a [bug in `nix`'s cc-wrapper](https://github.com/NixOS/nixpkgs/issues/154203) that causes it to error when attempting to sign the output "file" (`/dev/null`). Thus, `lld` and `gold` are just not used by Bazel in these cases since the compiler invocation fails.
### The Verbose Output
Alternatively, here is the output of `clang++ -xc++ - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=lld -v` in such a shell:
<details> <summary>Click to expand (grep for "lld")</summary>
```
clang version 13.0.0
Target: x86_64-unknown-linux-gnu
Thread model: posix
InstalledDir: /nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin
Found candidate GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0
Found candidate GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0
Selected GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0
Candidate multilib: .;@m64
Selected multilib: .;@m64
"/nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin/clang-13" -cc1 -triple x86_64-unknown-linux-gnu -emit-obj --mrelax-relocations -disable-free -disable-llvm-verifier -discard-value-names -main-file-name - -mrelocation-model pic -pic-level 2 -fhalf-no-semantic-interposition -mframe-pointer=none -fmath-errno -fno-rounding-math -mconstructor-aliases -munwind-tables -target-cpu x86-64 -tune-cpu generic -debugger-tuning=gdb -v -fcoverage-compilation-dir=/run/user/12064113/tmp.sqsgMmvY2k -nostdsysteminc -resource-dir /nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root -idirafter /nix/store/sr2bm5pp15dhb80w16vrwyhgh40kd4yf-glibc-2.33-59-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0 -isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu -D _FORTIFY_SOURCE=2 -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0 -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/x86_64-unknown-linux-gnu -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/backward -internal-isystem /nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root/include -O2 -Wformat -Wformat-security -Werror=format-security -fdeprecated-macro -fdebug-compilation-dir=/run/user/12064113/tmp.sqsgMmvY2k -ferror-limit 19 -fwrapv -stack-protector 2 -stack-protector-buffer-size 4 -fgnuc-version=4.2.1 -fcxx-exceptions -fexceptions -fcolor-diagnostics -vectorize-loops -vectorize-slp -faddrsig -D__GCC_HAVE_DWARF2_CFI_ASM=1 -o /run/user/12064113/--158e23.o -x c++ -
clang -cc1 version 13.0.0 based upon LLVM 13.0.0 default target x86_64-unknown-linux-gnu
ignoring duplicate directory "/nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include"
ignoring duplicate directory "/nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include"
ignoring duplicate directory "/nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include"
ignoring duplicate directory "/nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include"
ignoring duplicate directory "/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0"
ignoring duplicate directory "/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu"
#include "..." search starts here:
#include <...> search starts here:
/nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include
/nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include
/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0
/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu
/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/backward
/nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root/include
/nix/store/sr2bm5pp15dhb80w16vrwyhgh40kd4yf-glibc-2.33-59-dev/include
End of search list.
"/nix/store/kcg1vn8a2xf0fzd789fq16mww4wqvg0n-llvm-binutils-wrapper-13.0.0/bin/ld.lld" -z relro --hash-style=gnu --hash-style=both --enable-new-dtags --eh-frame-hdr -m elf_x86_64 -o /dev/null /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crt1.o /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crti.o /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0/crtbegin.o -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0 -L/nix/store/lwkpcfphv5nwymi0bvnvfga6q4p72v8i-gcc-10.3.0-lib/x86_64-unknown-linux-gnu/lib -L/nix/store/svzwv5yhcp03rysmc0lh5922a6f7qgp8-clang-13.0.0-lib/lib -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0 -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../lib64 -L/lib/../lib64 -L/usr/lib/../lib64 -L/nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin/../lib -L/lib -L/usr/lib -dynamic-linker=/nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/ld-linux-x86-64.so.2 /run/user/12064113/--158e23.o --start-lib --end-lib -rpath /nix/store/4ibh63jd2lzx03mrhjn9km8ihha9wq55-shell/lib64 -rpath /nix/store/4ibh63jd2lzx03mrhjn9km8ihha9wq55-shell/lib -lstdc++ -lm -lgcc_s -lgcc -lc -lgcc_s -lgcc /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0/crtend.o /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crtn.o
```
</details>
## Potential Solutions
A trivial fix would be to search for the linker name plus a space or end quote (i.e. `lld `/`lld"`/`lld'`, `gold `/`gold"`/`gold'`) or, for `lld` specifically, to search for `.lld` (afaik all the flavors of `lld` *for unix platforms* are `<something>.lld`: `ld64.lld`, `ld.lld`). This _seems_ like it'd fix this issue without relying on compiler-specific output.
Another potential solution is to only use the _last_ line of the verbose output. However, I am not sure we can rely on this always being the linker invocation, especially for compilers that aren't `clang`.
Ideally we'd just ask the compiler for the path to the linker but I don't know of any way to do so that's well supported across different compilers.
If any of these solutions are deemed a reasonable fix (or if a better path forward is proposed), I'm happy to make a PR.
## Workarounds
To anyone facing a similar issue, my workaround for now was to simply wrap the `lldb` derivation and have it use another name:
```nix
with nixpkgs; stdenvNoCC.mkDerivation {
pname = "llvm-debugger";
inherit (lldb) version outputs;
lldb_outputs = lldb.all;
src = "${lldb}";
disallowedReferences = [lldb];
buildPhase = ''
i=0
lldb_outs=($lldb_outputs)
outs=($outputs)
for o in "''${outs[@]}"; do
echo "[$o]" "''${lldb_outs[$i]}" to "''${!o}"
# can't symlink because then the paths used for `LDFLAGS` still
# contain "lldb":
cp -R "''${lldb_outs[$i]}" "''${!o}"
: $((++i))
done
# Remove some lingering references to the actual lldb package so that it's
# not registered as a runtime dep of this package.
scrub() {
local i=0
for o in "''${outs[@]}"; do
substituteInPlace "$1" \
--replace "''${lldb_outs[$i]}" "''${!o}"
: $((++i))
done
}
scrub "$out/bin/lldb"
scrub "$dev/nix-support/propagated-build-inputs"
'';
installPhase = "true";
}
# Alternatively the below will work too but will cause you to rebuild `lldb` from source:
# nixpkgs.lldb.overrideDerivation (o: { name = "llvm-debugger"; })
```
|
1.0
|
`@bazel_tools`'s `unix_cc_configure.bzl` breaks when other link opts contain `lld` - ## Problem
When creating [`@local_config_cc` during autoconfiguration](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/cc_configure.bzl#L125), `unix_cc_configure.bzl`'s [`_find_linker_path`](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L174-L218) is used to parse [`-v` command line output from a compiler invocation](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L197) to figure out if [`lld`/`gold` are present](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L421-L422) and what their paths are.
The parsing essentially looks for the [first space separated occurrence of the linker name](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L208-L210) in the [first line containing the linker name](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L205-L207) in the verbose output and then [strips spaces and quotes from it](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L213).
The [compiler invocation](https://github.com/bazelbuild/bazel/blob/0ba4caa5fc4bfd2d7f9af2a1653ef5b4bd83d176/tools/cpp/unix_cc_configure.bzl#L186-L198) is essentially: `$CXX -xc - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=$linker`.
All together that's essentially:
```bash
$CXX -xc++ - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=$linker -v \
|& grep $linker \
| xargs -n 1 \
| grep $linker \
| tr -d ' ''"'"'" \
| head -1
```
The problem is that this can potentially isolate things other than the path to the linker.
I ran into this issue when using a [nix-shell](https://nixos.org/manual/nix/stable/command-ref/nix-shell.html) with `lldb` present; `nix` by default [wraps](https://github.com/NixOS/nixpkgs/tree/master/pkgs/build-support/cc-wrapper) its compilers (including the compiler it exposes as `$CC` and `$CXX` which Bazel's cc toolchain autoconfiguration picks up by default to make `@local_config_cc`) in a shell script that adds the contents of `$NIX_LDFLAGS` (and other env vars) to the actual compiler invocation.
In its verbose output, `clang` prints out the actual compiler invocation; `_find_linker_path` interprets the `-L` include that `nix`'s wrapper adds to the invocation for `lldb`'s library files to be the linker path.
This manifests in compile errors on `cc_binary`s that look like:
```
INFO: Found 1 target...
ERROR: .../BUILD.bazel:127:10: Linking <some_target> [for host] failed: (Exit 1): clang failed: error executing command /nix/store/q52j3nyvc8947za806109xrxaz4dqdzf-clang-wrapper-13.0.0/bin/clang @bazel-out/host/bin/external/some/target.params
Use --sandbox_debug to see verbose messages from the sandbox
clang-13: error: invalid linker name in argument '-fuse-ld=/nix/store/izgkyvmb4m35pary1blnpypa9l9j059y-lldb-13.0.0-dev/include'
```
## To Reproduce
### Using a `nix-shell`
In `nix-shell -p bazelisk -p lldb -p llvmPackages_13.bintools -p llvmPackages_13.clang`:
```bash
cd $(mktemp -d)
touch WORKSPACE
echo "int main() { return 3; }" >> main.cc
echo "5.0.0" >> .bazelversion
cat <<EOF > BUILD.bazel
cc_binary(
name = "test",
srcs = ["main.cc"],
)
EOF
export BAZEL_USE_CPP_ONLY_TOOLCHAIN=1
bazelisk build --action_env=CC=clang --action_env=CXX=clang++ //:test -s
````
Note the use of `bazelisk` to get Bazel 5.0+; [this commit](https://github.com/bazelbuild/bazel/commit/00e30ca5968d42b4a1e42327fa683debc1063b89) introduces the use of `lld` in autoconfiguration and it landed in Bazel 5.0.
Also note that this is currently not reproducible on macOS on arm64, [even with `BAZEL_USE_CPP_ONLY_TOOLCHAIN=1`](https://github.com/bazelbuild/bazel/blob/be21f194ae00d1a21e8e36a6fbb30be9e449086c/tools/cpp/cc_configure.bzl#L121-L125), because of a [bug in `nix`'s cc-wrapper](https://github.com/NixOS/nixpkgs/issues/154203) that causes it to error when attempting to sign the output "file" (`/dev/null`). Thus, `lld` and `gold` are just not used by Bazel in these cases since the compiler invocation fails.
### The Verbose Output
Alternatively, here is the output of `clang++ -xc++ - <<<"int main() {}" -o /dev/null -Wl,--start-lib -Wl,--end-lib -fuse-ld=lld -v` in such a shell:
<details> <summary>Click to expand (grep for "lld")</summary>
```
clang version 13.0.0
Target: x86_64-unknown-linux-gnu
Thread model: posix
InstalledDir: /nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin
Found candidate GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0
Found candidate GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0
Selected GCC installation: /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0
Candidate multilib: .;@m64
Selected multilib: .;@m64
"/nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin/clang-13" -cc1 -triple x86_64-unknown-linux-gnu -emit-obj --mrelax-relocations -disable-free -disable-llvm-verifier -discard-value-names -main-file-name - -mrelocation-model pic -pic-level 2 -fhalf-no-semantic-interposition -mframe-pointer=none -fmath-errno -fno-rounding-math -mconstructor-aliases -munwind-tables -target-cpu x86-64 -tune-cpu generic -debugger-tuning=gdb -v -fcoverage-compilation-dir=/run/user/12064113/tmp.sqsgMmvY2k -nostdsysteminc -resource-dir /nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root -idirafter /nix/store/sr2bm5pp15dhb80w16vrwyhgh40kd4yf-glibc-2.33-59-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include -isystem /nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include -isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0 -isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu -D _FORTIFY_SOURCE=2 -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0 -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/x86_64-unknown-linux-gnu -internal-isystem /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/backward -internal-isystem /nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root/include -O2 -Wformat -Wformat-security -Werror=format-security -fdeprecated-macro -fdebug-compilation-dir=/run/user/12064113/tmp.sqsgMmvY2k -ferror-limit 19 -fwrapv -stack-protector 2 -stack-protector-buffer-size 4 -fgnuc-version=4.2.1 -fcxx-exceptions -fexceptions -fcolor-diagnostics -vectorize-loops -vectorize-slp -faddrsig -D__GCC_HAVE_DWARF2_CFI_ASM=1 -o /run/user/12064113/--158e23.o -x c++ -
clang -cc1 version 13.0.0 based upon LLVM 13.0.0 default target x86_64-unknown-linux-gnu
ignoring duplicate directory "/nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include"
ignoring duplicate directory "/nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include"
ignoring duplicate directory "/nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include"
ignoring duplicate directory "/nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include"
ignoring duplicate directory "/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0"
ignoring duplicate directory "/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu"
#include "..." search starts here:
#include <...> search starts here:
/nix/store/vbbk1rqm9na1czcggw25s6j4z95jx9wl-lldb-13.0.0-dev/include
/nix/store/mp5y1kdgslrn1nxpk3s04dlgqd8almh3-compiler-rt-libc-13.0.0-dev/include
/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0
/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/include/c++/10.3.0/x86_64-unknown-linux-gnu
/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../include/c++/10.3.0/backward
/nix/store/pbfj9cqpg9m1x64jq9fj8q4fardqmk2q-clang-wrapper-13.0.0/resource-root/include
/nix/store/sr2bm5pp15dhb80w16vrwyhgh40kd4yf-glibc-2.33-59-dev/include
End of search list.
"/nix/store/kcg1vn8a2xf0fzd789fq16mww4wqvg0n-llvm-binutils-wrapper-13.0.0/bin/ld.lld" -z relro --hash-style=gnu --hash-style=both --enable-new-dtags --eh-frame-hdr -m elf_x86_64 -o /dev/null /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crt1.o /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crti.o /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0/crtbegin.o -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/lhimspyyv7cg448gzdjzin05jm4xisp9-lldb-13.0.0-lib/lib -L/nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0 -L/nix/store/lwkpcfphv5nwymi0bvnvfga6q4p72v8i-gcc-10.3.0-lib/x86_64-unknown-linux-gnu/lib -L/nix/store/svzwv5yhcp03rysmc0lh5922a6f7qgp8-clang-13.0.0-lib/lib -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0 -L/nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib64/gcc/x86_64-unknown-linux-gnu/10.3.0/../../../../lib64 -L/lib/../lib64 -L/usr/lib/../lib64 -L/nix/store/qrmngdzjyvl4qdr049dwnslrh462am3z-clang-13.0.0/bin/../lib -L/lib -L/usr/lib -dynamic-linker=/nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/ld-linux-x86-64.so.2 /run/user/12064113/--158e23.o --start-lib --end-lib -rpath /nix/store/4ibh63jd2lzx03mrhjn9km8ihha9wq55-shell/lib64 -rpath /nix/store/4ibh63jd2lzx03mrhjn9km8ihha9wq55-shell/lib -lstdc++ -lm -lgcc_s -lgcc -lc -lgcc_s -lgcc /nix/store/mlfy9ifzszg7z2q6aiblvm5qkfn3bmwb-gcc-10.3.0/lib/gcc/x86_64-unknown-linux-gnu/10.3.0/crtend.o /nix/store/s9qbqh7gzacs7h68b2jfmn9l6q4jwfjz-glibc-2.33-59/lib/crtn.o
```
</details>
## Potential Solutions
A trivial fix would be to search for the linker name plus a space or end quote (i.e. `lld `/`lld"`/`lld'`, `gold `/`gold"`/`gold'`) or, for `lld` specifically, to search for `.lld` (afaik all the flavors of `lld` *for unix platforms* are `<something>.lld`: `ld64.lld`, `ld.lld`). This _seems_ like it'd fix this issue without relying on compiler-specific output.
Another potential solution is to only use the _last_ line of the verbose output. However, I am not sure we can rely on this always being the linker invocation, especially for compilers that aren't `clang`.
Ideally we'd just ask the compiler for the path to the linker but I don't know of any way to do so that's well supported across different compilers.
If any of these solutions are deemed a reasonable fix (or if a better path forward is proposed), I'm happy to make a PR.
## Workarounds
To anyone facing a similar issue, my workaround for now was to simply wrap the `lldb` derivation and have it use another name:
```nix
with nixpkgs; stdenvNoCC.mkDerivation {
pname = "llvm-debugger";
inherit (lldb) version outputs;
lldb_outputs = lldb.all;
src = "${lldb}";
disallowedReferences = [lldb];
buildPhase = ''
i=0
lldb_outs=($lldb_outputs)
outs=($outputs)
for o in "''${outs[@]}"; do
echo "[$o]" "''${lldb_outs[$i]}" to "''${!o}"
# can't symlink because then the paths used for `LDFLAGS` still
# contain "lldb":
cp -R "''${lldb_outs[$i]}" "''${!o}"
: $((++i))
done
# Remove some lingering references to the actual lldb package so that it's
# not registered as a runtime dep of this package.
scrub() {
local i=0
for o in "''${outs[@]}"; do
substituteInPlace "$1" \
--replace "''${lldb_outs[$i]}" "''${!o}"
: $((++i))
done
}
scrub "$out/bin/lldb"
scrub "$dev/nix-support/propagated-build-inputs"
'';
installPhase = "true";
}
# Alternatively the below will work too but will cause you to rebuild `lldb` from source:
# nixpkgs.lldb.overrideDerivation (o: { name = "llvm-debugger"; })
```
|
process
|
bazel tools s unix cc configure bzl breaks when other link opts contain lld problem when creating unix cc configure bzl s is used to parse to figure out if and what their paths are the parsing essentially looks for the in the in the verbose output and then the is essentially cxx xc int main o dev null wl start lib wl end lib fuse ld linker all together that s essentially bash cxx xc int main o dev null wl start lib wl end lib fuse ld linker v grep linker xargs n grep linker tr d head the problem is that this can potentially isolate things other than the path to the linker i ran into this issue when using a with lldb present nix by default its compilers including the compiler it exposes as cc and cxx which bazel s cc toolchain autoconfiguration picks up by default to make local config cc in a shell script that adds the contents of nix ldflags and other env vars to the actual compiler invocation in its verbose output clang prints out the actual compiler invocation find linker path interprets the l include that nix s wrapper adds to the invocation for lldb s library files to be the linker path this manifests in compile errors on cc binary s that look like info found target error build bazel linking failed exit clang failed error executing command nix store clang wrapper bin clang bazel out host bin external some target params use sandbox debug to see verbose messages from the sandbox clang error invalid linker name in argument fuse ld nix store lldb dev include to reproduce using a nix shell in nix shell p bazelisk p lldb p llvmpackages bintools p llvmpackages clang bash cd mktemp d touch workspace echo int main return main cc echo bazelversion cat build bazel cc binary name test srcs eof export bazel use cpp only toolchain bazelisk build action env cc clang action env cxx clang test s note the use of bazelisk to get bazel introduces the use of lld in autoconfiguration and it landed in bazel also note that this is currently not reproducible on macos on because of a that causes it to error when attempting to sign the output file dev null thus lld and gold are just not used by bazel in these cases since the compiler invocation fails the verbose output alternatively here is the output of clang xc int main o dev null wl start lib wl end lib fuse ld lld v in such a shell click to expand grep for lld clang version target unknown linux gnu thread model posix installeddir nix store clang bin found candidate gcc installation nix store gcc lib gcc unknown linux gnu found candidate gcc installation nix store gcc gcc unknown linux gnu selected gcc installation nix store gcc gcc unknown linux gnu candidate multilib selected multilib nix store clang bin clang triple unknown linux gnu emit obj mrelax relocations disable free disable llvm verifier discard value names main file name mrelocation model pic pic level fhalf no semantic interposition mframe pointer none fmath errno fno rounding math mconstructor aliases munwind tables target cpu tune cpu generic debugger tuning gdb v fcoverage compilation dir run user tmp nostdsysteminc resource dir nix store clang wrapper resource root idirafter nix store glibc dev include isystem nix store lldb dev include isystem nix store compiler rt libc dev include isystem nix store lldb dev include isystem nix store compiler rt libc dev include isystem nix store lldb dev include isystem nix store compiler rt libc dev include isystem nix store gcc include c isystem nix store gcc include c unknown linux gnu d fortify source internal isystem nix store gcc gcc unknown linux gnu include c internal isystem nix store gcc gcc unknown linux gnu include c unknown linux gnu internal isystem nix store gcc gcc unknown linux gnu include c backward internal isystem nix store clang wrapper resource root include wformat wformat security werror format security fdeprecated macro fdebug compilation dir run user tmp ferror limit fwrapv stack protector stack protector buffer size fgnuc version fcxx exceptions fexceptions fcolor diagnostics vectorize loops vectorize slp faddrsig d gcc have cfi asm o run user o x c clang version based upon llvm default target unknown linux gnu ignoring duplicate directory nix store lldb dev include ignoring duplicate directory nix store compiler rt libc dev include ignoring duplicate directory nix store lldb dev include ignoring duplicate directory nix store compiler rt libc dev include ignoring duplicate directory nix store gcc include c ignoring duplicate directory nix store gcc include c unknown linux gnu include search starts here include search starts here nix store lldb dev include nix store compiler rt libc dev include nix store gcc include c nix store gcc include c unknown linux gnu nix store gcc gcc unknown linux gnu include c backward nix store clang wrapper resource root include nix store glibc dev include end of search list nix store llvm binutils wrapper bin ld lld z relro hash style gnu hash style both enable new dtags eh frame hdr m elf o dev null nix store glibc lib o nix store glibc lib crti o nix store gcc lib gcc unknown linux gnu crtbegin o l nix store lldb lib lib l nix store lldb lib lib l nix store lldb lib lib l nix store lldb lib lib l nix store glibc lib l nix store gcc lib gcc unknown linux gnu l nix store gcc lib unknown linux gnu lib l nix store clang lib lib l nix store gcc gcc unknown linux gnu l nix store gcc gcc unknown linux gnu l lib l usr lib l nix store clang bin lib l lib l usr lib dynamic linker nix store glibc lib ld linux so run user o start lib end lib rpath nix store shell rpath nix store shell lib lstdc lm lgcc s lgcc lc lgcc s lgcc nix store gcc lib gcc unknown linux gnu crtend o nix store glibc lib crtn o potential solutions a trivial fix would be to search for the linker name plus a space or end quote i e lld lld lld gold gold gold or for lld specifically to search for lld afaik all the flavors of lld for unix platforms are lld lld ld lld this seems like it d fix this issue without relying on compiler specific output another potential solution is to only use the last line of the verbose output however i am not sure we can rely on this always being the linker invocation especially for compilers that aren t clang ideally we d just ask the compiler for the path to the linker but i don t know of any way to do so that s well supported across different compilers if any of these solutions are deemed a reasonable fix or if a better path forward is proposed i m happy to make a pr workarounds to anyone facing a similar issue my workaround for now was to simply wrap the lldb derivation and have it use another name nix with nixpkgs stdenvnocc mkderivation pname llvm debugger inherit lldb version outputs lldb outputs lldb all src lldb disallowedreferences buildphase i lldb outs lldb outputs outs outputs for o in outs do echo lldb outs to o can t symlink because then the paths used for ldflags still contain lldb cp r lldb outs o i done remove some lingering references to the actual lldb package so that it s not registered as a runtime dep of this package scrub local i for o in outs do substituteinplace replace lldb outs o i done scrub out bin lldb scrub dev nix support propagated build inputs installphase true alternatively the below will work too but will cause you to rebuild lldb from source nixpkgs lldb overridederivation o name llvm debugger
| 1
|
11,181
| 13,957,695,473
|
IssuesEvent
|
2020-10-24 08:11:33
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
RO: new harvesting
|
Geoportal Harvesting process RO - Romania
|
Dear Angelo,
Can we have a new harvesting of our national geoportal, please?
We've made some changes in order to link the resources.
Best regards,
Simona Bunea
|
1.0
|
RO: new harvesting - Dear Angelo,
Can we have a new harvesting of our national geoportal, please?
We've made some changes in order to link the resources.
Best regards,
Simona Bunea
|
process
|
ro new harvesting dear angelo can we have a new harvesting of our national geoportal please we ve made some changes in order to link the resources best regards simona bunea
| 1
|
19,647
| 26,006,252,622
|
IssuesEvent
|
2022-12-20 19:41:13
|
hashgraph/hedera-mirror-node
|
https://api.github.com/repos/hashgraph/hedera-mirror-node
|
closed
|
Release automation set wrong VERSION
|
bug regression process
|
### Description
The recent changes to release automation github workflow causes wrong version set and the subsequent issues
- other workflows triggered by the new tag fails because the version is `$VERSION`
- the create SNAPSHOT PR job fails for the same reason
### Steps to reproduce
Check the release automation workflow run for 0.71.0-beta1 and the corresponding triggered workflow runs
### Additional context
_No response_
### Hedera network
other
### Version
0.71.0-SNAPSHOT
### Operating system
None
|
1.0
|
Release automation set wrong VERSION - ### Description
The recent changes to release automation github workflow causes wrong version set and the subsequent issues
- other workflows triggered by the new tag fails because the version is `$VERSION`
- the create SNAPSHOT PR job fails for the same reason
### Steps to reproduce
Check the release automation workflow run for 0.71.0-beta1 and the corresponding triggered workflow runs
### Additional context
_No response_
### Hedera network
other
### Version
0.71.0-SNAPSHOT
### Operating system
None
|
process
|
release automation set wrong version description the recent changes to release automation github workflow causes wrong version set and the subsequent issues other workflows triggered by the new tag fails because the version is version the create snapshot pr job fails for the same reason steps to reproduce check the release automation workflow run for and the corresponding triggered workflow runs additional context no response hedera network other version snapshot operating system none
| 1
|
14,582
| 17,703,494,332
|
IssuesEvent
|
2021-08-25 03:08:39
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
New Term - MaterialCitation
|
Term - add Class - new normative Process - complete
|
## New term
* Submitter: Plazi (Donat Agosti, @myrmoteras)
* Justification (why is this term necessary?): Specimens are the basis for taxonomic research and are cited in taxonomic treatments and other works. Increasingly these material citations are extracted from publications and submitted as part of data sets to GBIF and reused in studies. Currently GBIF includes 33,199 datasets derived from taxonomic publications and 408,021 material citations which are labeled as occurrences. An estimate of 45,000 species are only represented in GBIF through material citations from publications, mainly covering new species that are extracted from the literature just after the publishing time and thus cut short the time to enter taxonomic names and re-submit datasets to GBIF. This leads to confusion and discussions (https://discourse.gbif.org/t/basisofrecord-for-plazi-datasets/2238/16 ) which need to be resolved, not least because material citations are not specimens per se, but the citation of a specimen. Furthermore, material citations can be part of a specimen, a specimen, or groups of specimens. They can be very verbose or very cursory.
* Proponents (at least two independent parties who need this term).
Demand requirement. There are four main groups: the publishers of taxonomic and other scholarly publications; the data converters; the re-users of the data, e.g. data aggregators and scientists analysing data.
Three publishers already make use of material citations: Pensoft GmbH, the Science Press of the Muséum d’histoire naturelle Paris, and the CETAF publishing group.
As a data converter, Plazi so far produces [400,000 material citations](https://www.gbif.org/publisher/7ce8aef0-9e92-11dc-8738-b8a03c50a862) from 33,000 publications and submits them to GBIF. This includes data from xx scientific journals. GBIF is importing material citations as occurrence data and would like to separate these data from other occurrence data.
Scientists used data originating from taxonomic publications in over 320 scientific articles.
Furthermore there is a demand raised by GBIF users to separate material citations from specimen occurrence data (see above).
There is clearly a demand to introduce this new class.
Efficacy requirement: Within this community there is a consensus that this new class will accomplish the desired outcome. The equivalent term material-citation (https://terms.tdwg.org/wiki/tp:material-citation) in the TaxPub Journal Article Tag Suit is already used in the production of over 30 scholarly journals (eg. http://plazi.org/resources/schemas-and-ontologies/taxpub/ ) .
Stability requirement: Since this is a new class, this will not interfere with existing implementations but rather contribute to resolve a well known issue.
Proposed attributes of the new term:
* Term name (in lowerCamelCase): MaterialCitation (modified by @tucotuco)
* Class (e.g. Location, Taxon): None (modified by @tucotuco - the proposal is for a Class, not a property)
* Definition of the term: A reference to or citation of one, a part of, or multiple specimens in scholarly publications. (modified by @tucotuco)
* Usage comments (recommendations regarding content, etc.): This class constitutes a new value for the controlled vocabulary in the recommendations for basisOfRecord. When importing Darwin Core Archives of literature-based datasets to GBIF, the basisOfRecord should be changed from “Occurrence”, "PreservedSpecimen" or "Literature" to “MaterialCitation”.
* Examples:
* the citation of a physical specimen from a scientific collection in a taxonomic treatment in a scientific publication.
* the citation of a group of physical specimens, such as paratypes in a taxonomic treatment in a scientific publication.
* the occurrence mentioned in a field note book
* Refines (identifier of the broader term this term refines, if applicable): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD, if applicable): Not in ABCD.
BioCASe/ABCD provides for a slightly different set of values:
```
<xs:enumeration value="PreservedSpecimen"/>
<xs:enumeration value="LivingSpecimen"/>
<xs:enumeration value="FossileSpecimen"/>
<xs:enumeration value="OtherSpecimen"/>
<xs:enumeration value="HumanObservation"/>
<xs:enumeration value="MachineObservation"/>
<xs:enumeration value="DrawingOrPhotograph"/>
<xs:enumeration value="MultimediaObject"/>
<xs:enumeration value="AbsenceObservation"/>
```
|
1.0
|
New Term - MaterialCitation - ## New term
* Submitter: Plazi (Donat Agosti, @myrmoteras)
* Justification (why is this term necessary?): Specimens are the basis for taxonomic research and are cited in taxonomic treatments and other works. Increasingly these material citations are extracted from publications and submitted as part of data sets to GBIF and reused in studies. Currently GBIF includes 33,199 datasets derived from taxonomic publications and 408,021 material citations which are labeled as occurrences. An estimate of 45,000 species are only represented in GBIF through material citations from publications, mainly covering new species that are extracted from the literature just after the publishing time and thus cut short the time to enter taxonomic names and re-submit datasets to GBIF. This leads to confusion and discussions (https://discourse.gbif.org/t/basisofrecord-for-plazi-datasets/2238/16 ) which need to be resolved, not least because material citations are not specimens per se, but the citation of a specimen. Furthermore, material citations can be part of a specimen, a specimen, or groups of specimens. They can be very verbose or very cursory.
* Proponents (at least two independent parties who need this term).
Demand requirement. There are four main groups: the publishers of taxonomic and other scholarly publications; the data converters; the re-users of the data, e.g. data aggregators and scientists analysing data.
Three publishers already make use of material citations: Pensoft GmbH, the Science Press of the Muséum d’histoire naturelle Paris, and the CETAF publishing group.
As a data converter, Plazi so far produces [400,000 material citations](https://www.gbif.org/publisher/7ce8aef0-9e92-11dc-8738-b8a03c50a862) from 33,000 publications and submits them to GBIF. This includes data from xx scientific journals. GBIF is importing material citations as occurrence data and would like to separate these data from other occurrence data.
Scientists used data originating from taxonomic publications in over 320 scientific articles.
Furthermore there is a demand raised by GBIF users to separate material citations from specimen occurrence data (see above).
There is clearly a demand to introduce this new class.
Efficacy requirement: Within this community there is a consensus that this new class will accomplish the desired outcome. The equivalent term material-citation (https://terms.tdwg.org/wiki/tp:material-citation) in the TaxPub Journal Article Tag Suit is already used in the production of over 30 scholarly journals (eg. http://plazi.org/resources/schemas-and-ontologies/taxpub/ ) .
Stability requirement: Since this is a new class, this will not interfere with existing implementations but rather contribute to resolve a well known issue.
Proposed attributes of the new term:
* Term name (in lowerCamelCase): MaterialCitation (modified by @tucotuco)
* Class (e.g. Location, Taxon): None (modified by @tucotuco - the proposal is for a Class, not a property)
* Definition of the term: A reference to or citation of one, a part of, or multiple specimens in scholarly publications. (modified by @tucotuco)
* Usage comments (recommendations regarding content, etc.): This class constitutes a new value for the controlled vocabulary in the recommendations for basisOfRecord. When importing Darwin Core Archives of literature-based datasets to GBIF, the basisOfRecord should be changed from “Occurrence”, "PreservedSpecimen" or "Literature" to “MaterialCitation”.
* Examples:
* the citation of a physical specimen from a scientific collection in a taxonomic treatment in a scientific publication.
* the citation of a group of physical specimens, such as paratypes in a taxonomic treatment in a scientific publication.
* the occurrence mentioned in a field note book
* Refines (identifier of the broader term this term refines, if applicable): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD, if applicable): Not in ABCD.
BioCASe/ABCD provides for a slightly different set of values:
```
<xs:enumeration value="PreservedSpecimen"/>
<xs:enumeration value="LivingSpecimen"/>
<xs:enumeration value="FossileSpecimen"/>
<xs:enumeration value="OtherSpecimen"/>
<xs:enumeration value="HumanObservation"/>
<xs:enumeration value="MachineObservation"/>
<xs:enumeration value="DrawingOrPhotograph"/>
<xs:enumeration value="MultimediaObject"/>
<xs:enumeration value="AbsenceObservation"/>
```
|
process
|
new term materialcitation new term submitter plazi donat agosti myrmoteras justification why is this term necessary specimens are the basis for taxonomic research and are cited in taxonomic treatments and other works increasingly these material citations are extracted from publications and submitted as part of data sets to gbif and reused in studies currently gbif includes datasets derived from taxonomic publications and material citations which are labeled as occurrences an estimate of species are only represented in gbif through material citations from publications mainly covering new species that are extracted from the literature just after the publishing time and thus cut short the time to enter taxonomic names and re submit datasets to gbif this leads to confusion and discussions which need to be resolved not least because material citations are not specimens per se but the citation of a specimen furthermore material citations can be part of a specimen a specimen or groups of specimens they can be very verbose or very cursory proponents at least two independent parties who need this term demand requirement there are four main groups the publishers of taxonomic and other scholarly publications the data converters the re users of the data e g data aggregators and scientists analysing data three publishers already make use of material citations pensoft gmbh the science press of the muséum d’histoire naturelle paris and the cetaf publishing group as a data converter plazi so far produces from publications and submits them to gbif this includes data from xx scientific journals gbif is importing material citations as occurrence data and would like to separate these data from other occurrence data scientists used data originating from taxonomic publications in over scientific articles furthermore there is a demand raised by gbif users to separate material citations from specimen occurrence data see above there is clearly a demand to introduce this new class efficacy requirement within this community there is a consensus that this new class will accomplish the desired outcome the equivalent term material citation in the taxpub journal article tag suit is already used in the production of over scholarly journals eg stability requirement since this is a new class this will not interfere with existing implementations but rather contribute to resolve a well known issue proposed attributes of the new term term name in lowercamelcase materialcitation modified by tucotuco class e g location taxon none modified by tucotuco the proposal is for a class not a property definition of the term a reference to or citation of one a part of or multiple specimens in scholarly publications modified by tucotuco usage comments recommendations regarding content etc this class constitutes a new value for the controlled vocabulary in the recommendations for basisofrecord when importing darwin core archives of literature based datasets to gbif the basisofrecord should be changed from “occurrence” preservedspecimen or literature to “materialcitation” examples the citation of a physical specimen from a scientific collection in a taxonomic treatment in a scientific publication the citation of a group of physical specimens such as paratypes in a taxonomic treatment in a scientific publication the occurrence mentioned in a field note book refines identifier of the broader term this term refines if applicable none replaces identifier of the existing term that would be deprecated and replaced by this term if applicable none abcd xpath of the equivalent term in abcd if applicable not in abcd biocase abcd provides for a slightly different set of values
| 1
|
556,206
| 16,477,594,493
|
IssuesEvent
|
2021-05-24 07:46:28
|
edwisely-ai/Tech-Bridge
|
https://api.github.com/repos/edwisely-ai/Tech-Bridge
|
opened
|
Dark Mode Issue the App (Student)
|
Criticality High Priority High
|
whenever dark mode is activated the students cannot see the Questions.
This is a serious issue in the Front End while Writing exams, This should be a Priority and must be resolved immediately
Note: The Thing Should be Checked with Faculty App
|
1.0
|
Dark Mode Issue the App (Student) - whenever dark mode is activated the students cannot see the Questions.
This is a serious issue in the Front End while Writing exams, This should be a Priority and must be resolved immediately
Note: The Thing Should be Checked with Faculty App
|
non_process
|
dark mode issue the app student whenever dark mode is activated the students cannot see the questions this is a serious issue in the front end while writing exams this should be a priority and must be resolved immediately note the thing should be checked with faculty app
| 0
|
65,464
| 8,815,302,082
|
IssuesEvent
|
2018-12-29 16:50:59
|
KoolTheba/arcade-game-clone
|
https://api.github.com/repos/KoolTheba/arcade-game-clone
|
opened
|
Basic functionality
|
:book: documentation
|
In this game you have a Player and Enemies (bugs). The goal of the player is to reach the water, without colliding into any one of the enemies.
- [ ] The player can move left, right, up and down
- [ ] The enemies move at varying speeds on the paved block portion of the game board
- [ ] Once a the player collides with an enemy, the game is reset and the player moves back to the starting square
- [ ] Once the player reaches the water (i.e., the top of the game board), the game is won
|
1.0
|
Basic functionality - In this game you have a Player and Enemies (bugs). The goal of the player is to reach the water, without colliding into any one of the enemies.
- [ ] The player can move left, right, up and down
- [ ] The enemies move at varying speeds on the paved block portion of the game board
- [ ] Once a the player collides with an enemy, the game is reset and the player moves back to the starting square
- [ ] Once the player reaches the water (i.e., the top of the game board), the game is won
|
non_process
|
basic functionality in this game you have a player and enemies bugs the goal of the player is to reach the water without colliding into any one of the enemies the player can move left right up and down the enemies move at varying speeds on the paved block portion of the game board once a the player collides with an enemy the game is reset and the player moves back to the starting square once the player reaches the water i e the top of the game board the game is won
| 0
|
15,835
| 20,022,557,495
|
IssuesEvent
|
2022-02-01 17:43:55
|
EKGF/ekg-mm
|
https://api.github.com/repos/EKGF/ekg-mm
|
closed
|
Github Kanban is a linear process?
|
ekg-mm-process
|
@jgeluk
You are interpreting Kanban your own unique way and with this you are
reinventing the OODA loop.
Look it up.
I have been using and teaching this stuff for 40 years all around the world
and I know how it works.
GitHub is just another implementation of Kanban. And now you want still
another implementation. Kanban is a linear process.
OODA is a linear process.
Both have decades of research and experienced practitioners. Why would we
not want to ride on their coat tails?
Sincerely,
_Originally posted by @DennisWisnosky in https://github.com/EKGF/ekg-mm/issues/16#issuecomment-754864156_
|
1.0
|
Github Kanban is a linear process? - @jgeluk
You are interpreting Kanban your own unique way and with this you are
reinventing the OODA loop.
Look it up.
I have been using and teaching this stuff for 40 years all around the world
and I know how it works.
GitHub is just another implementation of Kanban. And now you want still
another implementation. Kanban is a linear process.
OODA is a linear process.
Both have decades of research and experienced practitioners. Why would we
not want to ride on their coat tails?
Sincerely,
_Originally posted by @DennisWisnosky in https://github.com/EKGF/ekg-mm/issues/16#issuecomment-754864156_
|
process
|
github kanban is a linear process jgeluk you are interpreting kanban your own unique way and with this you are reinventing the ooda loop look it up i have been using and teaching this stuff for years all around the world and i know how it works github is just another implementation of kanban and now you want still another implementation kanban is a linear process ooda is a linear process both have decades of research and experienced practitioners why would we not want to ride on their coat tails sincerely originally posted by denniswisnosky in
| 1
|
11,837
| 14,656,376,218
|
IssuesEvent
|
2020-12-28 13:17:17
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
closed
|
window.location constructor properties shadowed incorrectly
|
AREA: client FREQUENCY: level 1 SYSTEM: client side processing TYPE: bug
|
Sample code:
```
console.log('-1->', window.location.constructor.name);
console.log('-2->', window.location.constructor.toString());
console.log('-3->', Function.prototype.toString.apply(window.location.constructor));
console.log('-4->', window.location instanceof Location);
```
When run in a real browser:
-1-> Location
-2-> function Location() { [native code] }
-3-> function Location() { [native code] }
-4-> true
When run under Testcafe:
-1-> g
-2-> function(e,t,r){var n=this ...
-3-> function(e,t,r){var n=this ...
-4-> false
Issue is that many websites (e.g. Google login) use a check like this as part of botguard and change their behaviour as a result.
Can testcafe-hammerhead be improved to "stealth" these changes?
|
1.0
|
window.location constructor properties shadowed incorrectly - Sample code:
```
console.log('-1->', window.location.constructor.name);
console.log('-2->', window.location.constructor.toString());
console.log('-3->', Function.prototype.toString.apply(window.location.constructor));
console.log('-4->', window.location instanceof Location);
```
When run in a real browser:
-1-> Location
-2-> function Location() { [native code] }
-3-> function Location() { [native code] }
-4-> true
When run under Testcafe:
-1-> g
-2-> function(e,t,r){var n=this ...
-3-> function(e,t,r){var n=this ...
-4-> false
Issue is that many websites (e.g. Google login) use a check like this as part of botguard and change their behaviour as a result.
Can testcafe-hammerhead be improved to "stealth" these changes?
|
process
|
window location constructor properties shadowed incorrectly sample code console log window location constructor name console log window location constructor tostring console log function prototype tostring apply window location constructor console log window location instanceof location when run in a real browser location function location function location true when run under testcafe g function e t r var n this function e t r var n this false issue is that many websites e g google login use a check like this as part of botguard and change their behaviour as a result can testcafe hammerhead be improved to stealth these changes
| 1
|
41,899
| 10,821,362,696
|
IssuesEvent
|
2019-11-08 18:29:40
|
tensorflow/tensorflow
|
https://api.github.com/repos/tensorflow/tensorflow
|
closed
|
Tensorflow
|
subtype:centos type:build/install
|
<em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Scientific Linux 7.6, but also tested on Centos 7.6
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: NA
- TensorFlow installed from (source or binary): source
- TensorFlow version: 1.12.0
- Python version: 2.6
- Installed using virtualenv? pip? conda?: manually compiling using bazel
- Bazel version (if compiling from source): 1.16.1, but also tried 1.15.0 and 1.18.0
- GCC/Compiler version (if compiling from source): 4.8.5
- CUDA/cuDNN version: Cuda 9 CuDNN 7
- GPU model and memory: Compute Cluster: multpile Tesla K80 (24GB) or Tesla P100 (16GB).
**Describe the problem**
During compilation I'm encountering the following ERROR messages:
[205 / 208] Compiling tensorflow/contrib/image/kernels/adjust_hsv_in_yiq_op_gpu.cu.cc; 10s local
ERROR: /tmp/Tensorflow/PACKAGES/tensorflow/tensorflow/contrib/image/BUILD:115:1: undeclared inclusion(s) in rule '//tensorflow/contrib/image:python/ops/_distort_image_ops_gpu':
this rule is missing dependency declarations for the following files included by 'tensorflow/contrib/image/kernels/adjust_hsv_in_yiq_op_gpu.cu.cc':
'/tmp/bazel/userid_output/external/eigen_archive/Eigen/Core'
'/tmp/bazel/userid_output/external/eigen_archive/Eigen/src/Core/util/DisableStupidWarnings.h'
'/tmp/bazel/userid_output/external/eigen_archive/Eigen/src/Core/util/ReenableStupidWarnings.h
However, when I add ---verbose_failures to bazel and execute the failing compilation job manually to debug the issue, the compilation is executed correctly. So I'm assuming only some check of bazel is failing, although the compilation itself could be executed correctly.
Restarting the build (after my manually compilation) will also continue, until the next job where Eigen is used again. The build is failing with a similar error. Previous versions, including 1.11.0, are not encountering this issue. I could not find any noticeable changes in the Eigen build files for bazel or related to Eigen. Therefore I'm currently out of ideas.....
**Provide the exact sequence of commands / steps that you executed before running into the problem**
bazel --output_base=/tmp/bazel/userid_output build --jobs 12 -c opt --copt=-mavx2 --copt=-O --copt=-msse4.2 --copt=-mfma -config=cuda //tensorflow/tools/pip_package:build_pip_package
(I'm using bazel output_base because my home directory is on NFS)
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
1.0
|
Tensorflow - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Scientific Linux 7.6, but also tested on Centos 7.6
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: NA
- TensorFlow installed from (source or binary): source
- TensorFlow version: 1.12.0
- Python version: 2.6
- Installed using virtualenv? pip? conda?: manually compiling using bazel
- Bazel version (if compiling from source): 1.16.1, but also tried 1.15.0 and 1.18.0
- GCC/Compiler version (if compiling from source): 4.8.5
- CUDA/cuDNN version: Cuda 9 CuDNN 7
- GPU model and memory: Compute Cluster: multpile Tesla K80 (24GB) or Tesla P100 (16GB).
**Describe the problem**
During compilation I'm encountering the following ERROR messages:
[205 / 208] Compiling tensorflow/contrib/image/kernels/adjust_hsv_in_yiq_op_gpu.cu.cc; 10s local
ERROR: /tmp/Tensorflow/PACKAGES/tensorflow/tensorflow/contrib/image/BUILD:115:1: undeclared inclusion(s) in rule '//tensorflow/contrib/image:python/ops/_distort_image_ops_gpu':
this rule is missing dependency declarations for the following files included by 'tensorflow/contrib/image/kernels/adjust_hsv_in_yiq_op_gpu.cu.cc':
'/tmp/bazel/userid_output/external/eigen_archive/Eigen/Core'
'/tmp/bazel/userid_output/external/eigen_archive/Eigen/src/Core/util/DisableStupidWarnings.h'
'/tmp/bazel/userid_output/external/eigen_archive/Eigen/src/Core/util/ReenableStupidWarnings.h
However, when I add ---verbose_failures to bazel and execute the failing compilation job manually to debug the issue, the compilation is executed correctly. So I'm assuming only some check of bazel is failing, although the compilation itself could be executed correctly.
Restarting the build (after my manually compilation) will also continue, until the next job where Eigen is used again. The build is failing with a similar error. Previous versions, including 1.11.0, are not encountering this issue. I could not find any noticeable changes in the Eigen build files for bazel or related to Eigen. Therefore I'm currently out of ideas.....
**Provide the exact sequence of commands / steps that you executed before running into the problem**
bazel --output_base=/tmp/bazel/userid_output build --jobs 12 -c opt --copt=-mavx2 --copt=-O --copt=-msse4.2 --copt=-mfma -config=cuda //tensorflow/tools/pip_package:build_pip_package
(I'm using bazel output_base because my home directory is on NFS)
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
non_process
|
tensorflow please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution e g linux ubuntu scientific linux but also tested on centos mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device na tensorflow installed from source or binary source tensorflow version python version installed using virtualenv pip conda manually compiling using bazel bazel version if compiling from source but also tried and gcc compiler version if compiling from source cuda cudnn version cuda cudnn gpu model and memory compute cluster multpile tesla or tesla describe the problem during compilation i m encountering the following error messages compiling tensorflow contrib image kernels adjust hsv in yiq op gpu cu cc local error tmp tensorflow packages tensorflow tensorflow contrib image build undeclared inclusion s in rule tensorflow contrib image python ops distort image ops gpu this rule is missing dependency declarations for the following files included by tensorflow contrib image kernels adjust hsv in yiq op gpu cu cc tmp bazel userid output external eigen archive eigen core tmp bazel userid output external eigen archive eigen src core util disablestupidwarnings h tmp bazel userid output external eigen archive eigen src core util reenablestupidwarnings h however when i add verbose failures to bazel and execute the failing compilation job manually to debug the issue the compilation is executed correctly so i m assuming only some check of bazel is failing although the compilation itself could be executed correctly restarting the build after my manually compilation will also continue until the next job where eigen is used again the build is failing with a similar error previous versions including are not encountering this issue i could not find any noticeable changes in the eigen build files for bazel or related to eigen therefore i m currently out of ideas provide the exact sequence of commands steps that you executed before running into the problem bazel output base tmp bazel userid output build jobs c opt copt copt o copt copt mfma config cuda tensorflow tools pip package build pip package i m using bazel output base because my home directory is on nfs any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached
| 0
|
685,152
| 23,445,481,238
|
IssuesEvent
|
2022-08-15 19:09:28
|
UniversityDAO/udao
|
https://api.github.com/repos/UniversityDAO/udao
|
closed
|
Refresh Issue
|
priority
|
Right now, as of the reading dataflow rework, we only load all the data onto the page ONCE and that's as soon as a user presses "Launch App" on the landing page. Once they do so, it directs them to a "Loading" page where we then fetch all the data. Once the data is fetched and done loading, we automatically redirect to the dashboard page.
The problem with this is what if we're on the dashboard/grants/proposals pages, and the user refreshes the page, the page ends up being empty. It's empty because once the page refreshes the entire app has to re-render, and once it re-renders the user will be left on whatever page they refreshed on, meaning they essentially 'bypassed' the launch app button. This is a bit of an issue.
One solution would be to have an event listener for page refresh, and upon refresh we redirect them to the loading page. where data will be loaded up. The issue with this solution is that if somebody is on the grants page and refreshes, once the data loads back in they're going to be on the dashboard page. Not an app-breaking problem but still annoying from a user standpoint.
|
1.0
|
Refresh Issue - Right now, as of the reading dataflow rework, we only load all the data onto the page ONCE and that's as soon as a user presses "Launch App" on the landing page. Once they do so, it directs them to a "Loading" page where we then fetch all the data. Once the data is fetched and done loading, we automatically redirect to the dashboard page.
The problem with this is what if we're on the dashboard/grants/proposals pages, and the user refreshes the page, the page ends up being empty. It's empty because once the page refreshes the entire app has to re-render, and once it re-renders the user will be left on whatever page they refreshed on, meaning they essentially 'bypassed' the launch app button. This is a bit of an issue.
One solution would be to have an event listener for page refresh, and upon refresh we redirect them to the loading page. where data will be loaded up. The issue with this solution is that if somebody is on the grants page and refreshes, once the data loads back in they're going to be on the dashboard page. Not an app-breaking problem but still annoying from a user standpoint.
|
non_process
|
refresh issue right now as of the reading dataflow rework we only load all the data onto the page once and that s as soon as a user presses launch app on the landing page once they do so it directs them to a loading page where we then fetch all the data once the data is fetched and done loading we automatically redirect to the dashboard page the problem with this is what if we re on the dashboard grants proposals pages and the user refreshes the page the page ends up being empty it s empty because once the page refreshes the entire app has to re render and once it re renders the user will be left on whatever page they refreshed on meaning they essentially bypassed the launch app button this is a bit of an issue one solution would be to have an event listener for page refresh and upon refresh we redirect them to the loading page where data will be loaded up the issue with this solution is that if somebody is on the grants page and refreshes once the data loads back in they re going to be on the dashboard page not an app breaking problem but still annoying from a user standpoint
| 0
|
9,970
| 13,016,342,927
|
IssuesEvent
|
2020-07-26 05:57:39
|
DrFrankenstein/avpipe
|
https://api.github.com/repos/DrFrankenstein/avpipe
|
opened
|
Source: In/Out Markers
|
Source processing
|
As a user, I want to set "in" and "out" markers on a source so that I can cut out the parts I don't want.
The "out" marker should only be available for sources that have a set duration.
FFmpeg options: `-t`, `-to`, `-ss`, `-sseof`.
|
1.0
|
Source: In/Out Markers - As a user, I want to set "in" and "out" markers on a source so that I can cut out the parts I don't want.
The "out" marker should only be available for sources that have a set duration.
FFmpeg options: `-t`, `-to`, `-ss`, `-sseof`.
|
process
|
source in out markers as a user i want to set in and out markers on a source so that i can cut out the parts i don t want the out marker should only be available for sources that have a set duration ffmpeg options t to ss sseof
| 1
|
21,156
| 28,132,167,129
|
IssuesEvent
|
2023-04-01 01:31:44
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
[MLv2] You should be able to add `:expression`s as order-bys
|
Type:Bug .Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
|
`lib/order-by` currently errors if you try to use it with an expression reference.
|
1.0
|
[MLv2] You should be able to add `:expression`s as order-bys - `lib/order-by` currently errors if you try to use it with an expression reference.
|
process
|
you should be able to add expression s as order bys lib order by currently errors if you try to use it with an expression reference
| 1
|
13,879
| 16,654,715,446
|
IssuesEvent
|
2021-06-05 10:01:41
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] Responsive issue in Change password screen > UI issues
|
Bug P2 Participant manager Process: Fixed Process: Tested dev
|
Responsive issue in the Change password screen > UI issues

|
2.0
|
[PM] Responsive issue in Change password screen > UI issues - Responsive issue in the Change password screen > UI issues

|
process
|
responsive issue in change password screen ui issues responsive issue in the change password screen ui issues
| 1
|
356
| 2,794,528,579
|
IssuesEvent
|
2015-05-11 17:10:29
|
linguisticteam/resource-central
|
https://api.github.com/repos/linguisticteam/resource-central
|
closed
|
US1: Initial sketch of UI for adding resources
|
Work in Process
|
Initial very rough sketch of the UI for adding new resources.

|
1.0
|
US1: Initial sketch of UI for adding resources - Initial very rough sketch of the UI for adding new resources.

|
process
|
initial sketch of ui for adding resources initial very rough sketch of the ui for adding new resources
| 1
|
51,120
| 13,618,142,278
|
IssuesEvent
|
2020-09-23 18:05:15
|
znanstvenikumeni/competitionmanager
|
https://api.github.com/repos/znanstvenikumeni/competitionmanager
|
closed
|
User id isn't being logged for AccountSecurity/LoginFail/WrongPassword errors
|
affects competitors affects jury members affects mentors affects organisers bug security
|
As stated in the title, user ID isn't being logged for AccountSecurity/LoginFail/WrongPassword errors, which should be fixed as it disables account-level login attempt blocking and makes debugging and security harder.

|
True
|
User id isn't being logged for AccountSecurity/LoginFail/WrongPassword errors - As stated in the title, user ID isn't being logged for AccountSecurity/LoginFail/WrongPassword errors, which should be fixed as it disables account-level login attempt blocking and makes debugging and security harder.

|
non_process
|
user id isn t being logged for accountsecurity loginfail wrongpassword errors as stated in the title user id isn t being logged for accountsecurity loginfail wrongpassword errors which should be fixed as it disables account level login attempt blocking and makes debugging and security harder
| 0
|
15,405
| 5,956,115,111
|
IssuesEvent
|
2017-05-28 13:52:00
|
Framstag/libosmscout
|
https://api.github.com/repos/Framstag/libosmscout
|
closed
|
Compiler warnings in GenCoordDat.cpp
|
build enhancement help wanted
|
See https://ci.appveyor.com/project/Framstag/libosmscout/build/job/644tpk6kyk8u0q3h
```
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::FindDuplicateCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const':
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:108:16: warning: unused variable 'oldUpperLimit' [-Wunused-variable]
Id oldUpperLimit=currentUpperLimit;
^~~~~~~~~~~~~
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::StoreCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const':
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:283:19: warning: unused variable 'oldUpperLimit' [-Wunused-variable]
OSMId oldUpperLimit=currentUpperLimit;
^~~~~~~~~~~~~
```
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/44932709-compiler-warnings-in-gencoorddat-cpp?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
1.0
|
Compiler warnings in GenCoordDat.cpp - See https://ci.appveyor.com/project/Framstag/libosmscout/build/job/644tpk6kyk8u0q3h
```
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::FindDuplicateCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const':
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:108:16: warning: unused variable 'oldUpperLimit' [-Wunused-variable]
Id oldUpperLimit=currentUpperLimit;
^~~~~~~~~~~~~
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::StoreCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const':
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:283:19: warning: unused variable 'oldUpperLimit' [-Wunused-variable]
OSMId oldUpperLimit=currentUpperLimit;
^~~~~~~~~~~~~
```
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/44932709-compiler-warnings-in-gencoorddat-cpp?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
non_process
|
compiler warnings in gencoorddat cpp see c projects libosmscout libosmscout import src osmscout import gencoorddat cpp in member function bool osmscout coorddatagenerator findduplicatecoordinates const osmscout typeconfig const osmscout importparameter osmscout progress std unordered map const c projects libosmscout libosmscout import src osmscout import gencoorddat cpp warning unused variable oldupperlimit id oldupperlimit currentupperlimit c projects libosmscout libosmscout import src osmscout import gencoorddat cpp in member function bool osmscout coorddatagenerator storecoordinates const osmscout typeconfig const osmscout importparameter osmscout progress std unordered map const c projects libosmscout libosmscout import src osmscout import gencoorddat cpp warning unused variable oldupperlimit osmid oldupperlimit currentupperlimit want to back this issue we accept bounties via
| 0
|
21,972
| 30,466,569,085
|
IssuesEvent
|
2023-07-17 10:47:35
|
mikerae/sarah-rae-illustrations
|
https://api.github.com/repos/mikerae/sarah-rae-illustrations
|
opened
|
USER STORY:Stripe: Migrate to using fully integrating Stripe
|
Could Have Epic Admin and Store Management Site Owner Order Processing & Delivery
|
As **site owner** I can **fully integrate with Stripe** so that **use Stripe's powerful invoicing, reporting and analytics**.
### Acceptance Criteria
- All CRUD functionality on products is integrated with Stripe
- All CRUD functionality on Users is integrated with Stripe
- Stripe Analytics are accessable by Site owner
### Tasks
- [ ] Migrate from cardElement to paymentElement
- [ ] Integrate Product Id:
- [ ] Integrate Price ID:
- [ ] Integrate Users with Customers:
- [ ] Integrate Users home and billing address on payments
- [ ] set up Analytics
- [ ] display analytics
|
1.0
|
USER STORY:Stripe: Migrate to using fully integrating Stripe - As **site owner** I can **fully integrate with Stripe** so that **use Stripe's powerful invoicing, reporting and analytics**.
### Acceptance Criteria
- All CRUD functionality on products is integrated with Stripe
- All CRUD functionality on Users is integrated with Stripe
- Stripe Analytics are accessable by Site owner
### Tasks
- [ ] Migrate from cardElement to paymentElement
- [ ] Integrate Product Id:
- [ ] Integrate Price ID:
- [ ] Integrate Users with Customers:
- [ ] Integrate Users home and billing address on payments
- [ ] set up Analytics
- [ ] display analytics
|
process
|
user story stripe migrate to using fully integrating stripe as site owner i can fully integrate with stripe so that use stripe s powerful invoicing reporting and analytics acceptance criteria all crud functionality on products is integrated with stripe all crud functionality on users is integrated with stripe stripe analytics are accessable by site owner tasks migrate from cardelement to paymentelement integrate product id integrate price id integrate users with customers integrate users home and billing address on payments set up analytics display analytics
| 1
|
2,907
| 5,889,722,895
|
IssuesEvent
|
2017-05-17 13:33:28
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
Changing Color on selected filter
|
enhancement inprocess
|
I want to change the color of selected filter to a different color.
For example see below. I selected filter for Run to Trial-Run-I . I would like to change color in text insider selected filter to something more visible color.

Is there any way to pass color for selected?
|
1.0
|
Changing Color on selected filter - I want to change the color of selected filter to a different color.
For example see below. I selected filter for Run to Trial-Run-I . I would like to change color in text insider selected filter to something more visible color.

Is there any way to pass color for selected?
|
process
|
changing color on selected filter i want to change the color of selected filter to a different color for example see below i selected filter for run to trial run i i would like to change color in text insider selected filter to something more visible color is there any way to pass color for selected
| 1
|
2,918
| 5,914,478,679
|
IssuesEvent
|
2017-05-22 03:04:28
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
investigate flaky sequential/test-benchmark-child-process on Windows
|
benchmark child_process test windows
|
* **Version**: v8.0.0-pre
* **Platform**: win2008r2
* **Subsystem**: test
<!-- Enter your issue details below this comment. -->
`sequential/test-benchmark-child-process ` is still failing sometimes flaky on Windows in CI. I'll open a PR to mark it as flaky. This issue is for trying to locate the problem and a solution.
When the test succeeds, it seems to take just a few seconds.
https://ci.nodejs.org/job/node-test-binary-windows/RUN_SUBSET=3,VS_VERSION=vs2015,label=win2008r2/7865/console
```console
ok 356 sequential/test-benchmark-child-process
---
duration_ms: 1.764
```
When it fails, it's a timeout.
https://ci.nodejs.org/job/node-test-binary-windows/7867/RUN_SUBSET=3,VS_VERSION=vs2015,label=win2008r2/console
```console
not ok 356 sequential/test-benchmark-child-process
---
duration_ms: 60.70
severity: fail
stack: |-
timeout
```
This would suggest a race condition or something else causing a child process to hang or something. And that might be the cause. But...
Interestingly, a stress test where the five benchmarks that this test calls were all split out into individual tests, succeeded but each test took around 30 seconds to run. Wha??!! I know! (Only other change in those tests is the `dur` option for the benchmarks was increased from 0 to 0.1. Well, that, and that this test was run on win2016 so maybe the results are completely irrelevant? I don't know.)
https://ci.nodejs.org/job/node-stress-single-test/1161/nodes=win2016/console:
```console
ok 1 sequential/test-benchmark-child-process-exec-stdout
---
duration_ms: 3.158
...
ok 2 sequential/test-benchmark-child-process-params
---
duration_ms: 36.735
...
ok 3 sequential/test-benchmark-child-process-read-ipc
---
duration_ms: 30.116
...
ok 4 sequential/test-benchmark-child-process-read
---
duration_ms: 31.844
...
ok 5 sequential/test-benchmark-child-process-spawn-echo
---
duration_ms: 31.661
```
So I'm not sure what's going on here. Maybe it can be worked out by someone more comfortable testing and debugging on Windows or someone more deeply familiar with child_process and/or our benchmarking code. @nodejs/platform-windows @nodejs/benchmarking @mscdex @cjihrig @bnoordhuis @nodejs/testing
|
1.0
|
investigate flaky sequential/test-benchmark-child-process on Windows - * **Version**: v8.0.0-pre
* **Platform**: win2008r2
* **Subsystem**: test
<!-- Enter your issue details below this comment. -->
`sequential/test-benchmark-child-process ` is still failing sometimes flaky on Windows in CI. I'll open a PR to mark it as flaky. This issue is for trying to locate the problem and a solution.
When the test succeeds, it seems to take just a few seconds.
https://ci.nodejs.org/job/node-test-binary-windows/RUN_SUBSET=3,VS_VERSION=vs2015,label=win2008r2/7865/console
```console
ok 356 sequential/test-benchmark-child-process
---
duration_ms: 1.764
```
When it fails, it's a timeout.
https://ci.nodejs.org/job/node-test-binary-windows/7867/RUN_SUBSET=3,VS_VERSION=vs2015,label=win2008r2/console
```console
not ok 356 sequential/test-benchmark-child-process
---
duration_ms: 60.70
severity: fail
stack: |-
timeout
```
This would suggest a race condition or something else causing a child process to hang or something. And that might be the cause. But...
Interestingly, a stress test where the five benchmarks that this test calls were all split out into individual tests, succeeded but each test took around 30 seconds to run. Wha??!! I know! (Only other change in those tests is the `dur` option for the benchmarks was increased from 0 to 0.1. Well, that, and that this test was run on win2016 so maybe the results are completely irrelevant? I don't know.)
https://ci.nodejs.org/job/node-stress-single-test/1161/nodes=win2016/console:
```console
ok 1 sequential/test-benchmark-child-process-exec-stdout
---
duration_ms: 3.158
...
ok 2 sequential/test-benchmark-child-process-params
---
duration_ms: 36.735
...
ok 3 sequential/test-benchmark-child-process-read-ipc
---
duration_ms: 30.116
...
ok 4 sequential/test-benchmark-child-process-read
---
duration_ms: 31.844
...
ok 5 sequential/test-benchmark-child-process-spawn-echo
---
duration_ms: 31.661
```
So I'm not sure what's going on here. Maybe it can be worked out by someone more comfortable testing and debugging on Windows or someone more deeply familiar with child_process and/or our benchmarking code. @nodejs/platform-windows @nodejs/benchmarking @mscdex @cjihrig @bnoordhuis @nodejs/testing
|
process
|
investigate flaky sequential test benchmark child process on windows version pre platform subsystem test sequential test benchmark child process is still failing sometimes flaky on windows in ci i ll open a pr to mark it as flaky this issue is for trying to locate the problem and a solution when the test succeeds it seems to take just a few seconds console ok sequential test benchmark child process duration ms when it fails it s a timeout console not ok sequential test benchmark child process duration ms severity fail stack timeout this would suggest a race condition or something else causing a child process to hang or something and that might be the cause but interestingly a stress test where the five benchmarks that this test calls were all split out into individual tests succeeded but each test took around seconds to run wha i know only other change in those tests is the dur option for the benchmarks was increased from to well that and that this test was run on so maybe the results are completely irrelevant i don t know console ok sequential test benchmark child process exec stdout duration ms ok sequential test benchmark child process params duration ms ok sequential test benchmark child process read ipc duration ms ok sequential test benchmark child process read duration ms ok sequential test benchmark child process spawn echo duration ms so i m not sure what s going on here maybe it can be worked out by someone more comfortable testing and debugging on windows or someone more deeply familiar with child process and or our benchmarking code nodejs platform windows nodejs benchmarking mscdex cjihrig bnoordhuis nodejs testing
| 1
|
66,887
| 8,973,689,742
|
IssuesEvent
|
2019-01-29 21:44:24
|
PegaSysEng/pantheon
|
https://api.github.com/repos/PegaSysEng/pantheon
|
closed
|
Update installation to include libsodium dependency
|
doc next release documentation
|
### Acceptance Criteria
Document Pantheon dependency on Orion which depends on libsodium in installation docs.
|
1.0
|
Update installation to include libsodium dependency - ### Acceptance Criteria
Document Pantheon dependency on Orion which depends on libsodium in installation docs.
|
non_process
|
update installation to include libsodium dependency acceptance criteria document pantheon dependency on orion which depends on libsodium in installation docs
| 0
|
4,154
| 7,103,700,373
|
IssuesEvent
|
2018-01-16 06:48:15
|
log2timeline/plaso
|
https://api.github.com/repos/log2timeline/plaso
|
closed
|
add preprocess plugin to detect Linux operating system information
|
enhancement preprocessing
|
- [x] add preprocess pluging to detect Linux operating system information (also see #928 and #929)
- https://github.com/ForensicArtifacts/artifacts/pull/230
- https://codereview.appspot.com/322490043
- `/etc/system-release`
```
Fedora release 26 (Twenty Six)
```
- [x] redefine processor / knowledge base to continue and use most appropriate match
- https://codereview.appspot.com/339180043
- [x] alternative `/etc/os-release`
- https://codereview.appspot.com/339180043
```
NAME=Fedora
VERSION="26 (Workstation Edition)"
ID=fedora
VERSION_ID=26
PRETTY_NAME="Fedora 26 (Workstation Edition)"
ANSI_COLOR="0;34"
CPE_NAME="cpe:/o:fedoraproject:fedora:26"
HOME_URL="https://fedoraproject.org/"
BUG_REPORT_URL="https://bugzilla.redhat.com/"
REDHAT_BUGZILLA_PRODUCT="Fedora"
REDHAT_BUGZILLA_PRODUCT_VERSION=26
REDHAT_SUPPORT_PRODUCT="Fedora"
REDHAT_SUPPORT_PRODUCT_VERSION=26
PRIVACY_POLICY_URL=https://fedoraproject.org/wiki/Legal:PrivacyPolicy
VARIANT="Workstation Edition"
VARIANT_ID=workstation
```
- [x] alternative `/etc/lsb-release`
- https://codereview.appspot.com/339180043
- [x] check if operating system information is stored
- https://codereview.appspot.com/338350043
```
***************************** System configuration *****************************
Hostname : TEST
Operating system : N/A
Operating system product : N/A
Operating system version : N/A
Code page : cp1252
Keyboard layout : N/A
Time zone : UTC
--------------------------------------------------------------------------------
```
* [x] fall back to `/etc/issue` or `/etc/issue.net` for older releases, for example
* https://github.com/ForensicArtifacts/artifacts/pull/244
* https://codereview.appspot.com/337430043/
```
Debian GNU/Linux 5.0 \n \l
```
This could be unreliable e.g. Fedora:
```
\S
Kernel \r on an \m (\l)
```
|
1.0
|
add preprocess plugin to detect Linux operating system information - - [x] add preprocess pluging to detect Linux operating system information (also see #928 and #929)
- https://github.com/ForensicArtifacts/artifacts/pull/230
- https://codereview.appspot.com/322490043
- `/etc/system-release`
```
Fedora release 26 (Twenty Six)
```
- [x] redefine processor / knowledge base to continue and use most appropriate match
- https://codereview.appspot.com/339180043
- [x] alternative `/etc/os-release`
- https://codereview.appspot.com/339180043
```
NAME=Fedora
VERSION="26 (Workstation Edition)"
ID=fedora
VERSION_ID=26
PRETTY_NAME="Fedora 26 (Workstation Edition)"
ANSI_COLOR="0;34"
CPE_NAME="cpe:/o:fedoraproject:fedora:26"
HOME_URL="https://fedoraproject.org/"
BUG_REPORT_URL="https://bugzilla.redhat.com/"
REDHAT_BUGZILLA_PRODUCT="Fedora"
REDHAT_BUGZILLA_PRODUCT_VERSION=26
REDHAT_SUPPORT_PRODUCT="Fedora"
REDHAT_SUPPORT_PRODUCT_VERSION=26
PRIVACY_POLICY_URL=https://fedoraproject.org/wiki/Legal:PrivacyPolicy
VARIANT="Workstation Edition"
VARIANT_ID=workstation
```
- [x] alternative `/etc/lsb-release`
- https://codereview.appspot.com/339180043
- [x] check if operating system information is stored
- https://codereview.appspot.com/338350043
```
***************************** System configuration *****************************
Hostname : TEST
Operating system : N/A
Operating system product : N/A
Operating system version : N/A
Code page : cp1252
Keyboard layout : N/A
Time zone : UTC
--------------------------------------------------------------------------------
```
* [x] fall back to `/etc/issue` or `/etc/issue.net` for older releases, for example
* https://github.com/ForensicArtifacts/artifacts/pull/244
* https://codereview.appspot.com/337430043/
```
Debian GNU/Linux 5.0 \n \l
```
This could be unreliable e.g. Fedora:
```
\S
Kernel \r on an \m (\l)
```
|
process
|
add preprocess plugin to detect linux operating system information add preprocess pluging to detect linux operating system information also see and etc system release fedora release twenty six redefine processor knowledge base to continue and use most appropriate match alternative etc os release name fedora version workstation edition id fedora version id pretty name fedora workstation edition ansi color cpe name cpe o fedoraproject fedora home url bug report url redhat bugzilla product fedora redhat bugzilla product version redhat support product fedora redhat support product version privacy policy url variant workstation edition variant id workstation alternative etc lsb release check if operating system information is stored system configuration hostname test operating system n a operating system product n a operating system version n a code page keyboard layout n a time zone utc fall back to etc issue or etc issue net for older releases for example debian gnu linux n l this could be unreliable e g fedora s kernel r on an m l
| 1
|
16,774
| 3,560,795,584
|
IssuesEvent
|
2016-01-23 09:52:18
|
kefirfromperm/kefirbb
|
https://api.github.com/repos/kefirfromperm/kefirbb
|
opened
|
Do tests refactoring
|
test
|
It's needed to do refactoring of tests. There are to many double codes and it is not clear which text was parsed on failure.
|
1.0
|
Do tests refactoring - It's needed to do refactoring of tests. There are to many double codes and it is not clear which text was parsed on failure.
|
non_process
|
do tests refactoring it s needed to do refactoring of tests there are to many double codes and it is not clear which text was parsed on failure
| 0
|
12,111
| 14,740,473,049
|
IssuesEvent
|
2021-01-07 09:08:43
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
SAB Error - CC Processing
|
anc-process anp-0.5 ant-bug has attachment
|
In GitLab by @kdjstudios on Nov 9, 2018, 11:56
**Submitted by:** "Arianna Screen" <arianna.screen@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-11-09-55722/conversation
**Server:** Internal
**Client/Site:** Santa Rosa
**Account:** NA
**Issue:**
We are experiencing an error with processing credit card payments in SA Billing. Please see attached.

|
1.0
|
SAB Error - CC Processing - In GitLab by @kdjstudios on Nov 9, 2018, 11:56
**Submitted by:** "Arianna Screen" <arianna.screen@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-11-09-55722/conversation
**Server:** Internal
**Client/Site:** Santa Rosa
**Account:** NA
**Issue:**
We are experiencing an error with processing credit card payments in SA Billing. Please see attached.

|
process
|
sab error cc processing in gitlab by kdjstudios on nov submitted by arianna screen helpdesk server internal client site santa rosa account na issue we are experiencing an error with processing credit card payments in sa billing please see attached uploads sab cc payment error png
| 1
|
188,486
| 14,445,999,233
|
IssuesEvent
|
2020-12-08 00:13:54
|
kalexmills/github-vet-tests-dec2020
|
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
|
closed
|
paul-lee-attorney/fabric-ca-1.4.7-gm: cmd/fabric-ca-client/command/main_test.go; 3 LoC
|
fresh test tiny
|
Found a possible issue in [paul-lee-attorney/fabric-ca-1.4.7-gm](https://www.github.com/paul-lee-attorney/fabric-ca-1.4.7-gm) at [cmd/fabric-ca-client/command/main_test.go](https://github.com/paul-lee-attorney/fabric-ca-1.4.7-gm/blob/c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578/cmd/fabric-ca-client/command/main_test.go#L152-L154)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to e at line 153 may start a goroutine
[Click here to see the code in its original context.](https://github.com/paul-lee-attorney/fabric-ca-1.4.7-gm/blob/c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578/cmd/fabric-ca-client/command/main_test.go#L152-L154)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, e := range errCases {
extraArgErrorTest(&e, t)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578
|
1.0
|
paul-lee-attorney/fabric-ca-1.4.7-gm: cmd/fabric-ca-client/command/main_test.go; 3 LoC -
Found a possible issue in [paul-lee-attorney/fabric-ca-1.4.7-gm](https://www.github.com/paul-lee-attorney/fabric-ca-1.4.7-gm) at [cmd/fabric-ca-client/command/main_test.go](https://github.com/paul-lee-attorney/fabric-ca-1.4.7-gm/blob/c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578/cmd/fabric-ca-client/command/main_test.go#L152-L154)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to e at line 153 may start a goroutine
[Click here to see the code in its original context.](https://github.com/paul-lee-attorney/fabric-ca-1.4.7-gm/blob/c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578/cmd/fabric-ca-client/command/main_test.go#L152-L154)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, e := range errCases {
extraArgErrorTest(&e, t)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: c7ad827cf9aca0c8fdd47a9348a987b7bc7c9578
|
non_process
|
paul lee attorney fabric ca gm cmd fabric ca client command main test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to e at line may start a goroutine click here to show the line s of go which triggered the analyzer go for e range errcases extraargerrortest e t leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 0
|
13,887
| 16,654,864,392
|
IssuesEvent
|
2021-06-05 10:36:08
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] Responsive issue > UI is broken for toaster messages
|
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
|
UI is broken for all the toaster messages
[Note: Issue should be fixed for all the toaster messages]

|
3.0
|
[PM] Responsive issue > UI is broken for toaster messages - UI is broken for all the toaster messages
[Note: Issue should be fixed for all the toaster messages]

|
process
|
responsive issue ui is broken for toaster messages ui is broken for all the toaster messages
| 1
|
344,820
| 10,349,642,356
|
IssuesEvent
|
2019-09-04 23:18:43
|
oslc-op/jira-migration-landfill
|
https://api.github.com/repos/oslc-op/jira-migration-landfill
|
closed
|
Example 3 in oslc-core-overview is incorrect
|
Core: Main Spec Jira: formatting Priority: High Xtra: Jira
|
Example 3 in the OSLC Core Overview is wrong:
Example 3: Example URL with oslc.prefix
[http://example.com/bugs/4242?oslc.prefix=foaf=&oslc.properties=foaf:lastName](http://example.com/bugs/4242?oslc.prefix=foaf=&oslc.properties=foaf:lastName),...
It omits the required Uri\\\_ref\\\_esc for foaf.
It should be:
Example 3: Example URL with oslc.prefix
[http://example.com/bugs/4242?oslc.prefix=foaf=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2F&oslc.properties=foaf:lastName](http://example.com/bugs/4242?oslc.prefix=foaf=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2F&oslc.properties=foaf:lastName),...
---
_Migrated from https://issues.oasis-open.org/browse/OSLCCORE-152 (opened by @oslc-bot; previously assigned to @jamsden)_
|
1.0
|
Example 3 in oslc-core-overview is incorrect - Example 3 in the OSLC Core Overview is wrong:
Example 3: Example URL with oslc.prefix
[http://example.com/bugs/4242?oslc.prefix=foaf=&oslc.properties=foaf:lastName](http://example.com/bugs/4242?oslc.prefix=foaf=&oslc.properties=foaf:lastName),...
It omits the required Uri\\\_ref\\\_esc for foaf.
It should be:
Example 3: Example URL with oslc.prefix
[http://example.com/bugs/4242?oslc.prefix=foaf=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2F&oslc.properties=foaf:lastName](http://example.com/bugs/4242?oslc.prefix=foaf=http%3A%2F%2Fxmlns.com%2Ffoaf%2F0.1%2F&oslc.properties=foaf:lastName),...
---
_Migrated from https://issues.oasis-open.org/browse/OSLCCORE-152 (opened by @oslc-bot; previously assigned to @jamsden)_
|
non_process
|
example in oslc core overview is incorrect example in the oslc core overview is wrong example example url with oslc prefix it omits the required uri ref esc for foaf it should be example example url with oslc prefix migrated from opened by oslc bot previously assigned to jamsden
| 0
|
9,957
| 12,990,502,135
|
IssuesEvent
|
2020-07-23 00:13:56
|
parcel-bundler/parcel
|
https://api.github.com/repos/parcel-bundler/parcel
|
closed
|
using posthtml without options kills php tags
|
:bug: Bug HTML Preprocessing Stale
|
in src/packagers/HTMLPackager.js, this code:
```js
async addAsset(asset) {
let html = asset.generated.html || '';
// Find child bundles that have JS or CSS sibling bundles,
// add them to the head so they are loaded immediately.
let siblingBundles = Array.from(this.bundle.childBundles)
.reduce((p, b) => p.concat([...b.siblingBundles.values()]), [])
.filter(b => b.type === 'css' || b.type === 'js');
if (siblingBundles.length > 0) {
html = posthtml(
this.insertSiblingBundles.bind(this, siblingBundles)
).process(html, {sync: true}).html;
}
await this.write(html);
}
```
calls posthtml().process with a options `{sync: true}` and that causes it to strip out unknown directives.
do you mind if i load up the .posthtmlrc options and add sync: true to them and use that here? or any other suggestion?
My goal is to get php to not be clobbered by parcel's html asset handler
|
1.0
|
using posthtml without options kills php tags - in src/packagers/HTMLPackager.js, this code:
```js
async addAsset(asset) {
let html = asset.generated.html || '';
// Find child bundles that have JS or CSS sibling bundles,
// add them to the head so they are loaded immediately.
let siblingBundles = Array.from(this.bundle.childBundles)
.reduce((p, b) => p.concat([...b.siblingBundles.values()]), [])
.filter(b => b.type === 'css' || b.type === 'js');
if (siblingBundles.length > 0) {
html = posthtml(
this.insertSiblingBundles.bind(this, siblingBundles)
).process(html, {sync: true}).html;
}
await this.write(html);
}
```
calls posthtml().process with a options `{sync: true}` and that causes it to strip out unknown directives.
do you mind if i load up the .posthtmlrc options and add sync: true to them and use that here? or any other suggestion?
My goal is to get php to not be clobbered by parcel's html asset handler
|
process
|
using posthtml without options kills php tags in src packagers htmlpackager js this code js async addasset asset let html asset generated html find child bundles that have js or css sibling bundles add them to the head so they are loaded immediately let siblingbundles array from this bundle childbundles reduce p b p concat filter b b type css b type js if siblingbundles length html posthtml this insertsiblingbundles bind this siblingbundles process html sync true html await this write html calls posthtml process with a options sync true and that causes it to strip out unknown directives do you mind if i load up the posthtmlrc options and add sync true to them and use that here or any other suggestion my goal is to get php to not be clobbered by parcel s html asset handler
| 1
|
2,154
| 5,005,712,689
|
IssuesEvent
|
2016-12-12 11:39:14
|
Alfresco/alfresco-ng2-components
|
https://api.github.com/repos/Alfresco/alfresco-ng2-components
|
closed
|
Cannot scroll large forms
|
browser: firefox browser: safari bug comp: activiti-processList
|
If a large form is attached to start event/user task can only scroll form using up and down arrows on keyboard, scrollbar within browser is disabled/not accessible.
Issue only in Firefox and Safari, this maybe browser limitation
|
1.0
|
Cannot scroll large forms - If a large form is attached to start event/user task can only scroll form using up and down arrows on keyboard, scrollbar within browser is disabled/not accessible.
Issue only in Firefox and Safari, this maybe browser limitation
|
process
|
cannot scroll large forms if a large form is attached to start event user task can only scroll form using up and down arrows on keyboard scrollbar within browser is disabled not accessible issue only in firefox and safari this maybe browser limitation
| 1
|
9,058
| 4,389,957,380
|
IssuesEvent
|
2016-08-09 00:29:26
|
jeff1evesque/machine-learning
|
https://api.github.com/repos/jeff1evesque/machine-learning
|
opened
|
Remove 'vagrant_implement' instances
|
build enhancement
|
We need to remove all instances of `$vagrant_mounted = $hiera_general['vagrant_implement']`, from any puppet manifests, and make necessary adjustments within `setup_tables.py`. The former puppet variable was created when only a singled puppet environment existed, to contain the following (current) puppet environments:
- docker
- vagrant
|
1.0
|
Remove 'vagrant_implement' instances - We need to remove all instances of `$vagrant_mounted = $hiera_general['vagrant_implement']`, from any puppet manifests, and make necessary adjustments within `setup_tables.py`. The former puppet variable was created when only a singled puppet environment existed, to contain the following (current) puppet environments:
- docker
- vagrant
|
non_process
|
remove vagrant implement instances we need to remove all instances of vagrant mounted hiera general from any puppet manifests and make necessary adjustments within setup tables py the former puppet variable was created when only a singled puppet environment existed to contain the following current puppet environments docker vagrant
| 0
|
12,293
| 14,850,813,958
|
IssuesEvent
|
2021-01-18 05:32:43
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] Sites tab > Contents are not displaying in the sites tab : getting error message
|
Blocker Bug P0 Participant manager Process: Fixed Process: Tested dev
|
Sites tab > Contents are not displaying in the sites tab : getting an error message

|
2.0
|
[PM] Sites tab > Contents are not displaying in the sites tab : getting error message - Sites tab > Contents are not displaying in the sites tab : getting an error message

|
process
|
sites tab contents are not displaying in the sites tab getting error message sites tab contents are not displaying in the sites tab getting an error message
| 1
|
18,817
| 24,718,365,315
|
IssuesEvent
|
2022-10-20 08:48:17
|
hermes-hmc/workflow
|
https://api.github.com/repos/hermes-hmc/workflow
|
opened
|
Implement author deduplication based on config
|
enhancement 2️ processing
|
- **Requires:** #39
Author deduplication (including email mapping) should be part of the processing step.
|
1.0
|
Implement author deduplication based on config - - **Requires:** #39
Author deduplication (including email mapping) should be part of the processing step.
|
process
|
implement author deduplication based on config requires author deduplication including email mapping should be part of the processing step
| 1
|
15,086
| 18,795,513,924
|
IssuesEvent
|
2021-11-08 21:47:38
|
googleapis/python-bigquery
|
https://api.github.com/repos/googleapis/python-bigquery
|
closed
|
type check with mypy as well
|
api: bigquery type: process
|
https://issues.apache.org/jira/browse/BEAM-12975 -- https://github.com/googleapis/python-bigquery/pull/976 appears to have broken Apache BEAM, which uses mypy for type checking.
This would also help with Mac developers, as I've been unable to run pytype locally due to some missing wheels.
|
1.0
|
type check with mypy as well - https://issues.apache.org/jira/browse/BEAM-12975 -- https://github.com/googleapis/python-bigquery/pull/976 appears to have broken Apache BEAM, which uses mypy for type checking.
This would also help with Mac developers, as I've been unable to run pytype locally due to some missing wheels.
|
process
|
type check with mypy as well appears to have broken apache beam which uses mypy for type checking this would also help with mac developers as i ve been unable to run pytype locally due to some missing wheels
| 1
|
375,195
| 26,151,579,617
|
IssuesEvent
|
2022-12-30 14:28:31
|
FilledStacks/stacked
|
https://api.github.com/repos/FilledStacks/stacked
|
closed
|
Update Get Started to explain the generated Code
|
documentation
|
At the moment the Get Started section of the documentation lacks an explanation of the output of the cli.
Since we have some counter code in the generated code we should go over explaining how the view and viewmodel works by using the generated code as an example.
|
1.0
|
Update Get Started to explain the generated Code - At the moment the Get Started section of the documentation lacks an explanation of the output of the cli.
Since we have some counter code in the generated code we should go over explaining how the view and viewmodel works by using the generated code as an example.
|
non_process
|
update get started to explain the generated code at the moment the get started section of the documentation lacks an explanation of the output of the cli since we have some counter code in the generated code we should go over explaining how the view and viewmodel works by using the generated code as an example
| 0
|
4,842
| 7,736,175,388
|
IssuesEvent
|
2018-05-27 23:29:37
|
vital-software/scala-redox
|
https://api.github.com/repos/vital-software/scala-redox
|
opened
|
Set up CI to perform releases
|
process: ci stage: backlog type: devops
|
Similar to what we've done with our other Scala libs, this library should support versioning the project during the CI build of master commits: https://github.com/vital-software/doc-dna/blob/deb774269b7c56c44fe72aab68e7532dcdcb8fc9/.buildkite/pipeline.yml#L27-L34
The CI system should have Sonatype access, and use it to publish artifacts.
|
1.0
|
Set up CI to perform releases - Similar to what we've done with our other Scala libs, this library should support versioning the project during the CI build of master commits: https://github.com/vital-software/doc-dna/blob/deb774269b7c56c44fe72aab68e7532dcdcb8fc9/.buildkite/pipeline.yml#L27-L34
The CI system should have Sonatype access, and use it to publish artifacts.
|
process
|
set up ci to perform releases similar to what we ve done with our other scala libs this library should support versioning the project during the ci build of master commits the ci system should have sonatype access and use it to publish artifacts
| 1
|
87,744
| 17,368,201,987
|
IssuesEvent
|
2021-07-30 10:15:08
|
Serrin/Celestra
|
https://api.github.com/repos/Serrin/Celestra
|
closed
|
Changes in v4.4.3
|
closed - done or fixed code code - CUT code - ESM documentation maintenance type - enhancement
|
1. Documentation and pdf fixes.
2. Add a new function: `findLast(<collection>,<callback>);`
3. Add an alias: `contains(<collection>,<value>);` -> `includes(<collection>,<value>);`
4. Fix the description of these functions:
````javascript
arrayRemove(<array>,<value>[,all=false]);
arrayCycle(<collection>[,n=100]);
arrayRepeat(<value>[,n=100]);
iterRange([start=0[,step=1[,end=Infinity]]]);
iterCycle(<iter>[,n=Infinity]);
iterRepeat(<value>[,n=Infinity]);
take(<collection>[,n=1]);
takeRight(<collection>[,n=1]);
drop(<collection>[,n=1]);
dropRight(<collection>[,n=1]);
setUnion(<collection1>[,collectionN]);
randomString([length[,specChar=false]]);
javaHash(<data>[,hexa=false]);
````
5. Remove the description of these removed function in __celestra.html__ and __readme.md__ and collect these function names in a new line of the __Collections__ table:
````javascript
forOf(<collection>,<callback>);
mapOf(<collection>,<callback>);
sizeOf(<collection>);
filterOf(<collection>,<callback>);
hasOf(<collection>,<value>);
findOf(<collection>,<callback>);
everyOf(<collection>,<callback>);
someOf(<collection>,<callback>);
noneOf(<collection>,<callback>);
firstOf(<collection>);
lastOf(<collection>);
sliceOf(<collection>[,begin[,end]]);
reverseOf(<collection>);
sortOf(<collection>);
reduceOf(<collection>,<callback>[,initialvalue]);
concatOf(<collection1>[,collectionN]);
flatOf(<collection>);
enumerateOf(<collection>);
joinOf(<collection>[,separator=","]);
takeOf(<collection>[,n]);
dropOf(<collection>[,n]);
````
6. __CUT v0.8.19__
- Replace the __unittest.dev.html__, __unittest.min.html__ and __unittest.esm.html__ files with the __unittest.html__
|
3.0
|
Changes in v4.4.3 - 1. Documentation and pdf fixes.
2. Add a new function: `findLast(<collection>,<callback>);`
3. Add an alias: `contains(<collection>,<value>);` -> `includes(<collection>,<value>);`
4. Fix the description of these functions:
````javascript
arrayRemove(<array>,<value>[,all=false]);
arrayCycle(<collection>[,n=100]);
arrayRepeat(<value>[,n=100]);
iterRange([start=0[,step=1[,end=Infinity]]]);
iterCycle(<iter>[,n=Infinity]);
iterRepeat(<value>[,n=Infinity]);
take(<collection>[,n=1]);
takeRight(<collection>[,n=1]);
drop(<collection>[,n=1]);
dropRight(<collection>[,n=1]);
setUnion(<collection1>[,collectionN]);
randomString([length[,specChar=false]]);
javaHash(<data>[,hexa=false]);
````
5. Remove the description of these removed function in __celestra.html__ and __readme.md__ and collect these function names in a new line of the __Collections__ table:
````javascript
forOf(<collection>,<callback>);
mapOf(<collection>,<callback>);
sizeOf(<collection>);
filterOf(<collection>,<callback>);
hasOf(<collection>,<value>);
findOf(<collection>,<callback>);
everyOf(<collection>,<callback>);
someOf(<collection>,<callback>);
noneOf(<collection>,<callback>);
firstOf(<collection>);
lastOf(<collection>);
sliceOf(<collection>[,begin[,end]]);
reverseOf(<collection>);
sortOf(<collection>);
reduceOf(<collection>,<callback>[,initialvalue]);
concatOf(<collection1>[,collectionN]);
flatOf(<collection>);
enumerateOf(<collection>);
joinOf(<collection>[,separator=","]);
takeOf(<collection>[,n]);
dropOf(<collection>[,n]);
````
6. __CUT v0.8.19__
- Replace the __unittest.dev.html__, __unittest.min.html__ and __unittest.esm.html__ files with the __unittest.html__
|
non_process
|
changes in documentation and pdf fixes add a new function findlast add an alias contains includes fix the description of these functions javascript arrayremove arraycycle arrayrepeat iterrange itercycle iterrepeat take takeright drop dropright setunion randomstring javahash remove the description of these removed function in celestra html and readme md and collect these function names in a new line of the collections table javascript forof mapof sizeof filterof hasof findof everyof someof noneof firstof lastof sliceof reverseof sortof reduceof concatof flatof enumerateof joinof takeof dropof cut replace the unittest dev html unittest min html and unittest esm html files with the unittest html
| 0
|
15,013
| 18,724,271,103
|
IssuesEvent
|
2021-11-03 14:51:39
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
closed
|
Autogen: fix coverage landmines.
|
api: datastore api: pubsub api: bigtable api: logging api: vision api: monitoring api: translation api: speech api: spanner api: clouderrorreporting api: language api: cloudtrace api: videointelligence api: firestore type: process api: dataproc api: container api: bigquerydatatransfer api: dlp api: texttospeech api: redis api: cloudiot api: cloudtasks api: automl api: cloudkms api: cloudasset api: oslogin api: websecurityscanner api: iam api: irm api: talent api: webrisk api: datacatalog api: phishingprotection api: datalabeling api: cloudscheduler api: securitycenter
|
The `unit_cov_level` numbers in the various `noxfile.py` files are landmines (e.g., PR #8147 adds a `# pragma: NO COVER` and actually trips the check, because the total number of lines checked in BQDT drops!) The numbers don't represent any **real** goals (why 79% in one library or 95% in another?), but are merely "high water marks." The real goal is 100% coverage across all unit test runs, which we **don't** hit for many of the autogen-only libraries.
To address the first issue, I would argue that we should just drop the coverage checks inside the `unit` / `default` sessions.
To address the second issue, we need to apply a change like the one in #7413 which actually tests the "shim" modules for the autogen-only libraries (and excludes the namespace package files).
Both of these changes need to happen in the generator / templates (dropping the `--cov-fail-under` bit in the `unit` / `default` session, tweaking the `.coveragerc`, adding a testcase module for each generated shim module).
|
1.0
|
Autogen: fix coverage landmines. - The `unit_cov_level` numbers in the various `noxfile.py` files are landmines (e.g., PR #8147 adds a `# pragma: NO COVER` and actually trips the check, because the total number of lines checked in BQDT drops!) The numbers don't represent any **real** goals (why 79% in one library or 95% in another?), but are merely "high water marks." The real goal is 100% coverage across all unit test runs, which we **don't** hit for many of the autogen-only libraries.
To address the first issue, I would argue that we should just drop the coverage checks inside the `unit` / `default` sessions.
To address the second issue, we need to apply a change like the one in #7413 which actually tests the "shim" modules for the autogen-only libraries (and excludes the namespace package files).
Both of these changes need to happen in the generator / templates (dropping the `--cov-fail-under` bit in the `unit` / `default` session, tweaking the `.coveragerc`, adding a testcase module for each generated shim module).
|
process
|
autogen fix coverage landmines the unit cov level numbers in the various noxfile py files are landmines e g pr adds a pragma no cover and actually trips the check because the total number of lines checked in bqdt drops the numbers don t represent any real goals why in one library or in another but are merely high water marks the real goal is coverage across all unit test runs which we don t hit for many of the autogen only libraries to address the first issue i would argue that we should just drop the coverage checks inside the unit default sessions to address the second issue we need to apply a change like the one in which actually tests the shim modules for the autogen only libraries and excludes the namespace package files both of these changes need to happen in the generator templates dropping the cov fail under bit in the unit default session tweaking the coveragerc adding a testcase module for each generated shim module
| 1
|
134,673
| 19,299,530,891
|
IssuesEvent
|
2021-12-13 02:26:17
|
bookey-dev/bookey.bug
|
https://api.github.com/repos/bookey-dev/bookey.bug
|
closed
|
输入框为空状态下按钮高亮显示
|
P5 platform: ios easy fix v1.9 design
|
机型:iphone SE
步骤:1.启动Bookey,登陆首页

2.sign up
结果:输入框为空状态下next按钮高亮显示
期望:建议为空状态下,操作按钮暗淡色,输入后高亮显示,与android保持一致
|
1.0
|
输入框为空状态下按钮高亮显示 - 机型:iphone SE
步骤:1.启动Bookey,登陆首页

2.sign up
结果:输入框为空状态下next按钮高亮显示
期望:建议为空状态下,操作按钮暗淡色,输入后高亮显示,与android保持一致
|
non_process
|
输入框为空状态下按钮高亮显示 机型:iphone se 步骤: 启动bookey,登陆首页 sign up 结果:输入框为空状态下next按钮高亮显示 期望:建议为空状态下,操作按钮暗淡色,输入后高亮显示,与android保持一致
| 0
|
745
| 3,214,363,065
|
IssuesEvent
|
2015-10-07 01:11:32
|
grafeo/grafeo
|
https://api.github.com/repos/grafeo/grafeo
|
closed
|
Read and Write PPM
|
Component: Image Processing feature request priority: high
|
- `image_read_ppm`
- `image_write_ppm`
- link `image_read` and `image_write` with PPM versions
|
1.0
|
Read and Write PPM - - `image_read_ppm`
- `image_write_ppm`
- link `image_read` and `image_write` with PPM versions
|
process
|
read and write ppm image read ppm image write ppm link image read and image write with ppm versions
| 1
|
122,418
| 26,126,189,129
|
IssuesEvent
|
2022-12-28 18:56:17
|
shelcia/mocker
|
https://api.github.com/repos/shelcia/mocker
|
closed
|
Fix code scanning alert - Database query built from user-controlled sources
|
medium codepeak22
|
<!-- Warning: The suggested title contains the alert rule name. This can expose security information. -->
Tracking issue for:
- [ ] https://github.com/shelcia/mocker/security/code-scanning/1
|
1.0
|
Fix code scanning alert - Database query built from user-controlled sources - <!-- Warning: The suggested title contains the alert rule name. This can expose security information. -->
Tracking issue for:
- [ ] https://github.com/shelcia/mocker/security/code-scanning/1
|
non_process
|
fix code scanning alert database query built from user controlled sources tracking issue for
| 0
|
18,155
| 24,192,984,880
|
IssuesEvent
|
2022-09-23 19:41:56
|
google/android-fhir
|
https://api.github.com/repos/google/android-fhir
|
reopened
|
HAPI Structures, Java11, and Android API levels
|
process
|
This is an issue @vitorpamplona raised in the process of implementing #1403.
PR #1603 introduces new dependencies such as CQL evalutor, CQL engine, and CQL translator. They use HAPI version 6.0.1 which is compiled using Java11. There are certain Java11 APIs that are not available on older Android versions, especially pre Android API 26. For example, `java.lang.reflect.Method.getParameterCount()`.
As a result, with this PR, the workflow library will crash on Android API level 24. Please note, however, that although the API `java.lang.reflect.Method.getParameterCount()` is used in HAPI, we do not invoke code paths in HAPI using this API through our usage in the FHIR Engine and SDC library. Only in the workflow library in this PR. In other words, when I tested the FHIR Engine and SDC library with HAPI 6 on Android API level 24, there's no crash caused by the missing of this API.
Android API levels 26+ support Java11 APIs. So the above PR works fine in Android API 26+. This means that we will need to update the min api level for the workflow library to API 26.
At the moment, our gradle files are written so that all libraries share the same min api level. This will need to be changed so we do not raise API level above 24 for users of FHIR engine and SDC library.
cc: @joiskash
|
1.0
|
HAPI Structures, Java11, and Android API levels - This is an issue @vitorpamplona raised in the process of implementing #1403.
PR #1603 introduces new dependencies such as CQL evalutor, CQL engine, and CQL translator. They use HAPI version 6.0.1 which is compiled using Java11. There are certain Java11 APIs that are not available on older Android versions, especially pre Android API 26. For example, `java.lang.reflect.Method.getParameterCount()`.
As a result, with this PR, the workflow library will crash on Android API level 24. Please note, however, that although the API `java.lang.reflect.Method.getParameterCount()` is used in HAPI, we do not invoke code paths in HAPI using this API through our usage in the FHIR Engine and SDC library. Only in the workflow library in this PR. In other words, when I tested the FHIR Engine and SDC library with HAPI 6 on Android API level 24, there's no crash caused by the missing of this API.
Android API levels 26+ support Java11 APIs. So the above PR works fine in Android API 26+. This means that we will need to update the min api level for the workflow library to API 26.
At the moment, our gradle files are written so that all libraries share the same min api level. This will need to be changed so we do not raise API level above 24 for users of FHIR engine and SDC library.
cc: @joiskash
|
process
|
hapi structures and android api levels this is an issue vitorpamplona raised in the process of implementing pr introduces new dependencies such as cql evalutor cql engine and cql translator they use hapi version which is compiled using there are certain apis that are not available on older android versions especially pre android api for example java lang reflect method getparametercount as a result with this pr the workflow library will crash on android api level please note however that although the api java lang reflect method getparametercount is used in hapi we do not invoke code paths in hapi using this api through our usage in the fhir engine and sdc library only in the workflow library in this pr in other words when i tested the fhir engine and sdc library with hapi on android api level there s no crash caused by the missing of this api android api levels support apis so the above pr works fine in android api this means that we will need to update the min api level for the workflow library to api at the moment our gradle files are written so that all libraries share the same min api level this will need to be changed so we do not raise api level above for users of fhir engine and sdc library cc joiskash
| 1
|
10,088
| 13,044,162,001
|
IssuesEvent
|
2020-07-29 03:47:28
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `SubTimeDateTimeNull` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `SubTimeDateTimeNull` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `SubTimeDateTimeNull` from TiDB -
## Description
Port the scalar function `SubTimeDateTimeNull` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function subtimedatetimenull from tidb description port the scalar function subtimedatetimenull from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
10,203
| 13,066,579,076
|
IssuesEvent
|
2020-07-30 22:01:04
|
googleapis/google-auth-library-ruby
|
https://api.github.com/repos/googleapis/google-auth-library-ruby
|
closed
|
New release enquiry
|
type: process
|
Hi there,
I am looking to use the recent changes that have been merged to master:
https://github.com/googleapis/google-auth-library-ruby/commit/48c689aa93bfe81c5a6ae23362d86fc25ba098cf
What is the typical process of a change being merged and then a release being made? Currently, we could track master branch, but this will not work with projects that use gemspec.
|
1.0
|
New release enquiry - Hi there,
I am looking to use the recent changes that have been merged to master:
https://github.com/googleapis/google-auth-library-ruby/commit/48c689aa93bfe81c5a6ae23362d86fc25ba098cf
What is the typical process of a change being merged and then a release being made? Currently, we could track master branch, but this will not work with projects that use gemspec.
|
process
|
new release enquiry hi there i am looking to use the recent changes that have been merged to master what is the typical process of a change being merged and then a release being made currently we could track master branch but this will not work with projects that use gemspec
| 1
|
162,706
| 20,241,557,365
|
IssuesEvent
|
2022-02-14 09:46:27
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
[Security Solution] Not able to create an EQL rule due to validation error
|
bug impact:critical fixed Team: SecuritySolution Team:Detection Rules v7.17.1
|
**Describe the bug:**
- Not able to create an EQL rule due to validation error
**Kibana/Elasticsearch Stack version:**
- 7.17 latest branch (45be56082f7338428b62ebd1c67b222f470be0ce)
- 7.17.1 latest snapshot
**Steps to reproduce:**
1. Navigate to the rules page
2. Click on `create new rule`
2. Select `Event Correlation type`
3. Enter a valid EQL query
**Current behavior:**
- A validation error is returned
- You cannot proceed with the rule creation
- The rule cannot be created
<img width="1034" alt="Screenshot 2022-02-09 at 12 19 39" src="https://user-images.githubusercontent.com/17427073/153189817-8109045d-0515-4794-8a1c-3ec65f33bbd3.png">
<img width="769" alt="Screenshot 2022-02-09 at 12 19 55" src="https://user-images.githubusercontent.com/17427073/153189858-ba2a4bd0-f78e-44b5-a752-5d6edcf61414.png">
**Expected behavior:**
- No validation error is displayed
- The rule can be correctly created
**Additional information:**
- The displayed error:
````
{"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]"}],"type":"illegal_argument_exception","reason":"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]"},"status":400}
````
````
{
"name": "Error",
"message": "{\"error\":{\"root_cause\":[{\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"}],\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"},\"status\":400}",
"stack": "Error: {\"error\":{\"root_cause\":[{\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"}],\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"},\"status\":400}\n at u (https://upgrades.kb.us-central1.gcp.qa.cld.elstc.co:9243/46574/bundles/plugin/securitySolution/8.0.0/securitySolution.chunk.12.js:3:16800)\n at async f (https://upgrades.kb.us-central1.gcp.qa.cld.elstc.co:9243/46574/bundles/plugin/securitySolution/8.0.0/securitySolution.chunk.12.js:3:17361)"
}
````
- After the upgrade of a 7.17 EQL rule to 7.17.1 version, the rule still works and generate alerts
- After the upgrade of a 7.17 EQL rule to 7.17.1 version, the same error validation error is displayed when trying to edit the rule. So the rule cannot be edited.
|
True
|
[Security Solution] Not able to create an EQL rule due to validation error - **Describe the bug:**
- Not able to create an EQL rule due to validation error
**Kibana/Elasticsearch Stack version:**
- 7.17 latest branch (45be56082f7338428b62ebd1c67b222f470be0ce)
- 7.17.1 latest snapshot
**Steps to reproduce:**
1. Navigate to the rules page
2. Click on `create new rule`
2. Select `Event Correlation type`
3. Enter a valid EQL query
**Current behavior:**
- A validation error is returned
- You cannot proceed with the rule creation
- The rule cannot be created
<img width="1034" alt="Screenshot 2022-02-09 at 12 19 39" src="https://user-images.githubusercontent.com/17427073/153189817-8109045d-0515-4794-8a1c-3ec65f33bbd3.png">
<img width="769" alt="Screenshot 2022-02-09 at 12 19 55" src="https://user-images.githubusercontent.com/17427073/153189858-ba2a4bd0-f78e-44b5-a752-5d6edcf61414.png">
**Expected behavior:**
- No validation error is displayed
- The rule can be correctly created
**Additional information:**
- The displayed error:
````
{"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]"}],"type":"illegal_argument_exception","reason":"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]"},"status":400}
````
````
{
"name": "Error",
"message": "{\"error\":{\"root_cause\":[{\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"}],\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"},\"status\":400}",
"stack": "Error: {\"error\":{\"root_cause\":[{\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"}],\"type\":\"illegal_argument_exception\",\"reason\":\"request [/apm-*-transaction*,traces-apm*,auditbeat-*,endgame-*,filebeat-*,logs-*,packetbeat-*,winlogbeat-*/_eql/search] contains unrecognized parameter: [enable_fields_emulation]\"},\"status\":400}\n at u (https://upgrades.kb.us-central1.gcp.qa.cld.elstc.co:9243/46574/bundles/plugin/securitySolution/8.0.0/securitySolution.chunk.12.js:3:16800)\n at async f (https://upgrades.kb.us-central1.gcp.qa.cld.elstc.co:9243/46574/bundles/plugin/securitySolution/8.0.0/securitySolution.chunk.12.js:3:17361)"
}
````
- After the upgrade of a 7.17 EQL rule to 7.17.1 version, the rule still works and generate alerts
- After the upgrade of a 7.17 EQL rule to 7.17.1 version, the same error validation error is displayed when trying to edit the rule. So the rule cannot be edited.
|
non_process
|
not able to create an eql rule due to validation error describe the bug not able to create an eql rule due to validation error kibana elasticsearch stack version latest branch latest snapshot steps to reproduce navigate to the rules page click on create new rule select event correlation type enter a valid eql query current behavior a validation error is returned you cannot proceed with the rule creation the rule cannot be created img width alt screenshot at src img width alt screenshot at src expected behavior no validation error is displayed the rule can be correctly created additional information the displayed error error root cause contains unrecognized parameter type illegal argument exception reason request contains unrecognized parameter status name error message error root cause contains unrecognized parameter type illegal argument exception reason request contains unrecognized parameter status stack error error root cause contains unrecognized parameter type illegal argument exception reason request contains unrecognized parameter status n at u at async f after the upgrade of a eql rule to version the rule still works and generate alerts after the upgrade of a eql rule to version the same error validation error is displayed when trying to edit the rule so the rule cannot be edited
| 0
|
284,484
| 8,742,809,063
|
IssuesEvent
|
2018-12-12 17:20:25
|
K-Ho/code
|
https://api.github.com/repos/K-Ho/code
|
opened
|
Fixed Initial sizing/ position of Graph Network
|
UI/UX high priority
|
Make the graph network render larger by default and always in an open space (center of screen)
|
1.0
|
Fixed Initial sizing/ position of Graph Network - Make the graph network render larger by default and always in an open space (center of screen)
|
non_process
|
fixed initial sizing position of graph network make the graph network render larger by default and always in an open space center of screen
| 0
|
189,344
| 22,047,021,634
|
IssuesEvent
|
2022-05-30 03:43:46
|
dpteam/RK3188_TABLET
|
https://api.github.com/repos/dpteam/RK3188_TABLET
|
closed
|
WS-2021-0279 (Medium) detected in randomv3.0.66 - autoclosed
|
security vulnerability
|
## WS-2021-0279 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>randomv3.0.66</b></p></summary>
<p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git>https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/tree-log.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Linux Kernel in versions v2.6.29-rc1 to v5.12.9 there is an error handling in fixup_inode_link_counts which could lead to memory leak.
<p>Publish Date: 2021-06-25
<p>URL: <a href=https://github.com/gregkh/linux/commit/4cd303735bdfacd115ee20a6f3235b0084924174>WS-2021-0279</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1000819">https://osv.dev/vulnerability/UVI-2021-1000819</a></p>
<p>Release Date: 2021-06-25</p>
<p>Fix Resolution: v5.12.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2021-0279 (Medium) detected in randomv3.0.66 - autoclosed - ## WS-2021-0279 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>randomv3.0.66</b></p></summary>
<p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git>https://git.kernel.org/pub/scm/linux/kernel/git/tytso/random.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/tree-log.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Linux Kernel in versions v2.6.29-rc1 to v5.12.9 there is an error handling in fixup_inode_link_counts which could lead to memory leak.
<p>Publish Date: 2021-06-25
<p>URL: <a href=https://github.com/gregkh/linux/commit/4cd303735bdfacd115ee20a6f3235b0084924174>WS-2021-0279</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1000819">https://osv.dev/vulnerability/UVI-2021-1000819</a></p>
<p>Release Date: 2021-06-25</p>
<p>Fix Resolution: v5.12.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium detected in autoclosed ws medium severity vulnerability vulnerable library library home page a href found in head commit a href found in base branch master vulnerable source files fs btrfs tree log c vulnerability details linux kernel in versions to there is an error handling in fixup inode link counts which could lead to memory leak publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
36,979
| 15,110,609,649
|
IssuesEvent
|
2021-02-08 19:27:52
|
microsoft/BotFramework-Composer
|
https://api.github.com/repos/microsoft/BotFramework-Composer
|
closed
|
End conversation by the Bot
|
Bot Services Support Type: Question customer-replied-to customer-reported
|
I have created a Bot which I have attached to Omnichannel for customer services (D365 extension). When a user starts a conversation with he Bot through any channel (e.g. Teams) the conversation goes for ever with the Bot.
I am not able to end up the conversation in Omnichannel for customer service as there is no standard Action in the Bot Framework Composer to end the conversation.
Could you advise what is the easiest way to end up the conversation by the Bot so that it ends in Omnichannel for customer services (D365) as well?
|
1.0
|
End conversation by the Bot - I have created a Bot which I have attached to Omnichannel for customer services (D365 extension). When a user starts a conversation with he Bot through any channel (e.g. Teams) the conversation goes for ever with the Bot.
I am not able to end up the conversation in Omnichannel for customer service as there is no standard Action in the Bot Framework Composer to end the conversation.
Could you advise what is the easiest way to end up the conversation by the Bot so that it ends in Omnichannel for customer services (D365) as well?
|
non_process
|
end conversation by the bot i have created a bot which i have attached to omnichannel for customer services extension when a user starts a conversation with he bot through any channel e g teams the conversation goes for ever with the bot i am not able to end up the conversation in omnichannel for customer service as there is no standard action in the bot framework composer to end the conversation could you advise what is the easiest way to end up the conversation by the bot so that it ends in omnichannel for customer services as well
| 0
|
51,938
| 13,690,550,323
|
IssuesEvent
|
2020-09-30 14:30:43
|
mrgadgil/test
|
https://api.github.com/repos/mrgadgil/test
|
opened
|
HIGH severity finding reported by IBM Security Advisor
|
IBM Security Advisor
|
**Source**: Security Advisor
**Finding**: 15c57fd8bfbff988b328b55b167a1689/providers/security-advisor/occurrences/ata-1601476234889
**Severity**: HIGH
[View in Security Advisor Dashboard](https://cloud.ibm.com/security-advisor#/findings?id=15c57fd8bfbff988b328b55b167a1689/providers/security-advisor/occurrences/ata-1601476234889®ion=us-south)
|
True
|
HIGH severity finding reported by IBM Security Advisor - **Source**: Security Advisor
**Finding**: 15c57fd8bfbff988b328b55b167a1689/providers/security-advisor/occurrences/ata-1601476234889
**Severity**: HIGH
[View in Security Advisor Dashboard](https://cloud.ibm.com/security-advisor#/findings?id=15c57fd8bfbff988b328b55b167a1689/providers/security-advisor/occurrences/ata-1601476234889®ion=us-south)
|
non_process
|
high severity finding reported by ibm security advisor source security advisor finding providers security advisor occurrences ata severity high
| 0
|
13,394
| 15,866,825,496
|
IssuesEvent
|
2021-04-08 16:10:14
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
Improve error for invalid permissions on Linux
|
bug/2-confirmed kind/bug process/candidate team/client
|
I'm running into this on ubuntu 20.04 while having `prisma@2.20.1` installed globally.
`prisma db push --preview-feature`
```sh
Environment variables loaded from prisma/.env
Prisma schema loaded from prisma/schema.prisma
Datasource "db": PostgreSQL database "debug-migrate", schema "public" at "playground-database-postgres.cluster-clfeqqifnebj.eu-west-1.rds.amazonaws.com:5432"
PostgreSQL database debug-migrate created at playground-database-postgres.cluster-clfeqqifnebj.eu-west-1.rds.amazonaws.com:5432
🚀 Your database is now in sync with your schema. Done in 2.54s
Running generate... (Use --skip-generate to skip the generators)
Error: EACCES: permission denied, copyfile '/home/millsp/.cache/prisma/master/60ba6551f29b17d7d6ce479e5733c70d9c00860e/debian-openssl-1.1.x/query-engine' -> '/usr/local/lib/node_modules/prisma/query-engine-debian-openssl-1.1.x'
```
|
1.0
|
Improve error for invalid permissions on Linux - I'm running into this on ubuntu 20.04 while having `prisma@2.20.1` installed globally.
`prisma db push --preview-feature`
```sh
Environment variables loaded from prisma/.env
Prisma schema loaded from prisma/schema.prisma
Datasource "db": PostgreSQL database "debug-migrate", schema "public" at "playground-database-postgres.cluster-clfeqqifnebj.eu-west-1.rds.amazonaws.com:5432"
PostgreSQL database debug-migrate created at playground-database-postgres.cluster-clfeqqifnebj.eu-west-1.rds.amazonaws.com:5432
🚀 Your database is now in sync with your schema. Done in 2.54s
Running generate... (Use --skip-generate to skip the generators)
Error: EACCES: permission denied, copyfile '/home/millsp/.cache/prisma/master/60ba6551f29b17d7d6ce479e5733c70d9c00860e/debian-openssl-1.1.x/query-engine' -> '/usr/local/lib/node_modules/prisma/query-engine-debian-openssl-1.1.x'
```
|
process
|
improve error for invalid permissions on linux i m running into this on ubuntu while having prisma installed globally prisma db push preview feature sh environment variables loaded from prisma env prisma schema loaded from prisma schema prisma datasource db postgresql database debug migrate schema public at playground database postgres cluster clfeqqifnebj eu west rds amazonaws com postgresql database debug migrate created at playground database postgres cluster clfeqqifnebj eu west rds amazonaws com 🚀 your database is now in sync with your schema done in running generate use skip generate to skip the generators error eacces permission denied copyfile home millsp cache prisma master debian openssl x query engine usr local lib node modules prisma query engine debian openssl x
| 1
|
791,756
| 27,874,745,612
|
IssuesEvent
|
2023-03-21 15:25:39
|
pastas/pastas
|
https://api.github.com/repos/pastas/pastas
|
closed
|
[ENHANCEMENT] allow ffill and bfill as fill_before and fill_after options for time series
|
enhancement priority 2
|
**Describe the proposed enhancement**
Allow the following settings dictionary:
```python
settings = {
"sample_up": "interpolate",
"sample_down": "mean",
"fill_before": "bfill", # <- new option
"fill_after": "ffill", # <- new option
"fill_nan": "interpolate",
}
```
I know this is the same as getting the first/last value from the timeseries and entering that into the settings dictionary for fill before/after, but I like not having to think about it :).
When would you ever use this?
I used it for a pumping test, where I used a reference head time series (with no pumping influence) as input for my time series models, that in some cases needed to be extended slightly into the past/future. In my case, extending the first/last values as fill values worked pretty well.
|
1.0
|
[ENHANCEMENT] allow ffill and bfill as fill_before and fill_after options for time series - **Describe the proposed enhancement**
Allow the following settings dictionary:
```python
settings = {
"sample_up": "interpolate",
"sample_down": "mean",
"fill_before": "bfill", # <- new option
"fill_after": "ffill", # <- new option
"fill_nan": "interpolate",
}
```
I know this is the same as getting the first/last value from the timeseries and entering that into the settings dictionary for fill before/after, but I like not having to think about it :).
When would you ever use this?
I used it for a pumping test, where I used a reference head time series (with no pumping influence) as input for my time series models, that in some cases needed to be extended slightly into the past/future. In my case, extending the first/last values as fill values worked pretty well.
|
non_process
|
allow ffill and bfill as fill before and fill after options for time series describe the proposed enhancement allow the following settings dictionary python settings sample up interpolate sample down mean fill before bfill new option fill after ffill new option fill nan interpolate i know this is the same as getting the first last value from the timeseries and entering that into the settings dictionary for fill before after but i like not having to think about it when would you ever use this i used it for a pumping test where i used a reference head time series with no pumping influence as input for my time series models that in some cases needed to be extended slightly into the past future in my case extending the first last values as fill values worked pretty well
| 0
|
10,197
| 13,064,284,901
|
IssuesEvent
|
2020-07-30 17:50:32
|
GetTerminus/terminus-oss
|
https://api.github.com/repos/GetTerminus/terminus-oss
|
closed
|
Split NGX-Tools repo and move parts into monorepo
|
Goal: Process Improvement Type: chore
|
- [ ] Follow existing endpoints for package separation?
- [ ] Publish to existing NPM package
- [ ] Move any applicable issues
|
1.0
|
Split NGX-Tools repo and move parts into monorepo - - [ ] Follow existing endpoints for package separation?
- [ ] Publish to existing NPM package
- [ ] Move any applicable issues
|
process
|
split ngx tools repo and move parts into monorepo follow existing endpoints for package separation publish to existing npm package move any applicable issues
| 1
|
9,176
| 12,226,503,509
|
IssuesEvent
|
2020-05-03 11:14:24
|
gfleetwood/asteres
|
https://api.github.com/repos/gfleetwood/asteres
|
opened
|
AdvancerTechnologies/MyoWare_MuscleSensor (35062662)
|
Processing engineering
|
https://github.com/AdvancerTechnologies/MyoWare_MuscleSensor
Example code and documentation for the MyoWare™ Muscle Sensor
|
1.0
|
AdvancerTechnologies/MyoWare_MuscleSensor (35062662) - https://github.com/AdvancerTechnologies/MyoWare_MuscleSensor
Example code and documentation for the MyoWare™ Muscle Sensor
|
process
|
advancertechnologies myoware musclesensor example code and documentation for the myoware™ muscle sensor
| 1
|
20,628
| 27,300,060,398
|
IssuesEvent
|
2023-02-24 00:39:09
|
googleapis/google-cloudevents-python
|
https://api.github.com/repos/googleapis/google-cloudevents-python
|
closed
|
Warning: a recent release failed
|
type: process status: will not fix api: eventarc
|
The following release PRs may have failed:
* #190 - The release job was triggered, but has not reported back success.
* #187 - The release job was triggered, but has not reported back success.
* #185 - The release job was triggered, but has not reported back success.
* #183 - The release job was triggered, but has not reported back success.
* #181 - The release job was triggered, but has not reported back success.
* #136 - The release job was triggered, but has not reported back success.
|
1.0
|
Warning: a recent release failed - The following release PRs may have failed:
* #190 - The release job was triggered, but has not reported back success.
* #187 - The release job was triggered, but has not reported back success.
* #185 - The release job was triggered, but has not reported back success.
* #183 - The release job was triggered, but has not reported back success.
* #181 - The release job was triggered, but has not reported back success.
* #136 - The release job was triggered, but has not reported back success.
|
process
|
warning a recent release failed the following release prs may have failed the release job was triggered but has not reported back success the release job was triggered but has not reported back success the release job was triggered but has not reported back success the release job was triggered but has not reported back success the release job was triggered but has not reported back success the release job was triggered but has not reported back success
| 1
|
2,996
| 5,970,835,671
|
IssuesEvent
|
2017-05-31 00:03:08
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
Desktop: System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper failed with "Xunit.Sdk.EqualException"
|
area-System.ServiceProcess test-run-desktop
|
Failed test: System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper
Detail: https://ci.dot.net/job/dotnet_corefx/job/master/job/outerloop_netfx_windows_nt_debug/66/testReport/System.ServiceProcess.Tests/ServiceControllerTests/ConstructWithServiceName_ToUpper/
Configuration: outerloop_netfx_windows_nt_debug
MESSAGE:
~~~
Assert.Equal() Failure
↓ (pos 3)
Expected: 749b386a-fd41-4bed-951f-75d4a705f7d9
Actual: 749B386A-FD41-4BED-951F-75D4A705F7D9
↑ (pos 3)
~~~
STACK TRACE:
~~~
at System.ServiceProcess.Tests.ServiceControllerTests.AssertExpectedProperties(ServiceController testServiceController)
in D:\j\workspace\outerloop_net---903ddde6\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.Tests\ServiceControllerTests.cs:line 92
at System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper()
in D:\j\workspace\outerloop_net---903ddde6\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.Tests\ServiceControllerTests.cs:line 109
~~~
|
1.0
|
Desktop: System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper failed with "Xunit.Sdk.EqualException" - Failed test: System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper
Detail: https://ci.dot.net/job/dotnet_corefx/job/master/job/outerloop_netfx_windows_nt_debug/66/testReport/System.ServiceProcess.Tests/ServiceControllerTests/ConstructWithServiceName_ToUpper/
Configuration: outerloop_netfx_windows_nt_debug
MESSAGE:
~~~
Assert.Equal() Failure
↓ (pos 3)
Expected: 749b386a-fd41-4bed-951f-75d4a705f7d9
Actual: 749B386A-FD41-4BED-951F-75D4A705F7D9
↑ (pos 3)
~~~
STACK TRACE:
~~~
at System.ServiceProcess.Tests.ServiceControllerTests.AssertExpectedProperties(ServiceController testServiceController)
in D:\j\workspace\outerloop_net---903ddde6\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.Tests\ServiceControllerTests.cs:line 92
at System.ServiceProcess.Tests.ServiceControllerTests.ConstructWithServiceName_ToUpper()
in D:\j\workspace\outerloop_net---903ddde6\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.Tests\ServiceControllerTests.cs:line 109
~~~
|
process
|
desktop system serviceprocess tests servicecontrollertests constructwithservicename toupper failed with xunit sdk equalexception failed test system serviceprocess tests servicecontrollertests constructwithservicename toupper detail configuration outerloop netfx windows nt debug message assert equal failure ↓ pos expected actual ↑ pos stack trace at system serviceprocess tests servicecontrollertests assertexpectedproperties servicecontroller testservicecontroller in d j workspace outerloop net src system serviceprocess servicecontroller tests system serviceprocess servicecontroller tests servicecontrollertests cs line at system serviceprocess tests servicecontrollertests constructwithservicename toupper in d j workspace outerloop net src system serviceprocess servicecontroller tests system serviceprocess servicecontroller tests servicecontrollertests cs line
| 1
|
174,427
| 27,637,871,856
|
IssuesEvent
|
2023-03-10 15:44:41
|
coder/coder
|
https://api.github.com/repos/coder/coder
|
closed
|
Discussion: How should we handle joins?
|
chore design
|
# Problem
Our current use of SQLc uses almost 0 joins throughout all of our queries. This causes excessive db round trips. The function [`workspaceData`][workspaceData] does 9 database calls to completely populate the returned workspace from the api. This can be reduced with [joins][workspace-joined].
When `dbauthz` is enabled by default, this compounds the issue as many objects require fetching related objects to run authorization. Examples: [workspace builds][build-rbac], [template versions][version-rbac], jobs, build parameters, ... .
[build-rbac]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/database/dbauthz/querier.go#L1241-L1251
[version-rbac]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/database/dbauthz/querier.go#L628-L643
[workspaceData]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/workspaces.go#L1047-L1047
[workspace-joined]: https://github.com/coder/coder/blob/6adacd74fa79ae2c424e2144afc65c6e5994510c/coderd/database/sqlxqueries/workspace.gosql#L2-L75
# Solutions [(this PR talks about both)](https://github.com/coder/coder/pull/6371)
Allowing **queries to leverage joins** can reduce db round trips. The issue is how to handle these. We currently use SQLc, which would create a new model for each query and it becomes cumbersome. Additionally, we often have multiple `SELECT` queries for each datatype.
I have 2 proposals.
## Keeping SQLc [(originally investigated here)](https://github.com/coder/coder/issues/2201)
If we want to keep SQLc, the best way we can do this is with **views**. I am suggesting non-materialized views. So essentially these are saved queries in our postgres database that we can reference from SQLc.
Because views are saved in postgres, they require the same migration maintenance a table would.
https://github.com/coder/coder/blob/c3218f71cab8d571c2a95bca08efe2816a41fa86/coderd/database/migrations/000102_workspace_build_view.up.sql#L1-L14
A view is a "table" for SQLc, so a model is generated for it. In our `.sql` files, we just reference the view.
https://github.com/coder/coder/blob/c3218f71cab8d571c2a95bca08efe2816a41fa86/coderd/database/queries/workspacebuilds.sql#L1-L9
## Abandon SQLc and use SQLx + Go Templates [(an implementation here)](https://github.com/coder/coder/pull/6356)
We use SQLc to generate Golang code from our sql. But this code isn't actually that complex in Golang if we use SQLx. And then we can use Go templates to build dynamic queries.
A very basic implementation of this makes our `*.sql` files look like this. Note the highlighting will be a bit off as we are mixing SQL and Go templates.
```sql
{{ define "workspace_builds_rbac" }}
(
SELECT
workspace_builds.*,
workspaces.organization_id AS organization_id,
workspaces.owner_id AS workspace_owner_id
FROM
workspace_builds
INNER JOIN
workspaces ON workspace_builds.workspace_id = workspaces.id
)
{{ end }}
-- To use the template above
{{ define "GetWorkspaceBuildByID" }}
SELECT
*
FROM
{{ template "workspace_builds_rbac" }}
WHERE
id = @build_id
{{ end }}
```
I investigated [IDE Highlighting][IDE-High], but the tl;dr is that just using `sql` highlighting is likely the best. It will never be perfect with Go templates :cry:.
### How to use template query in golang?
I made a package called `sqxqueriers` that handles templates and keeping our `@param` named parameters for easier to read sqlc. You can see that [here](sqlxqueries).
To call a query you made, you can use a generic function called sqlxGet for fetching one row:
https://github.com/coder/coder/blob/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/sqlx.go#L10
The type embeds the sqlc `WorkspaceBuild` type. We would keep SQLc for generating models. We'd just move queries to this SQLx. The `db` struct tags are used for matching to columns. Obviously the query can be handled more manually if the tags do not match.
https://github.com/coder/coder/blob/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/modelqueries.go#L183-L191
[IDE-High]: https://github.com/coder/coder/blob/6adacd74fa79ae2c424e2144afc65c6e5994510c/coderd/database/sqlxqueries/README.md
[sqlxqueries]: https://github.com/coder/coder/tree/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/sqlxqueries
# Comparison
| | SQLc Views | SQLx + Templates |
|--------------------------|--------------------------------|------------------|
| SQL highlighting | ✅ | ⚠️ (kinda) |
| Golang Code | ✅ | ✅ |
| Golang Simple Types | ✅ | ✅ |
| Golang Joined Types | No embeds, duplicated | Uses Embeds |
| Supports Dynamic Queries | ❌ (gross strings replace) | ✅ |
| No Migrations | ❌ (migration to maintain view) | ✅ |
## Highligting
SQL template highlighting in vscode isn't that bad. In Goland it's pretty terrible.
## Golang Code
SQLx + Templates requires more infrastructure code to support, but this also gives us opportunity to add in features. SQLc is slow to update and add features we need.
## Joined Types
SQLc joined types are duplicated structs with identical fields. We can do anonymous embeds for template joined types.
I would really like to see `dbauthz` not adhere to `db.Store` and then we can do some better type handling at this layer to make types more consistent. You cannot insert into a view, so when updating or inserting data, you cannot return the joined data. Meaning 2 types will exist. A `thin` and a `joined`.
## Dynamic Queries
We currently use `CASE WHEN` statements for dynamic where clauses. This works fine, but is only supported in `WHERE`. Things not supported currently:
- Conditional sort order (ASC vs DESC)
- Conditional sort column
- Conditional updates, only update fields provided. [Caused bug in v1](https://github.com/coder/v1/issues/13767)
|
1.0
|
Discussion: How should we handle joins? - # Problem
Our current use of SQLc uses almost 0 joins throughout all of our queries. This causes excessive db round trips. The function [`workspaceData`][workspaceData] does 9 database calls to completely populate the returned workspace from the api. This can be reduced with [joins][workspace-joined].
When `dbauthz` is enabled by default, this compounds the issue as many objects require fetching related objects to run authorization. Examples: [workspace builds][build-rbac], [template versions][version-rbac], jobs, build parameters, ... .
[build-rbac]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/database/dbauthz/querier.go#L1241-L1251
[version-rbac]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/database/dbauthz/querier.go#L628-L643
[workspaceData]: https://github.com/coder/coder/blob/25b05ed8a401cb80cea8e7cdebe58a0c8f9f7f58/coderd/workspaces.go#L1047-L1047
[workspace-joined]: https://github.com/coder/coder/blob/6adacd74fa79ae2c424e2144afc65c6e5994510c/coderd/database/sqlxqueries/workspace.gosql#L2-L75
# Solutions [(this PR talks about both)](https://github.com/coder/coder/pull/6371)
Allowing **queries to leverage joins** can reduce db round trips. The issue is how to handle these. We currently use SQLc, which would create a new model for each query and it becomes cumbersome. Additionally, we often have multiple `SELECT` queries for each datatype.
I have 2 proposals.
## Keeping SQLc [(originally investigated here)](https://github.com/coder/coder/issues/2201)
If we want to keep SQLc, the best way we can do this is with **views**. I am suggesting non-materialized views. So essentially these are saved queries in our postgres database that we can reference from SQLc.
Because views are saved in postgres, they require the same migration maintenance a table would.
https://github.com/coder/coder/blob/c3218f71cab8d571c2a95bca08efe2816a41fa86/coderd/database/migrations/000102_workspace_build_view.up.sql#L1-L14
A view is a "table" for SQLc, so a model is generated for it. In our `.sql` files, we just reference the view.
https://github.com/coder/coder/blob/c3218f71cab8d571c2a95bca08efe2816a41fa86/coderd/database/queries/workspacebuilds.sql#L1-L9
## Abandon SQLc and use SQLx + Go Templates [(an implementation here)](https://github.com/coder/coder/pull/6356)
We use SQLc to generate Golang code from our sql. But this code isn't actually that complex in Golang if we use SQLx. And then we can use Go templates to build dynamic queries.
A very basic implementation of this makes our `*.sql` files look like this. Note the highlighting will be a bit off as we are mixing SQL and Go templates.
```sql
{{ define "workspace_builds_rbac" }}
(
SELECT
workspace_builds.*,
workspaces.organization_id AS organization_id,
workspaces.owner_id AS workspace_owner_id
FROM
workspace_builds
INNER JOIN
workspaces ON workspace_builds.workspace_id = workspaces.id
)
{{ end }}
-- To use the template above
{{ define "GetWorkspaceBuildByID" }}
SELECT
*
FROM
{{ template "workspace_builds_rbac" }}
WHERE
id = @build_id
{{ end }}
```
I investigated [IDE Highlighting][IDE-High], but the tl;dr is that just using `sql` highlighting is likely the best. It will never be perfect with Go templates :cry:.
### How to use template query in golang?
I made a package called `sqxqueriers` that handles templates and keeping our `@param` named parameters for easier to read sqlc. You can see that [here](sqlxqueries).
To call a query you made, you can use a generic function called sqlxGet for fetching one row:
https://github.com/coder/coder/blob/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/sqlx.go#L10
The type embeds the sqlc `WorkspaceBuild` type. We would keep SQLc for generating models. We'd just move queries to this SQLx. The `db` struct tags are used for matching to columns. Obviously the query can be handled more manually if the tags do not match.
https://github.com/coder/coder/blob/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/modelqueries.go#L183-L191
[IDE-High]: https://github.com/coder/coder/blob/6adacd74fa79ae2c424e2144afc65c6e5994510c/coderd/database/sqlxqueries/README.md
[sqlxqueries]: https://github.com/coder/coder/tree/e9f0711c020a0b47e0f06f15b93099d9314be95c/coderd/database/sqlxqueries
# Comparison
| | SQLc Views | SQLx + Templates |
|--------------------------|--------------------------------|------------------|
| SQL highlighting | ✅ | ⚠️ (kinda) |
| Golang Code | ✅ | ✅ |
| Golang Simple Types | ✅ | ✅ |
| Golang Joined Types | No embeds, duplicated | Uses Embeds |
| Supports Dynamic Queries | ❌ (gross strings replace) | ✅ |
| No Migrations | ❌ (migration to maintain view) | ✅ |
## Highligting
SQL template highlighting in vscode isn't that bad. In Goland it's pretty terrible.
## Golang Code
SQLx + Templates requires more infrastructure code to support, but this also gives us opportunity to add in features. SQLc is slow to update and add features we need.
## Joined Types
SQLc joined types are duplicated structs with identical fields. We can do anonymous embeds for template joined types.
I would really like to see `dbauthz` not adhere to `db.Store` and then we can do some better type handling at this layer to make types more consistent. You cannot insert into a view, so when updating or inserting data, you cannot return the joined data. Meaning 2 types will exist. A `thin` and a `joined`.
## Dynamic Queries
We currently use `CASE WHEN` statements for dynamic where clauses. This works fine, but is only supported in `WHERE`. Things not supported currently:
- Conditional sort order (ASC vs DESC)
- Conditional sort column
- Conditional updates, only update fields provided. [Caused bug in v1](https://github.com/coder/v1/issues/13767)
|
non_process
|
discussion how should we handle joins problem our current use of sqlc uses almost joins throughout all of our queries this causes excessive db round trips the function does database calls to completely populate the returned workspace from the api this can be reduced with when dbauthz is enabled by default this compounds the issue as many objects require fetching related objects to run authorization examples jobs build parameters solutions allowing queries to leverage joins can reduce db round trips the issue is how to handle these we currently use sqlc which would create a new model for each query and it becomes cumbersome additionally we often have multiple select queries for each datatype i have proposals keeping sqlc if we want to keep sqlc the best way we can do this is with views i am suggesting non materialized views so essentially these are saved queries in our postgres database that we can reference from sqlc because views are saved in postgres they require the same migration maintenance a table would a view is a table for sqlc so a model is generated for it in our sql files we just reference the view abandon sqlc and use sqlx go templates we use sqlc to generate golang code from our sql but this code isn t actually that complex in golang if we use sqlx and then we can use go templates to build dynamic queries a very basic implementation of this makes our sql files look like this note the highlighting will be a bit off as we are mixing sql and go templates sql define workspace builds rbac select workspace builds workspaces organization id as organization id workspaces owner id as workspace owner id from workspace builds inner join workspaces on workspace builds workspace id workspaces id end to use the template above define getworkspacebuildbyid select from template workspace builds rbac where id build id end i investigated but the tl dr is that just using sql highlighting is likely the best it will never be perfect with go templates cry how to use template query in golang i made a package called sqxqueriers that handles templates and keeping our param named parameters for easier to read sqlc you can see that sqlxqueries to call a query you made you can use a generic function called sqlxget for fetching one row the type embeds the sqlc workspacebuild type we would keep sqlc for generating models we d just move queries to this sqlx the db struct tags are used for matching to columns obviously the query can be handled more manually if the tags do not match comparison sqlc views sqlx templates sql highlighting ✅ ⚠️ kinda golang code ✅ ✅ golang simple types ✅ ✅ golang joined types no embeds duplicated uses embeds supports dynamic queries ❌ gross strings replace ✅ no migrations ❌ migration to maintain view ✅ highligting sql template highlighting in vscode isn t that bad in goland it s pretty terrible golang code sqlx templates requires more infrastructure code to support but this also gives us opportunity to add in features sqlc is slow to update and add features we need joined types sqlc joined types are duplicated structs with identical fields we can do anonymous embeds for template joined types i would really like to see dbauthz not adhere to db store and then we can do some better type handling at this layer to make types more consistent you cannot insert into a view so when updating or inserting data you cannot return the joined data meaning types will exist a thin and a joined dynamic queries we currently use case when statements for dynamic where clauses this works fine but is only supported in where things not supported currently conditional sort order asc vs desc conditional sort column conditional updates only update fields provided
| 0
|
125,235
| 4,954,634,826
|
IssuesEvent
|
2016-12-01 18:09:09
|
orcidee/rpgconmanager
|
https://api.github.com/repos/orcidee/rpgconmanager
|
closed
|
Griser le nb de tables
|
bug Priority : 1
|
Il faudrait griser/rendre visiblement non-éditable le champs nombre de tables quand la partie a été vérifiée et validée.
|
1.0
|
Griser le nb de tables - Il faudrait griser/rendre visiblement non-éditable le champs nombre de tables quand la partie a été vérifiée et validée.
|
non_process
|
griser le nb de tables il faudrait griser rendre visiblement non éditable le champs nombre de tables quand la partie a été vérifiée et validée
| 0
|
5,761
| 8,598,978,418
|
IssuesEvent
|
2018-11-15 23:47:47
|
w3c/w3process
|
https://api.github.com/repos/w3c/w3process
|
closed
|
Does a spec need to be Rec-ready to be a CR? Align defn with process
|
Agenda+ PendingReview Process2019Candidate
|
https://www.w3.org/2018/Process-20180201/#candidate-rec sets out requirements for advancement to CR. It does not require any statement of reasons why a CR may [update: instating omitted text from original issue filing, with apologies] _not_ advance to Rec. [/update]
However https://www.w3.org/2018/Process-20180201/#RecsCR defines what a CR is and adds informatively:
> Note: Candidate Recommendations are expected to be acceptable as Recommendations. Announcement of a different next step should include the reasons why the change in expectations comes at so late a stage.
This (first part) is not tested in the requirements for advancing to CR, and in fact it is not unusual for substantive changes to be made after first CR, on the path to Rec. The second part is very reasonable, however it omits an important real world case, which is non-advancement for an unreasonably long period.
As a minimum, the two sections should be aligned so that the tests for advancement clearly verify the requirements.
My preference for achieving the alignment would be to remove or modify the expectation from the note, since I don't think it reflects reality or is particularly helpful. It is obvious enough that a CR is a step along the path to Rec and that there are extra steps needed if making substantive changes between CR and Rec. Conversely, if the substantive content of a CR is acceptable as a Rec, there are good reasons why we do not just publish it as a Rec: the "acceptability" depends not on the document in itself, but rather on the demonstration that implementers and the AC and Director agree that it is acceptable. So the statement of expectation doesn't seem to be clear enough.
I would propose:
> Note: Candidate Recommendations are expected to be advanced to Recommendation. Announcement of a different next step should include the reasons why the change in expectations comes at so late a stage.
Handling the second problem, omission of what should happen if a document stays in CR for an unreasonably long time, is harder to fix and probably needs some discussion. Automatically revert the CR to a Note? I imagine that would be unacceptable to some people, but it might focus attention on tests and implementation reports... If it is going to be picked up for discussion, probably better for someone to raise a new issue for tracking purposes.
|
1.0
|
Does a spec need to be Rec-ready to be a CR? Align defn with process - https://www.w3.org/2018/Process-20180201/#candidate-rec sets out requirements for advancement to CR. It does not require any statement of reasons why a CR may [update: instating omitted text from original issue filing, with apologies] _not_ advance to Rec. [/update]
However https://www.w3.org/2018/Process-20180201/#RecsCR defines what a CR is and adds informatively:
> Note: Candidate Recommendations are expected to be acceptable as Recommendations. Announcement of a different next step should include the reasons why the change in expectations comes at so late a stage.
This (first part) is not tested in the requirements for advancing to CR, and in fact it is not unusual for substantive changes to be made after first CR, on the path to Rec. The second part is very reasonable, however it omits an important real world case, which is non-advancement for an unreasonably long period.
As a minimum, the two sections should be aligned so that the tests for advancement clearly verify the requirements.
My preference for achieving the alignment would be to remove or modify the expectation from the note, since I don't think it reflects reality or is particularly helpful. It is obvious enough that a CR is a step along the path to Rec and that there are extra steps needed if making substantive changes between CR and Rec. Conversely, if the substantive content of a CR is acceptable as a Rec, there are good reasons why we do not just publish it as a Rec: the "acceptability" depends not on the document in itself, but rather on the demonstration that implementers and the AC and Director agree that it is acceptable. So the statement of expectation doesn't seem to be clear enough.
I would propose:
> Note: Candidate Recommendations are expected to be advanced to Recommendation. Announcement of a different next step should include the reasons why the change in expectations comes at so late a stage.
Handling the second problem, omission of what should happen if a document stays in CR for an unreasonably long time, is harder to fix and probably needs some discussion. Automatically revert the CR to a Note? I imagine that would be unacceptable to some people, but it might focus attention on tests and implementation reports... If it is going to be picked up for discussion, probably better for someone to raise a new issue for tracking purposes.
|
process
|
does a spec need to be rec ready to be a cr align defn with process sets out requirements for advancement to cr it does not require any statement of reasons why a cr may not advance to rec however defines what a cr is and adds informatively note candidate recommendations are expected to be acceptable as recommendations announcement of a different next step should include the reasons why the change in expectations comes at so late a stage this first part is not tested in the requirements for advancing to cr and in fact it is not unusual for substantive changes to be made after first cr on the path to rec the second part is very reasonable however it omits an important real world case which is non advancement for an unreasonably long period as a minimum the two sections should be aligned so that the tests for advancement clearly verify the requirements my preference for achieving the alignment would be to remove or modify the expectation from the note since i don t think it reflects reality or is particularly helpful it is obvious enough that a cr is a step along the path to rec and that there are extra steps needed if making substantive changes between cr and rec conversely if the substantive content of a cr is acceptable as a rec there are good reasons why we do not just publish it as a rec the acceptability depends not on the document in itself but rather on the demonstration that implementers and the ac and director agree that it is acceptable so the statement of expectation doesn t seem to be clear enough i would propose note candidate recommendations are expected to be advanced to recommendation announcement of a different next step should include the reasons why the change in expectations comes at so late a stage handling the second problem omission of what should happen if a document stays in cr for an unreasonably long time is harder to fix and probably needs some discussion automatically revert the cr to a note i imagine that would be unacceptable to some people but it might focus attention on tests and implementation reports if it is going to be picked up for discussion probably better for someone to raise a new issue for tracking purposes
| 1
|
17,230
| 22,917,655,768
|
IssuesEvent
|
2022-07-17 07:34:08
|
streamnative/flink
|
https://api.github.com/repos/streamnative/flink
|
closed
|
[SQL Connector] create default database when PulsarCatalog is created if default database is not present.
|
compute/data-processing
|
Currently, the default database might not be present if users didn't create the database explicitly. We should create the database upon PulsarCatalog creation
|
1.0
|
[SQL Connector] create default database when PulsarCatalog is created if default database is not present. - Currently, the default database might not be present if users didn't create the database explicitly. We should create the database upon PulsarCatalog creation
|
process
|
create default database when pulsarcatalog is created if default database is not present currently the default database might not be present if users didn t create the database explicitly we should create the database upon pulsarcatalog creation
| 1
|
2,208
| 5,049,113,886
|
IssuesEvent
|
2016-12-20 15:02:43
|
CERNDocumentServer/cds
|
https://api.github.com/repos/CERNDocumentServer/cds
|
closed
|
webhooks: cancel tasks
|
avc_processing review
|
When canceling an event that **is still running** we need to stop all the celery tasks associated with it and also delete any file (on disk and/or db) that was created.
- Downloader:
- [x] Stop download task and delete the `ObjectVersion` and the file on disk
- AVCWorflow:
- [x] If the master file was (is being) downloaded by the event, stop the task if still running and delete the `ObjectVersion` and the file on disk
- [x] Delete any extracted metadata that might have being added to the deposit and also any `ObjectVersionTag` that might have being created (specially if the file was upload via http)
- [x] Stop all the Sorenson jobs
- [x] Delete any `ObjectVersion` that the transcode task might have created, if any of the jobs have finished we also need to delete the file from disk
|
1.0
|
webhooks: cancel tasks - When canceling an event that **is still running** we need to stop all the celery tasks associated with it and also delete any file (on disk and/or db) that was created.
- Downloader:
- [x] Stop download task and delete the `ObjectVersion` and the file on disk
- AVCWorflow:
- [x] If the master file was (is being) downloaded by the event, stop the task if still running and delete the `ObjectVersion` and the file on disk
- [x] Delete any extracted metadata that might have being added to the deposit and also any `ObjectVersionTag` that might have being created (specially if the file was upload via http)
- [x] Stop all the Sorenson jobs
- [x] Delete any `ObjectVersion` that the transcode task might have created, if any of the jobs have finished we also need to delete the file from disk
|
process
|
webhooks cancel tasks when canceling an event that is still running we need to stop all the celery tasks associated with it and also delete any file on disk and or db that was created downloader stop download task and delete the objectversion and the file on disk avcworflow if the master file was is being downloaded by the event stop the task if still running and delete the objectversion and the file on disk delete any extracted metadata that might have being added to the deposit and also any objectversiontag that might have being created specially if the file was upload via http stop all the sorenson jobs delete any objectversion that the transcode task might have created if any of the jobs have finished we also need to delete the file from disk
| 1
|
142,301
| 19,089,396,360
|
IssuesEvent
|
2021-11-29 10:22:03
|
tharun453/samples
|
https://api.github.com/repos/tharun453/samples
|
opened
|
CVE-2021-23358 (High) detected in underscore-1.7.0.tgz
|
security vulnerability
|
## CVE-2021-23358 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore-1.7.0.tgz</b></p></summary>
<p>JavaScript's functional programming helper library.</p>
<p>Library home page: <a href="https://registry.npmjs.org/underscore/-/underscore-1.7.0.tgz">https://registry.npmjs.org/underscore/-/underscore-1.7.0.tgz</a></p>
<p>Path to dependency file: samples/core/tutorials/buggyamb/BuggyAmb/wwwroot/scripts/jquery-ui-1.12.1/package.json</p>
<p>Path to vulnerable library: samples/core/tutorials/buggyamb/BuggyAmb/wwwroot/scripts/jquery-ui-1.12.1/node_modules/underscore/package.json</p>
<p>
Dependency Hierarchy:
- grunt-0.4.5.tgz (Root Library)
- js-yaml-2.0.5.tgz
- argparse-0.1.16.tgz
- :x: **underscore-1.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tharun453/samples/commit/0d7f1931b9759c22f0469b959114a5d94f8f92e4">0d7f1931b9759c22f0469b959114a5d94f8f92e4</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package underscore from 1.13.0-0 and before 1.13.0-2, from 1.3.2 and before 1.12.1 are vulnerable to Arbitrary Code Injection via the template function, particularly when a variable property is passed as an argument as it is not sanitized.
<p>Publish Date: 2021-03-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23358>CVE-2021-23358</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358</a></p>
<p>Release Date: 2021-03-29</p>
<p>Fix Resolution: underscore - 1.12.1,1.13.0-2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23358 (High) detected in underscore-1.7.0.tgz - ## CVE-2021-23358 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore-1.7.0.tgz</b></p></summary>
<p>JavaScript's functional programming helper library.</p>
<p>Library home page: <a href="https://registry.npmjs.org/underscore/-/underscore-1.7.0.tgz">https://registry.npmjs.org/underscore/-/underscore-1.7.0.tgz</a></p>
<p>Path to dependency file: samples/core/tutorials/buggyamb/BuggyAmb/wwwroot/scripts/jquery-ui-1.12.1/package.json</p>
<p>Path to vulnerable library: samples/core/tutorials/buggyamb/BuggyAmb/wwwroot/scripts/jquery-ui-1.12.1/node_modules/underscore/package.json</p>
<p>
Dependency Hierarchy:
- grunt-0.4.5.tgz (Root Library)
- js-yaml-2.0.5.tgz
- argparse-0.1.16.tgz
- :x: **underscore-1.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tharun453/samples/commit/0d7f1931b9759c22f0469b959114a5d94f8f92e4">0d7f1931b9759c22f0469b959114a5d94f8f92e4</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package underscore from 1.13.0-0 and before 1.13.0-2, from 1.3.2 and before 1.12.1 are vulnerable to Arbitrary Code Injection via the template function, particularly when a variable property is passed as an argument as it is not sanitized.
<p>Publish Date: 2021-03-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23358>CVE-2021-23358</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358</a></p>
<p>Release Date: 2021-03-29</p>
<p>Fix Resolution: underscore - 1.12.1,1.13.0-2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in underscore tgz cve high severity vulnerability vulnerable library underscore tgz javascript s functional programming helper library library home page a href path to dependency file samples core tutorials buggyamb buggyamb wwwroot scripts jquery ui package json path to vulnerable library samples core tutorials buggyamb buggyamb wwwroot scripts jquery ui node modules underscore package json dependency hierarchy grunt tgz root library js yaml tgz argparse tgz x underscore tgz vulnerable library found in head commit a href found in base branch main vulnerability details the package underscore from and before from and before are vulnerable to arbitrary code injection via the template function particularly when a variable property is passed as an argument as it is not sanitized publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution underscore step up your open source security game with whitesource
| 0
|
3,903
| 6,823,007,495
|
IssuesEvent
|
2017-11-07 22:06:15
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
closed
|
whenBlock search by date seems broken
|
status-inprocess tools-whenBlock type-bug
|
When I run this (following the command help):
./whenBlock 2017-03-02
We obtain this strange error. Notice that I am using the date returned for block 0.
I could not make it work with any date, most of the times it returns this error
-------------------------------------------------------------------------------------------------------------------
whenBlock argc: 2 [1:2017-03-02]
whenBlock 2017-03-02
Could not open the mini-block database: /home/carlos/.quickBlocks/cache/miniBlocks.bin.
Usage: whenBlock [-a|-l|-v|-h] date / block
Purpose: Finds the nearest block before the JSON-formatted date, or the nearest date before the given block.
Where:
date / block one of the special values listed below or YYYY-MM-DD-[HH[:MM[:SS]]] or a blockNumber (required)
-a (--alone) show the found block or found date unadorned (useful for scripting)
-l (--list) list special blocks timestamps and dates
-v (--verbose) set verbose level. Either -v, --verbose or -v:n where 'n' is level
-h (--help) display this help screen
Powered by QuickBlocks
|
1.0
|
whenBlock search by date seems broken - When I run this (following the command help):
./whenBlock 2017-03-02
We obtain this strange error. Notice that I am using the date returned for block 0.
I could not make it work with any date, most of the times it returns this error
-------------------------------------------------------------------------------------------------------------------
whenBlock argc: 2 [1:2017-03-02]
whenBlock 2017-03-02
Could not open the mini-block database: /home/carlos/.quickBlocks/cache/miniBlocks.bin.
Usage: whenBlock [-a|-l|-v|-h] date / block
Purpose: Finds the nearest block before the JSON-formatted date, or the nearest date before the given block.
Where:
date / block one of the special values listed below or YYYY-MM-DD-[HH[:MM[:SS]]] or a blockNumber (required)
-a (--alone) show the found block or found date unadorned (useful for scripting)
-l (--list) list special blocks timestamps and dates
-v (--verbose) set verbose level. Either -v, --verbose or -v:n where 'n' is level
-h (--help) display this help screen
Powered by QuickBlocks
|
process
|
whenblock search by date seems broken when i run this following the command help whenblock we obtain this strange error notice that i am using the date returned for block i could not make it work with any date most of the times it returns this error whenblock argc whenblock could not open the mini block database home carlos quickblocks cache miniblocks bin usage whenblock date block purpose finds the nearest block before the json formatted date or the nearest date before the given block where date block one of the special values listed below or yyyy mm dd or a blocknumber required a alone show the found block or found date unadorned useful for scripting l list list special blocks timestamps and dates v verbose set verbose level either v verbose or v n where n is level h help display this help screen powered by quickblocks
| 1
|
273,128
| 20,772,510,979
|
IssuesEvent
|
2022-03-16 07:05:01
|
DLR-RM/stable-baselines3
|
https://api.github.com/repos/DLR-RM/stable-baselines3
|
closed
|
The code problem of Custom Feature Extractor
|
documentation
|
**Important Note: We do not do technical support, nor consulting** and don't answer personal questions per email.
Please post your question on the [RL Discord](https://discord.com/invite/xhfNqQv), [Reddit](https://www.reddit.com/r/reinforcementlearning/) or [Stack Overflow](https://stackoverflow.com/) in that case.
### 📚 Documentation
####problem
In [this code](https://github.com/DLR-RM/stable-baselines3/blob/master/docs/guide/custom_policy.rst#custom-feature-extractor)
1:what's the function of PART"with torch.no_grad...." ? why use torch.no_grad to compute "n_flatten"
2: what's the meaning of n_flatten ?
### Checklist
- [x] I have read the [documentation](https://stable-baselines3.readthedocs.io/en/master/) (**required**)
- [x] I have checked that there is no similar [issue](https://github.com/DLR-RM/stable-baselines3/issues) in the repo (**required**)
###
<!--- This Template is an edited version of the one from https://github.com/pytorch/pytorch -->
|
1.0
|
The code problem of Custom Feature Extractor - **Important Note: We do not do technical support, nor consulting** and don't answer personal questions per email.
Please post your question on the [RL Discord](https://discord.com/invite/xhfNqQv), [Reddit](https://www.reddit.com/r/reinforcementlearning/) or [Stack Overflow](https://stackoverflow.com/) in that case.
### 📚 Documentation
####problem
In [this code](https://github.com/DLR-RM/stable-baselines3/blob/master/docs/guide/custom_policy.rst#custom-feature-extractor)
1:what's the function of PART"with torch.no_grad...." ? why use torch.no_grad to compute "n_flatten"
2: what's the meaning of n_flatten ?
### Checklist
- [x] I have read the [documentation](https://stable-baselines3.readthedocs.io/en/master/) (**required**)
- [x] I have checked that there is no similar [issue](https://github.com/DLR-RM/stable-baselines3/issues) in the repo (**required**)
###
<!--- This Template is an edited version of the one from https://github.com/pytorch/pytorch -->
|
non_process
|
the code problem of custom feature extractor important note we do not do technical support nor consulting and don t answer personal questions per email please post your question on the or in that case 📚 documentation problem in what s the function of part with torch no grad why use torch no grad to compute n flatten what s the meaning of n flatten checklist i have read the required i have checked that there is no similar in the repo required
| 0
|
5,264
| 8,057,651,126
|
IssuesEvent
|
2018-08-02 15:57:34
|
GoogleCloudPlatform/google-cloud-python
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
|
closed
|
BigQuery: 'Client.update_dataset' systest error, 412 PreconditionFailed
|
api: bigquery backend flaky testing type: process
|
First error in [this CI run](https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/7126):
```python
_______________________ TestBigQuery.test_update_dataset _______________________
self = <tests.system.TestBigQuery testMethod=test_update_dataset>
def test_update_dataset(self):
dataset = self.temp_dataset(_make_dataset_id('update_dataset'))
self.assertTrue(_dataset_exists(dataset))
self.assertIsNone(dataset.friendly_name)
self.assertIsNone(dataset.description)
self.assertEquals(dataset.labels, {})
dataset.friendly_name = 'Friendly'
dataset.description = 'Description'
dataset.labels = {'priority': 'high', 'color': 'blue'}
ds2 = Config.CLIENT.update_dataset(
dataset,
> ('friendly_name', 'description', 'labels'))
tests/system.py:189:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google/cloud/bigquery/client.py:387: in update_dataset
headers=headers)
...
if not 200 <= response.status_code < 300:
> raise exceptions.from_http_response(response)
E PreconditionFailed: 412 PATCH https://www.googleapis.com/bigquery/v2/projects/precise-truck-742/datasets/update_dataset_7126_1531263324: Precondition Failed
../.nox/sys-2-7/lib/python2.7/site-packages/google/cloud/_http.py:293: PreconditionFailed
```
|
1.0
|
BigQuery: 'Client.update_dataset' systest error, 412 PreconditionFailed - First error in [this CI run](https://circleci.com/gh/GoogleCloudPlatform/google-cloud-python/7126):
```python
_______________________ TestBigQuery.test_update_dataset _______________________
self = <tests.system.TestBigQuery testMethod=test_update_dataset>
def test_update_dataset(self):
dataset = self.temp_dataset(_make_dataset_id('update_dataset'))
self.assertTrue(_dataset_exists(dataset))
self.assertIsNone(dataset.friendly_name)
self.assertIsNone(dataset.description)
self.assertEquals(dataset.labels, {})
dataset.friendly_name = 'Friendly'
dataset.description = 'Description'
dataset.labels = {'priority': 'high', 'color': 'blue'}
ds2 = Config.CLIENT.update_dataset(
dataset,
> ('friendly_name', 'description', 'labels'))
tests/system.py:189:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google/cloud/bigquery/client.py:387: in update_dataset
headers=headers)
...
if not 200 <= response.status_code < 300:
> raise exceptions.from_http_response(response)
E PreconditionFailed: 412 PATCH https://www.googleapis.com/bigquery/v2/projects/precise-truck-742/datasets/update_dataset_7126_1531263324: Precondition Failed
../.nox/sys-2-7/lib/python2.7/site-packages/google/cloud/_http.py:293: PreconditionFailed
```
|
process
|
bigquery client update dataset systest error preconditionfailed first error in python testbigquery test update dataset self def test update dataset self dataset self temp dataset make dataset id update dataset self asserttrue dataset exists dataset self assertisnone dataset friendly name self assertisnone dataset description self assertequals dataset labels dataset friendly name friendly dataset description description dataset labels priority high color blue config client update dataset dataset friendly name description labels tests system py google cloud bigquery client py in update dataset headers headers if not response status code raise exceptions from http response response e preconditionfailed patch precondition failed nox sys lib site packages google cloud http py preconditionfailed
| 1
|
14,793
| 18,066,914,025
|
IssuesEvent
|
2021-09-20 20:21:34
|
Jeffail/benthos
|
https://api.github.com/repos/Jeffail/benthos
|
closed
|
Support dynamic queries in sql processors
|
enhancement processors effort: lower
|
Hello, could you support dynamic queries also for SQL processors like [AWS dynamo db] (https://www.benthos.dev/docs/components/processors/aws_dynamodb_partiql/) does?
Otherwise, the user has to write multiple configs for different queries.
|
1.0
|
Support dynamic queries in sql processors - Hello, could you support dynamic queries also for SQL processors like [AWS dynamo db] (https://www.benthos.dev/docs/components/processors/aws_dynamodb_partiql/) does?
Otherwise, the user has to write multiple configs for different queries.
|
process
|
support dynamic queries in sql processors hello could you support dynamic queries also for sql processors like does otherwise the user has to write multiple configs for different queries
| 1
|
4,395
| 7,286,431,786
|
IssuesEvent
|
2018-02-23 09:43:36
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[needs-docs]Rename Node Editor and Node Tool (fixes #17807)
|
Automatic new feature Easy Processing Screenshots
|
Original commit: https://github.com/qgis/QGIS/commit/5461d3c52ace8f7afe72b9215a0b0532574294e2 by DelazJ
Unfortunately this naughty coder did not write a description... :-(
|
1.0
|
[needs-docs]Rename Node Editor and Node Tool (fixes #17807) - Original commit: https://github.com/qgis/QGIS/commit/5461d3c52ace8f7afe72b9215a0b0532574294e2 by DelazJ
Unfortunately this naughty coder did not write a description... :-(
|
process
|
rename node editor and node tool fixes original commit by delazj unfortunately this naughty coder did not write a description
| 1
|
204,336
| 7,086,994,848
|
IssuesEvent
|
2018-01-11 16:25:31
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
Do stuff with cloud provider's DNS
|
lifecycle/stale priority/backlog sig/network
|
In #9682, @pnovotnak wrote:
_I'd like to create resources that are accessible within my private network at my cloud provider (though outside my Kubernetes cluster), but not to the outside internet, that are resolvable via DNS within the cluster. Ironically it seems like endpoint IPs fit the bill (although they change and are not resolvable) whereas service IPs, which are stable and what, are not accessible from outside the cluster._
to which @thockin responded:
_Reaching up into the cloud-provider's DNS is in the cards for sometime after 1.0_
|
1.0
|
Do stuff with cloud provider's DNS - In #9682, @pnovotnak wrote:
_I'd like to create resources that are accessible within my private network at my cloud provider (though outside my Kubernetes cluster), but not to the outside internet, that are resolvable via DNS within the cluster. Ironically it seems like endpoint IPs fit the bill (although they change and are not resolvable) whereas service IPs, which are stable and what, are not accessible from outside the cluster._
to which @thockin responded:
_Reaching up into the cloud-provider's DNS is in the cards for sometime after 1.0_
|
non_process
|
do stuff with cloud provider s dns in pnovotnak wrote i d like to create resources that are accessible within my private network at my cloud provider though outside my kubernetes cluster but not to the outside internet that are resolvable via dns within the cluster ironically it seems like endpoint ips fit the bill although they change and are not resolvable whereas service ips which are stable and what are not accessible from outside the cluster to which thockin responded reaching up into the cloud provider s dns is in the cards for sometime after
| 0
|
11,169
| 13,957,694,554
|
IssuesEvent
|
2020-10-24 08:11:17
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
SE: Harvesting Request
|
Geoportal Harvesting process SE - Sweden
|
Hi!
After some updates by our contributers we would neeed a new harvesting of the Swedish node.
Have a great weekend!
Björn Olofsson, The Swedish Geoportal
|
1.0
|
SE: Harvesting Request - Hi!
After some updates by our contributers we would neeed a new harvesting of the Swedish node.
Have a great weekend!
Björn Olofsson, The Swedish Geoportal
|
process
|
se harvesting request hi after some updates by our contributers we would neeed a new harvesting of the swedish node have a great weekend bj ouml rn olofsson the swedish geoportal
| 1
|
11,888
| 14,681,300,946
|
IssuesEvent
|
2020-12-31 12:49:28
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
Studies tab > Loader icon is missing at the bottom of the page
|
Bug P1 Participant manager Process: Dev Process: Fixed Process: Tested QA
|
AR : Studies tab > Loader icon is missing at the bottom of the page
ER : Loader icon should be present
(Note : Only 10 sets of data should load at a time)

|
3.0
|
Studies tab > Loader icon is missing at the bottom of the page - AR : Studies tab > Loader icon is missing at the bottom of the page
ER : Loader icon should be present
(Note : Only 10 sets of data should load at a time)

|
process
|
studies tab loader icon is missing at the bottom of the page ar studies tab loader icon is missing at the bottom of the page er loader icon should be present note only sets of data should load at a time
| 1
|
821,601
| 30,828,067,633
|
IssuesEvent
|
2023-08-01 21:57:45
|
dotCMS/core
|
https://api.github.com/repos/dotCMS/core
|
closed
|
Design System: Implement `button` component
|
Team : Lunik Type : New Functionality Triage OKR : Core Features Priority : 2 High
|
### Parent Issue
#25355
### User Story
As a stakeholder, I want you to implement the button component new design for primeng and dojo widgets.
### Acceptance Criteria
1. Match the design provided with all the types and severities in primeng
2. Add Storybook stories for the buttons with the same structure of primeng
3. Implement the buttons in dojo
4. User the CSS variables accordingly
### Proposed Objective
Core Features
### Proposed Priority
Priority 3 - Average
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
[Figma](https://www.figma.com/file/7CzbMcGSbTIJjerWtcapAm/Design-System-2023?type=design&node-id=42%3A689&mode=dev)
### Assumptions & Initiation Needs
[_No response_](https://github.com/dotCMS/core/assets/751424/52512905-2ff6-46c3-8244-cb472ab1a83e)
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_
|
1.0
|
Design System: Implement `button` component - ### Parent Issue
#25355
### User Story
As a stakeholder, I want you to implement the button component new design for primeng and dojo widgets.
### Acceptance Criteria
1. Match the design provided with all the types and severities in primeng
2. Add Storybook stories for the buttons with the same structure of primeng
3. Implement the buttons in dojo
4. User the CSS variables accordingly
### Proposed Objective
Core Features
### Proposed Priority
Priority 3 - Average
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
[Figma](https://www.figma.com/file/7CzbMcGSbTIJjerWtcapAm/Design-System-2023?type=design&node-id=42%3A689&mode=dev)
### Assumptions & Initiation Needs
[_No response_](https://github.com/dotCMS/core/assets/751424/52512905-2ff6-46c3-8244-cb472ab1a83e)
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_
|
non_process
|
design system implement button component parent issue user story as a stakeholder i want you to implement the button component new design for primeng and dojo widgets acceptance criteria match the design provided with all the types and severities in primeng add storybook stories for the buttons with the same structure of primeng implement the buttons in dojo user the css variables accordingly proposed objective core features proposed priority priority average external links slack conversations support tickets figma designs etc assumptions initiation needs quality assurance notes workarounds no response sub tasks estimates no response
| 0
|
12,057
| 14,231,685,221
|
IssuesEvent
|
2020-11-18 09:54:27
|
ValveSoftware/Proton
|
https://api.github.com/repos/ValveSoftware/Proton
|
reopened
|
428: Shibuya Scramble (648580)
|
Game compatibility - Unofficial
|
# Compatibility Report
- Name of the game with compatibility issues: 428: Shibuya Scramble
- Steam AppID of the game: 648580
## System Information
- GPU: R9 390
- Driver/LLVM version: Mesa 20.2.2-2
- Kernel version: 5.9.8.zen1-1 (identical behavior with 5.9.8.arch1-1)
- Link to full system information report as [Gist](https://gist.github.com/): https://gist.github.com/hajnal-endot/8028dacae9ca79bac10d82aa842ccb93
- Proton version: 5.13-2
## I confirm:
- [yes ] that I haven't found an existing compatibility report for this game.
- [yes ] that I have checked whether there are updates for my system available.
<!-- Please add `PROTON_LOG=1 %command%` to the game's launch options and drag
and drop the generated `$HOME/steam-$APPID.log` into this issue report -->
[steam-648580.log](https://github.com/ValveSoftware/Proton/files/5557001/steam-648580.log)
## Symptoms <!-- What's the problem? -->
Game attempts to open in fullscreen, black screen appears in fullscreen, game immediately crashes to desktop.
## Reproduction
1. Open game through Steam.
2. Observe issue.
COMMENT: There's a good chance the answer is staring me in the face and I'm just really stupid. Others seem to have very little trouble with this game per ProtonDB, and this is a newly-configured system that I haven't ironed out all the issues with yet.
|
True
|
428: Shibuya Scramble (648580) - # Compatibility Report
- Name of the game with compatibility issues: 428: Shibuya Scramble
- Steam AppID of the game: 648580
## System Information
- GPU: R9 390
- Driver/LLVM version: Mesa 20.2.2-2
- Kernel version: 5.9.8.zen1-1 (identical behavior with 5.9.8.arch1-1)
- Link to full system information report as [Gist](https://gist.github.com/): https://gist.github.com/hajnal-endot/8028dacae9ca79bac10d82aa842ccb93
- Proton version: 5.13-2
## I confirm:
- [yes ] that I haven't found an existing compatibility report for this game.
- [yes ] that I have checked whether there are updates for my system available.
<!-- Please add `PROTON_LOG=1 %command%` to the game's launch options and drag
and drop the generated `$HOME/steam-$APPID.log` into this issue report -->
[steam-648580.log](https://github.com/ValveSoftware/Proton/files/5557001/steam-648580.log)
## Symptoms <!-- What's the problem? -->
Game attempts to open in fullscreen, black screen appears in fullscreen, game immediately crashes to desktop.
## Reproduction
1. Open game through Steam.
2. Observe issue.
COMMENT: There's a good chance the answer is staring me in the face and I'm just really stupid. Others seem to have very little trouble with this game per ProtonDB, and this is a newly-configured system that I haven't ironed out all the issues with yet.
|
non_process
|
shibuya scramble compatibility report name of the game with compatibility issues shibuya scramble steam appid of the game system information gpu driver llvm version mesa kernel version identical behavior with link to full system information report as proton version i confirm that i haven t found an existing compatibility report for this game that i have checked whether there are updates for my system available please add proton log command to the game s launch options and drag and drop the generated home steam appid log into this issue report symptoms game attempts to open in fullscreen black screen appears in fullscreen game immediately crashes to desktop reproduction open game through steam observe issue comment there s a good chance the answer is staring me in the face and i m just really stupid others seem to have very little trouble with this game per protondb and this is a newly configured system that i haven t ironed out all the issues with yet
| 0
|
4,990
| 7,822,175,201
|
IssuesEvent
|
2018-06-14 00:49:16
|
StrikeNP/trac_test
|
https://api.github.com/repos/StrikeNP/trac_test
|
closed
|
Add a function to convert.m to changed a pressure profile into altitude (Trac #4)
|
Migrated from Trac enhancement fasching@uwm.edu post_processing
|
Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/4
```json
{
"status": "closed",
"changetime": "2009-05-16T10:07:24",
"description": "Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.",
"reporter": "fasching@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1242468444000000",
"component": "post_processing",
"summary": "Add a function to convert.m to changed a pressure profile into altitude",
"priority": "minor",
"keywords": "conversion, MATLAB",
"time": "2009-05-01T21:20:08",
"milestone": "",
"owner": "fasching@uwm.edu",
"type": "enhancement"
}
```
|
1.0
|
Add a function to convert.m to changed a pressure profile into altitude (Trac #4) - Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/4
```json
{
"status": "closed",
"changetime": "2009-05-16T10:07:24",
"description": "Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.",
"reporter": "fasching@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1242468444000000",
"component": "post_processing",
"summary": "Add a function to convert.m to changed a pressure profile into altitude",
"priority": "minor",
"keywords": "conversion, MATLAB",
"time": "2009-05-01T21:20:08",
"milestone": "",
"owner": "fasching@uwm.edu",
"type": "enhancement"
}
```
|
process
|
add a function to convert m to changed a pressure profile into altitude trac add a function to convert m to changed a pressure profile into altitude this would be useful for cases that do not specify things in terms of altitude attachments migrated from json status closed changetime description add a function to convert m to changed a pressure profile into altitude this would be useful for cases that do not specify things in terms of altitude reporter fasching uwm edu cc resolution verified by v larson ts component post processing summary add a function to convert m to changed a pressure profile into altitude priority minor keywords conversion matlab time milestone owner fasching uwm edu type enhancement
| 1
|
81,617
| 15,781,626,709
|
IssuesEvent
|
2021-04-01 11:39:45
|
GTNewHorizons/GT-New-Horizons-Modpack
|
https://api.github.com/repos/GTNewHorizons/GT-New-Horizons-Modpack
|
closed
|
(Enderio) Inventory Panel Sorting Issues
|
Status: stale Type: Need Code changes Type: suggestion
|
_Using the EnderIO the inventory panel has options for how to sort its contents. However unlike in other systems such as an ME system it does not keep the sorting preference once left and re-entered, which can be quite annoying. Bart recommended me to make this ticket as its likely not a terribly hard fix._
|
1.0
|
(Enderio) Inventory Panel Sorting Issues - _Using the EnderIO the inventory panel has options for how to sort its contents. However unlike in other systems such as an ME system it does not keep the sorting preference once left and re-entered, which can be quite annoying. Bart recommended me to make this ticket as its likely not a terribly hard fix._
|
non_process
|
enderio inventory panel sorting issues using the enderio the inventory panel has options for how to sort its contents however unlike in other systems such as an me system it does not keep the sorting preference once left and re entered which can be quite annoying bart recommended me to make this ticket as its likely not a terribly hard fix
| 0
|
9,493
| 12,486,892,589
|
IssuesEvent
|
2020-05-31 05:35:16
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
TIN Interpolation - output raster format and size issues
|
Bug Feedback Processing
|
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). -->
**Describe the bug**
1) Produces an ArcInfo ASCII Grid, but defaults to tif extension.
2) The rows/cols and pixel sizes (x/y) are not respected - ex. with the default 0.10 size, the output raster reports:
Origin
283796,5.04147e+06
Pixel Size
0.09977998600000000107,-0.09982621899999999382
3) Break and structure lines are not respected - see https://github.com/qgis/QGIS/issues/27048
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error -->
**QGIS and OS versions**
QGIS version
3.8.3-Zanzibar
QGIS code revision
685d8b15d2
Compiled against Qt
5.11.2
Running against Qt
5.11.2
Compiled against GDAL/OGR
2.4.1
Running against GDAL/OGR
2.4.1
Compiled against GEOS
3.7.2-CAPI-1.11.0
Running against GEOS
3.7.2-CAPI-1.11.0 b55d2125
PostgreSQL Client Version
10.8
SpatiaLite Version
4.3.0
QWT Version
6.1.3
QScintilla2 Version
2.10.8
Compiled against PROJ
5.2.0
Running against PROJ
Rel. 5.2.0, September 15th, 2018
OS Version
Windows 10 (10.0)
**Additional context**
<!-- Add any other context about the problem here. -->
|
1.0
|
TIN Interpolation - output raster format and size issues - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). -->
**Describe the bug**
1) Produces an ArcInfo ASCII Grid, but defaults to tif extension.
2) The rows/cols and pixel sizes (x/y) are not respected - ex. with the default 0.10 size, the output raster reports:
Origin
283796,5.04147e+06
Pixel Size
0.09977998600000000107,-0.09982621899999999382
3) Break and structure lines are not respected - see https://github.com/qgis/QGIS/issues/27048
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error -->
**QGIS and OS versions**
QGIS version
3.8.3-Zanzibar
QGIS code revision
685d8b15d2
Compiled against Qt
5.11.2
Running against Qt
5.11.2
Compiled against GDAL/OGR
2.4.1
Running against GDAL/OGR
2.4.1
Compiled against GEOS
3.7.2-CAPI-1.11.0
Running against GEOS
3.7.2-CAPI-1.11.0 b55d2125
PostgreSQL Client Version
10.8
SpatiaLite Version
4.3.0
QWT Version
6.1.3
QScintilla2 Version
2.10.8
Compiled against PROJ
5.2.0
Running against PROJ
Rel. 5.2.0, September 15th, 2018
OS Version
Windows 10 (10.0)
**Additional context**
<!-- Add any other context about the problem here. -->
|
process
|
tin interpolation output raster format and size issues bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue if the issue concerns a third party plugin then it cannot be fixed by the qgis team please raise your issue in the dedicated bug tracker for that specific plugin as listed in the plugin s description describe the bug produces an arcinfo ascii grid but defaults to tif extension the rows cols and pixel sizes x y are not respected ex with the default size the output raster reports origin pixel size break and structure lines are not respected see how to reproduce steps sample datasets and qgis project file to reproduce the behavior screencasts or screenshots welcome go to click on scroll down to see error qgis and os versions qgis version zanzibar qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi postgresql client version spatialite version qwt version version compiled against proj running against proj rel september os version windows additional context
| 1
|
21,692
| 30,190,141,149
|
IssuesEvent
|
2023-07-04 14:45:50
|
UnitTestBot/UTBotJava
|
https://api.github.com/repos/UnitTestBot/UTBotJava
|
opened
|
Rerun Spring integration tests after minimizations with full context reset
|
ctg-enhancement comp-instrumented-process comp-spring
|
**Description**
Right now we don't fully reset Spring context between concrete executions when generating integration tests with fuzzer, because it's too time consuming (can take several seconds per reset). We do our best to reset relevant parts of context (e.g. reset relevant beans and rollback transactions), however that may still not be enough because, for example, database id generators are not rollbacked with the transaction.
Partial reset of Spring context may lead to generation of unreproducible tests that rely on some code from earlier concrete executions being executed before them.
**Action plan**
To cope we that, it is suggested to rerun tests that are left after test case minimization with full context reset and use results obtained from these reruns.
|
1.0
|
Rerun Spring integration tests after minimizations with full context reset - **Description**
Right now we don't fully reset Spring context between concrete executions when generating integration tests with fuzzer, because it's too time consuming (can take several seconds per reset). We do our best to reset relevant parts of context (e.g. reset relevant beans and rollback transactions), however that may still not be enough because, for example, database id generators are not rollbacked with the transaction.
Partial reset of Spring context may lead to generation of unreproducible tests that rely on some code from earlier concrete executions being executed before them.
**Action plan**
To cope we that, it is suggested to rerun tests that are left after test case minimization with full context reset and use results obtained from these reruns.
|
process
|
rerun spring integration tests after minimizations with full context reset description right now we don t fully reset spring context between concrete executions when generating integration tests with fuzzer because it s too time consuming can take several seconds per reset we do our best to reset relevant parts of context e g reset relevant beans and rollback transactions however that may still not be enough because for example database id generators are not rollbacked with the transaction partial reset of spring context may lead to generation of unreproducible tests that rely on some code from earlier concrete executions being executed before them action plan to cope we that it is suggested to rerun tests that are left after test case minimization with full context reset and use results obtained from these reruns
| 1
|
115,889
| 11,889,727,323
|
IssuesEvent
|
2020-03-28 15:10:11
|
samiha-rahman/soen390
|
https://api.github.com/repos/samiha-rahman/soen390
|
opened
|
US-59: Update Defect Tracking and Report Section for sprint 4
|
developer story documentation epic 10
|
As a developer, I need to update the defect tracking and report
|
1.0
|
US-59: Update Defect Tracking and Report Section for sprint 4 - As a developer, I need to update the defect tracking and report
|
non_process
|
us update defect tracking and report section for sprint as a developer i need to update the defect tracking and report
| 0
|
143,953
| 5,533,275,796
|
IssuesEvent
|
2017-03-21 12:57:39
|
sbpp/sourcebans-pp
|
https://api.github.com/repos/sbpp/sourcebans-pp
|
closed
|
Unable to login via steam
|
Priority: Medium Status: Abandoned Status: Completed Status: Review Needed Type: Bug
|
I have installed php-curl module and tried to update `steamopenid.php` and `includes/openid.php` to the latest version from github.
All that happens is I get taken to steam as I am supposed to, but when I am returned to sourcebans I am not logged in.
No errors on the page or in any log files I know of (php and nginx).
- SourceBans++ 1.5.4.7
- PHP 7.0.8-0
- Ubuntu 16.04.1 LTS
|
1.0
|
Unable to login via steam - I have installed php-curl module and tried to update `steamopenid.php` and `includes/openid.php` to the latest version from github.
All that happens is I get taken to steam as I am supposed to, but when I am returned to sourcebans I am not logged in.
No errors on the page or in any log files I know of (php and nginx).
- SourceBans++ 1.5.4.7
- PHP 7.0.8-0
- Ubuntu 16.04.1 LTS
|
non_process
|
unable to login via steam i have installed php curl module and tried to update steamopenid php and includes openid php to the latest version from github all that happens is i get taken to steam as i am supposed to but when i am returned to sourcebans i am not logged in no errors on the page or in any log files i know of php and nginx sourcebans php ubuntu lts
| 0
|
341,097
| 24,682,902,845
|
IssuesEvent
|
2022-10-18 23:38:18
|
apollographql/apollo-server
|
https://api.github.com/repos/apollographql/apollo-server
|
opened
|
Docs: reinstate serverless deployment guides
|
:memo: documentation
|
We should reintroduce (and update) our deployment guides for getting Apollo Server up and running in the popular serverless frameworks (lambda, azure, cloud, ...others?)
@glasser proposed a one-size-fits-most approach using @vendia/serverless here which we should lean in to:
https://github.com/apollo-server-integrations/apollo-server-integration-aws-lambda/issues/38
We can suggest more specific implementations secondarily or as an "advanced usage" (i.e. `@as-integrations/aws-lambda` for users who are concerned about bundle size).
<!--
- Some features can be built as plugins.
We encourage exploring the plugin API prior to opening a feature request:
https://www.apollographql.com/docs/apollo-server/integrations/plugins/
In the event that the plugin API doesn't allow you to build a feature, it
may be that expanding the plugin API *itself* is the best place for the
feature to be introduced! Consider this flexible solution when opening a
new feature request since it also unlocks new opportunities.
- Prior to opening a feature request, please search for existing requests.
If you find an existing feature that matches your needs, use the 👍 emote
to show your support for it. If the specifics of your use case are not
covered in the existing feature request but the idea seems similar enough,
please take the time to *add new conversation* which helps the feature's
design evolve.
- If you do not find any other existing requests for the feature you desire,
you should open a new feature request. Please take the time to help us
understand your use-case as precisely as possible. Be sure to demonstrate
that you've evaluated existing features and found them unsuitable and were
unable to implement the functionality with the plugin API.
Be flexible in your design and consider slight variations which might
necessitate a specific API design. We also hope you'll be willing to engage
in the on-going design discussion prior to opening a pull-request.
-->
|
1.0
|
Docs: reinstate serverless deployment guides - We should reintroduce (and update) our deployment guides for getting Apollo Server up and running in the popular serverless frameworks (lambda, azure, cloud, ...others?)
@glasser proposed a one-size-fits-most approach using @vendia/serverless here which we should lean in to:
https://github.com/apollo-server-integrations/apollo-server-integration-aws-lambda/issues/38
We can suggest more specific implementations secondarily or as an "advanced usage" (i.e. `@as-integrations/aws-lambda` for users who are concerned about bundle size).
<!--
- Some features can be built as plugins.
We encourage exploring the plugin API prior to opening a feature request:
https://www.apollographql.com/docs/apollo-server/integrations/plugins/
In the event that the plugin API doesn't allow you to build a feature, it
may be that expanding the plugin API *itself* is the best place for the
feature to be introduced! Consider this flexible solution when opening a
new feature request since it also unlocks new opportunities.
- Prior to opening a feature request, please search for existing requests.
If you find an existing feature that matches your needs, use the 👍 emote
to show your support for it. If the specifics of your use case are not
covered in the existing feature request but the idea seems similar enough,
please take the time to *add new conversation* which helps the feature's
design evolve.
- If you do not find any other existing requests for the feature you desire,
you should open a new feature request. Please take the time to help us
understand your use-case as precisely as possible. Be sure to demonstrate
that you've evaluated existing features and found them unsuitable and were
unable to implement the functionality with the plugin API.
Be flexible in your design and consider slight variations which might
necessitate a specific API design. We also hope you'll be willing to engage
in the on-going design discussion prior to opening a pull-request.
-->
|
non_process
|
docs reinstate serverless deployment guides we should reintroduce and update our deployment guides for getting apollo server up and running in the popular serverless frameworks lambda azure cloud others glasser proposed a one size fits most approach using vendia serverless here which we should lean in to we can suggest more specific implementations secondarily or as an advanced usage i e as integrations aws lambda for users who are concerned about bundle size some features can be built as plugins we encourage exploring the plugin api prior to opening a feature request in the event that the plugin api doesn t allow you to build a feature it may be that expanding the plugin api itself is the best place for the feature to be introduced consider this flexible solution when opening a new feature request since it also unlocks new opportunities prior to opening a feature request please search for existing requests if you find an existing feature that matches your needs use the 👍 emote to show your support for it if the specifics of your use case are not covered in the existing feature request but the idea seems similar enough please take the time to add new conversation which helps the feature s design evolve if you do not find any other existing requests for the feature you desire you should open a new feature request please take the time to help us understand your use case as precisely as possible be sure to demonstrate that you ve evaluated existing features and found them unsuitable and were unable to implement the functionality with the plugin api be flexible in your design and consider slight variations which might necessitate a specific api design we also hope you ll be willing to engage in the on going design discussion prior to opening a pull request
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.