Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
21,865
| 17,937,013,443
|
IssuesEvent
|
2021-09-10 16:37:27
|
CAIDA/ioda-ui
|
https://api.github.com/repos/CAIDA/ioda-ui
|
opened
|
[ Modal ] - General Bugs
|
bug usability
|
- [ ] api calls are made even if the data is available when a modal is re-opened
- [ ] The api call for that modals raw signal data is re-triggered when closing the modal for the first time
|
True
|
[ Modal ] - General Bugs - - [ ] api calls are made even if the data is available when a modal is re-opened
- [ ] The api call for that modals raw signal data is re-triggered when closing the modal for the first time
|
non_process
|
general bugs api calls are made even if the data is available when a modal is re opened the api call for that modals raw signal data is re triggered when closing the modal for the first time
| 0
|
6,011
| 8,820,260,733
|
IssuesEvent
|
2019-01-01 10:01:02
|
linnovate/root
|
https://api.github.com/repos/linnovate/root
|
closed
|
delete tags that related to deleted folder
|
2.0.6 Fixed Process bug
|
press on settings
crate a new folder and call it 101
press on documents
crate a new item and call it TEST_101
press on "choose folder" and select 101 folder
(the document is connected to 101 folder )
press on settings again
delete the the 101 folder
it shows that the document still connected to deleted folder (101)
|
1.0
|
delete tags that related to deleted folder - press on settings
crate a new folder and call it 101
press on documents
crate a new item and call it TEST_101
press on "choose folder" and select 101 folder
(the document is connected to 101 folder )
press on settings again
delete the the 101 folder
it shows that the document still connected to deleted folder (101)
|
process
|
delete tags that related to deleted folder press on settings crate a new folder and call it press on documents crate a new item and call it test press on choose folder and select folder the document is connected to folder press on settings again delete the the folder it shows that the document still connected to deleted folder
| 1
|
248,081
| 7,927,212,659
|
IssuesEvent
|
2018-07-06 07:06:19
|
wso2-extensions/siddhi-store-rdbms
|
https://api.github.com/repos/wso2-extensions/siddhi-store-rdbms
|
closed
|
Cannot use IS NULL in the ON condition of table UPDATE OR INSERT INTO
|
Priority/Highest Type/Bug
|
**Description:**
Consider following Siddhi query.
```sql
from EventStream
select event_id, event_timestamp, event_value
update or insert into ProcessedEventTable
set
ProcessedEventTable.event_timestamp = event_timestamp
on
ProcessedEventTable.event_id == event_id and
ProcessedEventTable.event_value is null;
```
Below is the [parametrized SQL query fragment](https://github.com/wso2-extensions/siddhi-store-rdbms/blob/v4.0.22/component/src/main/java/org/wso2/extension/siddhi/store/rdbms/RDBMSConditionVisitor.java#L348) generated for above `on` condition.
```sql
(((ProcessedEventTable.alert_id <> ? ))AND IS NULL ProcessedEventTable.event_value )
```
Which is incorrect. It should be
```sql
(((ProcessedEventTable.alert_id <> ? ))AND ProcessedEventTable.event_value IS NULL )
```
**Bug Fix Suggestion:**
In `RDBMSConditionVisitor`, `IS NULL` is appended in [`beginVisitIsNull`](https://github.com/wso2-extensions/siddhi-store-rdbms/blob/v4.0.22/component/src/main/java/org/wso2/extension/siddhi/store/rdbms/RDBMSConditionVisitor.java#L194) which is invoked before visiting to the variable. Instead it should be appended in the [`endVisitIsNull`](https://github.com/wso2-extensions/siddhi-store-rdbms/blob/v4.0.22/component/src/main/java/org/wso2/extension/siddhi/store/rdbms/RDBMSConditionVisitor.java#L198) which is invoked after visiting the variable.
**Affected Product Version:**
v4.0.22
|
1.0
|
Cannot use IS NULL in the ON condition of table UPDATE OR INSERT INTO - **Description:**
Consider following Siddhi query.
```sql
from EventStream
select event_id, event_timestamp, event_value
update or insert into ProcessedEventTable
set
ProcessedEventTable.event_timestamp = event_timestamp
on
ProcessedEventTable.event_id == event_id and
ProcessedEventTable.event_value is null;
```
Below is the [parametrized SQL query fragment](https://github.com/wso2-extensions/siddhi-store-rdbms/blob/v4.0.22/component/src/main/java/org/wso2/extension/siddhi/store/rdbms/RDBMSConditionVisitor.java#L348) generated for above `on` condition.
```sql
(((ProcessedEventTable.alert_id <> ? ))AND IS NULL ProcessedEventTable.event_value )
```
Which is incorrect. It should be
```sql
(((ProcessedEventTable.alert_id <> ? ))AND ProcessedEventTable.event_value IS NULL )
```
**Bug Fix Suggestion:**
In `RDBMSConditionVisitor`, `IS NULL` is appended in [`beginVisitIsNull`](https://github.com/wso2-extensions/siddhi-store-rdbms/blob/v4.0.22/component/src/main/java/org/wso2/extension/siddhi/store/rdbms/RDBMSConditionVisitor.java#L194) which is invoked before visiting to the variable. Instead it should be appended in the [`endVisitIsNull`](https://github.com/wso2-extensions/siddhi-store-rdbms/blob/v4.0.22/component/src/main/java/org/wso2/extension/siddhi/store/rdbms/RDBMSConditionVisitor.java#L198) which is invoked after visiting the variable.
**Affected Product Version:**
v4.0.22
|
non_process
|
cannot use is null in the on condition of table update or insert into description consider following siddhi query sql from eventstream select event id event timestamp event value update or insert into processedeventtable set processedeventtable event timestamp event timestamp on processedeventtable event id event id and processedeventtable event value is null below is the generated for above on condition sql processedeventtable alert id and is null processedeventtable event value which is incorrect it should be sql processedeventtable alert id and processedeventtable event value is null bug fix suggestion in rdbmsconditionvisitor is null is appended in which is invoked before visiting to the variable instead it should be appended in the which is invoked after visiting the variable affected product version
| 0
|
20,215
| 3,317,000,978
|
IssuesEvent
|
2015-11-06 19:32:19
|
contao/core
|
https://api.github.com/repos/contao/core
|
closed
|
symfony form and contao core
|
defect up for discussion
|
I injected the contents of a template twig in a Contao's article. Everything is going well.
But, when I inject a content with a form in Contao, Contao controls the token because he thinks it's a Contao form?
Can I avoid this?
|
1.0
|
symfony form and contao core - I injected the contents of a template twig in a Contao's article. Everything is going well.
But, when I inject a content with a form in Contao, Contao controls the token because he thinks it's a Contao form?
Can I avoid this?
|
non_process
|
symfony form and contao core i injected the contents of a template twig in a contao s article everything is going well but when i inject a content with a form in contao contao controls the token because he thinks it s a contao form can i avoid this
| 0
|
188,354
| 15,159,843,448
|
IssuesEvent
|
2021-02-12 05:30:38
|
giangpham-cncs/Capstone
|
https://api.github.com/repos/giangpham-cncs/Capstone
|
closed
|
Understanding how to configure IPSEC/VPN on Packet Tracer
|
documentation
|
Here is the instruction for IPSec/VPN site-to-site configuration
http://www.firewall.cx/cisco-technical-knowledgebase/cisco-routers/867-cisco-router-site-to-site-ipsec-vpn.html
|
1.0
|
Understanding how to configure IPSEC/VPN on Packet Tracer - Here is the instruction for IPSec/VPN site-to-site configuration
http://www.firewall.cx/cisco-technical-knowledgebase/cisco-routers/867-cisco-router-site-to-site-ipsec-vpn.html
|
non_process
|
understanding how to configure ipsec vpn on packet tracer here is the instruction for ipsec vpn site to site configuration
| 0
|
410,450
| 27,787,481,499
|
IssuesEvent
|
2023-03-17 05:27:08
|
gotenberg/gotenberg
|
https://api.github.com/repos/gotenberg/gotenberg
|
closed
|
landscape not working with libreoffice/convert xls -> pdf
|
documentation
|
How can one tell if landscape option is effectively being used by libreoffice ? I don't see any indication in the logs (cloud run image) and as a matter of fact, landscape is never applied.
See file attached for testing - instead of getting one page in landscape, I get 3 which is not usable.
[landscape.xlsx](https://github.com/gotenberg/gotenberg/files/10975499/landscape.xlsx)
Also, what tells the paper size to use ? I see it's defaulted (hard-coded ?) to `Letter` but ideally I rather pick `Legal` or `A3` to show even more on one page. `convertapi` uses libreoffice too and provides a lot of options, but is slow and single-processing, hence the goal to host our own converter.
|
1.0
|
landscape not working with libreoffice/convert xls -> pdf - How can one tell if landscape option is effectively being used by libreoffice ? I don't see any indication in the logs (cloud run image) and as a matter of fact, landscape is never applied.
See file attached for testing - instead of getting one page in landscape, I get 3 which is not usable.
[landscape.xlsx](https://github.com/gotenberg/gotenberg/files/10975499/landscape.xlsx)
Also, what tells the paper size to use ? I see it's defaulted (hard-coded ?) to `Letter` but ideally I rather pick `Legal` or `A3` to show even more on one page. `convertapi` uses libreoffice too and provides a lot of options, but is slow and single-processing, hence the goal to host our own converter.
|
non_process
|
landscape not working with libreoffice convert xls pdf how can one tell if landscape option is effectively being used by libreoffice i don t see any indication in the logs cloud run image and as a matter of fact landscape is never applied see file attached for testing instead of getting one page in landscape i get which is not usable also what tells the paper size to use i see it s defaulted hard coded to letter but ideally i rather pick legal or to show even more on one page convertapi uses libreoffice too and provides a lot of options but is slow and single processing hence the goal to host our own converter
| 0
|
93,154
| 8,402,001,852
|
IssuesEvent
|
2018-10-11 04:10:38
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
teamcity: failed test: TestDistSQLDrainingHosts
|
C-test-failure O-robot
|
The following tests appear to have failed on master (testrace): TestDistSQLDrainingHosts
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestDistSQLDrainingHosts).
[#958004](https://teamcity.cockroachdb.com/viewLog.html?buildId=958004):
```
TestDistSQLDrainingHosts
...grations have run
I181011 03:41:22.058674 15577 server/server.go:1587 [n2] serving sql connections
I181011 03:41:22.080221 15823 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:45503} Attrs: Locality: ServerVersion:2.1 BuildTag:v2.2.0-alpha.00000000-1563-g957506f StartedAt:1539229282035188528 LocalityAddress:[]} ClusterID:387e9153-0d78-4b08-9569-c723b47f0151 StartedAt:1539229282035188528 LastUp:1539229282035188528}
I181011 03:41:22.081034 15821 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version
I181011 03:41:22.091846 16063 sql/event_log.go:126 [n1,client=127.0.0.1:59208,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root}
I181011 03:41:22.097030 16063 sql/event_log.go:126 [n1,client=127.0.0.1:59208,user=root] Event: "create_table", target: 53, info: {TableName:test.public.nums Statement:CREATE TABLE test.public.nums (num INT) User:root}
I181011 03:41:22.104195 16063 storage/replica_command.go:298 [n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /Table/53/1/1 [r2]
I181011 03:41:22.129006 16063 storage/store_snapshot.go:615 [n1,s1,r2/1:/{Table/53/1/1-Max}] sending preemptive snapshot f0555eaf at applied index 11
I181011 03:41:22.129268 16063 storage/store_snapshot.go:657 [n1,s1,r2/1:/{Table/53/1/1-Max}] streamed snapshot to (n2,s2):?: kv pairs: 8, log entries: 1, rate-limit: 2.0 MiB/sec, 1ms
I181011 03:41:22.129638 16122 storage/replica_raftstorage.go:803 [n2,s2,r2/?:{-}] applying preemptive snapshot at index 11 (id=f0555eaf, encoded size=380, 1 rocksdb batches, 1 log entries)
I181011 03:41:22.129933 16122 storage/replica_raftstorage.go:809 [n2,s2,r2/?:/{Table/53/1/1-Max}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms]
I181011 03:41:22.130299 16063 storage/replica_command.go:812 [n1,s1,r2/1:/{Table/53/1/1-Max}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r2:/{Table/53/1/1-Max} [(n1,s1):1, next=2, gen=0]
I181011 03:41:22.133330 16063 storage/replica.go:3899 [n1,s1,r2/1:/{Table/53/1/1-Max}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3
I181011 03:41:22.136811 16008 storage/replica_proposal.go:211 [n2,s2,r2/2:/{Table/53/1/1-Max}] new range lease repl=(n2,s2):2 seq=3 start=1539229282.133721061,0 epo=1 pro=1539229282.133723405,0 following repl=(n1,s1):1 seq=2 start=1539229281.593972390,0 exp=1539229290.594737365,0 pro=1539229281.594758312,0
I181011 03:41:22.138172 16150 storage/replica_command.go:812 [n2,s2,r2/2:/{Table/53/1/1-Max}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r2:/{Table/53/1/1-Max} [(n1,s1):1, (n2,s2):2, next=3, gen=0]
I181011 03:41:22.147270 16150 storage/replica.go:3899 [n2,s2,r2/2:/{Table/53/1/1-Max}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n2,s2):2] next=3
I181011 03:41:22.149039 16112 storage/store.go:2744 [n1,replicaGC,s1,r2/1:/{Table/53/1/1-Max}] removing replica r2/1
I181011 03:41:22.149254 16112 storage/replica.go:878 [n1,replicaGC,s1,r2/1:/{Table/53/1/1-Max}] removed 9 (2+7) keys in 0ms [clear=0ms commit=0ms]
I181011 03:41:22.212104 16194 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-subscription
1 [async] closedts-rangefeed-subscriber
W181011 03:41:22.212267 16097 storage/raft_transport.go:583 [n2] while processing outgoing Raft queue to node 1: rpc error: code = Unavailable desc = transport is closing:
I181011 03:41:22.212310 16145 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-rangefeed-subscriber
W181011 03:41:22.215643 16128 storage/raft_transport.go:583 [n1] while processing outgoing Raft queue to node 2: rpc error: code = Unavailable desc = transport is closing:
I181011 03:41:22.215841 16194 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-subscription
TestDistSQLDrainingHosts
...oelG_F996L-RtYtnV59679rE5vnTt08PV9UXX_i-ujuLMIdKZ51JztRgFcp7KISNn6eleuN80ZWw3bgscxVS2dRkOoazWD36F6x_h_TfYLP8-AgAA__-zG6EE]]
I181011 03:46:22.490702 13760 sql/distsql_physical_planner_test.go:513 SucceedsSoon:
expected:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyskT1rwzAQhvf-inJTCoJETrpoSumUoXbJBx2KCap1GEMsmZMELcH_vdgaEodYTSGjTn7uef3qCNooTGWNFsQncGCQQM6gIVOgtYa6cfhopb5BzBhUuvGuG-cMCkMI4giucgcEAVv5dcA1SoU0nQEDhU5Wh351Q1Ut6WepfW2BQeadeEyNRshbBsa701LrZIkgeMtuF7-UJWEpnaFpMvS-Zrt0u19nH5vJ06grGXWdFF4bUkioBvvzNp5mMUyz2b3tV-l2suTjYeaDMPz2xvldG_9DfPaP87s2fsW1RtsYbfGi-eubZ92LoCoxPJ81ngp8J1P0mnDMeq4fKLQu3PJwWOlw1QU8h3kUTgYwv4STKPwcN8-j8CIOL_4VO28ffgMAAP__nC9YuA==]]
got:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyUkEFL9DAQhu_fr_h4TwqBbfeYk-JpL63UFQ8SJDZDKLSZMklAWfrfpc1BV1jR47yT533CnBDYUWMnitDPqGEUZuGeYmRZo_Lg4N6gK4UhzDmtsVHoWQj6hDSkkaBxtK8jdWQdya6CgqNkh3GrnWWYrLzfhDxFKLQ56f8NB4JZFDinz9KYrCfoelG_F996L-RtYtnV59679rE5vnTt08PV9UXX_i-ujuLMIdKZ51JztRgFcp7KISNn6eleuN80ZWw3bgscxVS2dRkOoazWD36F6x_h_TfYLP8-AgAA__-zG6EE]]
I181011 03:46:23.499440 13760 sql/distsql_physical_planner_test.go:513 SucceedsSoon:
expected:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyskT1rwzAQhvf-inJTCoJETrpoSumUoXbJBx2KCap1GEMsmZMELcH_vdgaEodYTSGjTn7uef3qCNooTGWNFsQncGCQQM6gIVOgtYa6cfhopb5BzBhUuvGuG-cMCkMI4giucgcEAVv5dcA1SoU0nQEDhU5Wh351Q1Ut6WepfW2BQeadeEyNRshbBsa701LrZIkgeMtuF7-UJWEpnaFpMvS-Zrt0u19nH5vJ06grGXWdFF4bUkioBvvzNp5mMUyz2b3tV-l2suTjYeaDMPz2xvldG_9DfPaP87s2fsW1RtsYbfGi-eubZ92LoCoxPJ81ngp8J1P0mnDMeq4fKLQu3PJwWOlw1QU8h3kUTgYwv4STKPwcN8-j8CIOL_4VO28ffgMAAP__nC9YuA==]]
got:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyUkEFL9DAQhu_fr_h4TwqBbfeYk-JpL63UFQ8SJDZDKLSZMklAWfrfpc1BV1jR47yT533CnBDYUWMnitDPqGEUZuGeYmRZo_Lg4N6gK4UhzDmtsVHoWQj6hDSkkaBxtK8jdWQdya6CgqNkh3GrnWWYrLzfhDxFKLQ56f8NB4JZFDinz9KYrCfoelG_F996L-RtYtnV59679rE5vnTt08PV9UXX_i-ujuLMIdKZ51JztRgFcp7KISNn6eleuN80ZWw3bgscxVS2dRkOoazWD36F6x_h_TfYLP8-AgAA__-zG6EE]]
I181011 03:46:24.517878 13760 sql/distsql_physical_planner_test.go:513 SucceedsSoon:
expected:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyskT1rwzAQhvf-inJTCoJETrpoSumUoXbJBx2KCap1GEMsmZMELcH_vdgaEodYTSGjTn7uef3qCNooTGWNFsQncGCQQM6gIVOgtYa6cfhopb5BzBhUuvGuG-cMCkMI4giucgcEAVv5dcA1SoU0nQEDhU5Wh351Q1Ut6WepfW2BQeadeEyNRshbBsa701LrZIkgeMtuF7-UJWEpnaFpMvS-Zrt0u19nH5vJ06grGXWdFF4bUkioBvvzNp5mMUyz2b3tV-l2suTjYeaDMPz2xvldG_9DfPaP87s2fsW1RtsYbfGi-eubZ92LoCoxPJ81ngp8J1P0mnDMeq4fKLQu3PJwWOlw1QU8h3kUTgYwv4STKPwcN8-j8CIOL_4VO28ffgMAAP__nC9YuA==]]
got:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyUkEFL9DAQhu_fr_h4TwqBbfeYk-JpL63UFQ8SJDZDKLSZMklAWfrfpc1BV1jR47yT533CnBDYUWMnitDPqGEUZuGeYmRZo_Lg4N6gK4UhzDmtsVHoWQj6hDSkkaBxtK8jdWQdya6CgqNkh3GrnWWYrLzfhDxFKLQ56f8NB4JZFDinz9KYrCfoelG_F996L-RtYtnV59679rE5vnTt08PV9UXX_i-ujuLMIdKZ51JztRgFcp7KISNn6eleuN80ZWw3bgscxVS2dRkOoazWD36F6x_h_TfYLP8-AgAA__-zG6EE]]
I181011 03:46:25.519313 14654 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] transport racer
1 [async] closedts-subscription
1 [async] closedts-rangefeed-subscriber
I181011 03:46:25.521933 14655 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-subscription
1 [async] closedts-rangefeed-subscriber
I181011 03:46:25.523375 13680 kv/transport_race.go:113 transport race promotion: ran 103 iterations on up to 480 requests
I181011 03:46:25.526344 14654 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-subscription
1 [async] closedts-rangefeed-subscriber
W181011 03:46:25.526529 14393 storage/raft_transport.go:583 [n2] while processing outgoing Raft queue to node 1: EOF:
W181011 03:46:25.526658 14269 storage/raft_transport.go:583 [n1] while processing outgoing Raft queue to node 2: EOF:
I181011 03:46:25.535384 14654 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-subscription
W181011 03:46:25.537977 14193 gossip/gossip.go:1496 [n2] no incoming or outgoing connections
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
teamcity: failed test: TestDistSQLDrainingHosts - The following tests appear to have failed on master (testrace): TestDistSQLDrainingHosts
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestDistSQLDrainingHosts).
[#958004](https://teamcity.cockroachdb.com/viewLog.html?buildId=958004):
```
TestDistSQLDrainingHosts
...grations have run
I181011 03:41:22.058674 15577 server/server.go:1587 [n2] serving sql connections
I181011 03:41:22.080221 15823 sql/event_log.go:126 [n2] Event: "node_join", target: 2, info: {Descriptor:{NodeID:2 Address:{NetworkField:tcp AddressField:127.0.0.1:45503} Attrs: Locality: ServerVersion:2.1 BuildTag:v2.2.0-alpha.00000000-1563-g957506f StartedAt:1539229282035188528 LocalityAddress:[]} ClusterID:387e9153-0d78-4b08-9569-c723b47f0151 StartedAt:1539229282035188528 LastUp:1539229282035188528}
I181011 03:41:22.081034 15821 server/server_update.go:67 [n2] no need to upgrade, cluster already at the newest version
I181011 03:41:22.091846 16063 sql/event_log.go:126 [n1,client=127.0.0.1:59208,user=root] Event: "create_database", target: 52, info: {DatabaseName:test Statement:CREATE DATABASE IF NOT EXISTS test User:root}
I181011 03:41:22.097030 16063 sql/event_log.go:126 [n1,client=127.0.0.1:59208,user=root] Event: "create_table", target: 53, info: {TableName:test.public.nums Statement:CREATE TABLE test.public.nums (num INT) User:root}
I181011 03:41:22.104195 16063 storage/replica_command.go:298 [n1,s1,r1/1:/M{in-ax}] initiating a split of this range at key /Table/53/1/1 [r2]
I181011 03:41:22.129006 16063 storage/store_snapshot.go:615 [n1,s1,r2/1:/{Table/53/1/1-Max}] sending preemptive snapshot f0555eaf at applied index 11
I181011 03:41:22.129268 16063 storage/store_snapshot.go:657 [n1,s1,r2/1:/{Table/53/1/1-Max}] streamed snapshot to (n2,s2):?: kv pairs: 8, log entries: 1, rate-limit: 2.0 MiB/sec, 1ms
I181011 03:41:22.129638 16122 storage/replica_raftstorage.go:803 [n2,s2,r2/?:{-}] applying preemptive snapshot at index 11 (id=f0555eaf, encoded size=380, 1 rocksdb batches, 1 log entries)
I181011 03:41:22.129933 16122 storage/replica_raftstorage.go:809 [n2,s2,r2/?:/{Table/53/1/1-Max}] applied preemptive snapshot in 0ms [clear=0ms batch=0ms entries=0ms commit=0ms]
I181011 03:41:22.130299 16063 storage/replica_command.go:812 [n1,s1,r2/1:/{Table/53/1/1-Max}] change replicas (ADD_REPLICA (n2,s2):2): read existing descriptor r2:/{Table/53/1/1-Max} [(n1,s1):1, next=2, gen=0]
I181011 03:41:22.133330 16063 storage/replica.go:3899 [n1,s1,r2/1:/{Table/53/1/1-Max}] proposing ADD_REPLICA((n2,s2):2): updated=[(n1,s1):1 (n2,s2):2] next=3
I181011 03:41:22.136811 16008 storage/replica_proposal.go:211 [n2,s2,r2/2:/{Table/53/1/1-Max}] new range lease repl=(n2,s2):2 seq=3 start=1539229282.133721061,0 epo=1 pro=1539229282.133723405,0 following repl=(n1,s1):1 seq=2 start=1539229281.593972390,0 exp=1539229290.594737365,0 pro=1539229281.594758312,0
I181011 03:41:22.138172 16150 storage/replica_command.go:812 [n2,s2,r2/2:/{Table/53/1/1-Max}] change replicas (REMOVE_REPLICA (n1,s1):1): read existing descriptor r2:/{Table/53/1/1-Max} [(n1,s1):1, (n2,s2):2, next=3, gen=0]
I181011 03:41:22.147270 16150 storage/replica.go:3899 [n2,s2,r2/2:/{Table/53/1/1-Max}] proposing REMOVE_REPLICA((n1,s1):1): updated=[(n2,s2):2] next=3
I181011 03:41:22.149039 16112 storage/store.go:2744 [n1,replicaGC,s1,r2/1:/{Table/53/1/1-Max}] removing replica r2/1
I181011 03:41:22.149254 16112 storage/replica.go:878 [n1,replicaGC,s1,r2/1:/{Table/53/1/1-Max}] removed 9 (2+7) keys in 0ms [clear=0ms commit=0ms]
I181011 03:41:22.212104 16194 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-subscription
1 [async] closedts-rangefeed-subscriber
W181011 03:41:22.212267 16097 storage/raft_transport.go:583 [n2] while processing outgoing Raft queue to node 1: rpc error: code = Unavailable desc = transport is closing:
I181011 03:41:22.212310 16145 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-rangefeed-subscriber
W181011 03:41:22.215643 16128 storage/raft_transport.go:583 [n1] while processing outgoing Raft queue to node 2: rpc error: code = Unavailable desc = transport is closing:
I181011 03:41:22.215841 16194 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-subscription
TestDistSQLDrainingHosts
...oelG_F996L-RtYtnV59679rE5vnTt08PV9UXX_i-ujuLMIdKZ51JztRgFcp7KISNn6eleuN80ZWw3bgscxVS2dRkOoazWD36F6x_h_TfYLP8-AgAA__-zG6EE]]
I181011 03:46:22.490702 13760 sql/distsql_physical_planner_test.go:513 SucceedsSoon:
expected:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyskT1rwzAQhvf-inJTCoJETrpoSumUoXbJBx2KCap1GEMsmZMELcH_vdgaEodYTSGjTn7uef3qCNooTGWNFsQncGCQQM6gIVOgtYa6cfhopb5BzBhUuvGuG-cMCkMI4giucgcEAVv5dcA1SoU0nQEDhU5Wh351Q1Ut6WepfW2BQeadeEyNRshbBsa701LrZIkgeMtuF7-UJWEpnaFpMvS-Zrt0u19nH5vJ06grGXWdFF4bUkioBvvzNp5mMUyz2b3tV-l2suTjYeaDMPz2xvldG_9DfPaP87s2fsW1RtsYbfGi-eubZ92LoCoxPJ81ngp8J1P0mnDMeq4fKLQu3PJwWOlw1QU8h3kUTgYwv4STKPwcN8-j8CIOL_4VO28ffgMAAP__nC9YuA==]]
got:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyUkEFL9DAQhu_fr_h4TwqBbfeYk-JpL63UFQ8SJDZDKLSZMklAWfrfpc1BV1jR47yT533CnBDYUWMnitDPqGEUZuGeYmRZo_Lg4N6gK4UhzDmtsVHoWQj6hDSkkaBxtK8jdWQdya6CgqNkh3GrnWWYrLzfhDxFKLQ56f8NB4JZFDinz9KYrCfoelG_F996L-RtYtnV59679rE5vnTt08PV9UXX_i-ujuLMIdKZ51JztRgFcp7KISNn6eleuN80ZWw3bgscxVS2dRkOoazWD36F6x_h_TfYLP8-AgAA__-zG6EE]]
I181011 03:46:23.499440 13760 sql/distsql_physical_planner_test.go:513 SucceedsSoon:
expected:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyskT1rwzAQhvf-inJTCoJETrpoSumUoXbJBx2KCap1GEMsmZMELcH_vdgaEodYTSGjTn7uef3qCNooTGWNFsQncGCQQM6gIVOgtYa6cfhopb5BzBhUuvGuG-cMCkMI4giucgcEAVv5dcA1SoU0nQEDhU5Wh351Q1Ut6WepfW2BQeadeEyNRshbBsa701LrZIkgeMtuF7-UJWEpnaFpMvS-Zrt0u19nH5vJ06grGXWdFF4bUkioBvvzNp5mMUyz2b3tV-l2suTjYeaDMPz2xvldG_9DfPaP87s2fsW1RtsYbfGi-eubZ92LoCoxPJ81ngp8J1P0mnDMeq4fKLQu3PJwWOlw1QU8h3kUTgYwv4STKPwcN8-j8CIOL_4VO28ffgMAAP__nC9YuA==]]
got:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyUkEFL9DAQhu_fr_h4TwqBbfeYk-JpL63UFQ8SJDZDKLSZMklAWfrfpc1BV1jR47yT533CnBDYUWMnitDPqGEUZuGeYmRZo_Lg4N6gK4UhzDmtsVHoWQj6hDSkkaBxtK8jdWQdya6CgqNkh3GrnWWYrLzfhDxFKLQ56f8NB4JZFDinz9KYrCfoelG_F996L-RtYtnV59679rE5vnTt08PV9UXX_i-ujuLMIdKZ51JztRgFcp7KISNn6eleuN80ZWw3bgscxVS2dRkOoazWD36F6x_h_TfYLP8-AgAA__-zG6EE]]
I181011 03:46:24.517878 13760 sql/distsql_physical_planner_test.go:513 SucceedsSoon:
expected:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyskT1rwzAQhvf-inJTCoJETrpoSumUoXbJBx2KCap1GEMsmZMELcH_vdgaEodYTSGjTn7uef3qCNooTGWNFsQncGCQQM6gIVOgtYa6cfhopb5BzBhUuvGuG-cMCkMI4giucgcEAVv5dcA1SoU0nQEDhU5Wh351Q1Ut6WepfW2BQeadeEyNRshbBsa701LrZIkgeMtuF7-UJWEpnaFpMvS-Zrt0u19nH5vJ06grGXWdFF4bUkioBvvzNp5mMUyz2b3tV-l2suTjYeaDMPz2xvldG_9DfPaP87s2fsW1RtsYbfGi-eubZ92LoCoxPJ81ngp8J1P0mnDMeq4fKLQu3PJwWOlw1QU8h3kUTgYwv4STKPwcN8-j8CIOL_4VO28ffgMAAP__nC9YuA==]]
got:[[https://cockroachdb.github.io/distsqlplan/decode.html#eJyUkEFL9DAQhu_fr_h4TwqBbfeYk-JpL63UFQ8SJDZDKLSZMklAWfrfpc1BV1jR47yT533CnBDYUWMnitDPqGEUZuGeYmRZo_Lg4N6gK4UhzDmtsVHoWQj6hDSkkaBxtK8jdWQdya6CgqNkh3GrnWWYrLzfhDxFKLQ56f8NB4JZFDinz9KYrCfoelG_F996L-RtYtnV59679rE5vnTt08PV9UXX_i-ujuLMIdKZ51JztRgFcp7KISNn6eleuN80ZWw3bgscxVS2dRkOoazWD36F6x_h_TfYLP8-AgAA__-zG6EE]]
I181011 03:46:25.519313 14654 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] transport racer
1 [async] closedts-subscription
1 [async] closedts-rangefeed-subscriber
I181011 03:46:25.521933 14655 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-subscription
1 [async] closedts-rangefeed-subscriber
I181011 03:46:25.523375 13680 kv/transport_race.go:113 transport race promotion: ran 103 iterations on up to 480 requests
I181011 03:46:25.526344 14654 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-subscription
1 [async] closedts-rangefeed-subscriber
W181011 03:46:25.526529 14393 storage/raft_transport.go:583 [n2] while processing outgoing Raft queue to node 1: EOF:
W181011 03:46:25.526658 14269 storage/raft_transport.go:583 [n1] while processing outgoing Raft queue to node 2: EOF:
I181011 03:46:25.535384 14654 util/stop/stopper.go:537 quiescing; tasks left:
1 [async] closedts-subscription
W181011 03:46:25.537977 14193 gossip/gossip.go:1496 [n2] no incoming or outgoing connections
```
Please assign, take a look and update the issue accordingly.
|
non_process
|
teamcity failed test testdistsqldraininghosts the following tests appear to have failed on master testrace testdistsqldraininghosts you may want to check testdistsqldraininghosts grations have run server server go serving sql connections sql event log go event node join target info descriptor nodeid address networkfield tcp addressfield attrs locality serverversion buildtag alpha startedat localityaddress clusterid startedat lastup server server update go no need to upgrade cluster already at the newest version sql event log go event create database target info databasename test statement create database if not exists test user root sql event log go event create table target info tablename test public nums statement create table test public nums num int user root storage replica command go initiating a split of this range at key table storage store snapshot go sending preemptive snapshot at applied index storage store snapshot go streamed snapshot to kv pairs log entries rate limit mib sec storage replica raftstorage go applying preemptive snapshot at index id encoded size rocksdb batches log entries storage replica raftstorage go applied preemptive snapshot in storage replica command go change replicas add replica read existing descriptor table max storage replica go proposing add replica updated next storage replica proposal go new range lease repl seq start epo pro following repl seq start exp pro storage replica command go change replicas remove replica read existing descriptor table max storage replica go proposing remove replica updated next storage store go removing replica storage replica go removed keys in util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber storage raft transport go while processing outgoing raft queue to node rpc error code unavailable desc transport is closing util stop stopper go quiescing tasks left closedts rangefeed subscriber storage raft transport go while processing outgoing raft queue to node rpc error code unavailable desc transport is closing util stop stopper go quiescing tasks left closedts subscription testdistsqldraininghosts oelg i h agaa sql distsql physical planner test go succeedssoon expected got sql distsql physical planner test go succeedssoon expected got sql distsql physical planner test go succeedssoon expected got util stop stopper go quiescing tasks left transport racer closedts subscription closedts rangefeed subscriber util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber kv transport race go transport race promotion ran iterations on up to requests util stop stopper go quiescing tasks left closedts subscription closedts rangefeed subscriber storage raft transport go while processing outgoing raft queue to node eof storage raft transport go while processing outgoing raft queue to node eof util stop stopper go quiescing tasks left closedts subscription gossip gossip go no incoming or outgoing connections please assign take a look and update the issue accordingly
| 0
|
6,069
| 8,902,735,348
|
IssuesEvent
|
2019-01-17 08:34:11
|
Juris-M/citeproc-js
|
https://api.github.com/repos/Juris-M/citeproc-js
|
closed
|
Additional output formats
|
fix in process
|
A user of citeproc-java has expressed the wish that the additional output formats I included in citeproc-java (AsciiDoc and FOP) should be moved to citeproc-js:
https://github.com/michel-kraemer/citeproc-java/issues/35
What do you think about this? Should I prepare a pull request?
|
1.0
|
Additional output formats - A user of citeproc-java has expressed the wish that the additional output formats I included in citeproc-java (AsciiDoc and FOP) should be moved to citeproc-js:
https://github.com/michel-kraemer/citeproc-java/issues/35
What do you think about this? Should I prepare a pull request?
|
process
|
additional output formats a user of citeproc java has expressed the wish that the additional output formats i included in citeproc java asciidoc and fop should be moved to citeproc js what do you think about this should i prepare a pull request
| 1
|
323,088
| 9,842,814,985
|
IssuesEvent
|
2019-06-18 10:06:16
|
WoWManiaUK/Blackwing-Lair
|
https://api.github.com/repos/WoWManiaUK/Blackwing-Lair
|
closed
|
[Spell] Mind Blast
|
Class Confirmed Fixed Confirmed Fixed in Dev Priority-High
|
ID 8092
Spellpower coefficient is wrong and damage is very low.
**How it should work:**
Spell has approx ~1500 base damage and should have a 110,4% spell power coefficient on cataclysm.
**How it works atm:** The base damage looks okay but the spell coefficient is calculating to be around 42% atm.
PS: high priority as shadow priests are currently worthless
|
1.0
|
[Spell] Mind Blast - ID 8092
Spellpower coefficient is wrong and damage is very low.
**How it should work:**
Spell has approx ~1500 base damage and should have a 110,4% spell power coefficient on cataclysm.
**How it works atm:** The base damage looks okay but the spell coefficient is calculating to be around 42% atm.
PS: high priority as shadow priests are currently worthless
|
non_process
|
mind blast id spellpower coefficient is wrong and damage is very low how it should work spell has approx base damage and should have a spell power coefficient on cataclysm how it works atm the base damage looks okay but the spell coefficient is calculating to be around atm ps high priority as shadow priests are currently worthless
| 0
|
230,386
| 18,666,838,471
|
IssuesEvent
|
2021-10-30 01:07:24
|
tsontario/k8y
|
https://api.github.com/repos/tsontario/k8y
|
closed
|
Missing REST unit tests
|
test coverage
|
#23 shipped with good coverage but missing explicit unit tests for some files/classes:
- [x] auth.rb _Edit: concrete auth strategies have as an entrypoint this code and are tested as integration/downstream-unit-tests_
- [x] config_validator.rb _Edit: implementation does not exist yet, will write tests when that happens_
- [x] config.rb _Edit: this is better served with integration tests_
- [x] connection _Edit: this is better served with integration tests_
- [x] transport
|
1.0
|
Missing REST unit tests - #23 shipped with good coverage but missing explicit unit tests for some files/classes:
- [x] auth.rb _Edit: concrete auth strategies have as an entrypoint this code and are tested as integration/downstream-unit-tests_
- [x] config_validator.rb _Edit: implementation does not exist yet, will write tests when that happens_
- [x] config.rb _Edit: this is better served with integration tests_
- [x] connection _Edit: this is better served with integration tests_
- [x] transport
|
non_process
|
missing rest unit tests shipped with good coverage but missing explicit unit tests for some files classes auth rb edit concrete auth strategies have as an entrypoint this code and are tested as integration downstream unit tests config validator rb edit implementation does not exist yet will write tests when that happens config rb edit this is better served with integration tests connection edit this is better served with integration tests transport
| 0
|
314,681
| 23,532,989,170
|
IssuesEvent
|
2022-08-19 17:16:43
|
strongdm/accessbot
|
https://api.github.com/repos/strongdm/accessbot
|
closed
|
Documentation about how to deploy on Fargate
|
documentation
|
**Is your feature request related to a problem? Please describe.**
Since there are some users using AccessBot on Fargate and it is not very straight foward to enable the "requests persistency" in this scenario, it would be nice to have some documentation explaining how to properly configure it.
|
1.0
|
Documentation about how to deploy on Fargate - **Is your feature request related to a problem? Please describe.**
Since there are some users using AccessBot on Fargate and it is not very straight foward to enable the "requests persistency" in this scenario, it would be nice to have some documentation explaining how to properly configure it.
|
non_process
|
documentation about how to deploy on fargate is your feature request related to a problem please describe since there are some users using accessbot on fargate and it is not very straight foward to enable the requests persistency in this scenario it would be nice to have some documentation explaining how to properly configure it
| 0
|
647
| 3,105,864,151
|
IssuesEvent
|
2015-08-31 23:29:48
|
pwittchen/ReactiveNetwork
|
https://api.github.com/repos/pwittchen/ReactiveNetwork
|
closed
|
Create GitHub release for 0.0.3 after updating README.md
|
release process
|
Initial release notes:
- removed `WifiManager.WIFI_STATE_CHANGED_ACTION` filter from `BroadcastReceiver` for observing connectivity (now we're observing only situation when device connects to the network or disconnects from the network - not situation when user turns WiFi on or off)
- added `UNDEFINED` element for `ConnectivityStatus`
- fixed bug causing emission of the same `ConnectivityStatus` twice
|
1.0
|
Create GitHub release for 0.0.3 after updating README.md - Initial release notes:
- removed `WifiManager.WIFI_STATE_CHANGED_ACTION` filter from `BroadcastReceiver` for observing connectivity (now we're observing only situation when device connects to the network or disconnects from the network - not situation when user turns WiFi on or off)
- added `UNDEFINED` element for `ConnectivityStatus`
- fixed bug causing emission of the same `ConnectivityStatus` twice
|
process
|
create github release for after updating readme md initial release notes removed wifimanager wifi state changed action filter from broadcastreceiver for observing connectivity now we re observing only situation when device connects to the network or disconnects from the network not situation when user turns wifi on or off added undefined element for connectivitystatus fixed bug causing emission of the same connectivitystatus twice
| 1
|
14,374
| 17,397,637,031
|
IssuesEvent
|
2021-08-02 15:14:15
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
conditional variables based on stageDependencies
|
cba devops-cicd-process/tech devops/prod support-request
|
Im trying to set up a pipeline that switches environment based on a stageDependencies variable.
Is this possible or do i need to set up 2 stages, 1 apply_autoapprove and 1 apply_manual?
Below pipeline errors with `Unrecognized value: 'stageDependencies'. Located at position 4 within expression: eq(stageDependencies.Plan.Terraform_Plan.outputs['opa.opa_autoapprove'], 'true').`
```yaml
resources:
repositories:
- repository: templates
type: git
stages:
- stage: Plan
jobs:
- job: Terraform_Plan
displayName: Terraform Plan - Publish a package if Infrastructure changes are identified
continueOnError: false
steps:
- template: azure-devops/templates/steps/terraform.yaml@templates
parameters:
version: $(terraform_version)
path: $(directory)
package_name: $(product)
terraform_init: true
terraform_plan: true
terraform_opa: true
- stage: Apply
dependsOn: Plan
variables:
- name: opa_autoapprove
value: $[stageDependencies.Plan.Terraform_Plan.outputs['opa.opa_autoapprove']]
- name: deployment_environment
${{ if eq(variables['opa_autoapprove'], 'true')}}:
value: Terraform_Apply_Autoapprove
${{ if ne(variables['opa_autoapprove'], 'true')}}:
value: Terraform_Apply
condition: |
and(
succeeded(),
ne(variables['Build.Reason'], 'PullRequest'),
eq(stageDependencies.Plan.outputs['Terraform_Plan.plan.plan_exitcode'], '2')
)
jobs:
# track deployments on the environment
- deployment: Terraform_Apply
displayName: Terraform Apply - Resources creation
environment: ${{variables['deployment_environment']}}
strategy:
# default deployment strategy
runOnce:
deploy:
steps:
- template: azure-devops/templates/steps/terraform.yaml@templates
parameters:
version: $(terraform_version)
package_name: $(product)
terraform_apply: true
```
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6
* Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18
* Content: [Expressions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops#dependencies)
* Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/expressions.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
conditional variables based on stageDependencies - Im trying to set up a pipeline that switches environment based on a stageDependencies variable.
Is this possible or do i need to set up 2 stages, 1 apply_autoapprove and 1 apply_manual?
Below pipeline errors with `Unrecognized value: 'stageDependencies'. Located at position 4 within expression: eq(stageDependencies.Plan.Terraform_Plan.outputs['opa.opa_autoapprove'], 'true').`
```yaml
resources:
repositories:
- repository: templates
type: git
stages:
- stage: Plan
jobs:
- job: Terraform_Plan
displayName: Terraform Plan - Publish a package if Infrastructure changes are identified
continueOnError: false
steps:
- template: azure-devops/templates/steps/terraform.yaml@templates
parameters:
version: $(terraform_version)
path: $(directory)
package_name: $(product)
terraform_init: true
terraform_plan: true
terraform_opa: true
- stage: Apply
dependsOn: Plan
variables:
- name: opa_autoapprove
value: $[stageDependencies.Plan.Terraform_Plan.outputs['opa.opa_autoapprove']]
- name: deployment_environment
${{ if eq(variables['opa_autoapprove'], 'true')}}:
value: Terraform_Apply_Autoapprove
${{ if ne(variables['opa_autoapprove'], 'true')}}:
value: Terraform_Apply
condition: |
and(
succeeded(),
ne(variables['Build.Reason'], 'PullRequest'),
eq(stageDependencies.Plan.outputs['Terraform_Plan.plan.plan_exitcode'], '2')
)
jobs:
# track deployments on the environment
- deployment: Terraform_Apply
displayName: Terraform Apply - Resources creation
environment: ${{variables['deployment_environment']}}
strategy:
# default deployment strategy
runOnce:
deploy:
steps:
- template: azure-devops/templates/steps/terraform.yaml@templates
parameters:
version: $(terraform_version)
package_name: $(product)
terraform_apply: true
```
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6
* Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18
* Content: [Expressions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops#dependencies)
* Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/expressions.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
conditional variables based on stagedependencies im trying to set up a pipeline that switches environment based on a stagedependencies variable is this possible or do i need to set up stages apply autoapprove and apply manual below pipeline errors with unrecognized value stagedependencies located at position within expression eq stagedependencies plan terraform plan outputs true yaml resources repositories repository templates type git stages stage plan jobs job terraform plan displayname terraform plan publish a package if infrastructure changes are identified continueonerror false steps template azure devops templates steps terraform yaml templates parameters version terraform version path directory package name product terraform init true terraform plan true terraform opa true stage apply dependson plan variables name opa autoapprove value name deployment environment if eq variables true value terraform apply autoapprove if ne variables true value terraform apply condition and succeeded ne variables pullrequest eq stagedependencies plan outputs jobs track deployments on the environment deployment terraform apply displayname terraform apply resources creation environment variables strategy default deployment strategy runonce deploy steps template azure devops templates steps terraform yaml templates parameters version terraform version package name product terraform apply true document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
15,757
| 19,911,828,092
|
IssuesEvent
|
2022-01-25 17:56:38
|
input-output-hk/high-assurance-legacy
|
https://api.github.com/repos/input-output-hk/high-assurance-legacy
|
closed
|
Turn abbreviations into definitions were sensible
|
language: isabelle topic: process calculus type: improvement
|
Our initial approach was to never use `definition` but always use `abbreviation` instead. However, we have started to turn abbreviations into defined entities. Our goal is to look through the remaining abbreviations and make those of them definitions that should be definitions according to Subsubsection 2.3.3 of [“Programming and Proving in Isabelle/HOL”][prog-prove]. In the course, we want to η-contract `rewrites` clauses where such η-contraction is safely possible as a result of turning abbreviations into defined entities.
[prog-prove]:
https://isabelle.in.tum.de/dist/Isabelle2019/doc/prog-prove.pdf
"Programming and Proving in Isabelle/HOL"
|
1.0
|
Turn abbreviations into definitions were sensible - Our initial approach was to never use `definition` but always use `abbreviation` instead. However, we have started to turn abbreviations into defined entities. Our goal is to look through the remaining abbreviations and make those of them definitions that should be definitions according to Subsubsection 2.3.3 of [“Programming and Proving in Isabelle/HOL”][prog-prove]. In the course, we want to η-contract `rewrites` clauses where such η-contraction is safely possible as a result of turning abbreviations into defined entities.
[prog-prove]:
https://isabelle.in.tum.de/dist/Isabelle2019/doc/prog-prove.pdf
"Programming and Proving in Isabelle/HOL"
|
process
|
turn abbreviations into definitions were sensible our initial approach was to never use definition but always use abbreviation instead however we have started to turn abbreviations into defined entities our goal is to look through the remaining abbreviations and make those of them definitions that should be definitions according to subsubsection of in the course we want to η contract rewrites clauses where such η contraction is safely possible as a result of turning abbreviations into defined entities programming and proving in isabelle hol
| 1
|
217,849
| 16,890,737,028
|
IssuesEvent
|
2021-06-23 08:56:47
|
jinseobhong/typescript.reactNative.template
|
https://api.github.com/repos/jinseobhong/typescript.reactNative.template
|
reopened
|
[TEST] Verify the native app build and run
|
enviroment is valid format test
|
# Test for implementation <a href="#test-for-implementation" id="test-for-implementation">#</a>
To test the implementation
1. [What kind of implementation is the test for](#what-kind-of-implementation-is-the-test-for)
- [Describe to What you are testing](#describe-to-what-you-are-testing)
- [Test goals](#test-goals)
- [Environment for test](#environment-for-test)
- [Tasks of test](#tasks-of-test)
2. [Additional context](#additional-context)
3. [Reference](#reference)
## What kind of implementation is the test for <a href="#what-kind-of-implementation-is-the-test-for" id="what-kind-of-implementation-is-the-test-for">#</a>
Please check the type of **test** and add label, See [here](../blob/master/CONTRIBUTING.md#how-to-create-issue-about-test-for-implementation) to see what types are available.
### Describe to What you are testing <a href="#describe-to-what-you-are-testing" id="describe-to-what-you-are-testing">#</a>
Verify the native app build and run
### Test goals <a href="#test-goals" id="test-goals">#</a>
- [x] Test for android
- [x] Building native app
- [x] Run native app
- [ ] Test for ios
- [ ] Building native app
- [ ] Run native app
### Environment for test <a href="#environment-for-test" id="environment-for-test">#</a>
If you need to write an environment to test, Write down it.
#### Environment for Android
- OS : [ e.g: Ubuntu 20.04 LTS, etc .. ]
- Virtual execution environment
- Java : [ e.g: openjdk 11.0.11 2021-04-20, etc .. ]
- Android Studio :
- Android SDK :
- Android SDK Platform :
- Android Virtual Device :
- Development tools
- Node : v16.3.0
- Package dependency manager :
- npm : 7.15.1
- yarn : 1.22.10
- Packages :
- dependencies :
- "react": "17.0.1"
- "react-native": "0.64.2"
- devDependencies :
- "@babel/core": "^7.12.9",
- "@babel/runtime": "^7.12.5",
- "@react-native-community/eslint-config": "^2.0.0",
- "babel-jest": "^26.6.3",
- "eslint": "7.14.0",
- "jest": "^26.6.3",
- "metro-react-native-babel-preset": "^0.64.0",
- "react-test-renderer": "17.0.1"
#### Environment for ios
- Mac OS :
- Virtual execution environment
- Xcode :
- CocoaPods :
- Development tools
- Watchman :
- Node : [ e.g: v16.3.0, etc .. ]
- Package dependency manager :
- npm : [ e.g: 7.15.1, etc .. ]
- yarn :[ e.g: 1.22.10, etc .. ]
- Packages :
- dependencies :
- something dependencies [ e.g: "react-native": "0.64.2", etc .. ]
-
- devDependencies :
- something devDependencies [ e.g: "@babel/core": "^7.12.9", etc .. ]
-
### Tasks of test <a href="#tasks-of-test" id="tasks-of-test">#</a>
- Building for android app
1. Run command `npm install` in root directory(example: `home/jinseobhong/typescript.reactNative.template/`)
2. Run command `npm start` or `react-native start` for running native app
3. Run command `npm android` or `react-native run-android` for building native app
- Building for ios app
1.
## Additional context <a href="#additional-context" id="additional-context">#</a>
Once this test is complete, you should integrate typescript
## Reference <a href="#reference" id="reference">#</a>
|
1.0
|
[TEST] Verify the native app build and run - # Test for implementation <a href="#test-for-implementation" id="test-for-implementation">#</a>
To test the implementation
1. [What kind of implementation is the test for](#what-kind-of-implementation-is-the-test-for)
- [Describe to What you are testing](#describe-to-what-you-are-testing)
- [Test goals](#test-goals)
- [Environment for test](#environment-for-test)
- [Tasks of test](#tasks-of-test)
2. [Additional context](#additional-context)
3. [Reference](#reference)
## What kind of implementation is the test for <a href="#what-kind-of-implementation-is-the-test-for" id="what-kind-of-implementation-is-the-test-for">#</a>
Please check the type of **test** and add label, See [here](../blob/master/CONTRIBUTING.md#how-to-create-issue-about-test-for-implementation) to see what types are available.
### Describe to What you are testing <a href="#describe-to-what-you-are-testing" id="describe-to-what-you-are-testing">#</a>
Verify the native app build and run
### Test goals <a href="#test-goals" id="test-goals">#</a>
- [x] Test for android
- [x] Building native app
- [x] Run native app
- [ ] Test for ios
- [ ] Building native app
- [ ] Run native app
### Environment for test <a href="#environment-for-test" id="environment-for-test">#</a>
If you need to write an environment to test, Write down it.
#### Environment for Android
- OS : [ e.g: Ubuntu 20.04 LTS, etc .. ]
- Virtual execution environment
- Java : [ e.g: openjdk 11.0.11 2021-04-20, etc .. ]
- Android Studio :
- Android SDK :
- Android SDK Platform :
- Android Virtual Device :
- Development tools
- Node : v16.3.0
- Package dependency manager :
- npm : 7.15.1
- yarn : 1.22.10
- Packages :
- dependencies :
- "react": "17.0.1"
- "react-native": "0.64.2"
- devDependencies :
- "@babel/core": "^7.12.9",
- "@babel/runtime": "^7.12.5",
- "@react-native-community/eslint-config": "^2.0.0",
- "babel-jest": "^26.6.3",
- "eslint": "7.14.0",
- "jest": "^26.6.3",
- "metro-react-native-babel-preset": "^0.64.0",
- "react-test-renderer": "17.0.1"
#### Environment for ios
- Mac OS :
- Virtual execution environment
- Xcode :
- CocoaPods :
- Development tools
- Watchman :
- Node : [ e.g: v16.3.0, etc .. ]
- Package dependency manager :
- npm : [ e.g: 7.15.1, etc .. ]
- yarn :[ e.g: 1.22.10, etc .. ]
- Packages :
- dependencies :
- something dependencies [ e.g: "react-native": "0.64.2", etc .. ]
-
- devDependencies :
- something devDependencies [ e.g: "@babel/core": "^7.12.9", etc .. ]
-
### Tasks of test <a href="#tasks-of-test" id="tasks-of-test">#</a>
- Building for android app
1. Run command `npm install` in root directory(example: `home/jinseobhong/typescript.reactNative.template/`)
2. Run command `npm start` or `react-native start` for running native app
3. Run command `npm android` or `react-native run-android` for building native app
- Building for ios app
1.
## Additional context <a href="#additional-context" id="additional-context">#</a>
Once this test is complete, you should integrate typescript
## Reference <a href="#reference" id="reference">#</a>
|
non_process
|
verify the native app build and run test for implementation to test the implementation what kind of implementation is the test for describe to what you are testing test goals environment for test tasks of test additional context reference what kind of implementation is the test for please check the type of test and add label see blob master contributing md how to create issue about test for implementation to see what types are available describe to what you are testing verify the native app build and run test goals test for android building native app run native app test for ios building native app run native app environment for test if you need to write an environment to test write down it environment for android os virtual execution environment java android studio android sdk android sdk platform android virtual device development tools node package dependency manager npm yarn packages dependencies react react native devdependencies babel core babel runtime react native community eslint config babel jest eslint jest metro react native babel preset react test renderer environment for ios mac os virtual execution environment xcode cocoapods development tools watchman node package dependency manager npm yarn packages dependencies something dependencies devdependencies something devdependencies tasks of test building for android app run command npm install in root directory example home jinseobhong typescript reactnative template run command npm start or react native start for running native app run command npm android or react native run android for building native app building for ios app additional context once this test is complete you should integrate typescript reference
| 0
|
344,312
| 10,342,824,946
|
IssuesEvent
|
2019-09-04 07:32:24
|
oSoc19/gentlestudent-web
|
https://api.github.com/repos/oSoc19/gentlestudent-web
|
closed
|
Not able to validate a learning opportunity
|
Priority: critical
|
When creating a learning opportunity with the gentlestudent@arteveldehs.be -account on Gentlestudent it seems I can afterwards not validate it. The "Accepteer"-button turns grey but the opportunity is not displayed.
|
1.0
|
Not able to validate a learning opportunity - When creating a learning opportunity with the gentlestudent@arteveldehs.be -account on Gentlestudent it seems I can afterwards not validate it. The "Accepteer"-button turns grey but the opportunity is not displayed.
|
non_process
|
not able to validate a learning opportunity when creating a learning opportunity with the gentlestudent arteveldehs be account on gentlestudent it seems i can afterwards not validate it the accepteer button turns grey but the opportunity is not displayed
| 0
|
632,497
| 20,198,854,229
|
IssuesEvent
|
2022-02-11 13:22:15
|
BEXIS2/Core
|
https://api.github.com/repos/BEXIS2/Core
|
closed
|
User Story 37: As developer I would like to know how to integrate semantic search in BEXIS
|
Priority: Low
|
Involving semantic issues in general to bexis (Metadata annotation, dataset annotation, semantic search, …)
|
1.0
|
User Story 37: As developer I would like to know how to integrate semantic search in BEXIS - Involving semantic issues in general to bexis (Metadata annotation, dataset annotation, semantic search, …)
|
non_process
|
user story as developer i would like to know how to integrate semantic search in bexis involving semantic issues in general to bexis metadata annotation dataset annotation semantic search …
| 0
|
41,063
| 5,331,658,497
|
IssuesEvent
|
2017-02-15 20:02:19
|
prestodb/presto
|
https://api.github.com/repos/prestodb/presto
|
opened
|
Flaky test TestMetadataManager.testMetadataIsClearedAfterQueryFinished
|
tests
|
```
testMetadataIsClearedAfterQueryFinished(com.facebook.presto.tests.TestMetadataManager) Time elapsed: 0.032 sec <<< FAILURE!
java.lang.AssertionError: expected [0] but found [1]
at com.facebook.presto.tests.TestMetadataManager.testMetadataIsClearedAfterQueryFinished(TestMetadataManager.java:48)
|
1.0
|
Flaky test TestMetadataManager.testMetadataIsClearedAfterQueryFinished - ```
testMetadataIsClearedAfterQueryFinished(com.facebook.presto.tests.TestMetadataManager) Time elapsed: 0.032 sec <<< FAILURE!
java.lang.AssertionError: expected [0] but found [1]
at com.facebook.presto.tests.TestMetadataManager.testMetadataIsClearedAfterQueryFinished(TestMetadataManager.java:48)
|
non_process
|
flaky test testmetadatamanager testmetadataisclearedafterqueryfinished testmetadataisclearedafterqueryfinished com facebook presto tests testmetadatamanager time elapsed sec failure java lang assertionerror expected but found at com facebook presto tests testmetadatamanager testmetadataisclearedafterqueryfinished testmetadatamanager java
| 0
|
29,520
| 5,640,841,214
|
IssuesEvent
|
2017-04-06 17:20:16
|
DanwareCreations/DotKEGG
|
https://api.github.com/repos/DanwareCreations/DotKEGG
|
closed
|
Add Contribute Section to Main README
|
documentation enhancement
|
Main README file should have a "Contribute" section, with details about how to clone the repository, set up the solution, test it, etc.
|
1.0
|
Add Contribute Section to Main README - Main README file should have a "Contribute" section, with details about how to clone the repository, set up the solution, test it, etc.
|
non_process
|
add contribute section to main readme main readme file should have a contribute section with details about how to clone the repository set up the solution test it etc
| 0
|
434,183
| 12,515,329,820
|
IssuesEvent
|
2020-06-03 07:28:15
|
canonical-web-and-design/build.snapcraft.io
|
https://api.github.com/repos/canonical-web-and-design/build.snapcraft.io
|
closed
|
Turn on arm64 builds
|
Priority: Medium
|
We support four kernels with Ubuntu Core: 32-bit and 64-bit x86, 32-bit and 64-bit ARM. Right now we build for 64-bit x86 and 32-bit ARM. We should enable 64-bit ARM builds so that snaps for the dragonboard 410c can be built.
I currently consider this a low priority.
|
1.0
|
Turn on arm64 builds - We support four kernels with Ubuntu Core: 32-bit and 64-bit x86, 32-bit and 64-bit ARM. Right now we build for 64-bit x86 and 32-bit ARM. We should enable 64-bit ARM builds so that snaps for the dragonboard 410c can be built.
I currently consider this a low priority.
|
non_process
|
turn on builds we support four kernels with ubuntu core bit and bit bit and bit arm right now we build for bit and bit arm we should enable bit arm builds so that snaps for the dragonboard can be built i currently consider this a low priority
| 0
|
167,147
| 26,461,960,452
|
IssuesEvent
|
2023-01-16 18:32:37
|
webstudio-is/webstudio-designer
|
https://api.github.com/repos/webstudio-is/webstudio-designer
|
closed
|
Allow clone project only through login
|
type:enhancement complexity:medium area:designer prio:1
|
As of now, the user can clone any project using the following link `https://alpha.webstudio.is/rest/project/clone/webstudiois`
We want to enforce login during clone operation.
|
1.0
|
Allow clone project only through login - As of now, the user can clone any project using the following link `https://alpha.webstudio.is/rest/project/clone/webstudiois`
We want to enforce login during clone operation.
|
non_process
|
allow clone project only through login as of now the user can clone any project using the following link we want to enforce login during clone operation
| 0
|
21,596
| 29,998,564,896
|
IssuesEvent
|
2023-06-26 07:44:44
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
@saltcorn/mobile-builder 0.8.7-beta.2 has 2 guarddog issues
|
npm-install-script npm-silent-process-execution
|
```{"npm-install-script":[{"code":" \"postinstall\": \"node ./docker/post-installer.js\"","location":"package/package.json:10","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const child = spawn(\"docker\", dArgs, {\n cwd: \".\",\n stdio: \"ignore\",\n detached: true,\n });","location":"package/docker/post-installer.js:22","message":"This package is silently executing another executable"}]}```
|
1.0
|
@saltcorn/mobile-builder 0.8.7-beta.2 has 2 guarddog issues - ```{"npm-install-script":[{"code":" \"postinstall\": \"node ./docker/post-installer.js\"","location":"package/package.json:10","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const child = spawn(\"docker\", dArgs, {\n cwd: \".\",\n stdio: \"ignore\",\n detached: true,\n });","location":"package/docker/post-installer.js:22","message":"This package is silently executing another executable"}]}```
|
process
|
saltcorn mobile builder beta has guarddog issues npm install script npm silent process execution
| 1
|
122,887
| 4,846,647,216
|
IssuesEvent
|
2016-11-10 12:32:08
|
bounswe/bounswe2016group7
|
https://api.github.com/repos/bounswe/bounswe2016group7
|
opened
|
Error Page
|
priority: low
|
An error page must be implemented to be shown in an error case. This page should show a message and may give option to the user for returning another page(homepage or previous page)
|
1.0
|
Error Page - An error page must be implemented to be shown in an error case. This page should show a message and may give option to the user for returning another page(homepage or previous page)
|
non_process
|
error page an error page must be implemented to be shown in an error case this page should show a message and may give option to the user for returning another page homepage or previous page
| 0
|
15,127
| 18,872,086,196
|
IssuesEvent
|
2021-11-13 11:09:23
|
GSG-CF04/To-Do-list-team3
|
https://api.github.com/repos/GSG-CF04/To-Do-list-team3
|
opened
|
Fixing a bug due to a problem with tasks array and local storage
|
bug in-process
|
-The tasks array doesn't take the right order when an element (task) is deleted
-When a task is added after the first reload, the local storage loses the initial data and only shows the new added tasks
|
1.0
|
Fixing a bug due to a problem with tasks array and local storage - -The tasks array doesn't take the right order when an element (task) is deleted
-When a task is added after the first reload, the local storage loses the initial data and only shows the new added tasks
|
process
|
fixing a bug due to a problem with tasks array and local storage the tasks array doesn t take the right order when an element task is deleted when a task is added after the first reload the local storage loses the initial data and only shows the new added tasks
| 1
|
13,863
| 16,620,295,384
|
IssuesEvent
|
2021-06-02 23:14:28
|
GoogleCloudPlatform/cloud-ops-sandbox
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
|
closed
|
Upgrade OTel Trace library version in services written in Python
|
api: cloudtrace lang: python priority: p2 type: process
|
Because finally [Cloud Trace exporter for Python](https://github.com/GoogleCloudPlatform/opentelemetry-operations-python/releases/tag/v1.0.0) reached 1.0.0 (stable), it's time to update the library version in the following services.
* emailservice
* recommendationservicce
|
1.0
|
Upgrade OTel Trace library version in services written in Python - Because finally [Cloud Trace exporter for Python](https://github.com/GoogleCloudPlatform/opentelemetry-operations-python/releases/tag/v1.0.0) reached 1.0.0 (stable), it's time to update the library version in the following services.
* emailservice
* recommendationservicce
|
process
|
upgrade otel trace library version in services written in python because finally reached stable it s time to update the library version in the following services emailservice recommendationservicce
| 1
|
85,194
| 15,736,634,469
|
IssuesEvent
|
2021-03-30 01:05:29
|
hellohaptik/chatbot_ner
|
https://api.github.com/repos/hellohaptik/chatbot_ner
|
closed
|
CVE-2021-3281 (Medium) detected in Django-1.11.29-py2.py3-none-any.whl - autoclosed
|
security vulnerability
|
## CVE-2021-3281 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-1.11.29-py2.py3-none-any.whl</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/49/49/178daa8725d29c475216259eb19e90b2aa0b8c0431af8c7e9b490ae6481d/Django-1.11.29-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/49/49/178daa8725d29c475216259eb19e90b2aa0b8c0431af8c7e9b490ae6481d/Django-1.11.29-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: chatbot_ner/datastore</p>
<p>Path to vulnerable library: chatbot_ner/datastore,chatbot_ner/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Django-1.11.29-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/hellohaptik/chatbot_ner/commit/1dffbf6325ccfcf4a65dbce5276d7cc4cf428abb">1dffbf6325ccfcf4a65dbce5276d7cc4cf428abb</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Django 2.2 before 2.2.18, 3.0 before 3.0.12, and 3.1 before 3.1.6, the django.utils.archive.extract method (used by "startapp --template" and "startproject --template") allows directory traversal via an archive with absolute paths or relative paths with dot segments.
<p>Publish Date: 2021-02-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3281>CVE-2021-3281</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.djangoproject.com/weblog/2021/feb/01/security-releases/">https://www.djangoproject.com/weblog/2021/feb/01/security-releases/</a></p>
<p>Release Date: 2021-02-02</p>
<p>Fix Resolution: 2.2.18,3.0.12,3.1.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-3281 (Medium) detected in Django-1.11.29-py2.py3-none-any.whl - autoclosed - ## CVE-2021-3281 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-1.11.29-py2.py3-none-any.whl</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/49/49/178daa8725d29c475216259eb19e90b2aa0b8c0431af8c7e9b490ae6481d/Django-1.11.29-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/49/49/178daa8725d29c475216259eb19e90b2aa0b8c0431af8c7e9b490ae6481d/Django-1.11.29-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: chatbot_ner/datastore</p>
<p>Path to vulnerable library: chatbot_ner/datastore,chatbot_ner/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Django-1.11.29-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/hellohaptik/chatbot_ner/commit/1dffbf6325ccfcf4a65dbce5276d7cc4cf428abb">1dffbf6325ccfcf4a65dbce5276d7cc4cf428abb</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Django 2.2 before 2.2.18, 3.0 before 3.0.12, and 3.1 before 3.1.6, the django.utils.archive.extract method (used by "startapp --template" and "startproject --template") allows directory traversal via an archive with absolute paths or relative paths with dot segments.
<p>Publish Date: 2021-02-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3281>CVE-2021-3281</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.djangoproject.com/weblog/2021/feb/01/security-releases/">https://www.djangoproject.com/weblog/2021/feb/01/security-releases/</a></p>
<p>Release Date: 2021-02-02</p>
<p>Fix Resolution: 2.2.18,3.0.12,3.1.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in django none any whl autoclosed cve medium severity vulnerability vulnerable library django none any whl a high level python web framework that encourages rapid development and clean pragmatic design library home page a href path to dependency file chatbot ner datastore path to vulnerable library chatbot ner datastore chatbot ner requirements txt dependency hierarchy x django none any whl vulnerable library found in head commit a href found in base branch develop vulnerability details in django before before and before the django utils archive extract method used by startapp template and startproject template allows directory traversal via an archive with absolute paths or relative paths with dot segments publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
126,647
| 17,091,709,370
|
IssuesEvent
|
2021-07-08 18:24:40
|
chanzuckerberg/napari-hub
|
https://api.github.com/repos/chanzuckerberg/napari-hub
|
opened
|
right edge of magnifying glass icon is slightly cut off in searchbar
|
Design P2 bug
|
*from June 2020 pre-launch bug bash*
**Created by**: lucy
## Description
upon loading page, i can see the tiniest cut off the icon in the search bar
## Info
**Bug source**: *unknown*
**Browser**: safari
**PR**: *none*
## Attachments
https://dl.airtable.com/.attachments/66cddfd45c5fb50c895e5af0cd549ad2/5ae641fc/Screenshot2021-06-16at9.48.37AM.png
|
1.0
|
right edge of magnifying glass icon is slightly cut off in searchbar -
*from June 2020 pre-launch bug bash*
**Created by**: lucy
## Description
upon loading page, i can see the tiniest cut off the icon in the search bar
## Info
**Bug source**: *unknown*
**Browser**: safari
**PR**: *none*
## Attachments
https://dl.airtable.com/.attachments/66cddfd45c5fb50c895e5af0cd549ad2/5ae641fc/Screenshot2021-06-16at9.48.37AM.png
|
non_process
|
right edge of magnifying glass icon is slightly cut off in searchbar from june pre launch bug bash created by lucy description upon loading page i can see the tiniest cut off the icon in the search bar info bug source unknown browser safari pr none attachments
| 0
|
14,887
| 18,288,852,924
|
IssuesEvent
|
2021-10-05 13:18:48
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM][manage apps] Participant invitation > Email template issue
|
Bug P1 Participant manager Process: Fixed Process: Tested dev
|
**AR:** Participant invitation email > UI breakage in the Email template.
**ER:** Participant invitation email should be according to the Email template.

|
2.0
|
[PM][manage apps] Participant invitation > Email template issue - **AR:** Participant invitation email > UI breakage in the Email template.
**ER:** Participant invitation email should be according to the Email template.

|
process
|
participant invitation email template issue ar participant invitation email ui breakage in the email template er participant invitation email should be according to the email template
| 1
|
661,498
| 22,057,590,434
|
IssuesEvent
|
2022-05-30 14:13:15
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.coinbase.com - site is not usable
|
browser-firefox priority-important engine-gecko
|
<!-- @browser: Firefox 100.0.2 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; rv:91.0) Gecko/20100101 Firefox/91.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/105070 -->
**URL**: https://www.coinbase.com
**Browser / Version**: Firefox 100.0.2
**Operating System**: macos m1
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Unable to login
**Steps to Reproduce**:
i try to login with my email and my password and the website says that there are some problems with the login and that i should try it later.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/5/c1e70ecb-5ef6-41db-a03d-d3dbe4fc1727.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.coinbase.com - site is not usable - <!-- @browser: Firefox 100.0.2 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; rv:91.0) Gecko/20100101 Firefox/91.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/105070 -->
**URL**: https://www.coinbase.com
**Browser / Version**: Firefox 100.0.2
**Operating System**: macos m1
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Unable to login
**Steps to Reproduce**:
i try to login with my email and my password and the website says that there are some problems with the login and that i should try it later.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/5/c1e70ecb-5ef6-41db-a03d-d3dbe4fc1727.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
site is not usable url browser version firefox operating system macos tested another browser yes chrome problem type site is not usable description unable to login steps to reproduce i try to login with my email and my password and the website says that there are some problems with the login and that i should try it later view the screenshot img alt screenshot src browser configuration none from with ❤️
| 0
|
773,898
| 27,175,396,973
|
IssuesEvent
|
2023-02-18 00:45:26
|
ubiquity/bounty-bot
|
https://api.github.com/repos/ubiquity/bounty-bot
|
closed
|
Enforce conventional commits
|
Price: 100 USD Priority: 0 (Normal) Time: <1 Day
|
We want the repository commit messages to follow the conventional commits pattern. More info: https://www.conventionalcommits.org/en/v1.0.0/
What should be done:
1. On commit (on a local machine) we should check that commit message follows the conventional commits pattern.
2. Add github action that checks commit messages in PR and fails if they do not follow the conventional commits pattern.
|
1.0
|
Enforce conventional commits - We want the repository commit messages to follow the conventional commits pattern. More info: https://www.conventionalcommits.org/en/v1.0.0/
What should be done:
1. On commit (on a local machine) we should check that commit message follows the conventional commits pattern.
2. Add github action that checks commit messages in PR and fails if they do not follow the conventional commits pattern.
|
non_process
|
enforce conventional commits we want the repository commit messages to follow the conventional commits pattern more info what should be done on commit on a local machine we should check that commit message follows the conventional commits pattern add github action that checks commit messages in pr and fails if they do not follow the conventional commits pattern
| 0
|
431,371
| 30,229,723,213
|
IssuesEvent
|
2023-07-06 05:40:49
|
ACCESS-Hive/access-hive.github.io
|
https://api.github.com/repos/ACCESS-Hive/access-hive.github.io
|
closed
|
Have `requirements` section on each page and change `access to gadi@NCI` name
|
documentation enhancement UI Improvements
|
## Description
We might need to have individual `requirements` section in each of the `running a model` sections for the users to first meet the prerequisites to run a model. Also, need to perhaps change the name of `Access to Gadi@NCI` heading to something else (perhaps **Setup**?).
## Why
For enhanced user experience of the hive website.
## Close Criteria
- [ ] Making sure the individual `requirements` sections are included wherever required.
- [ ] Change the `Access to Gadi@NCI` heading name to something more appropriate.
|
1.0
|
Have `requirements` section on each page and change `access to gadi@NCI` name - ## Description
We might need to have individual `requirements` section in each of the `running a model` sections for the users to first meet the prerequisites to run a model. Also, need to perhaps change the name of `Access to Gadi@NCI` heading to something else (perhaps **Setup**?).
## Why
For enhanced user experience of the hive website.
## Close Criteria
- [ ] Making sure the individual `requirements` sections are included wherever required.
- [ ] Change the `Access to Gadi@NCI` heading name to something more appropriate.
|
non_process
|
have requirements section on each page and change access to gadi nci name description we might need to have individual requirements section in each of the running a model sections for the users to first meet the prerequisites to run a model also need to perhaps change the name of access to gadi nci heading to something else perhaps setup why for enhanced user experience of the hive website close criteria making sure the individual requirements sections are included wherever required change the access to gadi nci heading name to something more appropriate
| 0
|
233,642
| 19,025,470,961
|
IssuesEvent
|
2021-11-24 02:33:15
|
carbon-design-system/carbon-for-ibm-dotcom
|
https://api.github.com/repos/carbon-design-system/carbon-for-ibm-dotcom
|
closed
|
[Test Scenario]: Table of contents - DDS consulting
|
priority: high package: react dev package: web components owner: Hybrid Cloud test: e2e
|
**Engineering info:**
- HC engineer: @andy-blum
- HC JIRA Ticket: https://jsw.ibm.com/browse/HC-2307
- DDS consulting engineer: @ariellalgilmore
----
### Component and test scenario title
Table of contents
### Test scenario steps
a. Check Table of contents sidebar and content is loaded, clickable
b. Verify on click of sidebar menus, it must redirect to actual content on the right side
c. Check the Table of contents sidebar Select State
d. Verify on scrolling page, Table of contents sidebar must remain same position whereas the sidebar Select state changes.
e. Customize theme from Carbon theme and verify gray 10, gray 90 and gray 100 theme.
|
1.0
|
[Test Scenario]: Table of contents - DDS consulting - **Engineering info:**
- HC engineer: @andy-blum
- HC JIRA Ticket: https://jsw.ibm.com/browse/HC-2307
- DDS consulting engineer: @ariellalgilmore
----
### Component and test scenario title
Table of contents
### Test scenario steps
a. Check Table of contents sidebar and content is loaded, clickable
b. Verify on click of sidebar menus, it must redirect to actual content on the right side
c. Check the Table of contents sidebar Select State
d. Verify on scrolling page, Table of contents sidebar must remain same position whereas the sidebar Select state changes.
e. Customize theme from Carbon theme and verify gray 10, gray 90 and gray 100 theme.
|
non_process
|
table of contents dds consulting engineering info hc engineer andy blum hc jira ticket dds consulting engineer ariellalgilmore component and test scenario title table of contents test scenario steps a check table of contents sidebar and content is loaded clickable b verify on click of sidebar menus it must redirect to actual content on the right side c check the table of contents sidebar select state d verify on scrolling page table of contents sidebar must remain same position whereas the sidebar select state changes e customize theme from carbon theme and verify gray gray and gray theme
| 0
|
8,386
| 11,551,156,578
|
IssuesEvent
|
2020-02-19 00:32:43
|
okTurtles/group-income-simple
|
https://api.github.com/repos/okTurtles/group-income-simple
|
opened
|
Add a security-related bug bounty program
|
App:Backend App:Frontend Kind:Community Kind:Process Note:Security
|
### Problem
Although we will do our best to ensure no such bugs ever exist, we cannot absolutely guarantee that no vulnerability every makes its way into our code or the code of one of our dependencies.
### Solution
After launch, start a bug bounty program that pays out for security-related vulnerabilities.
|
1.0
|
Add a security-related bug bounty program - ### Problem
Although we will do our best to ensure no such bugs ever exist, we cannot absolutely guarantee that no vulnerability every makes its way into our code or the code of one of our dependencies.
### Solution
After launch, start a bug bounty program that pays out for security-related vulnerabilities.
|
process
|
add a security related bug bounty program problem although we will do our best to ensure no such bugs ever exist we cannot absolutely guarantee that no vulnerability every makes its way into our code or the code of one of our dependencies solution after launch start a bug bounty program that pays out for security related vulnerabilities
| 1
|
6,136
| 8,998,586,178
|
IssuesEvent
|
2019-02-02 23:17:24
|
leg2015/Aagos
|
https://api.github.com/repos/leg2015/Aagos
|
opened
|
Python script DataCleanStat.py
|
data processing
|
Stat file cleaning script is not working for some reason. Should debug at some point.
|
1.0
|
Python script DataCleanStat.py - Stat file cleaning script is not working for some reason. Should debug at some point.
|
process
|
python script datacleanstat py stat file cleaning script is not working for some reason should debug at some point
| 1
|
18,730
| 24,624,883,171
|
IssuesEvent
|
2022-10-16 11:44:18
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
denoise (profiled) not working in 3.6 (jpg)
|
understood: incomplete reproduce: peculiar scope: image processing no-issue-activity
|
**Describe the bug/issue**
darktable 3.6 on Windows 10 x64. This happens on all JPG (4 I've tried).
I have a photo taken at ISO 160 that has quite a bit of noise in solid-colored areas that are out of the focus plane. Turning on denoise (profiled), DT says “found match for ISO 160”. But I don’t see any change toggling the module on and off. In 3.4, the module defaults would give a visible noise reduction.
**To Reproduce**
1. Open jpg in darkroom
2. exposure module, enable compensation for camera exposure
3. filmic rgb module, click auto tune levels
4. denoise (profiled). turn on. Observe message (e.g. found match for ISO 160). Observe no change to the display (i.e. still noisy).
At this point, changing the sliders doesn't change the display. Picking the camera model from the profile (ILCE-6400 iso 160) doesn't change the display. Changing the blend color space between scene and display doesn't change the display.
Setting the mode to wavelets and the color mode to RGB changes the histogram (slightly) but not the display. You can toggle denoise (profiled) on/off and see a slight change in the histogram, but not the display. Resetting the parameters to defaults reverts to the above behavior (i.e. toggling denoise (profile) doesn't affect the display or histogram).
**Expected behavior**
display shows noise reduction.
**Screenshots**

**Screencast**
**Which commit introduced the error**
_If possible, please try using `git bisect` to determine which commit introduced the issue and place the result here._
_A bisect is much appreciated and can significantly simplify the developer's job._
_HowTo: https://github.com/darktable-org/darktable/wiki#finding-bug-causes and https://www.youtube.com/watch?v=D7JJnLFOn4A_
**Platform**
Installed from the official Windows x64 binary.
* darktable version : 3.6.0
* OS : Windows 10
* Memory : 32 GB
* Graphics card : NVidia GeForce GTX 1660
* Graphics driver : NVidia 466.77
* OpenCL installed : Y? (see output below)
* OpenCL activated :
` C:\Program Files\darktable\bin>darktable-cltest.exe
0.126174 [opencl_init] opencl related configuration options:
0.128112 [opencl_init]
0.128827 [opencl_init] opencl: 1
0.129716 [opencl_init] opencl_scheduling_profile: 'default'
0.131286 [opencl_init] opencl_library: ''
0.132575 [opencl_init] opencl_memory_requirement: 768
0.134156 [opencl_init] opencl_memory_headroom: 400
0.135647 [opencl_init] opencl_device_priority: '*/!0,*/*/*/!0,*'
0.137591 [opencl_init] opencl_mandatory_timeout: 200
0.140182 [opencl_init] opencl_size_roundup: 16
0.141662 [opencl_init] opencl_async_pixelpipe: 0
0.143082 [opencl_init] opencl_synch_cache: active module
0.144900 [opencl_init] opencl_number_event_handles: 25
0.146486 [opencl_init] opencl_micro_nap: 1000
0.147401 [opencl_init] opencl_use_pinned_memory: 0
0.148802 [opencl_init] opencl_use_cpu_devices: 0
0.150333 [opencl_init] opencl_avoid_atomics: 0
0.151967 [opencl_init]
0.156789 [opencl_init] found opencl runtime library 'OpenCL.dll'
0.159011 [opencl_init] opencl library 'OpenCL.dll' found on your system and loaded
0.347240 [opencl_init] found 1 platform
0.348769 [opencl_init] found 1 device
0.349935 [opencl_init] device 0 `NVIDIA GeForce GTX 1660' has sm_20 support.
0.352173 [opencl_init] device 0 `NVIDIA GeForce GTX 1660' supports image sizes of 32768 x 32768
0.354941 [opencl_init] device 0 `NVIDIA GeForce GTX 1660' allows GPU memory allocations of up to 1536MB
[opencl_init] device 0: NVIDIA GeForce GTX 1660
GLOBAL_MEM_SIZE: 6144MB
MAX_WORK_GROUP_SIZE: 1024
MAX_WORK_ITEM_DIMENSIONS: 3
MAX_WORK_ITEM_SIZES: [ 1024 1024 64 ]
DRIVER_VERSION: 466.77
DEVICE_VERSION: OpenCL 3.0 CUDA`
**Additional context**
denoise (profiled) with JPG was working on DT 3.4.
|
1.0
|
denoise (profiled) not working in 3.6 (jpg) - **Describe the bug/issue**
darktable 3.6 on Windows 10 x64. This happens on all JPG (4 I've tried).
I have a photo taken at ISO 160 that has quite a bit of noise in solid-colored areas that are out of the focus plane. Turning on denoise (profiled), DT says “found match for ISO 160”. But I don’t see any change toggling the module on and off. In 3.4, the module defaults would give a visible noise reduction.
**To Reproduce**
1. Open jpg in darkroom
2. exposure module, enable compensation for camera exposure
3. filmic rgb module, click auto tune levels
4. denoise (profiled). turn on. Observe message (e.g. found match for ISO 160). Observe no change to the display (i.e. still noisy).
At this point, changing the sliders doesn't change the display. Picking the camera model from the profile (ILCE-6400 iso 160) doesn't change the display. Changing the blend color space between scene and display doesn't change the display.
Setting the mode to wavelets and the color mode to RGB changes the histogram (slightly) but not the display. You can toggle denoise (profiled) on/off and see a slight change in the histogram, but not the display. Resetting the parameters to defaults reverts to the above behavior (i.e. toggling denoise (profile) doesn't affect the display or histogram).
**Expected behavior**
display shows noise reduction.
**Screenshots**

**Screencast**
**Which commit introduced the error**
_If possible, please try using `git bisect` to determine which commit introduced the issue and place the result here._
_A bisect is much appreciated and can significantly simplify the developer's job._
_HowTo: https://github.com/darktable-org/darktable/wiki#finding-bug-causes and https://www.youtube.com/watch?v=D7JJnLFOn4A_
**Platform**
Installed from the official Windows x64 binary.
* darktable version : 3.6.0
* OS : Windows 10
* Memory : 32 GB
* Graphics card : NVidia GeForce GTX 1660
* Graphics driver : NVidia 466.77
* OpenCL installed : Y? (see output below)
* OpenCL activated :
` C:\Program Files\darktable\bin>darktable-cltest.exe
0.126174 [opencl_init] opencl related configuration options:
0.128112 [opencl_init]
0.128827 [opencl_init] opencl: 1
0.129716 [opencl_init] opencl_scheduling_profile: 'default'
0.131286 [opencl_init] opencl_library: ''
0.132575 [opencl_init] opencl_memory_requirement: 768
0.134156 [opencl_init] opencl_memory_headroom: 400
0.135647 [opencl_init] opencl_device_priority: '*/!0,*/*/*/!0,*'
0.137591 [opencl_init] opencl_mandatory_timeout: 200
0.140182 [opencl_init] opencl_size_roundup: 16
0.141662 [opencl_init] opencl_async_pixelpipe: 0
0.143082 [opencl_init] opencl_synch_cache: active module
0.144900 [opencl_init] opencl_number_event_handles: 25
0.146486 [opencl_init] opencl_micro_nap: 1000
0.147401 [opencl_init] opencl_use_pinned_memory: 0
0.148802 [opencl_init] opencl_use_cpu_devices: 0
0.150333 [opencl_init] opencl_avoid_atomics: 0
0.151967 [opencl_init]
0.156789 [opencl_init] found opencl runtime library 'OpenCL.dll'
0.159011 [opencl_init] opencl library 'OpenCL.dll' found on your system and loaded
0.347240 [opencl_init] found 1 platform
0.348769 [opencl_init] found 1 device
0.349935 [opencl_init] device 0 `NVIDIA GeForce GTX 1660' has sm_20 support.
0.352173 [opencl_init] device 0 `NVIDIA GeForce GTX 1660' supports image sizes of 32768 x 32768
0.354941 [opencl_init] device 0 `NVIDIA GeForce GTX 1660' allows GPU memory allocations of up to 1536MB
[opencl_init] device 0: NVIDIA GeForce GTX 1660
GLOBAL_MEM_SIZE: 6144MB
MAX_WORK_GROUP_SIZE: 1024
MAX_WORK_ITEM_DIMENSIONS: 3
MAX_WORK_ITEM_SIZES: [ 1024 1024 64 ]
DRIVER_VERSION: 466.77
DEVICE_VERSION: OpenCL 3.0 CUDA`
**Additional context**
denoise (profiled) with JPG was working on DT 3.4.
|
process
|
denoise profiled not working in jpg describe the bug issue darktable on windows this happens on all jpg i ve tried i have a photo taken at iso that has quite a bit of noise in solid colored areas that are out of the focus plane turning on denoise profiled dt says “found match for iso ” but i don’t see any change toggling the module on and off in the module defaults would give a visible noise reduction to reproduce open jpg in darkroom exposure module enable compensation for camera exposure filmic rgb module click auto tune levels denoise profiled turn on observe message e g found match for iso observe no change to the display i e still noisy at this point changing the sliders doesn t change the display picking the camera model from the profile ilce iso doesn t change the display changing the blend color space between scene and display doesn t change the display setting the mode to wavelets and the color mode to rgb changes the histogram slightly but not the display you can toggle denoise profiled on off and see a slight change in the histogram but not the display resetting the parameters to defaults reverts to the above behavior i e toggling denoise profile doesn t affect the display or histogram expected behavior display shows noise reduction screenshots screencast which commit introduced the error if possible please try using git bisect to determine which commit introduced the issue and place the result here a bisect is much appreciated and can significantly simplify the developer s job howto and platform installed from the official windows binary darktable version os windows memory gb graphics card nvidia geforce gtx graphics driver nvidia opencl installed y see output below opencl activated c program files darktable bin darktable cltest exe opencl related configuration options opencl opencl scheduling profile default opencl library opencl memory requirement opencl memory headroom opencl device priority opencl mandatory timeout opencl size roundup opencl async pixelpipe opencl synch cache active module opencl number event handles opencl micro nap opencl use pinned memory opencl use cpu devices opencl avoid atomics found opencl runtime library opencl dll opencl library opencl dll found on your system and loaded found platform found device device nvidia geforce gtx has sm support device nvidia geforce gtx supports image sizes of x device nvidia geforce gtx allows gpu memory allocations of up to device nvidia geforce gtx global mem size max work group size max work item dimensions max work item sizes driver version device version opencl cuda additional context denoise profiled with jpg was working on dt
| 1
|
27,406
| 5,003,178,616
|
IssuesEvent
|
2016-12-11 19:45:03
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
opened
|
GetGenericArguments returns null for generic type definition
|
defect
|
### Expected
```js
Array
```
### Actual
```js
null
```
### Steps To Reproduce
[Deck](http://deck.net/6d4b2e89d25d3b84dd6a74c03ed9b609)
```cs
public class Base<T, U> { }
public class Derived<V> : Base<V, V> { }
public class Program
{
public static void Main()
{
Type derivedType = typeof(Derived<>);
Console.WriteLine(derivedType.GetGenericArguments());
}
}
```
|
1.0
|
GetGenericArguments returns null for generic type definition - ### Expected
```js
Array
```
### Actual
```js
null
```
### Steps To Reproduce
[Deck](http://deck.net/6d4b2e89d25d3b84dd6a74c03ed9b609)
```cs
public class Base<T, U> { }
public class Derived<V> : Base<V, V> { }
public class Program
{
public static void Main()
{
Type derivedType = typeof(Derived<>);
Console.WriteLine(derivedType.GetGenericArguments());
}
}
```
|
non_process
|
getgenericarguments returns null for generic type definition expected js array actual js null steps to reproduce cs public class base public class derived base public class program public static void main type derivedtype typeof derived console writeline derivedtype getgenericarguments
| 0
|
278,740
| 8,649,749,808
|
IssuesEvent
|
2018-11-26 20:22:09
|
ansible/awx
|
https://api.github.com/repos/ansible/awx
|
closed
|
RFE: add edge conflict tooltip to workflow graph
|
component:ui priority:medium type:enhancement
|
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
- UI
##### SUMMARY
Workflow graph edge conflicts should have a tool tip or help text displayed when present that instructs the user to go all or nothing success/fail nodes or always nodes. Without it, it's hard to know how to immediately resolve.
##### ENVIRONMENT
##### STEPS TO REPRODUCE
Create a workflow like:
root -- on success --> node 1 -- always --> node 3
\-- on success --> node 2
delete node 1
##### EXPECTED RESULTS
The edge conflict text has a tooltip on hover that tells you to adjust the node connection type
##### ACTUAL RESULTS
No help
##### ADDITIONAL INFORMATION
|
1.0
|
RFE: add edge conflict tooltip to workflow graph - ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
- UI
##### SUMMARY
Workflow graph edge conflicts should have a tool tip or help text displayed when present that instructs the user to go all or nothing success/fail nodes or always nodes. Without it, it's hard to know how to immediately resolve.
##### ENVIRONMENT
##### STEPS TO REPRODUCE
Create a workflow like:
root -- on success --> node 1 -- always --> node 3
\-- on success --> node 2
delete node 1
##### EXPECTED RESULTS
The edge conflict text has a tooltip on hover that tells you to adjust the node connection type
##### ACTUAL RESULTS
No help
##### ADDITIONAL INFORMATION
|
non_process
|
rfe add edge conflict tooltip to workflow graph issue type feature idea component name ui summary workflow graph edge conflicts should have a tool tip or help text displayed when present that instructs the user to go all or nothing success fail nodes or always nodes without it it s hard to know how to immediately resolve environment steps to reproduce create a workflow like root on success node always node on success node delete node expected results the edge conflict text has a tooltip on hover that tells you to adjust the node connection type actual results no help additional information
| 0
|
15,802
| 19,989,274,006
|
IssuesEvent
|
2022-01-31 02:56:56
|
gradle/gradle
|
https://api.github.com/repos/gradle/gradle
|
closed
|
Annotation processing does not included Filer-generated class files on compilation classpath in Java 8
|
a:bug in:annotation-processing
|
### Expected Behavior
If an annotation processor generates `.class` files directly via `Filer#createClassFile`, it should be on the compilation classpath and be on the IDE indexing classpath. This way other code (generated or manual) can use the generated APIs.
These should be considered part of the source sets as well (treated as classes without sources) and indexable by the IDE in the same way generated .java source files are. Yes they would be red before an initial build, but be present by the time annotation processing is done. Much in the same way as, say, how Dagger will generate a `Dagger__Component.java` that will be red until it's generated in the first build.
### Current Behavior
If an annotation processor generates `.class` files directly via `Filer#createClassFile`, the created class does not appear to be included on the compilation classpath. It _is_ included in the final jar though. This is a problem because it prevents sources (generated or manually written) from accessing any APIs in the generated class file.
The class file APIs do not index in the IDE either. They partially index if `"$buildDir/classes/java/main/io/sweers/classgentesting/sample"` is added to dependencies via `files()`, but that does not fix the missing class file on compilation classpath and compilation will still fail. It also does not index in an importable way, oddly. It can only be fully qualified even if in the same package.
This also appears to not work even if a `.class` file is included manually in source sets (as in - not generated, just copied in). Not that that's a target use case, but probably a similar problem to how the Java compilation task views them.
### Context
The example is a small case, but this could be a general use improvement to any annotation processor by allowing them to generate bytecode directly when possible. This was an idea researched by the Dagger team at one point, and also a good solution for small-scope annotation processors that generate very simple code that could be done without adding to the compilation workload after processing is done ([dagger-reflect](https://github.com/JakeWharton/dagger-reflect/tree/master/codegen) is a good example of this).
### Steps to Reproduce
Repro gradle project: https://github.com/ZacSweers/copydynamic/tree/z/dynamicInvocationTesting/classgentesting. Run `./gradlew :classgentesting:sample:build` Basically look at `SimpleClass.java` in the sample project [here](https://github.com/ZacSweers/copydynamic/blob/z%2FdynamicInvocationTesting/classgentesting/sample/src/main/java/io/sweers/classgentesting/sample/SimpleClass.java).
After one attempted build (to generate the initial class file), it will link in the IDE but the build (and any subsequent) will not pass. Bytecode generation is in `ClassGenProcessor.java` using ASM.
The sample [build file](https://github.com/ZacSweers/copydynamic/blob/z%2FdynamicInvocationTesting/classgentesting/sample/build.gradle) also has the IDE indexing dependency trick.
### Your Environment
Build scan URL: https://scans.gradle.com/s/cjcyp3tny64q6s (but this gives me a message saying "This page does not exist. (or you do not have permission to view it)", so not sure what to do).
|
1.0
|
Annotation processing does not included Filer-generated class files on compilation classpath in Java 8 - ### Expected Behavior
If an annotation processor generates `.class` files directly via `Filer#createClassFile`, it should be on the compilation classpath and be on the IDE indexing classpath. This way other code (generated or manual) can use the generated APIs.
These should be considered part of the source sets as well (treated as classes without sources) and indexable by the IDE in the same way generated .java source files are. Yes they would be red before an initial build, but be present by the time annotation processing is done. Much in the same way as, say, how Dagger will generate a `Dagger__Component.java` that will be red until it's generated in the first build.
### Current Behavior
If an annotation processor generates `.class` files directly via `Filer#createClassFile`, the created class does not appear to be included on the compilation classpath. It _is_ included in the final jar though. This is a problem because it prevents sources (generated or manually written) from accessing any APIs in the generated class file.
The class file APIs do not index in the IDE either. They partially index if `"$buildDir/classes/java/main/io/sweers/classgentesting/sample"` is added to dependencies via `files()`, but that does not fix the missing class file on compilation classpath and compilation will still fail. It also does not index in an importable way, oddly. It can only be fully qualified even if in the same package.
This also appears to not work even if a `.class` file is included manually in source sets (as in - not generated, just copied in). Not that that's a target use case, but probably a similar problem to how the Java compilation task views them.
### Context
The example is a small case, but this could be a general use improvement to any annotation processor by allowing them to generate bytecode directly when possible. This was an idea researched by the Dagger team at one point, and also a good solution for small-scope annotation processors that generate very simple code that could be done without adding to the compilation workload after processing is done ([dagger-reflect](https://github.com/JakeWharton/dagger-reflect/tree/master/codegen) is a good example of this).
### Steps to Reproduce
Repro gradle project: https://github.com/ZacSweers/copydynamic/tree/z/dynamicInvocationTesting/classgentesting. Run `./gradlew :classgentesting:sample:build` Basically look at `SimpleClass.java` in the sample project [here](https://github.com/ZacSweers/copydynamic/blob/z%2FdynamicInvocationTesting/classgentesting/sample/src/main/java/io/sweers/classgentesting/sample/SimpleClass.java).
After one attempted build (to generate the initial class file), it will link in the IDE but the build (and any subsequent) will not pass. Bytecode generation is in `ClassGenProcessor.java` using ASM.
The sample [build file](https://github.com/ZacSweers/copydynamic/blob/z%2FdynamicInvocationTesting/classgentesting/sample/build.gradle) also has the IDE indexing dependency trick.
### Your Environment
Build scan URL: https://scans.gradle.com/s/cjcyp3tny64q6s (but this gives me a message saying "This page does not exist. (or you do not have permission to view it)", so not sure what to do).
|
process
|
annotation processing does not included filer generated class files on compilation classpath in java expected behavior if an annotation processor generates class files directly via filer createclassfile it should be on the compilation classpath and be on the ide indexing classpath this way other code generated or manual can use the generated apis these should be considered part of the source sets as well treated as classes without sources and indexable by the ide in the same way generated java source files are yes they would be red before an initial build but be present by the time annotation processing is done much in the same way as say how dagger will generate a dagger component java that will be red until it s generated in the first build current behavior if an annotation processor generates class files directly via filer createclassfile the created class does not appear to be included on the compilation classpath it is included in the final jar though this is a problem because it prevents sources generated or manually written from accessing any apis in the generated class file the class file apis do not index in the ide either they partially index if builddir classes java main io sweers classgentesting sample is added to dependencies via files but that does not fix the missing class file on compilation classpath and compilation will still fail it also does not index in an importable way oddly it can only be fully qualified even if in the same package this also appears to not work even if a class file is included manually in source sets as in not generated just copied in not that that s a target use case but probably a similar problem to how the java compilation task views them context the example is a small case but this could be a general use improvement to any annotation processor by allowing them to generate bytecode directly when possible this was an idea researched by the dagger team at one point and also a good solution for small scope annotation processors that generate very simple code that could be done without adding to the compilation workload after processing is done is a good example of this steps to reproduce repro gradle project run gradlew classgentesting sample build basically look at simpleclass java in the sample project after one attempted build to generate the initial class file it will link in the ide but the build and any subsequent will not pass bytecode generation is in classgenprocessor java using asm the sample also has the ide indexing dependency trick your environment build scan url but this gives me a message saying this page does not exist or you do not have permission to view it so not sure what to do
| 1
|
10,736
| 13,534,068,869
|
IssuesEvent
|
2020-09-16 04:46:24
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
nonsense code causes infinite regress in error reports
|
Bug Modeller Processing
|
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
When I type totally bogus code into the field calculator, I get an infinite number of error messages.
**How to Reproduce**
This is just like my bug https://github.com/qgis/QGIS/issues/37737
1. Go to Processing/Graphical Modeler
2. Click on Algorithms. Search "field ca". Double click Vector table/Field calculator.
3. Configure the calculator to make a new attribute, called 'snam2'.
4. Configure the calculator to have a totally rediculous formula. I used ' 5/0 + dkdkdkd and his is this -'.
5. Run the model.
What I expect: immediate syntax error, model will never execute, marked as invalid syntax in the XML, big red X on the screen when it's open in the graphical modeler, all because the formula '5/0 + dkdkdkd and his is this -' is not parsable by any rational computer language including the qgis one. Errors in parsing like this, 'syntax errors', can be easily detected by running the code through the parser: especially the inclusion of the '-' on the end without the second argument to the minus operator is detectable. Probably also the other words appearing in positions where keywords would be expected and aren't. And - what language allows a random long rambling sequence of keywords (or variables for that matter) like 'dkdkdkd and his is this'? Not SQL certainly, which I think is what this language is trying to implement: SQL has a very strict sequence of words that must appear in order.
What I get instead:
Unlike my 37737 bug, this one is caught when the code is run. However, it's caught too many times. In fact I get:
"OK. Execution took 19.039 s (1 outputs).
Prepare algorithm: qgis:fieldcalculator_1
Running Field calculator [2/2]
Input Parameters:
{ FIELD_LENGTH: 10, FIELD_NAME: 'snam2', FIELD_PRECISION: 3, FIELD_TYPE: 0, FORMULA: '5/0 + dkdkdkd and his is this -', INPUT: 'memory://MultiLineString?crs=USER:100000&field=fid:long(0,0)&field=cat:integer(0,0)&field=length:double(0,0)&field=localid:integer(0,0)&field=zero:integer(0,0)&field=prefix:string(14,0)&field=streetname:string(50,0)&field=ftype:string(13,0)&field=leftadd1:integer(0,0)&field=leftadd2:integer(0,0)&field=rgtadd1:integer(0,0)&field=rgtadd2:integer(0,0)&field=leftzip:integer(0,0)&field=rightzip:integer(0,0)&field=type:integer(0,0)&field=lcounty:string(4,0)&field=rcounty:string(4,0)&field=lcity:string(4,0)&field=rcity:string(4,0)&field=roadtypeid:integer(0,0)&field=emissionsingramsperhr:double(0,0)&field=DN:integer(0,0)&field=length_2:double(0,0)&uid={d628594b-4822-4073-b5f2-1cf3fc71a5b6}', NEW_FIELD: True, OUTPUT: 'TEMPORARY_OUTPUT' }
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed? [...]".
QGIS should report a meaningful error message and then abort after the first report. "SYNTAX ERROR LINE 1" is enough, one time.
**QGIS and OS versions**
- QGIS dev version fetched just a few days ago.
- OS is Debian 10
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
The about screen appears completely black with no controls or text.
**Additional context**
<!-- Add any other context about the problem here. -->
c.f. https://github.com/qgis/QGIS/issues/37737
|
1.0
|
nonsense code causes infinite regress in error reports - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
When I type totally bogus code into the field calculator, I get an infinite number of error messages.
**How to Reproduce**
This is just like my bug https://github.com/qgis/QGIS/issues/37737
1. Go to Processing/Graphical Modeler
2. Click on Algorithms. Search "field ca". Double click Vector table/Field calculator.
3. Configure the calculator to make a new attribute, called 'snam2'.
4. Configure the calculator to have a totally rediculous formula. I used ' 5/0 + dkdkdkd and his is this -'.
5. Run the model.
What I expect: immediate syntax error, model will never execute, marked as invalid syntax in the XML, big red X on the screen when it's open in the graphical modeler, all because the formula '5/0 + dkdkdkd and his is this -' is not parsable by any rational computer language including the qgis one. Errors in parsing like this, 'syntax errors', can be easily detected by running the code through the parser: especially the inclusion of the '-' on the end without the second argument to the minus operator is detectable. Probably also the other words appearing in positions where keywords would be expected and aren't. And - what language allows a random long rambling sequence of keywords (or variables for that matter) like 'dkdkdkd and his is this'? Not SQL certainly, which I think is what this language is trying to implement: SQL has a very strict sequence of words that must appear in order.
What I get instead:
Unlike my 37737 bug, this one is caught when the code is run. However, it's caught too many times. In fact I get:
"OK. Execution took 19.039 s (1 outputs).
Prepare algorithm: qgis:fieldcalculator_1
Running Field calculator [2/2]
Input Parameters:
{ FIELD_LENGTH: 10, FIELD_NAME: 'snam2', FIELD_PRECISION: 3, FIELD_TYPE: 0, FORMULA: '5/0 + dkdkdkd and his is this -', INPUT: 'memory://MultiLineString?crs=USER:100000&field=fid:long(0,0)&field=cat:integer(0,0)&field=length:double(0,0)&field=localid:integer(0,0)&field=zero:integer(0,0)&field=prefix:string(14,0)&field=streetname:string(50,0)&field=ftype:string(13,0)&field=leftadd1:integer(0,0)&field=leftadd2:integer(0,0)&field=rgtadd1:integer(0,0)&field=rgtadd2:integer(0,0)&field=leftzip:integer(0,0)&field=rightzip:integer(0,0)&field=type:integer(0,0)&field=lcounty:string(4,0)&field=rcounty:string(4,0)&field=lcity:string(4,0)&field=rcity:string(4,0)&field=roadtypeid:integer(0,0)&field=emissionsingramsperhr:double(0,0)&field=DN:integer(0,0)&field=length_2:double(0,0)&uid={d628594b-4822-4073-b5f2-1cf3fc71a5b6}', NEW_FIELD: True, OUTPUT: 'TEMPORARY_OUTPUT' }
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed?
No root node! Parsing failed? [...]".
QGIS should report a meaningful error message and then abort after the first report. "SYNTAX ERROR LINE 1" is enough, one time.
**QGIS and OS versions**
- QGIS dev version fetched just a few days ago.
- OS is Debian 10
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
The about screen appears completely black with no controls or text.
**Additional context**
<!-- Add any other context about the problem here. -->
c.f. https://github.com/qgis/QGIS/issues/37737
|
process
|
nonsense code causes infinite regress in error reports bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug when i type totally bogus code into the field calculator i get an infinite number of error messages how to reproduce this is just like my bug go to processing graphical modeler click on algorithms search field ca double click vector table field calculator configure the calculator to make a new attribute called configure the calculator to have a totally rediculous formula i used dkdkdkd and his is this run the model what i expect immediate syntax error model will never execute marked as invalid syntax in the xml big red x on the screen when it s open in the graphical modeler all because the formula dkdkdkd and his is this is not parsable by any rational computer language including the qgis one errors in parsing like this syntax errors can be easily detected by running the code through the parser especially the inclusion of the on the end without the second argument to the minus operator is detectable probably also the other words appearing in positions where keywords would be expected and aren t and what language allows a random long rambling sequence of keywords or variables for that matter like dkdkdkd and his is this not sql certainly which i think is what this language is trying to implement sql has a very strict sequence of words that must appear in order what i get instead unlike my bug this one is caught when the code is run however it s caught too many times in fact i get ok execution took s outputs prepare algorithm qgis fieldcalculator running field calculator input parameters field length field name field precision field type formula dkdkdkd and his is this input memory multilinestring crs user field fid long field cat integer field length double field localid integer field zero integer field prefix string field streetname string field ftype string field integer field integer field integer field integer field leftzip integer field rightzip integer field type integer field lcounty string field rcounty string field lcity string field rcity string field roadtypeid integer field emissionsingramsperhr double field dn integer field length double uid new field true output temporary output no root node parsing failed no root node parsing failed no root node parsing failed no root node parsing failed no root node parsing failed no root node parsing failed no root node parsing failed no root node parsing failed qgis should report a meaningful error message and then abort after the first report syntax error line is enough one time qgis and os versions qgis dev version fetched just a few days ago os is debian about click in the table ctrl a and then ctrl c finally paste here the about screen appears completely black with no controls or text additional context c f
| 1
|
242,128
| 26,257,117,563
|
IssuesEvent
|
2023-01-06 02:25:11
|
turkdevops/grafana
|
https://api.github.com/repos/turkdevops/grafana
|
closed
|
CVE-2015-9251 (Medium) detected in github.com/smartystreets/goConvey-v1.6.4 - autoclosed
|
security vulnerability
|
## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/smartystreets/goConvey-v1.6.4</b></p></summary>
<p>Go testing in the browser. Integrates with `go test`. Write behavioral tests in Go.</p>
<p>Library home page: <a href="https://proxy.golang.org/github.com/smartystreets/goconvey/@v/v1.6.4.zip">https://proxy.golang.org/github.com/smartystreets/goconvey/@v/v1.6.4.zip</a></p>
<p>
Dependency Hierarchy:
- :x: **github.com/smartystreets/goConvey-v1.6.4** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/a1c271764655c7e3ff81126d5929b8dda6170bf4">a1c271764655c7e3ff81126d5929b8dda6170bf4</a></p>
<p>Found in base branch: <b>datasource-meta</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - 3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2015-9251 (Medium) detected in github.com/smartystreets/goConvey-v1.6.4 - autoclosed - ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/smartystreets/goConvey-v1.6.4</b></p></summary>
<p>Go testing in the browser. Integrates with `go test`. Write behavioral tests in Go.</p>
<p>Library home page: <a href="https://proxy.golang.org/github.com/smartystreets/goconvey/@v/v1.6.4.zip">https://proxy.golang.org/github.com/smartystreets/goconvey/@v/v1.6.4.zip</a></p>
<p>
Dependency Hierarchy:
- :x: **github.com/smartystreets/goConvey-v1.6.4** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/grafana/commit/a1c271764655c7e3ff81126d5929b8dda6170bf4">a1c271764655c7e3ff81126d5929b8dda6170bf4</a></p>
<p>Found in base branch: <b>datasource-meta</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - 3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in github com smartystreets goconvey autoclosed cve medium severity vulnerability vulnerable library github com smartystreets goconvey go testing in the browser integrates with go test write behavioral tests in go library home page a href dependency hierarchy x github com smartystreets goconvey vulnerable library found in head commit a href found in base branch datasource meta vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with mend
| 0
|
1,248
| 3,785,063,667
|
IssuesEvent
|
2016-03-20 08:15:02
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Node + Spawn a child process once and then provide stdin's and read stdout's from it
|
child_process question
|
0
down vote
favorite
I'm spawning a child process using spawn-command npm package, i do this when the node server starts and then for every request i read the query value and hit the running child process with stdin. The stdout that comes out of the child process is an event stream and i add a listener to read the data and then will be sending the data to sse-channel instance.
Everything works the way it should for the first request, the issue arises for subsequent requests. When i console.log the data inside the stdout's on listener handler then i could see that the data is getting cloned. Meaning, same set of data is getting repeated twice for the second request and thrice for the third request etc... I was researching this issue and possible solution for this would be to create independent streams for each request so that this data cloning would be avoided in the child.stdout's on listener handler. But i'm not sure whether this could resolve the issue. Could some one please suggest a way to overcome this hurdle.
Code looks like below,
`var spawnCommand = require('spawn-command');`
`var cmd = 'path to the binary file'; `
` module.exports = function (app) {`
`var child = spawnCommand(cmd);`
`srvc = {`
` get: function(req, res) {`
` child.stdin.write('a json object in a format that is expected by binary' + '\n');`
` child.stdout.on('data', function() {`
` console.log(''+ data);`
`});`
`}`
`}`
`}`
As i mentioned earlier the child process would be spawned once when the node server starts and then the get() method would run for every request. So each time when a request runs, the stdin's would be sent to the binary and then the "on" listener would log the data in stdout. This works perfectly for the first request, but when a second request is made (with the first request still logging data) i could see that the data's for both first and second are getting cloned twice and for third request the data's for each request is getting cloned thrice and it increases exponentially.
On the other hand if i spawn a child everytime inside the get() method i can overcome this problem but unfortunately i cannot do that because there'll be 100's of requests and i cannot afford 100's of binary instance eating up the memory.
Hope i explained it better, any suggestions would be of great help.
My noder version is 5.x.x (even though it does not matter with this query)
|
1.0
|
Node + Spawn a child process once and then provide stdin's and read stdout's from it - 0
down vote
favorite
I'm spawning a child process using spawn-command npm package, i do this when the node server starts and then for every request i read the query value and hit the running child process with stdin. The stdout that comes out of the child process is an event stream and i add a listener to read the data and then will be sending the data to sse-channel instance.
Everything works the way it should for the first request, the issue arises for subsequent requests. When i console.log the data inside the stdout's on listener handler then i could see that the data is getting cloned. Meaning, same set of data is getting repeated twice for the second request and thrice for the third request etc... I was researching this issue and possible solution for this would be to create independent streams for each request so that this data cloning would be avoided in the child.stdout's on listener handler. But i'm not sure whether this could resolve the issue. Could some one please suggest a way to overcome this hurdle.
Code looks like below,
`var spawnCommand = require('spawn-command');`
`var cmd = 'path to the binary file'; `
` module.exports = function (app) {`
`var child = spawnCommand(cmd);`
`srvc = {`
` get: function(req, res) {`
` child.stdin.write('a json object in a format that is expected by binary' + '\n');`
` child.stdout.on('data', function() {`
` console.log(''+ data);`
`});`
`}`
`}`
`}`
As i mentioned earlier the child process would be spawned once when the node server starts and then the get() method would run for every request. So each time when a request runs, the stdin's would be sent to the binary and then the "on" listener would log the data in stdout. This works perfectly for the first request, but when a second request is made (with the first request still logging data) i could see that the data's for both first and second are getting cloned twice and for third request the data's for each request is getting cloned thrice and it increases exponentially.
On the other hand if i spawn a child everytime inside the get() method i can overcome this problem but unfortunately i cannot do that because there'll be 100's of requests and i cannot afford 100's of binary instance eating up the memory.
Hope i explained it better, any suggestions would be of great help.
My noder version is 5.x.x (even though it does not matter with this query)
|
process
|
node spawn a child process once and then provide stdin s and read stdout s from it down vote favorite i m spawning a child process using spawn command npm package i do this when the node server starts and then for every request i read the query value and hit the running child process with stdin the stdout that comes out of the child process is an event stream and i add a listener to read the data and then will be sending the data to sse channel instance everything works the way it should for the first request the issue arises for subsequent requests when i console log the data inside the stdout s on listener handler then i could see that the data is getting cloned meaning same set of data is getting repeated twice for the second request and thrice for the third request etc i was researching this issue and possible solution for this would be to create independent streams for each request so that this data cloning would be avoided in the child stdout s on listener handler but i m not sure whether this could resolve the issue could some one please suggest a way to overcome this hurdle code looks like below var spawncommand require spawn command var cmd path to the binary file module exports function app var child spawncommand cmd srvc get function req res child stdin write a json object in a format that is expected by binary n child stdout on data function console log data as i mentioned earlier the child process would be spawned once when the node server starts and then the get method would run for every request so each time when a request runs the stdin s would be sent to the binary and then the on listener would log the data in stdout this works perfectly for the first request but when a second request is made with the first request still logging data i could see that the data s for both first and second are getting cloned twice and for third request the data s for each request is getting cloned thrice and it increases exponentially on the other hand if i spawn a child everytime inside the get method i can overcome this problem but unfortunately i cannot do that because there ll be s of requests and i cannot afford s of binary instance eating up the memory hope i explained it better any suggestions would be of great help my noder version is x x even though it does not matter with this query
| 1
|
8,919
| 12,025,699,321
|
IssuesEvent
|
2020-04-12 10:36:53
|
nanoframework/Home
|
https://api.github.com/repos/nanoframework/Home
|
opened
|
MDP error with SerialCommunications lib
|
Area: CL-Windows.Devices.SerialCommunication Area: Metadata Processor Priority: High Type: Bug
|
#167 Details about Problem
When building the serial communications library with the new MDP, it fails.
**VS version**
2019
**VS extension version**
2019.1.8.10
## Detailed repro steps so we can see the same problem
1. Clone the repo
2. Open with Visual Studio
3. Change to release mode
4. Rebuild solution
## Screenshot

|
1.0
|
MDP error with SerialCommunications lib - #167 Details about Problem
When building the serial communications library with the new MDP, it fails.
**VS version**
2019
**VS extension version**
2019.1.8.10
## Detailed repro steps so we can see the same problem
1. Clone the repo
2. Open with Visual Studio
3. Change to release mode
4. Rebuild solution
## Screenshot

|
process
|
mdp error with serialcommunications lib details about problem when building the serial communications library with the new mdp it fails vs version vs extension version detailed repro steps so we can see the same problem clone the repo open with visual studio change to release mode rebuild solution screenshot
| 1
|
45,057
| 7,156,822,589
|
IssuesEvent
|
2018-01-26 17:35:43
|
riganti/dotvvm
|
https://api.github.com/repos/riganti/dotvvm
|
closed
|
Install Bootstrap - installs BS4, DotVVM Bootstrap uses BS3.
|
Documentation bug
|
Please kindly give instructions...
https://www.dotvvm.com/docs/tutorials/commercial-bootstrap-for-dotvvm/latest
This uses BS4, the NuGet updates the project to BS4.
|
1.0
|
Install Bootstrap - installs BS4, DotVVM Bootstrap uses BS3. - Please kindly give instructions...
https://www.dotvvm.com/docs/tutorials/commercial-bootstrap-for-dotvvm/latest
This uses BS4, the NuGet updates the project to BS4.
|
non_process
|
install bootstrap installs dotvvm bootstrap uses please kindly give instructions this uses the nuget updates the project to
| 0
|
497,164
| 14,364,856,018
|
IssuesEvent
|
2020-12-01 00:22:39
|
California-Planet-Search/radvel
|
https://api.github.com/repos/California-Planet-Search/radvel
|
closed
|
minAfactor documentation
|
help wanted priority:medium question
|
I believe for mcmc to converge, minAfactor of the chain has to go above the default value of 40 (as indicated by the summary pdf as well), because autocorrelation time will plateau when the number of steps is large, giving N/tau a larger number. But in the documentation page, it says "once the minimum autocorreclation factor is below minAfactor, this criterion for convergence is met." This seems confusing.
|
1.0
|
minAfactor documentation - I believe for mcmc to converge, minAfactor of the chain has to go above the default value of 40 (as indicated by the summary pdf as well), because autocorrelation time will plateau when the number of steps is large, giving N/tau a larger number. But in the documentation page, it says "once the minimum autocorreclation factor is below minAfactor, this criterion for convergence is met." This seems confusing.
|
non_process
|
minafactor documentation i believe for mcmc to converge minafactor of the chain has to go above the default value of as indicated by the summary pdf as well because autocorrelation time will plateau when the number of steps is large giving n tau a larger number but in the documentation page it says once the minimum autocorreclation factor is below minafactor this criterion for convergence is met this seems confusing
| 0
|
19,956
| 26,432,428,631
|
IssuesEvent
|
2023-01-15 00:38:18
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[INFRAESTRUTURA] [PROCESSOS] [PLENO] [SALVADOR] Analista Processos Pl na [STEFANINI]
|
SALVADOR INFRAESTRUTURA PLENO AGILE Certificação ITIL PROCESSOS COBIT EXCEL POWER BI ITIL METODOLOGIAS ÁGEIS HELP WANTED BPMN Stale
|
<!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## Descrição da vaga
- Analista Processos Pl
## Local
- Salvador
## Benefícios
- Informações diretamente com o responsável/ recrutador da vaga
## Requisitos
**Obrigatórios:**
- Superior Completo
- Conhecimento em mapeamento dos processos baseados em ITIL para escrita de novos processos e revisão dos existentes;
- Conhecimento da metodologia BPMN e das ferramentas de modelagem como Bizagi e Visio;
- Conhecimento em gerenciamento de projetos (Agile);
- Conhecimento em COBIT;
- Excel intermediário (desejável Avançado).
- Conhecimento em PowerBI
- Cursos: ITIL Foundation Certified (Exigido)
## Contratação
- a combinar
## Nossa empresa
- Nosso principal objetivo é ajudar você nos desafios do seu negócio. Presente em 41 países com mais de 25.000 funcionários, somos a quinta multinacional brasileira mais internacionalizada* e estamos entre as 100 maiores empresas de TI do mundo. Nossas equipes especializadas trabalham com você para encontrar as soluções ideais para seus desafios comerciais e impulsionar a inovação necessária para garantir que sua empresa prospere na era digital.
- Com nossa abrangente análise de negócios e ampla visão de mercado, criamos soluções personalizadas que proporcionam transformações digitais perfeitas, garantem resultados rápidos e geram impacto duradouro em toda a empresa.
- Investimos em um completo ecossistema de inovação para atender as principais verticais e, por isso, somos reconhecidos com várias premiações. Temos ofertas alinhadas às tendências de mercado como automação, cloud, Internet das Coisas (IoT), automatização, tecnologia cognitiva e User Experience (UX). Assim, você pode contar com nossas soluções que vão desde consultoria e marketing, mobilidade, campanhas personalizadas e inteligência artificial a soluções tradicionais como Service Desk, Field Service e outsourcing (BPO).
- Trabalhamos com mais de 500 clientes em serviços financeiros, manufatura, telecomunicações, serviços químicos, tecnologias e também no setor público.
- Conte conosco para transformar digitalmente sua empresa e aumentar as suas oportunidades de negócio. Ao nos escolher, você se beneficiará com nossa experiência global, recursos digitais robustos e compromisso com o sucesso de sua empresa, do início ao fim.
## Como se candidatar
- [Clique aqui para se candidatar](https://jobs.kenoby.com/vagasstefanini/job/analista-processos-pl/5f6cfe7b64060e562587cf03)
|
1.0
|
[INFRAESTRUTURA] [PROCESSOS] [PLENO] [SALVADOR] Analista Processos Pl na [STEFANINI] - <!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## Descrição da vaga
- Analista Processos Pl
## Local
- Salvador
## Benefícios
- Informações diretamente com o responsável/ recrutador da vaga
## Requisitos
**Obrigatórios:**
- Superior Completo
- Conhecimento em mapeamento dos processos baseados em ITIL para escrita de novos processos e revisão dos existentes;
- Conhecimento da metodologia BPMN e das ferramentas de modelagem como Bizagi e Visio;
- Conhecimento em gerenciamento de projetos (Agile);
- Conhecimento em COBIT;
- Excel intermediário (desejável Avançado).
- Conhecimento em PowerBI
- Cursos: ITIL Foundation Certified (Exigido)
## Contratação
- a combinar
## Nossa empresa
- Nosso principal objetivo é ajudar você nos desafios do seu negócio. Presente em 41 países com mais de 25.000 funcionários, somos a quinta multinacional brasileira mais internacionalizada* e estamos entre as 100 maiores empresas de TI do mundo. Nossas equipes especializadas trabalham com você para encontrar as soluções ideais para seus desafios comerciais e impulsionar a inovação necessária para garantir que sua empresa prospere na era digital.
- Com nossa abrangente análise de negócios e ampla visão de mercado, criamos soluções personalizadas que proporcionam transformações digitais perfeitas, garantem resultados rápidos e geram impacto duradouro em toda a empresa.
- Investimos em um completo ecossistema de inovação para atender as principais verticais e, por isso, somos reconhecidos com várias premiações. Temos ofertas alinhadas às tendências de mercado como automação, cloud, Internet das Coisas (IoT), automatização, tecnologia cognitiva e User Experience (UX). Assim, você pode contar com nossas soluções que vão desde consultoria e marketing, mobilidade, campanhas personalizadas e inteligência artificial a soluções tradicionais como Service Desk, Field Service e outsourcing (BPO).
- Trabalhamos com mais de 500 clientes em serviços financeiros, manufatura, telecomunicações, serviços químicos, tecnologias e também no setor público.
- Conte conosco para transformar digitalmente sua empresa e aumentar as suas oportunidades de negócio. Ao nos escolher, você se beneficiará com nossa experiência global, recursos digitais robustos e compromisso com o sucesso de sua empresa, do início ao fim.
## Como se candidatar
- [Clique aqui para se candidatar](https://jobs.kenoby.com/vagasstefanini/job/analista-processos-pl/5f6cfe7b64060e562587cf03)
|
process
|
analista processos pl na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na descrição da vaga analista processos pl local salvador benefícios informações diretamente com o responsável recrutador da vaga requisitos obrigatórios superior completo conhecimento em mapeamento dos processos baseados em itil para escrita de novos processos e revisão dos existentes conhecimento da metodologia bpmn e das ferramentas de modelagem como bizagi e visio conhecimento em gerenciamento de projetos agile conhecimento em cobit excel intermediário desejável avançado conhecimento em powerbi cursos itil foundation certified exigido contratação a combinar nossa empresa nosso principal objetivo é ajudar você nos desafios do seu negócio presente em países com mais de funcionários somos a quinta multinacional brasileira mais internacionalizada e estamos entre as maiores empresas de ti do mundo nossas equipes especializadas trabalham com você para encontrar as soluções ideais para seus desafios comerciais e impulsionar a inovação necessária para garantir que sua empresa prospere na era digital com nossa abrangente análise de negócios e ampla visão de mercado criamos soluções personalizadas que proporcionam transformações digitais perfeitas garantem resultados rápidos e geram impacto duradouro em toda a empresa investimos em um completo ecossistema de inovação para atender as principais verticais e por isso somos reconhecidos com várias premiações temos ofertas alinhadas às tendências de mercado como automação cloud internet das coisas iot automatização tecnologia cognitiva e user experience ux assim você pode contar com nossas soluções que vão desde consultoria e marketing mobilidade campanhas personalizadas e inteligência artificial a soluções tradicionais como service desk field service e outsourcing bpo trabalhamos com mais de clientes em serviços financeiros manufatura telecomunicações serviços químicos tecnologias e também no setor público conte conosco para transformar digitalmente sua empresa e aumentar as suas oportunidades de negócio ao nos escolher você se beneficiará com nossa experiência global recursos digitais robustos e compromisso com o sucesso de sua empresa do início ao fim como se candidatar
| 1
|
74,355
| 20,144,542,393
|
IssuesEvent
|
2022-02-09 05:20:18
|
envoyproxy/envoy
|
https://api.github.com/repos/envoyproxy/envoy
|
closed
|
feature request: better build @com_googlesource_chromium_v8:build
|
area/build help wanted area/wasm
|
This target is internal spawn lots of tasks within. However, this target is considered as a single-core task from the view of bazel scheduler.
Generally speaking, when you run `bazel build //source/exe:envoy` with N cores, N tasks will be scheduled and at most N clang is running. However, if the running task contains @com_googlesource_chromium_v8:build, ~2N clang is running, and the peak ram usage could be huge.
A workaround: you may want to run `bazel build @com_googlesource_chromium_v8//:build` first, exhausting the cpu and ram.
When the above task is completed, you can run `bazel build YOUR_REAL_TARGET` using the build cache of the above v8 build.
|
1.0
|
feature request: better build @com_googlesource_chromium_v8:build - This target is internal spawn lots of tasks within. However, this target is considered as a single-core task from the view of bazel scheduler.
Generally speaking, when you run `bazel build //source/exe:envoy` with N cores, N tasks will be scheduled and at most N clang is running. However, if the running task contains @com_googlesource_chromium_v8:build, ~2N clang is running, and the peak ram usage could be huge.
A workaround: you may want to run `bazel build @com_googlesource_chromium_v8//:build` first, exhausting the cpu and ram.
When the above task is completed, you can run `bazel build YOUR_REAL_TARGET` using the build cache of the above v8 build.
|
non_process
|
feature request better build com googlesource chromium build this target is internal spawn lots of tasks within however this target is considered as a single core task from the view of bazel scheduler generally speaking when you run bazel build source exe envoy with n cores n tasks will be scheduled and at most n clang is running however if the running task contains com googlesource chromium build clang is running and the peak ram usage could be huge a workaround you may want to run bazel build com googlesource chromium build first exhausting the cpu and ram when the above task is completed you can run bazel build your real target using the build cache of the above build
| 0
|
11,135
| 13,957,691,883
|
IssuesEvent
|
2020-10-24 08:10:32
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
LU: Unable to harvest from National Discovery Service
|
Geoportal Harvesting process LU - Luxembourg
|
Dear Jeff,
The INSPIRE Geoportal is having trouble harvesting from:
https://catalog.inspire.geoportail.lu/geonetwork/srv/eng/csw?SERVICE=CSW&VERSION=2.0.2&REQUEST=GetCapabilities
I suspect a connectivty problem.
The last time it could successfully harvest, the GetRecords endpoint was declared in the capabilities to be on port 443:
<ows:Operation name="GetRecords">
<ows:DCP>
<ows:HTTP>
<ows:Get xlink:href="https://catalog.inspire.geoportail.lu:443/geonetwork/srv/eng/csw"/>
<ows:Post xlink:href="https://catalog.inspire.geoportail.lu:443/geonetwork/srv/eng/csw"/>
</ows:HTTP>
However, now the capabilities declare port 8443, which is used by Tomcat for SSL connections
<ows:Operation name="GetRecords">
<ows:DCP>
<ows:HTTP>
<ows:Get xlink:href="https://catalog.inspire.geoportail.lu:8443/geonetwork/srv/eng/csw"/>
<ows:Post xlink:href="https://catalog.inspire.geoportail.lu:8443/geonetwork/srv/eng/csw"/>
</ows:HTTP>
</ows:DCP>
Best regards,
Angelo
|
1.0
|
LU: Unable to harvest from National Discovery Service - Dear Jeff,
The INSPIRE Geoportal is having trouble harvesting from:
https://catalog.inspire.geoportail.lu/geonetwork/srv/eng/csw?SERVICE=CSW&VERSION=2.0.2&REQUEST=GetCapabilities
I suspect a connectivty problem.
The last time it could successfully harvest, the GetRecords endpoint was declared in the capabilities to be on port 443:
<ows:Operation name="GetRecords">
<ows:DCP>
<ows:HTTP>
<ows:Get xlink:href="https://catalog.inspire.geoportail.lu:443/geonetwork/srv/eng/csw"/>
<ows:Post xlink:href="https://catalog.inspire.geoportail.lu:443/geonetwork/srv/eng/csw"/>
</ows:HTTP>
However, now the capabilities declare port 8443, which is used by Tomcat for SSL connections
<ows:Operation name="GetRecords">
<ows:DCP>
<ows:HTTP>
<ows:Get xlink:href="https://catalog.inspire.geoportail.lu:8443/geonetwork/srv/eng/csw"/>
<ows:Post xlink:href="https://catalog.inspire.geoportail.lu:8443/geonetwork/srv/eng/csw"/>
</ows:HTTP>
</ows:DCP>
Best regards,
Angelo
|
process
|
lu unable to harvest from national discovery service dear jeff the inspire geoportal is having trouble harvesting from i suspect a connectivty problem the last time it could successfully harvest the getrecords endpoint was declared in the capabilities to be on port lt ows operation name quot getrecords quot gt lt ows dcp gt lt ows http gt lt ows get xlink href quot lt ows post xlink href quot lt ows http gt however now the capabilities declare port which is used by tomcat for ssl connections lt ows operation name quot getrecords quot gt lt ows dcp gt lt ows http gt lt ows get xlink href quot lt ows post xlink href quot lt ows http gt lt ows dcp gt best regards angelo
| 1
|
17,257
| 23,039,962,444
|
IssuesEvent
|
2022-07-23 02:14:22
|
brucemiller/LaTeXML
|
https://api.github.com/repos/brucemiller/LaTeXML
|
closed
|
mathimages fails on matrix with more than 10 columns
|
bug postprocessing
|
LaTeXML 0.8.6 fails at `--mathimages` with the following test:
``` LaTeX
\documentclass{article}
\usepackage{amsmath}
\begin{document}
Default: 10 columns
\[
\begin{matrix}
1 & 2 & 3 & 4 & 5 & 6 & 7 & 8 & 9 & 10
\end{matrix},
\]
\setcounter{MaxMatrixCols}{11}
Eleven columns
\[
\begin{matrix}
1 & 2 & 3 & 4 & 5 & 6 & 7 & 8 & 9 & 10 & 11
\end{matrix},
\]
\end{document}
```
MathML works fine, so apparently the implementation doesn’t use any restriction on the column count.
In LaTeX, `amsmath` defines the counter and sets a default value:
``` TeX
\newcount\c@MaxMatrixCols \c@MaxMatrixCols=10
```
while I can’t find any place where `MaxMatrixCols` is defined or used in LaTeXML.
|
1.0
|
mathimages fails on matrix with more than 10 columns - LaTeXML 0.8.6 fails at `--mathimages` with the following test:
``` LaTeX
\documentclass{article}
\usepackage{amsmath}
\begin{document}
Default: 10 columns
\[
\begin{matrix}
1 & 2 & 3 & 4 & 5 & 6 & 7 & 8 & 9 & 10
\end{matrix},
\]
\setcounter{MaxMatrixCols}{11}
Eleven columns
\[
\begin{matrix}
1 & 2 & 3 & 4 & 5 & 6 & 7 & 8 & 9 & 10 & 11
\end{matrix},
\]
\end{document}
```
MathML works fine, so apparently the implementation doesn’t use any restriction on the column count.
In LaTeX, `amsmath` defines the counter and sets a default value:
``` TeX
\newcount\c@MaxMatrixCols \c@MaxMatrixCols=10
```
while I can’t find any place where `MaxMatrixCols` is defined or used in LaTeXML.
|
process
|
mathimages fails on matrix with more than columns latexml fails at mathimages with the following test latex documentclass article usepackage amsmath begin document default columns begin matrix end matrix setcounter maxmatrixcols eleven columns begin matrix end matrix end document mathml works fine so apparently the implementation doesn’t use any restriction on the column count in latex amsmath defines the counter and sets a default value tex newcount c maxmatrixcols c maxmatrixcols while i can’t find any place where maxmatrixcols is defined or used in latexml
| 1
|
13,527
| 10,314,046,636
|
IssuesEvent
|
2019-08-30 01:41:43
|
eclipse/vorto
|
https://api.github.com/repos/eclipse/vorto
|
opened
|
Bring up code/test coverage
|
Infrastructure
|
Current Situation: less than 50% coverage
Confirmations:
- Coverage is > 60%
- Coverage Report is available on SonarQube (Github)
- Maven build fails if coverage falls below 60%
|
1.0
|
Bring up code/test coverage - Current Situation: less than 50% coverage
Confirmations:
- Coverage is > 60%
- Coverage Report is available on SonarQube (Github)
- Maven build fails if coverage falls below 60%
|
non_process
|
bring up code test coverage current situation less than coverage confirmations coverage is coverage report is available on sonarqube github maven build fails if coverage falls below
| 0
|
3,845
| 6,808,538,318
|
IssuesEvent
|
2017-11-04 04:16:23
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
reopened
|
getBloom supports the hidden option --source and -s
|
status-inprocess tools-getBloom type-bug type-question
|
Looking at the test cases defined so far, I see that we support these options:
--source:cache
-s:node
--parity
These options are not documented at README.md. Do we keep them hidden for the moment?
|
1.0
|
getBloom supports the hidden option --source and -s - Looking at the test cases defined so far, I see that we support these options:
--source:cache
-s:node
--parity
These options are not documented at README.md. Do we keep them hidden for the moment?
|
process
|
getbloom supports the hidden option source and s looking at the test cases defined so far i see that we support these options source cache s node parity these options are not documented at readme md do we keep them hidden for the moment
| 1
|
812
| 3,287,454,855
|
IssuesEvent
|
2015-10-29 10:32:11
|
OmniLayer/omnicore
|
https://api.github.com/repos/OmniLayer/omnicore
|
opened
|
Reconsider project branch structure
|
process
|
Currently we're using the branch `omnicore-0.0.10` as only, and primary branch for this project.
It somehow makes sense at the moment, but after the release the name is no longer suitable.
Previously it was suggested to create a new branch, let's say `omnicore-0.0.11` afterwards, and continue the work there.
While this sounds somewhat reasonable, too, I'd like to reevaluate, whether this is our best option. The primary motivation, as far as I recall, was that using a new branch for each release cleanly seperates the code.
A similar effect is given with tags and releases though.
**I'd like to propose an alternative:**
1. there are two branches: `master` and `develop`
2. the `master` branch reflects "production-ready" code
3. the `develop` branch is the "integration branch" for development
4. new work is merged into the `develop` branch
5. once the `develop` work reaches a production-ready state, i.e. it's ready to be released, the `develop` branch is merged into `master`
6. finally a release is tagged
7. afterwards the work continues on the `develop` branch and the steps are repeated
A special note about Bitcoin Core: they maintain more than one branch in parallel, for once, to handle development and releases differently, and to preserve the ability to extend older version (i.e. backport changes). Neither of these properties seems to be suitable to us, and what I really dislike is that the branches can usually not be merged into each other (e.g. `0.11` won't merge into `0.10` cleanly).
What do you think?
Ping @msgilligan for further input.
|
1.0
|
Reconsider project branch structure - Currently we're using the branch `omnicore-0.0.10` as only, and primary branch for this project.
It somehow makes sense at the moment, but after the release the name is no longer suitable.
Previously it was suggested to create a new branch, let's say `omnicore-0.0.11` afterwards, and continue the work there.
While this sounds somewhat reasonable, too, I'd like to reevaluate, whether this is our best option. The primary motivation, as far as I recall, was that using a new branch for each release cleanly seperates the code.
A similar effect is given with tags and releases though.
**I'd like to propose an alternative:**
1. there are two branches: `master` and `develop`
2. the `master` branch reflects "production-ready" code
3. the `develop` branch is the "integration branch" for development
4. new work is merged into the `develop` branch
5. once the `develop` work reaches a production-ready state, i.e. it's ready to be released, the `develop` branch is merged into `master`
6. finally a release is tagged
7. afterwards the work continues on the `develop` branch and the steps are repeated
A special note about Bitcoin Core: they maintain more than one branch in parallel, for once, to handle development and releases differently, and to preserve the ability to extend older version (i.e. backport changes). Neither of these properties seems to be suitable to us, and what I really dislike is that the branches can usually not be merged into each other (e.g. `0.11` won't merge into `0.10` cleanly).
What do you think?
Ping @msgilligan for further input.
|
process
|
reconsider project branch structure currently we re using the branch omnicore as only and primary branch for this project it somehow makes sense at the moment but after the release the name is no longer suitable previously it was suggested to create a new branch let s say omnicore afterwards and continue the work there while this sounds somewhat reasonable too i d like to reevaluate whether this is our best option the primary motivation as far as i recall was that using a new branch for each release cleanly seperates the code a similar effect is given with tags and releases though i d like to propose an alternative there are two branches master and develop the master branch reflects production ready code the develop branch is the integration branch for development new work is merged into the develop branch once the develop work reaches a production ready state i e it s ready to be released the develop branch is merged into master finally a release is tagged afterwards the work continues on the develop branch and the steps are repeated a special note about bitcoin core they maintain more than one branch in parallel for once to handle development and releases differently and to preserve the ability to extend older version i e backport changes neither of these properties seems to be suitable to us and what i really dislike is that the branches can usually not be merged into each other e g won t merge into cleanly what do you think ping msgilligan for further input
| 1
|
19,692
| 26,047,037,655
|
IssuesEvent
|
2022-12-22 15:13:27
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
interesting behavior of conditional job dependencies
|
doc-enhancement devops/prod Pri2 devops-cicd-process/tech
|
Observed interesting behavior of conditional job dependencies, which I thought would be great addition to this page.
If a job does not have condition property set and has dependsOn property set to an array of jobs (fan-in scenario), some of which are conditional, it will be executed only when ALL and every parent jobs actually succeed. If some of those jobs end up "Skipped" because the condition did not satisfy, the dependant job does not run.
However, if we also include a condition property, then dependency is used only for sequencing purposes (as in completion) and the job gets to run whenever the condition is satisfied.
So the following job Z will run when all of the "parent" jobs complete (including Skipped, Cancelled, Failed, etc):
- job: Z
dependsOn:
- A
- B
condition: always()
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: d322215c-8025-4f21-0700-7dfa7dc5c46e
* Version Independent ID: 141fcdbb-8394-525b-bb29-eff9a693a9c4
* Content: [Stages in Azure Pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/stages?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/stages.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/stages.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
interesting behavior of conditional job dependencies - Observed interesting behavior of conditional job dependencies, which I thought would be great addition to this page.
If a job does not have condition property set and has dependsOn property set to an array of jobs (fan-in scenario), some of which are conditional, it will be executed only when ALL and every parent jobs actually succeed. If some of those jobs end up "Skipped" because the condition did not satisfy, the dependant job does not run.
However, if we also include a condition property, then dependency is used only for sequencing purposes (as in completion) and the job gets to run whenever the condition is satisfied.
So the following job Z will run when all of the "parent" jobs complete (including Skipped, Cancelled, Failed, etc):
- job: Z
dependsOn:
- A
- B
condition: always()
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: d322215c-8025-4f21-0700-7dfa7dc5c46e
* Version Independent ID: 141fcdbb-8394-525b-bb29-eff9a693a9c4
* Content: [Stages in Azure Pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/stages?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/stages.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/stages.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
interesting behavior of conditional job dependencies observed interesting behavior of conditional job dependencies which i thought would be great addition to this page if a job does not have condition property set and has dependson property set to an array of jobs fan in scenario some of which are conditional it will be executed only when all and every parent jobs actually succeed if some of those jobs end up skipped because the condition did not satisfy the dependant job does not run however if we also include a condition property then dependency is used only for sequencing purposes as in completion and the job gets to run whenever the condition is satisfied so the following job z will run when all of the parent jobs complete including skipped cancelled failed etc job z dependson a b condition always document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
12,626
| 15,015,989,391
|
IssuesEvent
|
2021-02-01 09:01:13
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
PM>No record displayed when loading
|
Bug P2 Participant manager Process: Fixed Process: Tested dev
|
**Describe the bug**
when the user navigates to a different tab, No record is displayed in the backend and then the result is displayed
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'PM having only a site
2. Click on Study or apps
3. No Record is displayed in the backend or when loading and then the result is displayed.
**Expected behavior**
On click on any tab, No record should not be displayed in the backend and instead display the actual result
**Screenshots**

|
2.0
|
PM>No record displayed when loading - **Describe the bug**
when the user navigates to a different tab, No record is displayed in the backend and then the result is displayed
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'PM having only a site
2. Click on Study or apps
3. No Record is displayed in the backend or when loading and then the result is displayed.
**Expected behavior**
On click on any tab, No record should not be displayed in the backend and instead display the actual result
**Screenshots**

|
process
|
pm no record displayed when loading describe the bug when the user navigates to a different tab no record is displayed in the backend and then the result is displayed to reproduce steps to reproduce the behavior go to pm having only a site click on study or apps no record is displayed in the backend or when loading and then the result is displayed expected behavior on click on any tab no record should not be displayed in the backend and instead display the actual result screenshots
| 1
|
65,530
| 19,565,572,576
|
IssuesEvent
|
2022-01-03 23:25:44
|
jMonkeyEngine/jmonkeyengine
|
https://api.github.com/repos/jMonkeyEngine/jmonkeyengine
|
closed
|
more serialization bugs in `RenderState`
|
defect
|
Discovered by code inspection while reviewing PR #1719:
```java
dfactorAlpha = ic.readEnum("dfactorRGB", BlendFunc.class, BlendFunc.One);
sfactorRGB = ic.readEnum("sfactorAlpha", BlendFunc.class, BlendFunc.One);
```
https://github.com/jMonkeyEngine/jmonkeyengine/blob/b17a2ef26bae43dfbfeb4ae66538f70b4e346d4a/jme3-core/src/main/java/com/jme3/material/RenderState.java#L559
Meanwhile the `write()` function has:
```java
oc.write(dfactorRGB, "dfactorRGB", dfactorRGB);
oc.write(sfactorAlpha, "sfactorAlpha", sfactorAlpha);
```
so at first glance one would expect serialization followed by de-serialization to interchange `dfactorAlpha` with `sfactorRGB`.
But it's worse than that, since both of the `oc.write()` invocations have `value == defVal`, which results in a no-op.
https://github.com/jMonkeyEngine/jmonkeyengine/blob/b17a2ef26bae43dfbfeb4ae66538f70b4e346d4a/jme3-core/src/plugins/java/com/jme3/export/binary/BinaryOutputCapsule.java#L976
So `read()` ends up setting the fields to `BlendFunc.One` no matter what the original setting was. And there are similar bugs for writing `sfactorRGB` and `dfactorAlpha`.
|
1.0
|
more serialization bugs in `RenderState` - Discovered by code inspection while reviewing PR #1719:
```java
dfactorAlpha = ic.readEnum("dfactorRGB", BlendFunc.class, BlendFunc.One);
sfactorRGB = ic.readEnum("sfactorAlpha", BlendFunc.class, BlendFunc.One);
```
https://github.com/jMonkeyEngine/jmonkeyengine/blob/b17a2ef26bae43dfbfeb4ae66538f70b4e346d4a/jme3-core/src/main/java/com/jme3/material/RenderState.java#L559
Meanwhile the `write()` function has:
```java
oc.write(dfactorRGB, "dfactorRGB", dfactorRGB);
oc.write(sfactorAlpha, "sfactorAlpha", sfactorAlpha);
```
so at first glance one would expect serialization followed by de-serialization to interchange `dfactorAlpha` with `sfactorRGB`.
But it's worse than that, since both of the `oc.write()` invocations have `value == defVal`, which results in a no-op.
https://github.com/jMonkeyEngine/jmonkeyengine/blob/b17a2ef26bae43dfbfeb4ae66538f70b4e346d4a/jme3-core/src/plugins/java/com/jme3/export/binary/BinaryOutputCapsule.java#L976
So `read()` ends up setting the fields to `BlendFunc.One` no matter what the original setting was. And there are similar bugs for writing `sfactorRGB` and `dfactorAlpha`.
|
non_process
|
more serialization bugs in renderstate discovered by code inspection while reviewing pr java dfactoralpha ic readenum dfactorrgb blendfunc class blendfunc one sfactorrgb ic readenum sfactoralpha blendfunc class blendfunc one meanwhile the write function has java oc write dfactorrgb dfactorrgb dfactorrgb oc write sfactoralpha sfactoralpha sfactoralpha so at first glance one would expect serialization followed by de serialization to interchange dfactoralpha with sfactorrgb but it s worse than that since both of the oc write invocations have value defval which results in a no op so read ends up setting the fields to blendfunc one no matter what the original setting was and there are similar bugs for writing sfactorrgb and dfactoralpha
| 0
|
70,244
| 3,321,590,819
|
IssuesEvent
|
2015-11-09 09:52:10
|
snaekobbi/sprints
|
https://api.github.com/repos/snaekobbi/sprints
|
closed
|
[4.3:37] The system shall support the braille code used in Denmark, Norway, Sweden and Switzerland for contracted braille.
|
0 - To do priority:1
|
### Requirement
[[4.3:37]](http://snaekobbi.github.io/requirements#4.3:37) The system shall support the braille code used in Denmark, Norway, Sweden and Switzerland for contracted braille.
## [:rewind:](https://github.com/snaekobbi/sprints/issues/123) sprint#4
### Tasks
- [ ] German (@egli)
- [ ] [[pipeline-mod-sbs#2] Integrate translation part of dtbook2sbsform.xsl](https://github.com/snaekobbi/pipeline-mod-sbs/issues/2)
- [ ] Norwegian (@usama49)
- [ ] Various [issues](https://github.com/snaekobbi/pipeline-mod-nlb/issues) (to be decided which exactly)
- [ ] execute system tests
- [ ] @mixa72
- [ ] @KariRudjord
[priority:1]: http://snaekobbi.github.io/sprints/priority_1.svg "priority:1"
[priority:2]: http://snaekobbi.github.io/sprints/priority_2.svg "priority:2"
[Ready for test]: http://snaekobbi.github.io/sprints/Ready_for_test.svg "Ready for text"
[PASS]: http://snaekobbi.github.io/sprints/PASS.svg "PASS"
[FAIL]: http://snaekobbi.github.io/sprints/FAIL.svg "FAIL"
<!---
@huboard:{"order":57.875,"milestone_order":123,"custom_state":""}
-->
|
1.0
|
[4.3:37] The system shall support the braille code used in Denmark, Norway, Sweden and Switzerland for contracted braille. - ### Requirement
[[4.3:37]](http://snaekobbi.github.io/requirements#4.3:37) The system shall support the braille code used in Denmark, Norway, Sweden and Switzerland for contracted braille.
## [:rewind:](https://github.com/snaekobbi/sprints/issues/123) sprint#4
### Tasks
- [ ] German (@egli)
- [ ] [[pipeline-mod-sbs#2] Integrate translation part of dtbook2sbsform.xsl](https://github.com/snaekobbi/pipeline-mod-sbs/issues/2)
- [ ] Norwegian (@usama49)
- [ ] Various [issues](https://github.com/snaekobbi/pipeline-mod-nlb/issues) (to be decided which exactly)
- [ ] execute system tests
- [ ] @mixa72
- [ ] @KariRudjord
[priority:1]: http://snaekobbi.github.io/sprints/priority_1.svg "priority:1"
[priority:2]: http://snaekobbi.github.io/sprints/priority_2.svg "priority:2"
[Ready for test]: http://snaekobbi.github.io/sprints/Ready_for_test.svg "Ready for text"
[PASS]: http://snaekobbi.github.io/sprints/PASS.svg "PASS"
[FAIL]: http://snaekobbi.github.io/sprints/FAIL.svg "FAIL"
<!---
@huboard:{"order":57.875,"milestone_order":123,"custom_state":""}
-->
|
non_process
|
the system shall support the braille code used in denmark norway sweden and switzerland for contracted braille requirement the system shall support the braille code used in denmark norway sweden and switzerland for contracted braille sprint tasks german egli integrate translation part of xsl norwegian various to be decided which exactly execute system tests karirudjord priority priority ready for text pass fail huboard order milestone order custom state
| 0
|
18,629
| 4,288,994,023
|
IssuesEvent
|
2016-07-17 20:34:55
|
cigalsace/mdedit
|
https://api.github.com/repos/cigalsace/mdedit
|
opened
|
Documenter l'interface utilisateur
|
documentation
|
Documenter l'interface utilisateur (à quoi serve chaque bouton et comment il fonctionne)
|
1.0
|
Documenter l'interface utilisateur - Documenter l'interface utilisateur (à quoi serve chaque bouton et comment il fonctionne)
|
non_process
|
documenter l interface utilisateur documenter l interface utilisateur à quoi serve chaque bouton et comment il fonctionne
| 0
|
6,246
| 9,204,822,240
|
IssuesEvent
|
2019-03-08 08:47:17
|
lutraconsulting/qgis-crayfish-plugin
|
https://api.github.com/repos/lutraconsulting/qgis-crayfish-plugin
|
closed
|
Rasterize timestep selection
|
critical bug processing
|
There's an issue concerning the output timestep selection of the Rasterize tool in Crayfish 3.1.0. The exported raster contains results of a different timestep than selected. Here is an exemplary dataset of BASMENT 2.8 solution files:
[example files (2dm, sol)](https://polybox.ethz.ch/index.php/s/bQ3g5d7IfD97dhH)
|
1.0
|
Rasterize timestep selection - There's an issue concerning the output timestep selection of the Rasterize tool in Crayfish 3.1.0. The exported raster contains results of a different timestep than selected. Here is an exemplary dataset of BASMENT 2.8 solution files:
[example files (2dm, sol)](https://polybox.ethz.ch/index.php/s/bQ3g5d7IfD97dhH)
|
process
|
rasterize timestep selection there s an issue concerning the output timestep selection of the rasterize tool in crayfish the exported raster contains results of a different timestep than selected here is an exemplary dataset of basment solution files
| 1
|
3,049
| 6,042,164,754
|
IssuesEvent
|
2017-06-11 10:13:57
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
Wrong column expanding row
|
bug inprocess
|
Hi there,
I am having an issue with having both expanded rows and select rows in my table. When I have the "hideSelectColumn" set to true and the "expandColumnVisible" is to true, there is an issue with the first column, of data not the expand button column, as it seems to be an event to expand the row, even if the TableHeaderColumn has the option "expandable" set to false. Below is some code to reproduce the error:
export default class ExpandRow extends React.Component {
constructor(props) {
super(props);
this.handleRowSelect = this.handleRowSelect.bind(this);
}
isExpandableRow(row) {
if (row.id < 3) return true;
else return false;
}
handleRowSelect(row, isSelected, e) {
e.preventDefault()
}
expandComponent(row) {
return (
<BSTable data={ row.expand } />
);
}
render() {
return (
<BootstrapTable data={ products }
options={{expandRowBgColor: 'rgb(242, 255, 163)', expandBy: 'column'}}
expandableRow={this.isExpandableRow}
expandComponent={this.expandComponent}
expandColumnOptions={{expandColumnVisible: true, expandColumnBeforeSelectColumn: false}}
selectRow={{ mode: 'checkbox', bgColor: 'pink', hideSelectColumn: true, clickToSelectAndEditCell: true, clickToExpand: this.handleRowSelect }}
search>
<TableHeaderColumn dataField='id' expandable={ false } isKey={ true }>Product ID</TableHeaderColumn>
<TableHeaderColumn dataField='name' expandable={ false }>Product Name</TableHeaderColumn>
<TableHeaderColumn dataField='price' expandable={ false }>Product Price</TableHeaderColumn>
</BootstrapTable>
);
}
}
Any thoughts of how to handle this issue? Much appreciated
|
1.0
|
Wrong column expanding row - Hi there,
I am having an issue with having both expanded rows and select rows in my table. When I have the "hideSelectColumn" set to true and the "expandColumnVisible" is to true, there is an issue with the first column, of data not the expand button column, as it seems to be an event to expand the row, even if the TableHeaderColumn has the option "expandable" set to false. Below is some code to reproduce the error:
export default class ExpandRow extends React.Component {
constructor(props) {
super(props);
this.handleRowSelect = this.handleRowSelect.bind(this);
}
isExpandableRow(row) {
if (row.id < 3) return true;
else return false;
}
handleRowSelect(row, isSelected, e) {
e.preventDefault()
}
expandComponent(row) {
return (
<BSTable data={ row.expand } />
);
}
render() {
return (
<BootstrapTable data={ products }
options={{expandRowBgColor: 'rgb(242, 255, 163)', expandBy: 'column'}}
expandableRow={this.isExpandableRow}
expandComponent={this.expandComponent}
expandColumnOptions={{expandColumnVisible: true, expandColumnBeforeSelectColumn: false}}
selectRow={{ mode: 'checkbox', bgColor: 'pink', hideSelectColumn: true, clickToSelectAndEditCell: true, clickToExpand: this.handleRowSelect }}
search>
<TableHeaderColumn dataField='id' expandable={ false } isKey={ true }>Product ID</TableHeaderColumn>
<TableHeaderColumn dataField='name' expandable={ false }>Product Name</TableHeaderColumn>
<TableHeaderColumn dataField='price' expandable={ false }>Product Price</TableHeaderColumn>
</BootstrapTable>
);
}
}
Any thoughts of how to handle this issue? Much appreciated
|
process
|
wrong column expanding row hi there i am having an issue with having both expanded rows and select rows in my table when i have the hideselectcolumn set to true and the expandcolumnvisible is to true there is an issue with the first column of data not the expand button column as it seems to be an event to expand the row even if the tableheadercolumn has the option expandable set to false below is some code to reproduce the error export default class expandrow extends react component constructor props super props this handlerowselect this handlerowselect bind this isexpandablerow row if row id return true else return false handlerowselect row isselected e e preventdefault expandcomponent row return render return bootstraptable data products options expandrowbgcolor rgb expandby column expandablerow this isexpandablerow expandcomponent this expandcomponent expandcolumnoptions expandcolumnvisible true expandcolumnbeforeselectcolumn false selectrow mode checkbox bgcolor pink hideselectcolumn true clicktoselectandeditcell true clicktoexpand this handlerowselect search product id product name product price any thoughts of how to handle this issue much appreciated
| 1
|
28,713
| 12,951,901,005
|
IssuesEvent
|
2020-07-19 18:30:22
|
Azure/azure-sdk-for-js
|
https://api.github.com/repos/Azure/azure-sdk-for-js
|
closed
|
MessageLockLostError :: when trying to do receiver.renewMessageLock
|
Client Service Bus customer-reported needs-author-feedback question
|
I have dead links on receivers while listening on to azure service bus subscription. While investigating the dead links, I found out, the receiver closes without retry on MessageLockLostError exception. To address that I tried the following:
`if (message.lockedUntilUtc.getTime() < new Date().getTime())
{
const newExpiryTimestamp = await receiver.renewMessageLock(message);
logger.info('lock token expired. new message lock duration is ::', newExpiryTimestamp, message.lockedUntilUtc);
}`
However, `receiver.renewMessageLock(message);` throws the same error.
`[connection-1] An error occured while auto renewing the message lock '278baf3b-8377-4305-9a2a-1be5797173a7' for message with id 'bba643a42f554dd490f295e203603224': { MessageLockLostError: The lock supplied is invalid. Either the lock expired, or the message has already been removed from the queue. Reference:531967e2-d3d4-4357-8149-7f7261912c24, TrackingId:76b38a81-dff6-f041-8ff0-0ff0a62f6819_B28, SystemTracker:###-###-dev:Topic:system.tasks|stager.tasks, Timestamp:2020-04-29T10:49:12`
I am left with no options now to do a `setInterval()`, check if `receiver.isClosed` or `receiver.isReceivingMessages()` and try and re-establish connections myself.
But this would have side effects with the sdks own timers, which am sure it has.
I really do not understand, why does the renewMessageLock throw error.
Been using service bus in Java as well as .Net. Never faced such errors.
Gone through all the relevant discussions on github and other such platforms. Would like to know what's the underlying issue.
my receivemode is 'PeekLock' by the way.
Please let me know if you need any more information.
Thanks.
|
1.0
|
MessageLockLostError :: when trying to do receiver.renewMessageLock - I have dead links on receivers while listening on to azure service bus subscription. While investigating the dead links, I found out, the receiver closes without retry on MessageLockLostError exception. To address that I tried the following:
`if (message.lockedUntilUtc.getTime() < new Date().getTime())
{
const newExpiryTimestamp = await receiver.renewMessageLock(message);
logger.info('lock token expired. new message lock duration is ::', newExpiryTimestamp, message.lockedUntilUtc);
}`
However, `receiver.renewMessageLock(message);` throws the same error.
`[connection-1] An error occured while auto renewing the message lock '278baf3b-8377-4305-9a2a-1be5797173a7' for message with id 'bba643a42f554dd490f295e203603224': { MessageLockLostError: The lock supplied is invalid. Either the lock expired, or the message has already been removed from the queue. Reference:531967e2-d3d4-4357-8149-7f7261912c24, TrackingId:76b38a81-dff6-f041-8ff0-0ff0a62f6819_B28, SystemTracker:###-###-dev:Topic:system.tasks|stager.tasks, Timestamp:2020-04-29T10:49:12`
I am left with no options now to do a `setInterval()`, check if `receiver.isClosed` or `receiver.isReceivingMessages()` and try and re-establish connections myself.
But this would have side effects with the sdks own timers, which am sure it has.
I really do not understand, why does the renewMessageLock throw error.
Been using service bus in Java as well as .Net. Never faced such errors.
Gone through all the relevant discussions on github and other such platforms. Would like to know what's the underlying issue.
my receivemode is 'PeekLock' by the way.
Please let me know if you need any more information.
Thanks.
|
non_process
|
messagelocklosterror when trying to do receiver renewmessagelock i have dead links on receivers while listening on to azure service bus subscription while investigating the dead links i found out the receiver closes without retry on messagelocklosterror exception to address that i tried the following if message lockeduntilutc gettime new date gettime const newexpirytimestamp await receiver renewmessagelock message logger info lock token expired new message lock duration is newexpirytimestamp message lockeduntilutc however receiver renewmessagelock message throws the same error an error occured while auto renewing the message lock for message with id messagelocklosterror the lock supplied is invalid either the lock expired or the message has already been removed from the queue reference trackingid systemtracker dev topic system tasks stager tasks timestamp i am left with no options now to do a setinterval check if receiver isclosed or receiver isreceivingmessages and try and re establish connections myself but this would have side effects with the sdks own timers which am sure it has i really do not understand why does the renewmessagelock throw error been using service bus in java as well as net never faced such errors gone through all the relevant discussions on github and other such platforms would like to know what s the underlying issue my receivemode is peeklock by the way please let me know if you need any more information thanks
| 0
|
22,224
| 30,772,579,920
|
IssuesEvent
|
2023-07-31 02:00:08
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Mon, 31 Jul 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
### Uncertainty-aware Unsupervised Multi-Object Tracking
- **Authors:** Kai Liu, Sheng Jin, Zhihang Fu, Ze Chen, Rongxin Jiang, Jieping Ye
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.15409
- **Pdf link:** https://arxiv.org/pdf/2307.15409
- **Abstract**
Without manually annotated identities, unsupervised multi-object trackers are inferior to learning reliable feature embeddings. It causes the similarity-based inter-frame association stage also be error-prone, where an uncertainty problem arises. The frame-by-frame accumulated uncertainty prevents trackers from learning the consistent feature embedding against time variation. To avoid this uncertainty problem, recent self-supervised techniques are adopted, whereas they failed to capture temporal relations. The interframe uncertainty still exists. In fact, this paper argues that though the uncertainty problem is inevitable, it is possible to leverage the uncertainty itself to improve the learned consistency in turn. Specifically, an uncertainty-based metric is developed to verify and rectify the risky associations. The resulting accurate pseudo-tracklets boost learning the feature consistency. And accurate tracklets can incorporate temporal information into spatial transformation. This paper proposes a tracklet-guided augmentation strategy to simulate tracklets' motion, which adopts a hierarchical uncertainty-based sampling mechanism for hard sample mining. The ultimate unsupervised MOT framework, namely U2MOT, is proven effective on MOT-Challenges and VisDrone-MOT benchmark. U2MOT achieves a SOTA performance among the published supervised and unsupervised trackers.
### Local and Global Information in Obstacle Detection on Railway Tracks
- **Authors:** Matthias Brucker, Andrei Cramariuc, Cornelius von Einem, Roland Siegwart, Cesar Cadena
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO)
- **Arxiv link:** https://arxiv.org/abs/2307.15478
- **Pdf link:** https://arxiv.org/pdf/2307.15478
- **Abstract**
Reliable obstacle detection on railways could help prevent collisions that result in injuries and potentially damage or derail the train. Unfortunately, generic object detectors do not have enough classes to account for all possible scenarios, and datasets featuring objects on railways are challenging to obtain. We propose utilizing a shallow network to learn railway segmentation from normal railway images. The limited receptive field of the network prevents overconfident predictions and allows the network to focus on the locally very distinct and repetitive patterns of the railway environment. Additionally, we explore the controlled inclusion of global information by learning to hallucinate obstacle-free images. We evaluate our method on a custom dataset featuring railway images with artificially augmented obstacles. Our proposed method outperforms other learning-based baseline methods.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### Combining transmission speckle photography and convolutional neural network for determination of fat content in cow's milk -- an exercise in classification of parameters of a complex suspension
- **Authors:** Kwasi Nyandey (1 and 2), Daniel Jakubczyk (1) ((1) Institute of Physics, Polish Academy of Sciences, Warsaw, Poland (2) Laser and Fibre Optics Centre, Department of Physics, School of Physical Sciences, College of Agriculture and Natural Sciences, University of Cape Coast, Cape Coast, Ghana)
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2307.15069
- **Pdf link:** https://arxiv.org/pdf/2307.15069
- **Abstract**
We have combined transmission speckle photography and machine learning for direct classification and recognition of milk fat content classes. Our aim was hinged on the fact that parameters of scattering particles (and the dispersion medium) can be linked to the intensity distribution (speckle) observed when coherent light is transmitted through a scattering medium. For milk, it is primarily the size distribution and concentration of fat globules, which constitutes the total fat content. Consequently, we trained convolutional neural network to recognise and classify laser speckle from different fat content classes (0.5, 1.5, 2.0 and 3.2%). We investigated four exposure-time protocols and obtained the highest performance for shorter exposure times, in which the intensity histograms are kept similar for all images and the most probable intensity in the speckle pattern is close to zero. Our neural network was able to recognize the milk fat content classes unambiguously and we obtained the highest test and independent classification accuracies of 100 and ~99% respectively. It indicates that the parameters of other complex realistic suspensions could be classified with similar methods.
### One-shot Joint Extraction, Registration and Segmentation of Neuroimaging Data
- **Authors:** Yao Su, Zhentian Qian, Lei Ma, Lifang He, Xiangnan Kong
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.15198
- **Pdf link:** https://arxiv.org/pdf/2307.15198
- **Abstract**
Brain extraction, registration and segmentation are indispensable preprocessing steps in neuroimaging studies. The aim is to extract the brain from raw imaging scans (i.e., extraction step), align it with a target brain image (i.e., registration step) and label the anatomical brain regions (i.e., segmentation step). Conventional studies typically focus on developing separate methods for the extraction, registration and segmentation tasks in a supervised setting. The performance of these methods is largely contingent on the quantity of training samples and the extent of visual inspections carried out by experts for error correction. Nevertheless, collecting voxel-level labels and performing manual quality control on high-dimensional neuroimages (e.g., 3D MRI) are expensive and time-consuming in many medical studies. In this paper, we study the problem of one-shot joint extraction, registration and segmentation in neuroimaging data, which exploits only one labeled template image (a.k.a. atlas) and a few unlabeled raw images for training. We propose a unified end-to-end framework, called JERS, to jointly optimize the extraction, registration and segmentation tasks, allowing feedback among them. Specifically, we use a group of extraction, registration and segmentation modules to learn the extraction mask, transformation and segmentation mask, where modules are interconnected and mutually reinforced by self-supervision. Empirical results on real-world datasets demonstrate that our proposed method performs exceptionally in the extraction, registration and segmentation tasks. Our code and data can be found at https://github.com/Anonymous4545/JERS
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### OAFuser: Towards Omni-Aperture Fusion for Light Field Semantic Segmentation of Road Scenes
- **Authors:** Fei Teng, Jiaming Zhang, Kunyu Peng, Kailun Yang, Yaonan Wang, Rainer Stiefelhagen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2307.15588
- **Pdf link:** https://arxiv.org/pdf/2307.15588
- **Abstract**
Light field cameras can provide rich angular and spatial information to enhance image semantic segmentation for scene understanding in the field of autonomous driving. However, the extensive angular information of light field cameras contains a large amount of redundant data, which is overwhelming for the limited hardware resource of intelligent vehicles. Besides, inappropriate compression leads to information corruption and data loss. To excavate representative information, we propose an Omni-Aperture Fusion model (OAFuser), which leverages dense context from the central view and discovers the angular information from sub-aperture images to generate a semantically-consistent result. To avoid feature loss during network propagation and simultaneously streamline the redundant information from the light field camera, we present a simple yet very effective Sub-Aperture Fusion Module (SAFM) to embed sub-aperture images into angular features without any additional memory cost. Furthermore, to address the mismatched spatial information across viewpoints, we present Center Angular Rectification Module (CARM) realized feature resorting and prevent feature occlusion caused by asymmetric information. Our proposed OAFuser achieves state-of-the-art performance on the UrbanLF-Real and -Syn datasets and sets a new record of 84.93% in mIoU on the UrbanLF-Real Extended dataset, with a gain of +4.53%. The source code of OAFuser will be made publicly available at https://github.com/FeiBryantkit/OAFuser.
## Keyword: RAW
### One-shot Joint Extraction, Registration and Segmentation of Neuroimaging Data
- **Authors:** Yao Su, Zhentian Qian, Lei Ma, Lifang He, Xiangnan Kong
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.15198
- **Pdf link:** https://arxiv.org/pdf/2307.15198
- **Abstract**
Brain extraction, registration and segmentation are indispensable preprocessing steps in neuroimaging studies. The aim is to extract the brain from raw imaging scans (i.e., extraction step), align it with a target brain image (i.e., registration step) and label the anatomical brain regions (i.e., segmentation step). Conventional studies typically focus on developing separate methods for the extraction, registration and segmentation tasks in a supervised setting. The performance of these methods is largely contingent on the quantity of training samples and the extent of visual inspections carried out by experts for error correction. Nevertheless, collecting voxel-level labels and performing manual quality control on high-dimensional neuroimages (e.g., 3D MRI) are expensive and time-consuming in many medical studies. In this paper, we study the problem of one-shot joint extraction, registration and segmentation in neuroimaging data, which exploits only one labeled template image (a.k.a. atlas) and a few unlabeled raw images for training. We propose a unified end-to-end framework, called JERS, to jointly optimize the extraction, registration and segmentation tasks, allowing feedback among them. Specifically, we use a group of extraction, registration and segmentation modules to learn the extraction mask, transformation and segmentation mask, where modules are interconnected and mutually reinforced by self-supervision. Empirical results on real-world datasets demonstrate that our proposed method performs exceptionally in the extraction, registration and segmentation tasks. Our code and data can be found at https://github.com/Anonymous4545/JERS
### Implicit neural representation for change detection
- **Authors:** Peter Naylor, Diego Di Carlo, Arianna Traviglia, Makoto Yamada, Marco Fiorucci
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2307.15428
- **Pdf link:** https://arxiv.org/pdf/2307.15428
- **Abstract**
Detecting changes that occurred in a pair of 3D airborne LiDAR point clouds, acquired at two different times over the same geographical area, is a challenging task because of unmatching spatial supports and acquisition system noise. Most recent attempts to detect changes on point clouds are based on supervised methods, which require large labelled data unavailable in real-world applications. To address these issues, we propose an unsupervised approach that comprises two components: Neural Field (NF) for continuous shape reconstruction and a Gaussian Mixture Model for categorising changes. NF offer a grid-agnostic representation to encode bi-temporal point clouds with unmatched spatial support that can be regularised to increase high-frequency details and reduce noise. The reconstructions at each timestamp are compared at arbitrary spatial scales, leading to a significant increase in detection capabilities. We apply our method to a benchmark dataset of simulated LiDAR point clouds for urban sprawling. The dataset offers different challenging scenarios with different resolutions, input modalities and noise levels, allowing a multi-scenario comparison of our method with the current state-of-the-art. We boast the previous methods on this dataset by a 10% margin in intersection over union metric. In addition, we apply our methods to a real-world scenario to identify illegal excavation (looting) of archaeological sites and confirm that they match findings from field experts.
## Keyword: raw image
### One-shot Joint Extraction, Registration and Segmentation of Neuroimaging Data
- **Authors:** Yao Su, Zhentian Qian, Lei Ma, Lifang He, Xiangnan Kong
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.15198
- **Pdf link:** https://arxiv.org/pdf/2307.15198
- **Abstract**
Brain extraction, registration and segmentation are indispensable preprocessing steps in neuroimaging studies. The aim is to extract the brain from raw imaging scans (i.e., extraction step), align it with a target brain image (i.e., registration step) and label the anatomical brain regions (i.e., segmentation step). Conventional studies typically focus on developing separate methods for the extraction, registration and segmentation tasks in a supervised setting. The performance of these methods is largely contingent on the quantity of training samples and the extent of visual inspections carried out by experts for error correction. Nevertheless, collecting voxel-level labels and performing manual quality control on high-dimensional neuroimages (e.g., 3D MRI) are expensive and time-consuming in many medical studies. In this paper, we study the problem of one-shot joint extraction, registration and segmentation in neuroimaging data, which exploits only one labeled template image (a.k.a. atlas) and a few unlabeled raw images for training. We propose a unified end-to-end framework, called JERS, to jointly optimize the extraction, registration and segmentation tasks, allowing feedback among them. Specifically, we use a group of extraction, registration and segmentation modules to learn the extraction mask, transformation and segmentation mask, where modules are interconnected and mutually reinforced by self-supervision. Empirical results on real-world datasets demonstrate that our proposed method performs exceptionally in the extraction, registration and segmentation tasks. Our code and data can be found at https://github.com/Anonymous4545/JERS
|
2.0
|
New submissions for Mon, 31 Jul 23 - ## Keyword: events
### Uncertainty-aware Unsupervised Multi-Object Tracking
- **Authors:** Kai Liu, Sheng Jin, Zhihang Fu, Ze Chen, Rongxin Jiang, Jieping Ye
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2307.15409
- **Pdf link:** https://arxiv.org/pdf/2307.15409
- **Abstract**
Without manually annotated identities, unsupervised multi-object trackers are inferior to learning reliable feature embeddings. It causes the similarity-based inter-frame association stage also be error-prone, where an uncertainty problem arises. The frame-by-frame accumulated uncertainty prevents trackers from learning the consistent feature embedding against time variation. To avoid this uncertainty problem, recent self-supervised techniques are adopted, whereas they failed to capture temporal relations. The interframe uncertainty still exists. In fact, this paper argues that though the uncertainty problem is inevitable, it is possible to leverage the uncertainty itself to improve the learned consistency in turn. Specifically, an uncertainty-based metric is developed to verify and rectify the risky associations. The resulting accurate pseudo-tracklets boost learning the feature consistency. And accurate tracklets can incorporate temporal information into spatial transformation. This paper proposes a tracklet-guided augmentation strategy to simulate tracklets' motion, which adopts a hierarchical uncertainty-based sampling mechanism for hard sample mining. The ultimate unsupervised MOT framework, namely U2MOT, is proven effective on MOT-Challenges and VisDrone-MOT benchmark. U2MOT achieves a SOTA performance among the published supervised and unsupervised trackers.
### Local and Global Information in Obstacle Detection on Railway Tracks
- **Authors:** Matthias Brucker, Andrei Cramariuc, Cornelius von Einem, Roland Siegwart, Cesar Cadena
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO)
- **Arxiv link:** https://arxiv.org/abs/2307.15478
- **Pdf link:** https://arxiv.org/pdf/2307.15478
- **Abstract**
Reliable obstacle detection on railways could help prevent collisions that result in injuries and potentially damage or derail the train. Unfortunately, generic object detectors do not have enough classes to account for all possible scenarios, and datasets featuring objects on railways are challenging to obtain. We propose utilizing a shallow network to learn railway segmentation from normal railway images. The limited receptive field of the network prevents overconfident predictions and allows the network to focus on the locally very distinct and repetitive patterns of the railway environment. Additionally, we explore the controlled inclusion of global information by learning to hallucinate obstacle-free images. We evaluate our method on a custom dataset featuring railway images with artificially augmented obstacles. Our proposed method outperforms other learning-based baseline methods.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### Combining transmission speckle photography and convolutional neural network for determination of fat content in cow's milk -- an exercise in classification of parameters of a complex suspension
- **Authors:** Kwasi Nyandey (1 and 2), Daniel Jakubczyk (1) ((1) Institute of Physics, Polish Academy of Sciences, Warsaw, Poland (2) Laser and Fibre Optics Centre, Department of Physics, School of Physical Sciences, College of Agriculture and Natural Sciences, University of Cape Coast, Cape Coast, Ghana)
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2307.15069
- **Pdf link:** https://arxiv.org/pdf/2307.15069
- **Abstract**
We have combined transmission speckle photography and machine learning for direct classification and recognition of milk fat content classes. Our aim was hinged on the fact that parameters of scattering particles (and the dispersion medium) can be linked to the intensity distribution (speckle) observed when coherent light is transmitted through a scattering medium. For milk, it is primarily the size distribution and concentration of fat globules, which constitutes the total fat content. Consequently, we trained convolutional neural network to recognise and classify laser speckle from different fat content classes (0.5, 1.5, 2.0 and 3.2%). We investigated four exposure-time protocols and obtained the highest performance for shorter exposure times, in which the intensity histograms are kept similar for all images and the most probable intensity in the speckle pattern is close to zero. Our neural network was able to recognize the milk fat content classes unambiguously and we obtained the highest test and independent classification accuracies of 100 and ~99% respectively. It indicates that the parameters of other complex realistic suspensions could be classified with similar methods.
### One-shot Joint Extraction, Registration and Segmentation of Neuroimaging Data
- **Authors:** Yao Su, Zhentian Qian, Lei Ma, Lifang He, Xiangnan Kong
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.15198
- **Pdf link:** https://arxiv.org/pdf/2307.15198
- **Abstract**
Brain extraction, registration and segmentation are indispensable preprocessing steps in neuroimaging studies. The aim is to extract the brain from raw imaging scans (i.e., extraction step), align it with a target brain image (i.e., registration step) and label the anatomical brain regions (i.e., segmentation step). Conventional studies typically focus on developing separate methods for the extraction, registration and segmentation tasks in a supervised setting. The performance of these methods is largely contingent on the quantity of training samples and the extent of visual inspections carried out by experts for error correction. Nevertheless, collecting voxel-level labels and performing manual quality control on high-dimensional neuroimages (e.g., 3D MRI) are expensive and time-consuming in many medical studies. In this paper, we study the problem of one-shot joint extraction, registration and segmentation in neuroimaging data, which exploits only one labeled template image (a.k.a. atlas) and a few unlabeled raw images for training. We propose a unified end-to-end framework, called JERS, to jointly optimize the extraction, registration and segmentation tasks, allowing feedback among them. Specifically, we use a group of extraction, registration and segmentation modules to learn the extraction mask, transformation and segmentation mask, where modules are interconnected and mutually reinforced by self-supervision. Empirical results on real-world datasets demonstrate that our proposed method performs exceptionally in the extraction, registration and segmentation tasks. Our code and data can be found at https://github.com/Anonymous4545/JERS
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### OAFuser: Towards Omni-Aperture Fusion for Light Field Semantic Segmentation of Road Scenes
- **Authors:** Fei Teng, Jiaming Zhang, Kunyu Peng, Kailun Yang, Yaonan Wang, Rainer Stiefelhagen
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2307.15588
- **Pdf link:** https://arxiv.org/pdf/2307.15588
- **Abstract**
Light field cameras can provide rich angular and spatial information to enhance image semantic segmentation for scene understanding in the field of autonomous driving. However, the extensive angular information of light field cameras contains a large amount of redundant data, which is overwhelming for the limited hardware resource of intelligent vehicles. Besides, inappropriate compression leads to information corruption and data loss. To excavate representative information, we propose an Omni-Aperture Fusion model (OAFuser), which leverages dense context from the central view and discovers the angular information from sub-aperture images to generate a semantically-consistent result. To avoid feature loss during network propagation and simultaneously streamline the redundant information from the light field camera, we present a simple yet very effective Sub-Aperture Fusion Module (SAFM) to embed sub-aperture images into angular features without any additional memory cost. Furthermore, to address the mismatched spatial information across viewpoints, we present Center Angular Rectification Module (CARM) realized feature resorting and prevent feature occlusion caused by asymmetric information. Our proposed OAFuser achieves state-of-the-art performance on the UrbanLF-Real and -Syn datasets and sets a new record of 84.93% in mIoU on the UrbanLF-Real Extended dataset, with a gain of +4.53%. The source code of OAFuser will be made publicly available at https://github.com/FeiBryantkit/OAFuser.
## Keyword: RAW
### One-shot Joint Extraction, Registration and Segmentation of Neuroimaging Data
- **Authors:** Yao Su, Zhentian Qian, Lei Ma, Lifang He, Xiangnan Kong
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.15198
- **Pdf link:** https://arxiv.org/pdf/2307.15198
- **Abstract**
Brain extraction, registration and segmentation are indispensable preprocessing steps in neuroimaging studies. The aim is to extract the brain from raw imaging scans (i.e., extraction step), align it with a target brain image (i.e., registration step) and label the anatomical brain regions (i.e., segmentation step). Conventional studies typically focus on developing separate methods for the extraction, registration and segmentation tasks in a supervised setting. The performance of these methods is largely contingent on the quantity of training samples and the extent of visual inspections carried out by experts for error correction. Nevertheless, collecting voxel-level labels and performing manual quality control on high-dimensional neuroimages (e.g., 3D MRI) are expensive and time-consuming in many medical studies. In this paper, we study the problem of one-shot joint extraction, registration and segmentation in neuroimaging data, which exploits only one labeled template image (a.k.a. atlas) and a few unlabeled raw images for training. We propose a unified end-to-end framework, called JERS, to jointly optimize the extraction, registration and segmentation tasks, allowing feedback among them. Specifically, we use a group of extraction, registration and segmentation modules to learn the extraction mask, transformation and segmentation mask, where modules are interconnected and mutually reinforced by self-supervision. Empirical results on real-world datasets demonstrate that our proposed method performs exceptionally in the extraction, registration and segmentation tasks. Our code and data can be found at https://github.com/Anonymous4545/JERS
### Implicit neural representation for change detection
- **Authors:** Peter Naylor, Diego Di Carlo, Arianna Traviglia, Makoto Yamada, Marco Fiorucci
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2307.15428
- **Pdf link:** https://arxiv.org/pdf/2307.15428
- **Abstract**
Detecting changes that occurred in a pair of 3D airborne LiDAR point clouds, acquired at two different times over the same geographical area, is a challenging task because of unmatching spatial supports and acquisition system noise. Most recent attempts to detect changes on point clouds are based on supervised methods, which require large labelled data unavailable in real-world applications. To address these issues, we propose an unsupervised approach that comprises two components: Neural Field (NF) for continuous shape reconstruction and a Gaussian Mixture Model for categorising changes. NF offer a grid-agnostic representation to encode bi-temporal point clouds with unmatched spatial support that can be regularised to increase high-frequency details and reduce noise. The reconstructions at each timestamp are compared at arbitrary spatial scales, leading to a significant increase in detection capabilities. We apply our method to a benchmark dataset of simulated LiDAR point clouds for urban sprawling. The dataset offers different challenging scenarios with different resolutions, input modalities and noise levels, allowing a multi-scenario comparison of our method with the current state-of-the-art. We boast the previous methods on this dataset by a 10% margin in intersection over union metric. In addition, we apply our methods to a real-world scenario to identify illegal excavation (looting) of archaeological sites and confirm that they match findings from field experts.
## Keyword: raw image
### One-shot Joint Extraction, Registration and Segmentation of Neuroimaging Data
- **Authors:** Yao Su, Zhentian Qian, Lei Ma, Lifang He, Xiangnan Kong
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2307.15198
- **Pdf link:** https://arxiv.org/pdf/2307.15198
- **Abstract**
Brain extraction, registration and segmentation are indispensable preprocessing steps in neuroimaging studies. The aim is to extract the brain from raw imaging scans (i.e., extraction step), align it with a target brain image (i.e., registration step) and label the anatomical brain regions (i.e., segmentation step). Conventional studies typically focus on developing separate methods for the extraction, registration and segmentation tasks in a supervised setting. The performance of these methods is largely contingent on the quantity of training samples and the extent of visual inspections carried out by experts for error correction. Nevertheless, collecting voxel-level labels and performing manual quality control on high-dimensional neuroimages (e.g., 3D MRI) are expensive and time-consuming in many medical studies. In this paper, we study the problem of one-shot joint extraction, registration and segmentation in neuroimaging data, which exploits only one labeled template image (a.k.a. atlas) and a few unlabeled raw images for training. We propose a unified end-to-end framework, called JERS, to jointly optimize the extraction, registration and segmentation tasks, allowing feedback among them. Specifically, we use a group of extraction, registration and segmentation modules to learn the extraction mask, transformation and segmentation mask, where modules are interconnected and mutually reinforced by self-supervision. Empirical results on real-world datasets demonstrate that our proposed method performs exceptionally in the extraction, registration and segmentation tasks. Our code and data can be found at https://github.com/Anonymous4545/JERS
|
process
|
new submissions for mon jul keyword events uncertainty aware unsupervised multi object tracking authors kai liu sheng jin zhihang fu ze chen rongxin jiang jieping ye subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract without manually annotated identities unsupervised multi object trackers are inferior to learning reliable feature embeddings it causes the similarity based inter frame association stage also be error prone where an uncertainty problem arises the frame by frame accumulated uncertainty prevents trackers from learning the consistent feature embedding against time variation to avoid this uncertainty problem recent self supervised techniques are adopted whereas they failed to capture temporal relations the interframe uncertainty still exists in fact this paper argues that though the uncertainty problem is inevitable it is possible to leverage the uncertainty itself to improve the learned consistency in turn specifically an uncertainty based metric is developed to verify and rectify the risky associations the resulting accurate pseudo tracklets boost learning the feature consistency and accurate tracklets can incorporate temporal information into spatial transformation this paper proposes a tracklet guided augmentation strategy to simulate tracklets motion which adopts a hierarchical uncertainty based sampling mechanism for hard sample mining the ultimate unsupervised mot framework namely is proven effective on mot challenges and visdrone mot benchmark achieves a sota performance among the published supervised and unsupervised trackers local and global information in obstacle detection on railway tracks authors matthias brucker andrei cramariuc cornelius von einem roland siegwart cesar cadena subjects computer vision and pattern recognition cs cv robotics cs ro arxiv link pdf link abstract reliable obstacle detection on railways could help prevent collisions that result in injuries and potentially damage or derail the train unfortunately generic object detectors do not have enough classes to account for all possible scenarios and datasets featuring objects on railways are challenging to obtain we propose utilizing a shallow network to learn railway segmentation from normal railway images the limited receptive field of the network prevents overconfident predictions and allows the network to focus on the locally very distinct and repetitive patterns of the railway environment additionally we explore the controlled inclusion of global information by learning to hallucinate obstacle free images we evaluate our method on a custom dataset featuring railway images with artificially augmented obstacles our proposed method outperforms other learning based baseline methods keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp combining transmission speckle photography and convolutional neural network for determination of fat content in cow s milk an exercise in classification of parameters of a complex suspension authors kwasi nyandey and daniel jakubczyk institute of physics polish academy of sciences warsaw poland laser and fibre optics centre department of physics school of physical sciences college of agriculture and natural sciences university of cape coast cape coast ghana subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract we have combined transmission speckle photography and machine learning for direct classification and recognition of milk fat content classes our aim was hinged on the fact that parameters of scattering particles and the dispersion medium can be linked to the intensity distribution speckle observed when coherent light is transmitted through a scattering medium for milk it is primarily the size distribution and concentration of fat globules which constitutes the total fat content consequently we trained convolutional neural network to recognise and classify laser speckle from different fat content classes and we investigated four exposure time protocols and obtained the highest performance for shorter exposure times in which the intensity histograms are kept similar for all images and the most probable intensity in the speckle pattern is close to zero our neural network was able to recognize the milk fat content classes unambiguously and we obtained the highest test and independent classification accuracies of and respectively it indicates that the parameters of other complex realistic suspensions could be classified with similar methods one shot joint extraction registration and segmentation of neuroimaging data authors yao su zhentian qian lei ma lifang he xiangnan kong subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract brain extraction registration and segmentation are indispensable preprocessing steps in neuroimaging studies the aim is to extract the brain from raw imaging scans i e extraction step align it with a target brain image i e registration step and label the anatomical brain regions i e segmentation step conventional studies typically focus on developing separate methods for the extraction registration and segmentation tasks in a supervised setting the performance of these methods is largely contingent on the quantity of training samples and the extent of visual inspections carried out by experts for error correction nevertheless collecting voxel level labels and performing manual quality control on high dimensional neuroimages e g mri are expensive and time consuming in many medical studies in this paper we study the problem of one shot joint extraction registration and segmentation in neuroimaging data which exploits only one labeled template image a k a atlas and a few unlabeled raw images for training we propose a unified end to end framework called jers to jointly optimize the extraction registration and segmentation tasks allowing feedback among them specifically we use a group of extraction registration and segmentation modules to learn the extraction mask transformation and segmentation mask where modules are interconnected and mutually reinforced by self supervision empirical results on real world datasets demonstrate that our proposed method performs exceptionally in the extraction registration and segmentation tasks our code and data can be found at keyword image signal processing there is no result keyword image signal process there is no result keyword compression oafuser towards omni aperture fusion for light field semantic segmentation of road scenes authors fei teng jiaming zhang kunyu peng kailun yang yaonan wang rainer stiefelhagen subjects computer vision and pattern recognition cs cv robotics cs ro image and video processing eess iv arxiv link pdf link abstract light field cameras can provide rich angular and spatial information to enhance image semantic segmentation for scene understanding in the field of autonomous driving however the extensive angular information of light field cameras contains a large amount of redundant data which is overwhelming for the limited hardware resource of intelligent vehicles besides inappropriate compression leads to information corruption and data loss to excavate representative information we propose an omni aperture fusion model oafuser which leverages dense context from the central view and discovers the angular information from sub aperture images to generate a semantically consistent result to avoid feature loss during network propagation and simultaneously streamline the redundant information from the light field camera we present a simple yet very effective sub aperture fusion module safm to embed sub aperture images into angular features without any additional memory cost furthermore to address the mismatched spatial information across viewpoints we present center angular rectification module carm realized feature resorting and prevent feature occlusion caused by asymmetric information our proposed oafuser achieves state of the art performance on the urbanlf real and syn datasets and sets a new record of in miou on the urbanlf real extended dataset with a gain of the source code of oafuser will be made publicly available at keyword raw one shot joint extraction registration and segmentation of neuroimaging data authors yao su zhentian qian lei ma lifang he xiangnan kong subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract brain extraction registration and segmentation are indispensable preprocessing steps in neuroimaging studies the aim is to extract the brain from raw imaging scans i e extraction step align it with a target brain image i e registration step and label the anatomical brain regions i e segmentation step conventional studies typically focus on developing separate methods for the extraction registration and segmentation tasks in a supervised setting the performance of these methods is largely contingent on the quantity of training samples and the extent of visual inspections carried out by experts for error correction nevertheless collecting voxel level labels and performing manual quality control on high dimensional neuroimages e g mri are expensive and time consuming in many medical studies in this paper we study the problem of one shot joint extraction registration and segmentation in neuroimaging data which exploits only one labeled template image a k a atlas and a few unlabeled raw images for training we propose a unified end to end framework called jers to jointly optimize the extraction registration and segmentation tasks allowing feedback among them specifically we use a group of extraction registration and segmentation modules to learn the extraction mask transformation and segmentation mask where modules are interconnected and mutually reinforced by self supervision empirical results on real world datasets demonstrate that our proposed method performs exceptionally in the extraction registration and segmentation tasks our code and data can be found at implicit neural representation for change detection authors peter naylor diego di carlo arianna traviglia makoto yamada marco fiorucci subjects computer vision and pattern recognition cs cv machine learning cs lg image and video processing eess iv arxiv link pdf link abstract detecting changes that occurred in a pair of airborne lidar point clouds acquired at two different times over the same geographical area is a challenging task because of unmatching spatial supports and acquisition system noise most recent attempts to detect changes on point clouds are based on supervised methods which require large labelled data unavailable in real world applications to address these issues we propose an unsupervised approach that comprises two components neural field nf for continuous shape reconstruction and a gaussian mixture model for categorising changes nf offer a grid agnostic representation to encode bi temporal point clouds with unmatched spatial support that can be regularised to increase high frequency details and reduce noise the reconstructions at each timestamp are compared at arbitrary spatial scales leading to a significant increase in detection capabilities we apply our method to a benchmark dataset of simulated lidar point clouds for urban sprawling the dataset offers different challenging scenarios with different resolutions input modalities and noise levels allowing a multi scenario comparison of our method with the current state of the art we boast the previous methods on this dataset by a margin in intersection over union metric in addition we apply our methods to a real world scenario to identify illegal excavation looting of archaeological sites and confirm that they match findings from field experts keyword raw image one shot joint extraction registration and segmentation of neuroimaging data authors yao su zhentian qian lei ma lifang he xiangnan kong subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract brain extraction registration and segmentation are indispensable preprocessing steps in neuroimaging studies the aim is to extract the brain from raw imaging scans i e extraction step align it with a target brain image i e registration step and label the anatomical brain regions i e segmentation step conventional studies typically focus on developing separate methods for the extraction registration and segmentation tasks in a supervised setting the performance of these methods is largely contingent on the quantity of training samples and the extent of visual inspections carried out by experts for error correction nevertheless collecting voxel level labels and performing manual quality control on high dimensional neuroimages e g mri are expensive and time consuming in many medical studies in this paper we study the problem of one shot joint extraction registration and segmentation in neuroimaging data which exploits only one labeled template image a k a atlas and a few unlabeled raw images for training we propose a unified end to end framework called jers to jointly optimize the extraction registration and segmentation tasks allowing feedback among them specifically we use a group of extraction registration and segmentation modules to learn the extraction mask transformation and segmentation mask where modules are interconnected and mutually reinforced by self supervision empirical results on real world datasets demonstrate that our proposed method performs exceptionally in the extraction registration and segmentation tasks our code and data can be found at
| 1
|
282,991
| 30,889,520,658
|
IssuesEvent
|
2023-08-04 02:51:02
|
maddyCode23/linux-4.1.15
|
https://api.github.com/repos/maddyCode23/linux-4.1.15
|
reopened
|
CVE-2016-6136 (Medium) detected in linux-stable-rtv4.1.33
|
Mend: dependency security vulnerability
|
## CVE-2016-6136 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/auditsc.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/auditsc.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Race condition in the audit_log_single_execve_arg function in kernel/auditsc.c in the Linux kernel through 4.7 allows local users to bypass intended character-set restrictions or disrupt system-call auditing by changing a certain string, aka a "double fetch" vulnerability.
<p>Publish Date: 2016-08-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-6136>CVE-2016-6136</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6136">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6136</a></p>
<p>Release Date: 2016-08-06</p>
<p>Fix Resolution: v4.8-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-6136 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2016-6136 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/auditsc.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/auditsc.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Race condition in the audit_log_single_execve_arg function in kernel/auditsc.c in the Linux kernel through 4.7 allows local users to bypass intended character-set restrictions or disrupt system-call auditing by changing a certain string, aka a "double fetch" vulnerability.
<p>Publish Date: 2016-08-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-6136>CVE-2016-6136</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6136">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-6136</a></p>
<p>Release Date: 2016-08-06</p>
<p>Fix Resolution: v4.8-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files kernel auditsc c kernel auditsc c vulnerability details race condition in the audit log single execve arg function in kernel auditsc c in the linux kernel through allows local users to bypass intended character set restrictions or disrupt system call auditing by changing a certain string aka a double fetch vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
524,749
| 15,222,712,361
|
IssuesEvent
|
2021-02-18 00:55:56
|
nlpsandbox/nlpsandbox
|
https://api.github.com/repos/nlpsandbox/nlpsandbox
|
closed
|
Add CONTRIBUTING.md to all repositories
|
Priority: Low
|
I'm currently working on a CONTRIBUTING.md template that we will be able to use for most repositories.
|
1.0
|
Add CONTRIBUTING.md to all repositories - I'm currently working on a CONTRIBUTING.md template that we will be able to use for most repositories.
|
non_process
|
add contributing md to all repositories i m currently working on a contributing md template that we will be able to use for most repositories
| 0
|
341,435
| 24,698,335,446
|
IssuesEvent
|
2022-10-19 13:41:22
|
pnp/pnpjs
|
https://api.github.com/repos/pnp/pnpjs
|
closed
|
Documentation link to missing registered observers doesn't work (anchor link typo)
|
type: bug status: fixed area: documentation
|
### Category
- [ ] Enhancement
- [ ] Bug
- [ ] Question
- [x] Documentation gap/issue
### Version
Please specify what version of the library you are using: [3.8.0]
Please specify what version(s) of SharePoint you are targeting: [Online]
### Expected / Desired Behavior / Question
When the issue about missing registered observer is raised in the browser console, it assumes to point to its specific article : https://pnp.github.io/pnpjs/queryable/queryable/#no-observers-registered-for-this-request
### Observed Behavior
The link mentioned in the browser console is this one: https://pnp.github.io/pnpjs/queryable/queryable/#No-observers-registered-for-this-request

There's a typo in the anchor link (it's supposed to be "no" instead of "No"). So when clicking on it, it redirects to the [@pnp/queryable/queryable](https://pnp.github.io/pnpjs/queryable/queryable/) global page instead of the dedicated part in it.
### Steps to Reproduce
* Declare a Timeline object (sp, sp-admin or graph) and try to use any endpoint without registering an observer
|
1.0
|
Documentation link to missing registered observers doesn't work (anchor link typo) - ### Category
- [ ] Enhancement
- [ ] Bug
- [ ] Question
- [x] Documentation gap/issue
### Version
Please specify what version of the library you are using: [3.8.0]
Please specify what version(s) of SharePoint you are targeting: [Online]
### Expected / Desired Behavior / Question
When the issue about missing registered observer is raised in the browser console, it assumes to point to its specific article : https://pnp.github.io/pnpjs/queryable/queryable/#no-observers-registered-for-this-request
### Observed Behavior
The link mentioned in the browser console is this one: https://pnp.github.io/pnpjs/queryable/queryable/#No-observers-registered-for-this-request

There's a typo in the anchor link (it's supposed to be "no" instead of "No"). So when clicking on it, it redirects to the [@pnp/queryable/queryable](https://pnp.github.io/pnpjs/queryable/queryable/) global page instead of the dedicated part in it.
### Steps to Reproduce
* Declare a Timeline object (sp, sp-admin or graph) and try to use any endpoint without registering an observer
|
non_process
|
documentation link to missing registered observers doesn t work anchor link typo category enhancement bug question documentation gap issue version please specify what version of the library you are using please specify what version s of sharepoint you are targeting expected desired behavior question when the issue about missing registered observer is raised in the browser console it assumes to point to its specific article observed behavior the link mentioned in the browser console is this one there s a typo in the anchor link it s supposed to be no instead of no so when clicking on it it redirects to the global page instead of the dedicated part in it steps to reproduce declare a timeline object sp sp admin or graph and try to use any endpoint without registering an observer
| 0
|
30,752
| 8,584,913,570
|
IssuesEvent
|
2018-11-14 00:48:45
|
DynamoRIO/drmemory
|
https://api.github.com/repos/DynamoRIO/drmemory
|
closed
|
cmake 2.8.12 policy CMP0022 warns about INTERFACE_LINK_LIBRARIES
|
Component-Build Migrated Priority-Medium
|
_From [priyankb...@gmail.com](https://code.google.com/u/113079830533410703845/) on March 21, 2014 18:35:33_
What steps will reproduce the problem? 1.cmake-gui .. on command line
2.configure
3.Visual Studio 10
4.Use Default Native Compiler
5.Finish What is the expected output? What do you see instead? Expected Output: Configuring done with no warning's
Result Output: Getting lots of warnings and then Configuring Done. What version of the product are you using? On what operating system? Using cmake version 2.8.12.2 using Qt 4.6.2 , Windows 8.1 x64 build Please provide any additional information below. Warning's:
CMake Warning (dev) at CMakeLists.txt:1428 (export):
Policy CMP0022 is not set: INTERFACE_LINK_LIBRARIES defines the link
interface. Run "cmake --help-policy CMP0022" for policy details. Use the
cmake_policy command to set the policy and suppress this warning.
Target "drsyscall" has an INTERFACE_LINK_LIBRARIES property which differs
from its LINK_INTERFACE_LIBRARIES properties.
INTERFACE_LINK_LIBRARIES:
```
dynamorio;drmgr;drcontainers;C:/drmemory-original/build/dynamorio/core/ntdll_imports.lib
```
LINK_INTERFACE_LIBRARIES:
Call Stack (most recent call first):
drsyscall/CMakeLists.txt:89 (export_target)
drsyscall/CMakeLists.txt:108 (export_drsyscall_target)
This warning is for project developers. Use -Wno-dev to suppress it.
CMake Warning (dev) in dynamorio/ext/drsyms/CMakeLists.txt:
Policy CMP0022 is not set: INTERFACE_LINK_LIBRARIES defines the link
interface. Run "cmake --help-policy CMP0022" for policy details. Use the
cmake_policy command to set the policy and suppress this warning.
Target "drsyms" has an INTERFACE_LINK_LIBRARIES property which differs from
its LINK_INTERFACE_LIBRARIES properties.
INTERFACE_LINK_LIBRARIES:
```
libcpmt;libcmt;dynamorio;dbghelp;C:/drmemory-original/build/dynamorio/ext/drsyms/dbghelp_imports.lib;drcontainers;C:/drmemory-original/dynamorio/ext/drsyms/libelftc-pecoff/lib32/dwarf.lib;C:/drmemory-original/dynamorio/ext/drsyms/libelftc-pecoff/lib32/elftc.lib
```
LINK_INTERFACE_LIBRARIES:
This warning is for project developers. Use -Wno-dev to suppress it.
CMake Warning (dev) at CMakeLists.txt:1428 (export):
Policy CMP0022 is not set: INTERFACE_LINK_LIBRARIES defines the link
interface. Run "cmake --help-policy CMP0022" for policy details. Use the
cmake_policy command to set the policy and suppress this warning.
Target "drsymcache" has an INTERFACE_LINK_LIBRARIES property which differs
from its LINK_INTERFACE_LIBRARIES properties.
INTERFACE_LINK_LIBRARIES:
```
dynamorio;drcontainers;drmgr;drsyms
```
LINK_INTERFACE_LIBRARIES:
Call Stack (most recent call first):
drsymcache/CMakeLists.txt:61 (export_target)
drsymcache/CMakeLists.txt:78 (export_drsymcache_target)
This warning is for project developers. Use -Wno-dev to suppress it.
CMake Warning (dev) at CMakeLists.txt:1428 (export):
Policy CMP0022 is not set: INTERFACE_LINK_LIBRARIES defines the link
interface. Run "cmake --help-policy CMP0022" for policy details. Use the
cmake_policy command to set the policy and suppress this warning.
Target "umbra" has an INTERFACE_LINK_LIBRARIES property which differs from
its LINK_INTERFACE_LIBRARIES properties.
INTERFACE_LINK_LIBRARIES:
```
dynamorio;drmgr;drcontainers;C:/drmemory-original/build/dynamorio/core/ntdll_imports.lib
```
LINK_INTERFACE_LIBRARIES:
Call Stack (most recent call first):
umbra/CMakeLists.txt:64 (export_target)
umbra/CMakeLists.txt:84 (export_umbra_target)
This warning is for project developers. Use -Wno-dev to suppress it.
_Original issue: http://code.google.com/p/drmemory/issues/detail?id=1481_
|
1.0
|
cmake 2.8.12 policy CMP0022 warns about INTERFACE_LINK_LIBRARIES - _From [priyankb...@gmail.com](https://code.google.com/u/113079830533410703845/) on March 21, 2014 18:35:33_
What steps will reproduce the problem? 1.cmake-gui .. on command line
2.configure
3.Visual Studio 10
4.Use Default Native Compiler
5.Finish What is the expected output? What do you see instead? Expected Output: Configuring done with no warning's
Result Output: Getting lots of warnings and then Configuring Done. What version of the product are you using? On what operating system? Using cmake version 2.8.12.2 using Qt 4.6.2 , Windows 8.1 x64 build Please provide any additional information below. Warning's:
CMake Warning (dev) at CMakeLists.txt:1428 (export):
Policy CMP0022 is not set: INTERFACE_LINK_LIBRARIES defines the link
interface. Run "cmake --help-policy CMP0022" for policy details. Use the
cmake_policy command to set the policy and suppress this warning.
Target "drsyscall" has an INTERFACE_LINK_LIBRARIES property which differs
from its LINK_INTERFACE_LIBRARIES properties.
INTERFACE_LINK_LIBRARIES:
```
dynamorio;drmgr;drcontainers;C:/drmemory-original/build/dynamorio/core/ntdll_imports.lib
```
LINK_INTERFACE_LIBRARIES:
Call Stack (most recent call first):
drsyscall/CMakeLists.txt:89 (export_target)
drsyscall/CMakeLists.txt:108 (export_drsyscall_target)
This warning is for project developers. Use -Wno-dev to suppress it.
CMake Warning (dev) in dynamorio/ext/drsyms/CMakeLists.txt:
Policy CMP0022 is not set: INTERFACE_LINK_LIBRARIES defines the link
interface. Run "cmake --help-policy CMP0022" for policy details. Use the
cmake_policy command to set the policy and suppress this warning.
Target "drsyms" has an INTERFACE_LINK_LIBRARIES property which differs from
its LINK_INTERFACE_LIBRARIES properties.
INTERFACE_LINK_LIBRARIES:
```
libcpmt;libcmt;dynamorio;dbghelp;C:/drmemory-original/build/dynamorio/ext/drsyms/dbghelp_imports.lib;drcontainers;C:/drmemory-original/dynamorio/ext/drsyms/libelftc-pecoff/lib32/dwarf.lib;C:/drmemory-original/dynamorio/ext/drsyms/libelftc-pecoff/lib32/elftc.lib
```
LINK_INTERFACE_LIBRARIES:
This warning is for project developers. Use -Wno-dev to suppress it.
CMake Warning (dev) at CMakeLists.txt:1428 (export):
Policy CMP0022 is not set: INTERFACE_LINK_LIBRARIES defines the link
interface. Run "cmake --help-policy CMP0022" for policy details. Use the
cmake_policy command to set the policy and suppress this warning.
Target "drsymcache" has an INTERFACE_LINK_LIBRARIES property which differs
from its LINK_INTERFACE_LIBRARIES properties.
INTERFACE_LINK_LIBRARIES:
```
dynamorio;drcontainers;drmgr;drsyms
```
LINK_INTERFACE_LIBRARIES:
Call Stack (most recent call first):
drsymcache/CMakeLists.txt:61 (export_target)
drsymcache/CMakeLists.txt:78 (export_drsymcache_target)
This warning is for project developers. Use -Wno-dev to suppress it.
CMake Warning (dev) at CMakeLists.txt:1428 (export):
Policy CMP0022 is not set: INTERFACE_LINK_LIBRARIES defines the link
interface. Run "cmake --help-policy CMP0022" for policy details. Use the
cmake_policy command to set the policy and suppress this warning.
Target "umbra" has an INTERFACE_LINK_LIBRARIES property which differs from
its LINK_INTERFACE_LIBRARIES properties.
INTERFACE_LINK_LIBRARIES:
```
dynamorio;drmgr;drcontainers;C:/drmemory-original/build/dynamorio/core/ntdll_imports.lib
```
LINK_INTERFACE_LIBRARIES:
Call Stack (most recent call first):
umbra/CMakeLists.txt:64 (export_target)
umbra/CMakeLists.txt:84 (export_umbra_target)
This warning is for project developers. Use -Wno-dev to suppress it.
_Original issue: http://code.google.com/p/drmemory/issues/detail?id=1481_
|
non_process
|
cmake policy warns about interface link libraries from on march what steps will reproduce the problem cmake gui on command line configure visual studio use default native compiler finish what is the expected output what do you see instead expected output configuring done with no warning s result output getting lots of warnings and then configuring done what version of the product are you using on what operating system using cmake version using qt windows build please provide any additional information below warning s cmake warning dev at cmakelists txt export policy is not set interface link libraries defines the link interface run cmake help policy for policy details use the cmake policy command to set the policy and suppress this warning target drsyscall has an interface link libraries property which differs from its link interface libraries properties interface link libraries dynamorio drmgr drcontainers c drmemory original build dynamorio core ntdll imports lib link interface libraries call stack most recent call first drsyscall cmakelists txt export target drsyscall cmakelists txt export drsyscall target this warning is for project developers use wno dev to suppress it cmake warning dev in dynamorio ext drsyms cmakelists txt policy is not set interface link libraries defines the link interface run cmake help policy for policy details use the cmake policy command to set the policy and suppress this warning target drsyms has an interface link libraries property which differs from its link interface libraries properties interface link libraries libcpmt libcmt dynamorio dbghelp c drmemory original build dynamorio ext drsyms dbghelp imports lib drcontainers c drmemory original dynamorio ext drsyms libelftc pecoff dwarf lib c drmemory original dynamorio ext drsyms libelftc pecoff elftc lib link interface libraries this warning is for project developers use wno dev to suppress it cmake warning dev at cmakelists txt export policy is not set interface link libraries defines the link interface run cmake help policy for policy details use the cmake policy command to set the policy and suppress this warning target drsymcache has an interface link libraries property which differs from its link interface libraries properties interface link libraries dynamorio drcontainers drmgr drsyms link interface libraries call stack most recent call first drsymcache cmakelists txt export target drsymcache cmakelists txt export drsymcache target this warning is for project developers use wno dev to suppress it cmake warning dev at cmakelists txt export policy is not set interface link libraries defines the link interface run cmake help policy for policy details use the cmake policy command to set the policy and suppress this warning target umbra has an interface link libraries property which differs from its link interface libraries properties interface link libraries dynamorio drmgr drcontainers c drmemory original build dynamorio core ntdll imports lib link interface libraries call stack most recent call first umbra cmakelists txt export target umbra cmakelists txt export umbra target this warning is for project developers use wno dev to suppress it original issue
| 0
|
21,453
| 29,489,705,482
|
IssuesEvent
|
2023-06-02 12:35:48
|
zammad/zammad
|
https://api.github.com/repos/zammad/zammad
|
closed
|
Can’t process email <ActiveRecord::RecordNotFound>
|
bug verified mail processing
|
<!--
Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓
Since november 15th we handle all requests, except real bugs, at our community board.
Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21
Please post:
- Feature requests
- Development questions
- Technical questions
on the board -> https://community.zammad.org !
If you think you hit a bug, please continue:
- Search existing issues and the CHANGELOG.md for your issue - there might be a solution already
- Make sure to use the latest version of Zammad if possible
- Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it!
- Please write the issue in english
- Don't remove the template - otherwise we will close the issue without further comments
- Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate
Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted).
* The upper textblock will be removed automatically when you submit your issue *
-->
### Infos:
* Used Zammad version: 2.8.0
* Used Zammad installation source: source
* Operating system: Debian
* Browser + version: Firefox (Win 10)
### Expected behavior:
* Tried to process an own test e-mail to Zammad which was desired to import correctly. e-mail import works just fine (e-mail is taken from IMAP), but can't be processed further actually.
### Actual behavior:
* Ticket e-mail can't be processed. Error message in scheduler-out.log says: "ERROR: Can't process email" and "ERROR: #<ActiveRecord::RecordNotFound: Couldn't find Ticket without an ID>"
* This seems to be an identical problem to the closed but (publicly unresolved) ticket [https://community.zammad.org/t/e-mails-cant-be-processed-recordnotfound/1152](https://community.zammad.org/t/e-mails-cant-be-processed-recordnotfound/1152) ! Can you tell me what the problem was that time?
Receiving e-mails worked on the vanilla Zammad setup. After (command-line) import from OTRS, e-mails can't be processed any longer in the same system. This is also true for e-mails to the native e-mail address assigned to mailbox cdo00-zammad, so filter rules are not the problem.
### Steps to reproduce the behavior:
* not working test e-mail is attached
By the way: I haven an active group named "AFP-Lookup" and an active filter for e-mail support@afp-lookup.com to be forwarded to this group:

Any e-mail to this address is forwarded to a collecting mailbox ("cdo00-zammad") for all support requests before. Zammad pulls all e-mails from that mailbox via IMAP.
Yes I'm sure this is a bug and no feature request or a general question.
(Copied from here, since it seems not to gain attention there. And error message clearly said: Open an issue on Github...)
[0da14456313e210db85b85e18db0eba8.eml.txt](https://github.com/zammad/zammad/files/2913974/0da14456313e210db85b85e18db0eba8.eml.txt)
|
1.0
|
Can’t process email <ActiveRecord::RecordNotFound> - <!--
Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓
Since november 15th we handle all requests, except real bugs, at our community board.
Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21
Please post:
- Feature requests
- Development questions
- Technical questions
on the board -> https://community.zammad.org !
If you think you hit a bug, please continue:
- Search existing issues and the CHANGELOG.md for your issue - there might be a solution already
- Make sure to use the latest version of Zammad if possible
- Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it!
- Please write the issue in english
- Don't remove the template - otherwise we will close the issue without further comments
- Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate
Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted).
* The upper textblock will be removed automatically when you submit your issue *
-->
### Infos:
* Used Zammad version: 2.8.0
* Used Zammad installation source: source
* Operating system: Debian
* Browser + version: Firefox (Win 10)
### Expected behavior:
* Tried to process an own test e-mail to Zammad which was desired to import correctly. e-mail import works just fine (e-mail is taken from IMAP), but can't be processed further actually.
### Actual behavior:
* Ticket e-mail can't be processed. Error message in scheduler-out.log says: "ERROR: Can't process email" and "ERROR: #<ActiveRecord::RecordNotFound: Couldn't find Ticket without an ID>"
* This seems to be an identical problem to the closed but (publicly unresolved) ticket [https://community.zammad.org/t/e-mails-cant-be-processed-recordnotfound/1152](https://community.zammad.org/t/e-mails-cant-be-processed-recordnotfound/1152) ! Can you tell me what the problem was that time?
Receiving e-mails worked on the vanilla Zammad setup. After (command-line) import from OTRS, e-mails can't be processed any longer in the same system. This is also true for e-mails to the native e-mail address assigned to mailbox cdo00-zammad, so filter rules are not the problem.
### Steps to reproduce the behavior:
* not working test e-mail is attached
By the way: I haven an active group named "AFP-Lookup" and an active filter for e-mail support@afp-lookup.com to be forwarded to this group:

Any e-mail to this address is forwarded to a collecting mailbox ("cdo00-zammad") for all support requests before. Zammad pulls all e-mails from that mailbox via IMAP.
Yes I'm sure this is a bug and no feature request or a general question.
(Copied from here, since it seems not to gain attention there. And error message clearly said: Open an issue on Github...)
[0da14456313e210db85b85e18db0eba8.eml.txt](https://github.com/zammad/zammad/files/2913974/0da14456313e210db85b85e18db0eba8.eml.txt)
|
process
|
can’t process email hi there thanks for filing an issue please ensure the following things before creating an issue thank you 🤓 since november we handle all requests except real bugs at our community board full explanation please post feature requests development questions technical questions on the board if you think you hit a bug please continue search existing issues and the changelog md for your issue there might be a solution already make sure to use the latest version of zammad if possible add the log production log file from your system attention make sure no confidential data is in it please write the issue in english don t remove the template otherwise we will close the issue without further comments ask questions about zammad configuration and usage at our mailinglist see note we always do our best unfortunately sometimes there are too many requests and we can t handle everything at once if you want to prioritize escalate your issue you can do so by means of a support contract see the upper textblock will be removed automatically when you submit your issue infos used zammad version used zammad installation source source operating system debian browser version firefox win expected behavior tried to process an own test e mail to zammad which was desired to import correctly e mail import works just fine e mail is taken from imap but can t be processed further actually actual behavior ticket e mail can t be processed error message in scheduler out log says error can t process email and error this seems to be an identical problem to the closed but publicly unresolved ticket can you tell me what the problem was that time receiving e mails worked on the vanilla zammad setup after command line import from otrs e mails can t be processed any longer in the same system this is also true for e mails to the native e mail address assigned to mailbox zammad so filter rules are not the problem steps to reproduce the behavior not working test e mail is attached by the way i haven an active group named afp lookup and an active filter for e mail support afp lookup com to be forwarded to this group any e mail to this address is forwarded to a collecting mailbox zammad for all support requests before zammad pulls all e mails from that mailbox via imap yes i m sure this is a bug and no feature request or a general question copied from here since it seems not to gain attention there and error message clearly said open an issue on github
| 1
|
16,785
| 21,970,912,354
|
IssuesEvent
|
2022-05-25 03:35:19
|
Tencent/tdesign-miniprogram
|
https://api.github.com/repos/Tencent/tdesign-miniprogram
|
closed
|
[Collapse] 无法通过t-class覆盖样式
|
bug processing
|
### tdesign-miniprogram 版本
0.12.0
### 重现链接
_No response_
### 重现步骤
```html
<t-collapse-panel t-class="demo-card__header" expand-icon value="{{0}}">
<t-cell title="111333" description="234234" slot="header"></t-cell>
<view slot="content">
此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容可自定义内容</view
>
</t-collapse-panel>
```
.demo-card__header {
height: 144rpx !important;
}
### 期望结果
应当可以用t-class成功覆盖?我尝试在externalClass中增加了't-collapse-panel__header',结果覆盖成功,但t-class就覆盖失败
### 实际结果
可以正常覆盖高度样式
### 框架版本
_No response_
### 浏览器版本
_No response_
### 系统版本
_No response_
### Node版本
_No response_
### 补充说明
_No response_
|
1.0
|
[Collapse] 无法通过t-class覆盖样式 - ### tdesign-miniprogram 版本
0.12.0
### 重现链接
_No response_
### 重现步骤
```html
<t-collapse-panel t-class="demo-card__header" expand-icon value="{{0}}">
<t-cell title="111333" description="234234" slot="header"></t-cell>
<view slot="content">
此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容可自定义内容</view
>
</t-collapse-panel>
```
.demo-card__header {
height: 144rpx !important;
}
### 期望结果
应当可以用t-class成功覆盖?我尝试在externalClass中增加了't-collapse-panel__header',结果覆盖成功,但t-class就覆盖失败
### 实际结果
可以正常覆盖高度样式
### 框架版本
_No response_
### 浏览器版本
_No response_
### 系统版本
_No response_
### Node版本
_No response_
### 补充说明
_No response_
|
process
|
无法通过t class覆盖样式 tdesign miniprogram 版本 重现链接 no response 重现步骤 html 此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容此处可自定义内容可自定义内容 view demo card header height important 期望结果 应当可以用t class成功覆盖?我尝试在externalclass中增加了 t collapse panel header ,结果覆盖成功,但t class就覆盖失败 实际结果 可以正常覆盖高度样式 框架版本 no response 浏览器版本 no response 系统版本 no response node版本 no response 补充说明 no response
| 1
|
2,076
| 4,891,592,205
|
IssuesEvent
|
2016-11-18 17:06:04
|
feds01/photo
|
https://api.github.com/repos/feds01/photo
|
closed
|
index data ---> table method too slow, needs re-work
|
slow-process
|
the index scan takes around 12 seconds to complete on an average size directory, but then the data transaction between table and index is far too slow; around 30 seconds
|
1.0
|
index data ---> table method too slow, needs re-work - the index scan takes around 12 seconds to complete on an average size directory, but then the data transaction between table and index is far too slow; around 30 seconds
|
process
|
index data table method too slow needs re work the index scan takes around seconds to complete on an average size directory but then the data transaction between table and index is far too slow around seconds
| 1
|
12,792
| 15,170,827,690
|
IssuesEvent
|
2021-02-13 00:27:03
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
Resize on export: can cause artifacts
|
no-issue-activity scope: image processing understood: unclear
|
<!-- IMPORTANT
Bug reports that do not make an effort to help the developers will be closed without notice.
Make sure that this bug has not already been opened and/or closed by searching the issues on GitHub, as duplicate bug reports will be closed.
A bug report simply stating that Darktable crashes is unhelpful, so please fill in most of the items below and provide detailed information.
-->
**Describe the bug**
Under these conditions I get artifacts in my output result:
* the picture contains pure, clipped white, (or is processed white enough)
* the picture needs to be resized down
* output is 8-bit format, PNG or JPEG
* "high quality resampling" is **enabled**
The artifacts will only appear on the white, clipped portions of the image.
**Expected behavior**
High quality resampling should work the same as "non"-high quality resampling. 16-bit output formats are also fine, tested PNG 16-bit and it's okay.
**Screenshots**
In the following image is the same exact image exported out to PNG 8 bit with and without HQ resize. The top left is without HQ resize, the bottom left is with HQ resize. On the right side I applied some heavy curve work (in gimp, using the exported files as source) on duplicated images to highlight the problem.

**Platform (please complete the following information):**
- Darktable Version: has been a problem for a long time. Witnessed with versions 2.6.2, 3.0.0 at least.
- OS: gentoo linux (darktable comes from portage 'master')
- opencl ENABLED
- nvidia, blob 435.21
**Additional context**
<!-- Add any other context about the problem here, for example:
- Can you reproduce with another Darktable version?
- Can you reproduce with a RAW or Jpeg or both?
- Are the steps above reproduce with a fresh edit (removing history)?
- Attach an XMP if this is necessary
- Did you compile Darktable yourself? If so which compiler was used, with what options?
- Is the issue still present using an empty/new config-dir
-->
|
1.0
|
Resize on export: can cause artifacts - <!-- IMPORTANT
Bug reports that do not make an effort to help the developers will be closed without notice.
Make sure that this bug has not already been opened and/or closed by searching the issues on GitHub, as duplicate bug reports will be closed.
A bug report simply stating that Darktable crashes is unhelpful, so please fill in most of the items below and provide detailed information.
-->
**Describe the bug**
Under these conditions I get artifacts in my output result:
* the picture contains pure, clipped white, (or is processed white enough)
* the picture needs to be resized down
* output is 8-bit format, PNG or JPEG
* "high quality resampling" is **enabled**
The artifacts will only appear on the white, clipped portions of the image.
**Expected behavior**
High quality resampling should work the same as "non"-high quality resampling. 16-bit output formats are also fine, tested PNG 16-bit and it's okay.
**Screenshots**
In the following image is the same exact image exported out to PNG 8 bit with and without HQ resize. The top left is without HQ resize, the bottom left is with HQ resize. On the right side I applied some heavy curve work (in gimp, using the exported files as source) on duplicated images to highlight the problem.

**Platform (please complete the following information):**
- Darktable Version: has been a problem for a long time. Witnessed with versions 2.6.2, 3.0.0 at least.
- OS: gentoo linux (darktable comes from portage 'master')
- opencl ENABLED
- nvidia, blob 435.21
**Additional context**
<!-- Add any other context about the problem here, for example:
- Can you reproduce with another Darktable version?
- Can you reproduce with a RAW or Jpeg or both?
- Are the steps above reproduce with a fresh edit (removing history)?
- Attach an XMP if this is necessary
- Did you compile Darktable yourself? If so which compiler was used, with what options?
- Is the issue still present using an empty/new config-dir
-->
|
process
|
resize on export can cause artifacts important bug reports that do not make an effort to help the developers will be closed without notice make sure that this bug has not already been opened and or closed by searching the issues on github as duplicate bug reports will be closed a bug report simply stating that darktable crashes is unhelpful so please fill in most of the items below and provide detailed information describe the bug under these conditions i get artifacts in my output result the picture contains pure clipped white or is processed white enough the picture needs to be resized down output is bit format png or jpeg high quality resampling is enabled the artifacts will only appear on the white clipped portions of the image expected behavior high quality resampling should work the same as non high quality resampling bit output formats are also fine tested png bit and it s okay screenshots in the following image is the same exact image exported out to png bit with and without hq resize the top left is without hq resize the bottom left is with hq resize on the right side i applied some heavy curve work in gimp using the exported files as source on duplicated images to highlight the problem platform please complete the following information darktable version has been a problem for a long time witnessed with versions at least os gentoo linux darktable comes from portage master opencl enabled nvidia blob additional context add any other context about the problem here for example can you reproduce with another darktable version can you reproduce with a raw or jpeg or both are the steps above reproduce with a fresh edit removing history attach an xmp if this is necessary did you compile darktable yourself if so which compiler was used with what options is the issue still present using an empty new config dir
| 1
|
9,121
| 12,197,759,715
|
IssuesEvent
|
2020-04-29 21:22:53
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
closed
|
Invalid indication for phast mods
|
Process Heating
|
Add error message to phast reports when modification is invalid and indicate which portion of modification is invalid
|
1.0
|
Invalid indication for phast mods - Add error message to phast reports when modification is invalid and indicate which portion of modification is invalid
|
process
|
invalid indication for phast mods add error message to phast reports when modification is invalid and indicate which portion of modification is invalid
| 1
|
18,825
| 24,724,750,548
|
IssuesEvent
|
2022-10-20 13:21:04
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
NTR inhibition of ectopic tissue mineralization
|
organism-level process
|
See https://github.com/geneontology/go-annotation/issues/4338
To annotate genes that increase levels of inorganic pyrophosphate, a mineralization inhibitor that prevents ossification/mineralization of non-bone tissue.
Parent could be: GO:0001894 tissue homeostasis
definition: A homeostatic process involved in the maintenance of non-mineral tissue, by preventing ectopic mineralization of non-bone tissue.
References:
PMID:21490328
PMID:30030150
|
1.0
|
NTR inhibition of ectopic tissue mineralization - See https://github.com/geneontology/go-annotation/issues/4338
To annotate genes that increase levels of inorganic pyrophosphate, a mineralization inhibitor that prevents ossification/mineralization of non-bone tissue.
Parent could be: GO:0001894 tissue homeostasis
definition: A homeostatic process involved in the maintenance of non-mineral tissue, by preventing ectopic mineralization of non-bone tissue.
References:
PMID:21490328
PMID:30030150
|
process
|
ntr inhibition of ectopic tissue mineralization see to annotate genes that increase levels of inorganic pyrophosphate a mineralization inhibitor that prevents ossification mineralization of non bone tissue parent could be go tissue homeostasis definition a homeostatic process involved in the maintenance of non mineral tissue by preventing ectopic mineralization of non bone tissue references pmid pmid
| 1
|
18,782
| 24,689,702,221
|
IssuesEvent
|
2022-10-19 07:36:47
|
Tencent/tdesign-miniprogram
|
https://api.github.com/repos/Tencent/tdesign-miniprogram
|
closed
|
[t-steps] 提示内存溢出
|
in process
|
### tdesign-miniprogram 版本
0.21.1
### 重现链接
https://t.wss.ink/f/9dq8e69vs97
### 重现步骤
使用t-steps组件,直接就提示内存溢出
steps-item中赋值parent,报错,提示内存溢出,自行注释掉,就不报错了
截图地址:https://t.wss.ink/f/9dq8e69vs97
使用开发者工具,没复现这种情况
### 期望结果
不报内存溢出
### 实际结果
输出: Maximum call stack size exceeded
### 框架版本
0.21.1
### 浏览器版本
谷歌浏览器 版本 103.0.5060.66(正式版本) (64 位)
### 系统版本
Windows
### Node版本
14.16.0
### 补充说明
在我项目中存在问题,单独代码片段没复现
|
1.0
|
[t-steps] 提示内存溢出 - ### tdesign-miniprogram 版本
0.21.1
### 重现链接
https://t.wss.ink/f/9dq8e69vs97
### 重现步骤
使用t-steps组件,直接就提示内存溢出
steps-item中赋值parent,报错,提示内存溢出,自行注释掉,就不报错了
截图地址:https://t.wss.ink/f/9dq8e69vs97
使用开发者工具,没复现这种情况
### 期望结果
不报内存溢出
### 实际结果
输出: Maximum call stack size exceeded
### 框架版本
0.21.1
### 浏览器版本
谷歌浏览器 版本 103.0.5060.66(正式版本) (64 位)
### 系统版本
Windows
### Node版本
14.16.0
### 补充说明
在我项目中存在问题,单独代码片段没复现
|
process
|
提示内存溢出 tdesign miniprogram 版本 重现链接 重现步骤 使用t steps组件,直接就提示内存溢出 steps item中赋值parent,报错,提示内存溢出,自行注释掉,就不报错了 截图地址: 使用开发者工具,没复现这种情况 期望结果 不报内存溢出 实际结果 输出: maximum call stack size exceeded 框架版本 浏览器版本 谷歌浏览器 版本 (正式版本) ( 位) 系统版本 windows node版本 补充说明 在我项目中存在问题,单独代码片段没复现
| 1
|
540,621
| 15,814,791,579
|
IssuesEvent
|
2021-04-05 10:03:24
|
AY2021S2-CS2113-F10-1/tp
|
https://api.github.com/repos/AY2021S2-CS2113-F10-1/tp
|
closed
|
[PE-D] Filter lease_remaining accepts all values
|
priority.High severity.High type.Bug
|
The user guide writes that the maximum lease is 99, but the user is able to input negative values or values more than 99. Although 'find' returns no flats, it may be more intuitive to the user to warn them which values are invalid.
<!--session: 1617437382415-f792c984-f41b-4a54-accb-acc47f679ba7-->
-------------
Labels: `severity.Low` `type.FeatureFlaw`
original: brynagoh/ped#2
|
1.0
|
[PE-D] Filter lease_remaining accepts all values - The user guide writes that the maximum lease is 99, but the user is able to input negative values or values more than 99. Although 'find' returns no flats, it may be more intuitive to the user to warn them which values are invalid.
<!--session: 1617437382415-f792c984-f41b-4a54-accb-acc47f679ba7-->
-------------
Labels: `severity.Low` `type.FeatureFlaw`
original: brynagoh/ped#2
|
non_process
|
filter lease remaining accepts all values the user guide writes that the maximum lease is but the user is able to input negative values or values more than although find returns no flats it may be more intuitive to the user to warn them which values are invalid labels severity low type featureflaw original brynagoh ped
| 0
|
101,524
| 12,692,214,208
|
IssuesEvent
|
2020-06-21 21:10:04
|
rubyforgood/casa
|
https://api.github.com/repos/rubyforgood/casa
|
opened
|
Volunteer shows inconsistently in search after name change
|
:clipboard: Supervisor :crown: Admin :paintbrush: Design :raised_hands: Volunteer Priority: High Type: Bug
|
**What is the problem, and what should happen instead?**
Admin user change the name of active volunteer `volunteer1@example.com` from nothing (empty) to `Kay Ulstrēd` and their name shows up attached to a case but does NOT show up in the list of volunteers (and it should)
**Screenshots of current behavior, if any**
<img width="1440" alt="Screen Shot 2020-06-21 at 2 05 50 PM" src="https://user-images.githubusercontent.com/578159/85235312-d9f89700-b3c8-11ea-90d5-53b08a6d2439.png">
|
1.0
|
Volunteer shows inconsistently in search after name change - **What is the problem, and what should happen instead?**
Admin user change the name of active volunteer `volunteer1@example.com` from nothing (empty) to `Kay Ulstrēd` and their name shows up attached to a case but does NOT show up in the list of volunteers (and it should)
**Screenshots of current behavior, if any**
<img width="1440" alt="Screen Shot 2020-06-21 at 2 05 50 PM" src="https://user-images.githubusercontent.com/578159/85235312-d9f89700-b3c8-11ea-90d5-53b08a6d2439.png">
|
non_process
|
volunteer shows inconsistently in search after name change what is the problem and what should happen instead admin user change the name of active volunteer example com from nothing empty to kay ulstrēd and their name shows up attached to a case but does not show up in the list of volunteers and it should screenshots of current behavior if any img width alt screen shot at pm src
| 0
|
37,646
| 15,352,969,380
|
IssuesEvent
|
2021-03-01 07:53:43
|
terraform-providers/terraform-provider-azurerm
|
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
|
closed
|
r/linux|windows_virtual_machine: should parse userAssignedIdentities insensitively to work around broken api
|
bug service/virtual-machine upstream-microsoft
|
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
After user managed identity was assigned to one of Azure VMs I noticed it breaking terraform and I couldn't run any action on it.
Error that I am getting is:
`Error: ID was missing the `userAssignedIdentities` element`
At first I thought that's because I didn't use `identity` block in my module. So I went to change that but error remained the same.
I tired to delete the instance from state and import it but without luck.
When deploying new VM with updated module everything was working as expected, I could add new identity, remove it altogether.
Then I checked output on [resources.azure.com](resources.azure.com) and noticed this difference in Identity block of VM instance.
Identity assigned via portal:
```json
"identity": {
"type": "UserAssigned",
"userAssignedIdentities": {
"/subscriptions/<subscriptionID>/resourceGroups/<rg_of_managed_identitiy>/providers/microsoft.managedidentity/userassignedidentities/ib-azure-metrics-exporter": {
"principalId": "<principalID>",
"clientId": "<clientID>"
}
}
}
```
Identity assigned via terraform/API:
```json
"identity": {
"type": "UserAssigned",
"userAssignedIdentities": {
"/subscriptions/<subscriptionID>/resourceGroups/<rg_of_managed_identitiy>/providers/Microsoft.ManagedIdentity/userAssignedIdentities/ib-azure-metrics-exporter": {
"principalId": "<principalID>",
"clientId": "<clientID>"
}
}
}
```
As you can see difference is in case of userAssignedIdentities and provider is doing check based on that string. Seems like another awful inconsistency from MS side but this will probably be fixed sooner.
### New or Affected Resource(s)
azurerm_linux_virtual_machine
azurerm_windows_virtual_machine
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/
--->
|
1.0
|
r/linux|windows_virtual_machine: should parse userAssignedIdentities insensitively to work around broken api - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
After user managed identity was assigned to one of Azure VMs I noticed it breaking terraform and I couldn't run any action on it.
Error that I am getting is:
`Error: ID was missing the `userAssignedIdentities` element`
At first I thought that's because I didn't use `identity` block in my module. So I went to change that but error remained the same.
I tired to delete the instance from state and import it but without luck.
When deploying new VM with updated module everything was working as expected, I could add new identity, remove it altogether.
Then I checked output on [resources.azure.com](resources.azure.com) and noticed this difference in Identity block of VM instance.
Identity assigned via portal:
```json
"identity": {
"type": "UserAssigned",
"userAssignedIdentities": {
"/subscriptions/<subscriptionID>/resourceGroups/<rg_of_managed_identitiy>/providers/microsoft.managedidentity/userassignedidentities/ib-azure-metrics-exporter": {
"principalId": "<principalID>",
"clientId": "<clientID>"
}
}
}
```
Identity assigned via terraform/API:
```json
"identity": {
"type": "UserAssigned",
"userAssignedIdentities": {
"/subscriptions/<subscriptionID>/resourceGroups/<rg_of_managed_identitiy>/providers/Microsoft.ManagedIdentity/userAssignedIdentities/ib-azure-metrics-exporter": {
"principalId": "<principalID>",
"clientId": "<clientID>"
}
}
}
```
As you can see difference is in case of userAssignedIdentities and provider is doing check based on that string. Seems like another awful inconsistency from MS side but this will probably be fixed sooner.
### New or Affected Resource(s)
azurerm_linux_virtual_machine
azurerm_windows_virtual_machine
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/
--->
|
non_process
|
r linux windows virtual machine should parse userassignedidentities insensitively to work around broken api community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description after user managed identity was assigned to one of azure vms i noticed it breaking terraform and i couldn t run any action on it error that i am getting is error id was missing the userassignedidentities element at first i thought that s because i didn t use identity block in my module so i went to change that but error remained the same i tired to delete the instance from state and import it but without luck when deploying new vm with updated module everything was working as expected i could add new identity remove it altogether then i checked output on resources azure com and noticed this difference in identity block of vm instance identity assigned via portal json identity type userassigned userassignedidentities subscriptions resourcegroups providers microsoft managedidentity userassignedidentities ib azure metrics exporter principalid clientid identity assigned via terraform api json identity type userassigned userassignedidentities subscriptions resourcegroups providers microsoft managedidentity userassignedidentities ib azure metrics exporter principalid clientid as you can see difference is in case of userassignedidentities and provider is doing check based on that string seems like another awful inconsistency from ms side but this will probably be fixed sooner new or affected resource s azurerm linux virtual machine azurerm windows virtual machine references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example
| 0
|
55,262
| 6,461,053,810
|
IssuesEvent
|
2017-08-16 06:57:54
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
reopened
|
Test: System.Net.Http.Functional.Tests.HttpClientHandlerTest/AllowAutoRedirect_True_ValidateNewMethodUsedOnRedirection(statusCode: 302, oldMethod: \"HEAD\", newMethod: \"HEAD\") failed with "Xunit.Sdk.FalseException"
|
area-System.Net.Http test-run-core
|
Opened on behalf of @Jiayili1
The test `System.Net.Http.Functional.Tests.HttpClientHandlerTest/AllowAutoRedirect_True_ValidateNewMethodUsedOnRedirection(statusCode: 302, oldMethod: \"HEAD\", newMethod: \"HEAD\")` has failed.
Faulted: System.AggregateException: One or more errors occurred. (An error occurred while sending the request.) ---> System.Net.Http.HttpRequestException: An error occurred while sending the request. ---> System.Net.Http.WinHttpException: The connection with the server was terminated abnormally\r
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()\r
at System.Threading.Tasks.RendezvousAwaitable`1.GetResult()\r
at System.Net.Http.WinHttpHandler.<StartRequest>d__105.MoveNext()\r
--- End of inner exception stack trace ---\r
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()\r
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)\r
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)\r
at System.Net.Http.HttpClient.<FinishSendAsyncBuffered>d__58.MoveNext()\r
--- End of inner exception stack trace ---\r
---> (Inner Exception #0) System.Net.Http.HttpRequestException: An error occurred while sending the request. ---> System.Net.Http.WinHttpException: The connection with the server was terminated abnormally\r
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()\r
at System.Threading.Tasks.RendezvousAwaitable`1.GetResult()\r
at System.Net.Http.WinHttpHandler.<StartRequest>d__105.MoveNext()\r
--- End of inner exception stack trace ---\r
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()\r
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)\r
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)\r
at System.Net.Http.HttpClient.<FinishSendAsyncBuffered>d__58.MoveNext()<---\r
\r
Expected: False\r
Actual: True
Stack Trace:
at System.Net.Http.Functional.Tests.HttpClientHandlerTest.<>c__DisplayClass46_1.<<AllowAutoRedirect_True_ValidateNewMethodUsedOnRedirection>b__0>d.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Net.Test.Common.LoopbackServer.<>c__DisplayClass3_0.<CreateServerAsync>b__0(Task t)
at System.Threading.Tasks.ContinuationTaskFromTask.InnerInvoke()
at System.Threading.Tasks.Task.<>c.<.cctor>b__279_1(Object obj)
at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
at System.Threading.Tasks.Task.ExecuteWithThreadLocal(Task& currentTaskSlot)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Net.Http.Functional.Tests.HttpClientHandlerTest.<AllowAutoRedirect_True_ValidateNewMethodUsedOnRedirection>d__46.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
Build : Master - 20170814.01 (Core Tests)
Failing configurations:
- Windows.10.Amd64-x86
- Release
Detail: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20170814.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FAllowAutoRedirect_True_ValidateNewMethodUsedOnRedirection(statusCode:%20302,%20oldMethod:%20%5C%22HEAD%5C%22,%20newMethod:%20%5C%22HEAD%5C%22)
|
1.0
|
Test: System.Net.Http.Functional.Tests.HttpClientHandlerTest/AllowAutoRedirect_True_ValidateNewMethodUsedOnRedirection(statusCode: 302, oldMethod: \"HEAD\", newMethod: \"HEAD\") failed with "Xunit.Sdk.FalseException" - Opened on behalf of @Jiayili1
The test `System.Net.Http.Functional.Tests.HttpClientHandlerTest/AllowAutoRedirect_True_ValidateNewMethodUsedOnRedirection(statusCode: 302, oldMethod: \"HEAD\", newMethod: \"HEAD\")` has failed.
Faulted: System.AggregateException: One or more errors occurred. (An error occurred while sending the request.) ---> System.Net.Http.HttpRequestException: An error occurred while sending the request. ---> System.Net.Http.WinHttpException: The connection with the server was terminated abnormally\r
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()\r
at System.Threading.Tasks.RendezvousAwaitable`1.GetResult()\r
at System.Net.Http.WinHttpHandler.<StartRequest>d__105.MoveNext()\r
--- End of inner exception stack trace ---\r
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()\r
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)\r
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)\r
at System.Net.Http.HttpClient.<FinishSendAsyncBuffered>d__58.MoveNext()\r
--- End of inner exception stack trace ---\r
---> (Inner Exception #0) System.Net.Http.HttpRequestException: An error occurred while sending the request. ---> System.Net.Http.WinHttpException: The connection with the server was terminated abnormally\r
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()\r
at System.Threading.Tasks.RendezvousAwaitable`1.GetResult()\r
at System.Net.Http.WinHttpHandler.<StartRequest>d__105.MoveNext()\r
--- End of inner exception stack trace ---\r
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()\r
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)\r
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)\r
at System.Net.Http.HttpClient.<FinishSendAsyncBuffered>d__58.MoveNext()<---\r
\r
Expected: False\r
Actual: True
Stack Trace:
at System.Net.Http.Functional.Tests.HttpClientHandlerTest.<>c__DisplayClass46_1.<<AllowAutoRedirect_True_ValidateNewMethodUsedOnRedirection>b__0>d.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Net.Test.Common.LoopbackServer.<>c__DisplayClass3_0.<CreateServerAsync>b__0(Task t)
at System.Threading.Tasks.ContinuationTaskFromTask.InnerInvoke()
at System.Threading.Tasks.Task.<>c.<.cctor>b__279_1(Object obj)
at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
at System.Threading.Tasks.Task.ExecuteWithThreadLocal(Task& currentTaskSlot)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Net.Http.Functional.Tests.HttpClientHandlerTest.<AllowAutoRedirect_True_ValidateNewMethodUsedOnRedirection>d__46.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
Build : Master - 20170814.01 (Core Tests)
Failing configurations:
- Windows.10.Amd64-x86
- Release
Detail: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20170814.01/workItem/System.Net.Http.Functional.Tests/analysis/xunit/System.Net.Http.Functional.Tests.HttpClientHandlerTest~2FAllowAutoRedirect_True_ValidateNewMethodUsedOnRedirection(statusCode:%20302,%20oldMethod:%20%5C%22HEAD%5C%22,%20newMethod:%20%5C%22HEAD%5C%22)
|
non_process
|
test system net http functional tests httpclienthandlertest allowautoredirect true validatenewmethodusedonredirection statuscode oldmethod head newmethod head failed with xunit sdk falseexception opened on behalf of the test system net http functional tests httpclienthandlertest allowautoredirect true validatenewmethodusedonredirection statuscode oldmethod head newmethod head has failed faulted system aggregateexception one or more errors occurred an error occurred while sending the request system net http httprequestexception an error occurred while sending the request system net http winhttpexception the connection with the server was terminated abnormally r at system runtime exceptionservices exceptiondispatchinfo throw r at system threading tasks rendezvousawaitable getresult r at system net http winhttphandler d movenext r end of inner exception stack trace r at system runtime exceptionservices exceptiondispatchinfo throw r at system runtime compilerservices taskawaiter throwfornonsuccess task task r at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task r at system net http httpclient d movenext r end of inner exception stack trace r inner exception system net http httprequestexception an error occurred while sending the request system net http winhttpexception the connection with the server was terminated abnormally r at system runtime exceptionservices exceptiondispatchinfo throw r at system threading tasks rendezvousawaitable getresult r at system net http winhttphandler d movenext r end of inner exception stack trace r at system runtime exceptionservices exceptiondispatchinfo throw r at system runtime compilerservices taskawaiter throwfornonsuccess task task r at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task r at system net http httpclient d movenext r r expected false r actual true stack trace at system net http functional tests httpclienthandlertest c b d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system net test common loopbackserver c b task t at system threading tasks continuationtaskfromtask innerinvoke at system threading tasks task c b object obj at system threading executioncontext run executioncontext executioncontext contextcallback callback object state at system threading tasks task executewiththreadlocal task currenttaskslot end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system net http functional tests httpclienthandlertest d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task build master core tests failing configurations windows release detail
| 0
|
10,744
| 13,540,426,047
|
IssuesEvent
|
2020-09-16 14:39:02
|
prisma/prisma-engines
|
https://api.github.com/repos/prisma/prisma-engines
|
opened
|
Datamodel validator capability to prevent a foreign key on pointing to a nullable field
|
process/candidate topic: SQL Server
|
SQL Server doesn't allow pointing a foreign key to a field that can be null.
Example:
```prisma
model ModelA {
id String @id @default(uuid())
u String? @unique
bs ModelB[]
}
model ModelB {
id String @id @default(uuid())
a_u String?
a ModelA? @relation(fields: [a_u], references: [u])
}
```
Here the field `u` in `ModelA` can be null, meaning the relation from `ModelB` will error out in SQL Server:
```
There are no primary or candidate keys in the referenced table 'ModelA' that match the referencing column list in the foreign key 'FK__ModelB__a_u__286302EC
```
This means due to the field being nullable, a normal unique constraint would not allow more than one null value per field. To go around that, we handle the uniqueness with an index that filters out null values, allowing multiple nulls to be stored. This kind of index is not possible in a foreign key, so to make it a bit easier for our users, we should have a new capability for all the other connectors to allow this, but leave SQL Server out. A nice validation error would be awesome too!
|
1.0
|
Datamodel validator capability to prevent a foreign key on pointing to a nullable field - SQL Server doesn't allow pointing a foreign key to a field that can be null.
Example:
```prisma
model ModelA {
id String @id @default(uuid())
u String? @unique
bs ModelB[]
}
model ModelB {
id String @id @default(uuid())
a_u String?
a ModelA? @relation(fields: [a_u], references: [u])
}
```
Here the field `u` in `ModelA` can be null, meaning the relation from `ModelB` will error out in SQL Server:
```
There are no primary or candidate keys in the referenced table 'ModelA' that match the referencing column list in the foreign key 'FK__ModelB__a_u__286302EC
```
This means due to the field being nullable, a normal unique constraint would not allow more than one null value per field. To go around that, we handle the uniqueness with an index that filters out null values, allowing multiple nulls to be stored. This kind of index is not possible in a foreign key, so to make it a bit easier for our users, we should have a new capability for all the other connectors to allow this, but leave SQL Server out. A nice validation error would be awesome too!
|
process
|
datamodel validator capability to prevent a foreign key on pointing to a nullable field sql server doesn t allow pointing a foreign key to a field that can be null example prisma model modela id string id default uuid u string unique bs modelb model modelb id string id default uuid a u string a modela relation fields references here the field u in modela can be null meaning the relation from modelb will error out in sql server there are no primary or candidate keys in the referenced table modela that match the referencing column list in the foreign key fk modelb a u this means due to the field being nullable a normal unique constraint would not allow more than one null value per field to go around that we handle the uniqueness with an index that filters out null values allowing multiple nulls to be stored this kind of index is not possible in a foreign key so to make it a bit easier for our users we should have a new capability for all the other connectors to allow this but leave sql server out a nice validation error would be awesome too
| 1
|
20,065
| 26,555,150,934
|
IssuesEvent
|
2023-01-20 11:19:50
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
--action_env ignored when `cfg = "exec"` is used
|
type: support / not a bug (process) team-Rules-CPP untriaged
|
### Description of the bug:
We need to pass some environment variables like "`$CPATH`" to the compiler when building TensorFlow with Bazel. This is cumbersome itself and has led to hard-to-debug issues like https://github.com/bazelbuild/bazel/issues/12059 in the past already.
Now we again see failures caused by action-env values not passed to the compiler invocation in TensorFlow 2.8.4 which I tracked down to https://github.com/tensorflow/tensorflow/commit/07cbc7bb0bf899aac2bee5e21e1ba4eb40038682 which changes `cfg = "host"` to `cfg = "exec"`
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
Build TensorFlow 2.8.4 with Bazel 4.2.2 passing `--action-env=CPATH` and observe that it is not passed to some compiler invocations resulting in e.g.:
> ```
> In file included from bazel-out/k8-opt-exec-50AE0418/bin/tensorflow/core/framework/dataset_options.pb.cc:4:
> bazel-out/k8-opt-exec-50AE0418/bin/tensorflow/core/framework/dataset_options.pb.h:10:10: fatal error: google/protobuf/port_def.inc: No such file or directory
> 10 | #include <google/protobuf/port_def.inc>
> ```
So can you provide information on how to use `--action-env` (or similar) in such circumstances?
An explanation on what is actually being done with the change to "exec" from "host" would also be very welcome. In our case we are not cross-compiling so host, target and build machine are all the same.
I would clearly classify this behavior as a bug because the [documentation](https://docs.bazel.build/versions/4.2.2/user-manual.html#flag--action_env) states:
> Specifies the set of environment variables available during the execution of all actions.
But obviously there are now actions where those are missing but they should be in "all actions"
### Which operating system are you running Bazel on?
REHL 7
|
1.0
|
--action_env ignored when `cfg = "exec"` is used - ### Description of the bug:
We need to pass some environment variables like "`$CPATH`" to the compiler when building TensorFlow with Bazel. This is cumbersome itself and has led to hard-to-debug issues like https://github.com/bazelbuild/bazel/issues/12059 in the past already.
Now we again see failures caused by action-env values not passed to the compiler invocation in TensorFlow 2.8.4 which I tracked down to https://github.com/tensorflow/tensorflow/commit/07cbc7bb0bf899aac2bee5e21e1ba4eb40038682 which changes `cfg = "host"` to `cfg = "exec"`
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
Build TensorFlow 2.8.4 with Bazel 4.2.2 passing `--action-env=CPATH` and observe that it is not passed to some compiler invocations resulting in e.g.:
> ```
> In file included from bazel-out/k8-opt-exec-50AE0418/bin/tensorflow/core/framework/dataset_options.pb.cc:4:
> bazel-out/k8-opt-exec-50AE0418/bin/tensorflow/core/framework/dataset_options.pb.h:10:10: fatal error: google/protobuf/port_def.inc: No such file or directory
> 10 | #include <google/protobuf/port_def.inc>
> ```
So can you provide information on how to use `--action-env` (or similar) in such circumstances?
An explanation on what is actually being done with the change to "exec" from "host" would also be very welcome. In our case we are not cross-compiling so host, target and build machine are all the same.
I would clearly classify this behavior as a bug because the [documentation](https://docs.bazel.build/versions/4.2.2/user-manual.html#flag--action_env) states:
> Specifies the set of environment variables available during the execution of all actions.
But obviously there are now actions where those are missing but they should be in "all actions"
### Which operating system are you running Bazel on?
REHL 7
|
process
|
action env ignored when cfg exec is used description of the bug we need to pass some environment variables like cpath to the compiler when building tensorflow with bazel this is cumbersome itself and has led to hard to debug issues like in the past already now we again see failures caused by action env values not passed to the compiler invocation in tensorflow which i tracked down to which changes cfg host to cfg exec what s the simplest easiest way to reproduce this bug please provide a minimal example if possible build tensorflow with bazel passing action env cpath and observe that it is not passed to some compiler invocations resulting in e g in file included from bazel out opt exec bin tensorflow core framework dataset options pb cc bazel out opt exec bin tensorflow core framework dataset options pb h fatal error google protobuf port def inc no such file or directory include so can you provide information on how to use action env or similar in such circumstances an explanation on what is actually being done with the change to exec from host would also be very welcome in our case we are not cross compiling so host target and build machine are all the same i would clearly classify this behavior as a bug because the states specifies the set of environment variables available during the execution of all actions but obviously there are now actions where those are missing but they should be in all actions which operating system are you running bazel on rehl
| 1
|
184,934
| 6,717,542,572
|
IssuesEvent
|
2017-10-14 22:39:37
|
ChalkyBrush/roshpit-bug-tracker
|
https://api.github.com/repos/ChalkyBrush/roshpit-bug-tracker
|
closed
|
non-English curated items on roshpit.ca.
|
bug: site enhancement priority: low
|
Curating items with non-English account leads to curated item page broke. (or even being unaviable)
Eye of Avernus

Ruinfall Skull Token

Magebane (they have now only 2 of each property variance curated)

Twilight

etc etc etc.
|
1.0
|
non-English curated items on roshpit.ca. - Curating items with non-English account leads to curated item page broke. (or even being unaviable)
Eye of Avernus

Ruinfall Skull Token

Magebane (they have now only 2 of each property variance curated)

Twilight

etc etc etc.
|
non_process
|
non english curated items on roshpit ca curating items with non english account leads to curated item page broke or even being unaviable eye of avernus ruinfall skull token magebane they have now only of each property variance curated twilight etc etc etc
| 0
|
143,832
| 22,162,585,447
|
IssuesEvent
|
2022-06-04 18:25:10
|
Facepunch/sbox-issues
|
https://api.github.com/repos/Facepunch/sbox-issues
|
opened
|
Logging.OnMessage shouldn't be internal
|
api design
|
### What it is?
Logging.OnMessage is internal, which is needlessly restrictive.
Access to the event when a new message is added allows you to write custom error handling code (such as also logging messages elsewhere, such as files or a database, reporting exceptions in your game-mode, et cetera).
### What should it be?
Logging.OnMessage should just lose the `internal` qualifier.
|
1.0
|
Logging.OnMessage shouldn't be internal - ### What it is?
Logging.OnMessage is internal, which is needlessly restrictive.
Access to the event when a new message is added allows you to write custom error handling code (such as also logging messages elsewhere, such as files or a database, reporting exceptions in your game-mode, et cetera).
### What should it be?
Logging.OnMessage should just lose the `internal` qualifier.
|
non_process
|
logging onmessage shouldn t be internal what it is logging onmessage is internal which is needlessly restrictive access to the event when a new message is added allows you to write custom error handling code such as also logging messages elsewhere such as files or a database reporting exceptions in your game mode et cetera what should it be logging onmessage should just lose the internal qualifier
| 0
|
1,161
| 3,644,118,731
|
IssuesEvent
|
2016-02-15 08:14:38
|
hbz/lobid-resources
|
https://api.github.com/repos/hbz/lobid-resources
|
closed
|
Set up automatic weekly indexing on gaia
|
processing review
|
For a start , the data should be indexed completely once a week. Target is gaia.
|
1.0
|
Set up automatic weekly indexing on gaia - For a start , the data should be indexed completely once a week. Target is gaia.
|
process
|
set up automatic weekly indexing on gaia for a start the data should be indexed completely once a week target is gaia
| 1
|
803,343
| 29,173,512,901
|
IssuesEvent
|
2023-05-19 05:32:31
|
yugabyte/yugabyte-db
|
https://api.github.com/repos/yugabyte/yugabyte-db
|
opened
|
[CDCSDK] DBZ: Add validation for consistent streaming to be run with EXPLICIT checkpoint only
|
kind/new-feature priority/low area/cdcsdk jira-originated
|
Jira Link: [DB-6610](https://yugabyte.atlassian.net/browse/DB-6610)
|
1.0
|
[CDCSDK] DBZ: Add validation for consistent streaming to be run with EXPLICIT checkpoint only - Jira Link: [DB-6610](https://yugabyte.atlassian.net/browse/DB-6610)
|
non_process
|
dbz add validation for consistent streaming to be run with explicit checkpoint only jira link
| 0
|
269,343
| 23,438,843,390
|
IssuesEvent
|
2022-08-15 13:02:05
|
status-im/status-desktop
|
https://api.github.com/repos/status-im/status-desktop
|
closed
|
The Profile Name field is jumping while typing a text there
|
onboarding tested pixel-perfect-issues
|
# Bug Report
## Description
## Steps to reproduce
1. Go to Sign up a new account
2. On the onboarding 'Your Profile' page start to type some text (more than 4 characters) in the field 'Display name'
#### Expected behavior
After typing the first 5 characters the field 'Display Name' is **not** jumping up because the error notification disappeared
#### Actual behavior
After typing the first 5 characters the field 'Display Name' is jumping up because of the error notification disappeared
### Additional Information
https://user-images.githubusercontent.com/14942081/181407635-d68f6328-b3f3-4dfe-97bd-028d2bc07f69.mov
- Status desktop version: https://ci.status.im/job/status-desktop/job/master/15/
- Operating System: macOS Monterey 12.3 Beta (21E5227a)
|
1.0
|
The Profile Name field is jumping while typing a text there - # Bug Report
## Description
## Steps to reproduce
1. Go to Sign up a new account
2. On the onboarding 'Your Profile' page start to type some text (more than 4 characters) in the field 'Display name'
#### Expected behavior
After typing the first 5 characters the field 'Display Name' is **not** jumping up because the error notification disappeared
#### Actual behavior
After typing the first 5 characters the field 'Display Name' is jumping up because of the error notification disappeared
### Additional Information
https://user-images.githubusercontent.com/14942081/181407635-d68f6328-b3f3-4dfe-97bd-028d2bc07f69.mov
- Status desktop version: https://ci.status.im/job/status-desktop/job/master/15/
- Operating System: macOS Monterey 12.3 Beta (21E5227a)
|
non_process
|
the profile name field is jumping while typing a text there bug report description steps to reproduce go to sign up a new account on the onboarding your profile page start to type some text more than characters in the field display name expected behavior after typing the first characters the field display name is not jumping up because the error notification disappeared actual behavior after typing the first characters the field display name is jumping up because of the error notification disappeared additional information status desktop version operating system macos monterey beta
| 0
|
344,213
| 24,802,138,826
|
IssuesEvent
|
2022-10-24 23:00:36
|
InstituteforDiseaseModeling/idmtools
|
https://api.github.com/repos/InstituteforDiseaseModeling/idmtools
|
closed
|
Creating a custom builder
|
Documentation Exclude from Changelog
|
Users can create their own Simulation Builders. We should document the API and how you can do this through an example/docs
|
1.0
|
Creating a custom builder - Users can create their own Simulation Builders. We should document the API and how you can do this through an example/docs
|
non_process
|
creating a custom builder users can create their own simulation builders we should document the api and how you can do this through an example docs
| 0
|
43,200
| 11,182,362,809
|
IssuesEvent
|
2019-12-31 08:23:23
|
cirosantilli/linux-kernel-module-cheat
|
https://api.github.com/repos/cirosantilli/linux-kernel-module-cheat
|
closed
|
ifup -a on aarch64 fails with segmentation fault in v3.0 (buildroot 2018.08, busybox, glibc)
|
bug buildroot fixed-upstream
|
I'm guessing this is caused by the recent switch to glibc from uclibc.
Cannot reproduce on pure upstream buildroot master 9152387703707bd1c8c49e7978ef47aec3f8baeb (post 2019.02) on Ubuntu 18.10 host:
make qemu_aarch64_virt_defconfig
make menuconfig
make BR2_JLEVEL="$(nproc)"
qemu-system-aarch64 \
-M virt \
-append "root=/dev/vda console=ttyAMA0" \
-cpu cortex-a57 \
-device virtio-blk-device,drive=hd0 \
-device virtio-net-device,netdev=eth0 \
-drive file=output/images/rootfs.ext4,if=none,format=raw,id=hd0 \
-kernel output/images/Image \
-netdev user,id=eth0 \
-nographic \
-smp 1
Edit: reproduce problem on tag: 2018.08 I can't find with `git log --grep` the commit that fixed it easily however, so I'll just update buildroot when new tag comes out.
|
1.0
|
ifup -a on aarch64 fails with segmentation fault in v3.0 (buildroot 2018.08, busybox, glibc) - I'm guessing this is caused by the recent switch to glibc from uclibc.
Cannot reproduce on pure upstream buildroot master 9152387703707bd1c8c49e7978ef47aec3f8baeb (post 2019.02) on Ubuntu 18.10 host:
make qemu_aarch64_virt_defconfig
make menuconfig
make BR2_JLEVEL="$(nproc)"
qemu-system-aarch64 \
-M virt \
-append "root=/dev/vda console=ttyAMA0" \
-cpu cortex-a57 \
-device virtio-blk-device,drive=hd0 \
-device virtio-net-device,netdev=eth0 \
-drive file=output/images/rootfs.ext4,if=none,format=raw,id=hd0 \
-kernel output/images/Image \
-netdev user,id=eth0 \
-nographic \
-smp 1
Edit: reproduce problem on tag: 2018.08 I can't find with `git log --grep` the commit that fixed it easily however, so I'll just update buildroot when new tag comes out.
|
non_process
|
ifup a on fails with segmentation fault in buildroot busybox glibc i m guessing this is caused by the recent switch to glibc from uclibc cannot reproduce on pure upstream buildroot master post on ubuntu host make qemu virt defconfig make menuconfig make jlevel nproc qemu system m virt append root dev vda console cpu cortex device virtio blk device drive device virtio net device netdev drive file output images rootfs if none format raw id kernel output images image netdev user id nographic smp edit reproduce problem on tag i can t find with git log grep the commit that fixed it easily however so i ll just update buildroot when new tag comes out
| 0
|
256,451
| 19,423,397,881
|
IssuesEvent
|
2021-12-21 00:10:14
|
microsoft/DirectXTK12
|
https://api.github.com/repos/microsoft/DirectXTK12
|
closed
|
Wiki Page - The Basic Game Loop - Link 404 ref
|
documentation
|
Wiki Page
https://github.com/microsoft/DirectXTK12/wiki/The-basic-game-loop
Setup Section, highlighted text ` VS 2017/2019/2022`
Points to:
https://github.com/walbourn/directx-vs-templates/raw/master/VSIX/Direct3DUWPGame.vsix
Which 404s.
Can we update the Wiki and point that link to a place that has the vsix? Would it be unwise to include a copy of the vsix in this project's codebase?
|
1.0
|
Wiki Page - The Basic Game Loop - Link 404 ref - Wiki Page
https://github.com/microsoft/DirectXTK12/wiki/The-basic-game-loop
Setup Section, highlighted text ` VS 2017/2019/2022`
Points to:
https://github.com/walbourn/directx-vs-templates/raw/master/VSIX/Direct3DUWPGame.vsix
Which 404s.
Can we update the Wiki and point that link to a place that has the vsix? Would it be unwise to include a copy of the vsix in this project's codebase?
|
non_process
|
wiki page the basic game loop link ref wiki page setup section highlighted text vs points to which can we update the wiki and point that link to a place that has the vsix would it be unwise to include a copy of the vsix in this project s codebase
| 0
|
16,805
| 22,047,994,540
|
IssuesEvent
|
2022-05-30 05:26:49
|
camunda/zeebe
|
https://api.github.com/repos/camunda/zeebe
|
closed
|
ZeebeDbInconsistentException in ColumnFamily DMN_DECISION_REQUIREMENTS
|
kind/bug scope/broker severity/high team/process-automation release/8.0.1 release/8.1.0-alpha1
|
**Describe the bug**
Found in error logs
https://console.cloud.google.com/errors/detail/CPDM9-CV9Nvk3wE;service=zeebe;time=P7D?project=camunda-cloud-240911
https://console.cloud.google.com/errors/detail/CLWTn7vY7pS04QE;service=zeebe;time=P7D?project=camunda-cloud-240911
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
**Log/Stacktrace**
<!-- If possible add the full stacktrace or Zeebe log which contains the issue. -->
<details><summary>Full Stacktrace</summary>
<p>
```
io.camunda.zeebe.db.ZeebeDbInconsistentException: Key DbLong{2251799813685350} in ColumnFamily DMN_DECISION_REQUIREMENTS already exists
at io.camunda.zeebe.db.impl.rocksdb.transaction.TransactionalColumnFamily.assertKeyDoesNotExist(TransactionalColumnFamily.java:273) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.TransactionalColumnFamily.lambda$insert$0(TransactionalColumnFamily.java:81) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.TransactionalColumnFamily.lambda$ensureInOpenTransaction$17(TransactionalColumnFamily.java:301) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.DefaultTransactionContext.runInTransaction(DefaultTransactionContext.java:33) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.TransactionalColumnFamily.ensureInOpenTransaction(TransactionalColumnFamily.java:300) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.TransactionalColumnFamily.insert(TransactionalColumnFamily.java:76) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.state.deployment.DbDecisionState.storeDecisionRequirements(DbDecisionState.java:163) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.state.appliers.DeploymentDistributedApplier.lambda$putDmnResourcesInState$0(DeploymentDistributedApplier.java:50) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at java.lang.Iterable.forEach(Unknown Source) ~[?:?]
at io.camunda.zeebe.engine.state.appliers.DeploymentDistributedApplier.putDmnResourcesInState(DeploymentDistributedApplier.java:45) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.state.appliers.DeploymentDistributedApplier.applyState(DeploymentDistributedApplier.java:39) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.state.appliers.DeploymentDistributedApplier.applyState(DeploymentDistributedApplier.java:23) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.state.appliers.EventAppliers.applyState(EventAppliers.java:239) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.writers.EventApplyingStateWriter.appendFollowUpEvent(EventApplyingStateWriter.java:36) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.deployment.distribute.DeploymentDistributeProcessor.processRecord(DeploymentDistributeProcessor.java:58) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.ProcessingStateMachine.lambda$processInTransaction$3(ProcessingStateMachine.java:300) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.ZeebeTransaction.run(ZeebeTransaction.java:84) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.ProcessingStateMachine.processInTransaction(ProcessingStateMachine.java:290) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.ProcessingStateMachine.processCommand(ProcessingStateMachine.java:253) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.ProcessingStateMachine.tryToReadNextRecord(ProcessingStateMachine.java:213) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.ProcessingStateMachine.readNextRecord(ProcessingStateMachine.java:189) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorJob.invoke(ActorJob.java:79) ~[zeebe-util-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorJob.execute(ActorJob.java:44) ~[zeebe-util-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorTask.execute(ActorTask.java:122) ~[zeebe-util-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorThread.executeCurrentTask(ActorThread.java:97) ~[zeebe-util-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorThread.doWork(ActorThread.java:80) ~[zeebe-util-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorThread.run(ActorThread.java:189) ~[zeebe-util-8.0.0.jar:8.0.0]
```
</p>
</details>
**Environment:**
- Zeebe Version: 8.0.0
|
1.0
|
ZeebeDbInconsistentException in ColumnFamily DMN_DECISION_REQUIREMENTS - **Describe the bug**
Found in error logs
https://console.cloud.google.com/errors/detail/CPDM9-CV9Nvk3wE;service=zeebe;time=P7D?project=camunda-cloud-240911
https://console.cloud.google.com/errors/detail/CLWTn7vY7pS04QE;service=zeebe;time=P7D?project=camunda-cloud-240911
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
**Log/Stacktrace**
<!-- If possible add the full stacktrace or Zeebe log which contains the issue. -->
<details><summary>Full Stacktrace</summary>
<p>
```
io.camunda.zeebe.db.ZeebeDbInconsistentException: Key DbLong{2251799813685350} in ColumnFamily DMN_DECISION_REQUIREMENTS already exists
at io.camunda.zeebe.db.impl.rocksdb.transaction.TransactionalColumnFamily.assertKeyDoesNotExist(TransactionalColumnFamily.java:273) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.TransactionalColumnFamily.lambda$insert$0(TransactionalColumnFamily.java:81) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.TransactionalColumnFamily.lambda$ensureInOpenTransaction$17(TransactionalColumnFamily.java:301) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.DefaultTransactionContext.runInTransaction(DefaultTransactionContext.java:33) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.TransactionalColumnFamily.ensureInOpenTransaction(TransactionalColumnFamily.java:300) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.TransactionalColumnFamily.insert(TransactionalColumnFamily.java:76) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.state.deployment.DbDecisionState.storeDecisionRequirements(DbDecisionState.java:163) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.state.appliers.DeploymentDistributedApplier.lambda$putDmnResourcesInState$0(DeploymentDistributedApplier.java:50) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at java.lang.Iterable.forEach(Unknown Source) ~[?:?]
at io.camunda.zeebe.engine.state.appliers.DeploymentDistributedApplier.putDmnResourcesInState(DeploymentDistributedApplier.java:45) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.state.appliers.DeploymentDistributedApplier.applyState(DeploymentDistributedApplier.java:39) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.state.appliers.DeploymentDistributedApplier.applyState(DeploymentDistributedApplier.java:23) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.state.appliers.EventAppliers.applyState(EventAppliers.java:239) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.writers.EventApplyingStateWriter.appendFollowUpEvent(EventApplyingStateWriter.java:36) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.deployment.distribute.DeploymentDistributeProcessor.processRecord(DeploymentDistributeProcessor.java:58) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.ProcessingStateMachine.lambda$processInTransaction$3(ProcessingStateMachine.java:300) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.db.impl.rocksdb.transaction.ZeebeTransaction.run(ZeebeTransaction.java:84) ~[zeebe-db-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.ProcessingStateMachine.processInTransaction(ProcessingStateMachine.java:290) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.ProcessingStateMachine.processCommand(ProcessingStateMachine.java:253) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.ProcessingStateMachine.tryToReadNextRecord(ProcessingStateMachine.java:213) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.engine.processing.streamprocessor.ProcessingStateMachine.readNextRecord(ProcessingStateMachine.java:189) ~[zeebe-workflow-engine-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorJob.invoke(ActorJob.java:79) ~[zeebe-util-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorJob.execute(ActorJob.java:44) ~[zeebe-util-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorTask.execute(ActorTask.java:122) ~[zeebe-util-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorThread.executeCurrentTask(ActorThread.java:97) ~[zeebe-util-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorThread.doWork(ActorThread.java:80) ~[zeebe-util-8.0.0.jar:8.0.0]
at io.camunda.zeebe.util.sched.ActorThread.run(ActorThread.java:189) ~[zeebe-util-8.0.0.jar:8.0.0]
```
</p>
</details>
**Environment:**
- Zeebe Version: 8.0.0
|
process
|
zeebedbinconsistentexception in columnfamily dmn decision requirements describe the bug found in error logs expected behavior log stacktrace full stacktrace io camunda zeebe db zeebedbinconsistentexception key dblong in columnfamily dmn decision requirements already exists at io camunda zeebe db impl rocksdb transaction transactionalcolumnfamily assertkeydoesnotexist transactionalcolumnfamily java at io camunda zeebe db impl rocksdb transaction transactionalcolumnfamily lambda insert transactionalcolumnfamily java at io camunda zeebe db impl rocksdb transaction transactionalcolumnfamily lambda ensureinopentransaction transactionalcolumnfamily java at io camunda zeebe db impl rocksdb transaction defaulttransactioncontext runintransaction defaulttransactioncontext java at io camunda zeebe db impl rocksdb transaction transactionalcolumnfamily ensureinopentransaction transactionalcolumnfamily java at io camunda zeebe db impl rocksdb transaction transactionalcolumnfamily insert transactionalcolumnfamily java at io camunda zeebe engine state deployment dbdecisionstate storedecisionrequirements dbdecisionstate java at io camunda zeebe engine state appliers deploymentdistributedapplier lambda putdmnresourcesinstate deploymentdistributedapplier java at java lang iterable foreach unknown source at io camunda zeebe engine state appliers deploymentdistributedapplier putdmnresourcesinstate deploymentdistributedapplier java at io camunda zeebe engine state appliers deploymentdistributedapplier applystate deploymentdistributedapplier java at io camunda zeebe engine state appliers deploymentdistributedapplier applystate deploymentdistributedapplier java at io camunda zeebe engine state appliers eventappliers applystate eventappliers java at io camunda zeebe engine processing streamprocessor writers eventapplyingstatewriter appendfollowupevent eventapplyingstatewriter java at io camunda zeebe engine processing deployment distribute deploymentdistributeprocessor processrecord deploymentdistributeprocessor java at io camunda zeebe engine processing streamprocessor processingstatemachine lambda processintransaction processingstatemachine java at io camunda zeebe db impl rocksdb transaction zeebetransaction run zeebetransaction java at io camunda zeebe engine processing streamprocessor processingstatemachine processintransaction processingstatemachine java at io camunda zeebe engine processing streamprocessor processingstatemachine processcommand processingstatemachine java at io camunda zeebe engine processing streamprocessor processingstatemachine trytoreadnextrecord processingstatemachine java at io camunda zeebe engine processing streamprocessor processingstatemachine readnextrecord processingstatemachine java at io camunda zeebe util sched actorjob invoke actorjob java at io camunda zeebe util sched actorjob execute actorjob java at io camunda zeebe util sched actortask execute actortask java at io camunda zeebe util sched actorthread executecurrenttask actorthread java at io camunda zeebe util sched actorthread dowork actorthread java at io camunda zeebe util sched actorthread run actorthread java environment zeebe version
| 1
|
252,634
| 19,044,621,250
|
IssuesEvent
|
2021-11-25 05:26:27
|
lex-world/Hash-Technolgies
|
https://api.github.com/repos/lex-world/Hash-Technolgies
|
closed
|
Link is not working on Footer
|
bug documentation help wanted good first issue
|
`<Link to="privacy-policy">Privacy Policy</Link>` is not working on package `react-location`
---
### Footer.jsx
```
import React from "react";
import "./style.scss";
import { Link } from "react-location";
export default function Footer() {
return (
<div className="footer__container">
<ul>
<li style={{ color: "#17203e", fontWeight: "600" }}>
Hash Technologies
</li>
<li>Jobs</li>
<li>
<Link to="privacy-policy">Privacy Policy</Link>
</li>
<li>Terms of Service</li>
</ul>
<div style={{ fontWeight: "500", color: "#7a7a7a" }}>
Copyright 2021, Hash Technologies
</div>
</div>
);
}
```
|
1.0
|
Link is not working on Footer - `<Link to="privacy-policy">Privacy Policy</Link>` is not working on package `react-location`
---
### Footer.jsx
```
import React from "react";
import "./style.scss";
import { Link } from "react-location";
export default function Footer() {
return (
<div className="footer__container">
<ul>
<li style={{ color: "#17203e", fontWeight: "600" }}>
Hash Technologies
</li>
<li>Jobs</li>
<li>
<Link to="privacy-policy">Privacy Policy</Link>
</li>
<li>Terms of Service</li>
</ul>
<div style={{ fontWeight: "500", color: "#7a7a7a" }}>
Copyright 2021, Hash Technologies
</div>
</div>
);
}
```
|
non_process
|
link is not working on footer privacy policy is not working on package react location footer jsx import react from react import style scss import link from react location export default function footer return hash technologies jobs privacy policy terms of service copyright hash technologies
| 0
|
4,469
| 7,333,018,847
|
IssuesEvent
|
2018-03-05 18:02:33
|
UKHomeOffice/dq-aws-transition
|
https://api.github.com/repos/UKHomeOffice/dq-aws-transition
|
closed
|
Test End-to-End load of S4 PARSED files from DQ Landing to Greenplum rpt_internal/rpt_external
|
Production S4 Processing
|
Test End-to-End load of S4 PARSED files from DQ Landing to Greenplum rpt_internal/rpt_external, ensuring all jobs succeed and record timings.
- [x] Turn On Data Transfer (S3 Landing to Data Ingest Win) on Data Ingest Windows
- [x] Record time taken to get all files into local directories
- [x] Run Sequence Check script against local files
- [x] Record time taken to get all files into local directories
- [x] Run Job_10_NRT > Job_13_poav_level… > Job_15_Field_Level_KPI
- [x] Record time taken per job
- [x] Run Job_60_NBTC_FMS
- [x] Record time taken per job
- [x] Run Job_30_INT
- [x] Record time taken per job
- [x] Run Job_40_EXT
- [x] Record time taken per job
- [x] Run Job_80_Backup…
- [x] Record time taken per job
|
1.0
|
Test End-to-End load of S4 PARSED files from DQ Landing to Greenplum rpt_internal/rpt_external - Test End-to-End load of S4 PARSED files from DQ Landing to Greenplum rpt_internal/rpt_external, ensuring all jobs succeed and record timings.
- [x] Turn On Data Transfer (S3 Landing to Data Ingest Win) on Data Ingest Windows
- [x] Record time taken to get all files into local directories
- [x] Run Sequence Check script against local files
- [x] Record time taken to get all files into local directories
- [x] Run Job_10_NRT > Job_13_poav_level… > Job_15_Field_Level_KPI
- [x] Record time taken per job
- [x] Run Job_60_NBTC_FMS
- [x] Record time taken per job
- [x] Run Job_30_INT
- [x] Record time taken per job
- [x] Run Job_40_EXT
- [x] Record time taken per job
- [x] Run Job_80_Backup…
- [x] Record time taken per job
|
process
|
test end to end load of parsed files from dq landing to greenplum rpt internal rpt external test end to end load of parsed files from dq landing to greenplum rpt internal rpt external ensuring all jobs succeed and record timings turn on data transfer landing to data ingest win on data ingest windows record time taken to get all files into local directories run sequence check script against local files record time taken to get all files into local directories run job nrt job poav level… job field level kpi record time taken per job run job nbtc fms record time taken per job run job int record time taken per job run job ext record time taken per job run job backup… record time taken per job
| 1
|
195,138
| 22,288,725,093
|
IssuesEvent
|
2022-06-12 03:11:38
|
mycomplexsoul/delta
|
https://api.github.com/repos/mycomplexsoul/delta
|
closed
|
CVE-2021-43138 (High) detected in async-1.5.2.tgz, async-2.6.3.tgz - autoclosed
|
security vulnerability
|
## CVE-2021-43138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>async-1.5.2.tgz</b>, <b>async-2.6.3.tgz</b></p></summary>
<p>
<details><summary><b>async-1.5.2.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-1.5.2.tgz">https://registry.npmjs.org/async/-/async-1.5.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/istanbul/node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- istanbul-0.4.5.tgz (Root Library)
- :x: **async-1.5.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>async-2.6.3.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-2.6.3.tgz">https://registry.npmjs.org/async/-/async-2.6.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-13.2.2.tgz (Root Library)
- webpack-dev-server-4.7.3.tgz
- portfinder-1.0.28.tgz
- :x: **async-2.6.3.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/mycomplexsoul/delta/commit/a3ff684f7a30fac8f67545d8e1ee5b93a74cbfe5">a3ff684f7a30fac8f67545d8e1ee5b93a74cbfe5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Async before 2.6.4 and 3.x before 3.2.2, a malicious user can obtain privileges via the mapValues() method, aka lib/internal/iterator.js createObjectIterator prototype pollution.
<p>Publish Date: 2022-04-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43138>CVE-2021-43138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p>
<p>Release Date: 2022-04-06</p>
<p>Fix Resolution (async): 2.6.4</p>
<p>Direct dependency fix Resolution (@angular-devkit/build-angular): 13.3.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-43138 (High) detected in async-1.5.2.tgz, async-2.6.3.tgz - autoclosed - ## CVE-2021-43138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>async-1.5.2.tgz</b>, <b>async-2.6.3.tgz</b></p></summary>
<p>
<details><summary><b>async-1.5.2.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-1.5.2.tgz">https://registry.npmjs.org/async/-/async-1.5.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/istanbul/node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- istanbul-0.4.5.tgz (Root Library)
- :x: **async-1.5.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>async-2.6.3.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-2.6.3.tgz">https://registry.npmjs.org/async/-/async-2.6.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-13.2.2.tgz (Root Library)
- webpack-dev-server-4.7.3.tgz
- portfinder-1.0.28.tgz
- :x: **async-2.6.3.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/mycomplexsoul/delta/commit/a3ff684f7a30fac8f67545d8e1ee5b93a74cbfe5">a3ff684f7a30fac8f67545d8e1ee5b93a74cbfe5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Async before 2.6.4 and 3.x before 3.2.2, a malicious user can obtain privileges via the mapValues() method, aka lib/internal/iterator.js createObjectIterator prototype pollution.
<p>Publish Date: 2022-04-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43138>CVE-2021-43138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p>
<p>Release Date: 2022-04-06</p>
<p>Fix Resolution (async): 2.6.4</p>
<p>Direct dependency fix Resolution (@angular-devkit/build-angular): 13.3.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in async tgz async tgz autoclosed cve high severity vulnerability vulnerable libraries async tgz async tgz async tgz higher order functions and common patterns for asynchronous code library home page a href path to dependency file package json path to vulnerable library node modules istanbul node modules async package json dependency hierarchy istanbul tgz root library x async tgz vulnerable library async tgz higher order functions and common patterns for asynchronous code library home page a href path to dependency file package json path to vulnerable library node modules async package json dependency hierarchy build angular tgz root library webpack dev server tgz portfinder tgz x async tgz vulnerable library found in head commit a href found in base branch master vulnerability details in async before and x before a malicious user can obtain privileges via the mapvalues method aka lib internal iterator js createobjectiterator prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution async direct dependency fix resolution angular devkit build angular step up your open source security game with mend
| 0
|
44,766
| 12,374,614,592
|
IssuesEvent
|
2020-05-19 02:07:21
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
opened
|
[ZOOM]: Section 103 - Modal windows MUST be usable at 200% to 400% zoom
|
508-defect-2 508-issue-mobile-design 508/Accessibility bah-section103
|
# [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
<!--
Enter an issue title using the format [ERROR TYPE]: Brief description of the problem
---
[SCREENREADER]: Edit buttons need aria-label for context
[KEYBOARD]: Add another user link will not receive keyboard focus
[AXE-CORE]: Heading levels should increase by one
[COGNITION]: Error messages should be more specific
[COLOR]: Blue button on blue background does not have sufficient contrast ratio
---
-->
<!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. -->
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️ Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
Several of the modal windows are expanding out of the viewport at 200-250% zoom. I saw this on a few of the longer modals like Yellow Ribbon and the college housing. Screenshot of Yellow Ribbon attached below.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Trevor_
## Acceptance Criteria
<!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. -->
- [ ] The modals do not break out of the viewport at any zoom between 200% and 400%
- [ ] No other zoom levels break or change their layout
## Environment
* Any browser, zoomed to 200% or more at 1280px width
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->
<img width="1274" alt="Screen Shot 2020-05-18 at 7 54 54 PM" src="https://user-images.githubusercontent.com/934879/82276657-7bfd0d80-994b-11ea-8085-32695592ce23.png">
|
1.0
|
[ZOOM]: Section 103 - Modal windows MUST be usable at 200% to 400% zoom - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
<!--
Enter an issue title using the format [ERROR TYPE]: Brief description of the problem
---
[SCREENREADER]: Edit buttons need aria-label for context
[KEYBOARD]: Add another user link will not receive keyboard focus
[AXE-CORE]: Heading levels should increase by one
[COGNITION]: Error messages should be more specific
[COLOR]: Blue button on blue background does not have sufficient contrast ratio
---
-->
<!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. -->
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️ Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
Several of the modal windows are expanding out of the viewport at 200-250% zoom. I saw this on a few of the longer modals like Yellow Ribbon and the college housing. Screenshot of Yellow Ribbon attached below.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket.
-->
**VFS Point of Contact:** _Trevor_
## Acceptance Criteria
<!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. -->
- [ ] The modals do not break out of the viewport at any zoom between 200% and 400%
- [ ] No other zoom levels break or change their layout
## Environment
* Any browser, zoomed to 200% or more at 1280px width
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->
<img width="1274" alt="Screen Shot 2020-05-18 at 7 54 54 PM" src="https://user-images.githubusercontent.com/934879/82276657-7bfd0d80-994b-11ea-8085-32695592ce23.png">
|
non_process
|
section modal windows must be usable at to zoom enter an issue title using the format brief description of the problem edit buttons need aria label for context add another user link will not receive keyboard focus heading levels should increase by one error messages should be more specific blue button on blue background does not have sufficient contrast ratio feedback framework ❗️ must for if the feedback must be applied ⚠️ should if the feedback is best practice ✔️ consider for suggestions enhancements description several of the modal windows are expanding out of the viewport at zoom i saw this on a few of the longer modals like yellow ribbon and the college housing screenshot of yellow ribbon attached below point of contact if this issue is being opened by a vfs team member please add a point of contact usually this is the same person who enters the issue ticket vfs point of contact trevor acceptance criteria the modals do not break out of the viewport at any zoom between and no other zoom levels break or change their layout environment any browser zoomed to or more at width screenshots or trace logs img width alt screen shot at pm src
| 0
|
314,198
| 23,510,824,083
|
IssuesEvent
|
2022-08-18 16:21:30
|
loft-sh/vcluster
|
https://api.github.com/repos/loft-sh/vcluster
|
closed
|
Error from server (Forbidden): namespaces is forbidden: User "system:serviceaccount:vcluster:default" cannot list resource "namespaces" in API group "" at the cluster scope
|
area/documentation
|
### What happened?
I am following the docs from CONTRIBUTING.md
Listing the namespaces, as suggested in CONTRIBUTING fails.
```
vcluster on main [!] via 🐹 v1.18.2
❯ devspace run dev
.... starts fine
```
```
vcluster-0:vcluster-dev$ kubectl get ns
Error from server (Forbidden): namespaces is forbidden: User "system:serviceaccount:vcluster:default" cannot list resource "namespaces" in API group "" at the cluster scope
```
### What did you expect to happen?
According to CONTRIBUTING.md getting the namespaces should work.
From the docs:
```
root@vcluster-0:/vcluster# kubectl get ns
NAME STATUS AGE
default Active 2m18s
kube-system Active 2m18s
kube-public Active 2m18s
kube-node-lease Active 2m18s
```
### How can we reproduce it (as minimally and precisely as possible)?
I am unsure if it if only on my machine, but maybe following the instructions reproduces this on other machines, too.
### Anything else we need to know?
_No response_
### Host cluster Kubernetes version
<details>
```console
$ kubectl version
Client Version: version.Info{Major:"1", Minor:"24", GitVersion:"v1.24.3", GitCommit:"aef86a93758dc3cb2c658dd9657ab4ad4afc21cb", GitTreeState:"clean", BuildDate:"2022-07-14T02:31:37Z", GoVersion:"go1.18.3", Compiler:"gc", Platform:"linux/amd64"}
Kustomize Version: v4.5.4
Server Version: version.Info{Major:"1", Minor:"23", GitVersion:"v1.23.3", GitCommit:"816c97ab8cff8a1c72eccca1026f7820e93e0d25", GitTreeState:"clean", BuildDate:"2022-01-25T21:19:12Z", GoVersion:"go1.17.6", Compiler:"gc", Platform:"linux/amd64"}
```
</details>
### Host cluster Kubernetes distribution
❯ minikube version
minikube version: v1.25.2
### vlcuster version
from git.
### Vcluster Kubernetes distribution(k3s(default)), k8s, k0s)
default
### OS and Arch
<details>
```
❯ cat /etc/os-release
PRETTY_NAME="Ubuntu 21.10"
NAME="Ubuntu"
VERSION_ID="21.10"
VERSION="21.10 (Impish Indri)"
VERSION_CODENAME=impish
ID=ubuntu
ID_LIKE=debian
HOME_URL="https://www.ubuntu.com/"
SUPPORT_URL="https://help.ubuntu.com/"
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/"
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy"
UBUNTU_CODENAME=impish
```
</details>
|
1.0
|
Error from server (Forbidden): namespaces is forbidden: User "system:serviceaccount:vcluster:default" cannot list resource "namespaces" in API group "" at the cluster scope - ### What happened?
I am following the docs from CONTRIBUTING.md
Listing the namespaces, as suggested in CONTRIBUTING fails.
```
vcluster on main [!] via 🐹 v1.18.2
❯ devspace run dev
.... starts fine
```
```
vcluster-0:vcluster-dev$ kubectl get ns
Error from server (Forbidden): namespaces is forbidden: User "system:serviceaccount:vcluster:default" cannot list resource "namespaces" in API group "" at the cluster scope
```
### What did you expect to happen?
According to CONTRIBUTING.md getting the namespaces should work.
From the docs:
```
root@vcluster-0:/vcluster# kubectl get ns
NAME STATUS AGE
default Active 2m18s
kube-system Active 2m18s
kube-public Active 2m18s
kube-node-lease Active 2m18s
```
### How can we reproduce it (as minimally and precisely as possible)?
I am unsure if it if only on my machine, but maybe following the instructions reproduces this on other machines, too.
### Anything else we need to know?
_No response_
### Host cluster Kubernetes version
<details>
```console
$ kubectl version
Client Version: version.Info{Major:"1", Minor:"24", GitVersion:"v1.24.3", GitCommit:"aef86a93758dc3cb2c658dd9657ab4ad4afc21cb", GitTreeState:"clean", BuildDate:"2022-07-14T02:31:37Z", GoVersion:"go1.18.3", Compiler:"gc", Platform:"linux/amd64"}
Kustomize Version: v4.5.4
Server Version: version.Info{Major:"1", Minor:"23", GitVersion:"v1.23.3", GitCommit:"816c97ab8cff8a1c72eccca1026f7820e93e0d25", GitTreeState:"clean", BuildDate:"2022-01-25T21:19:12Z", GoVersion:"go1.17.6", Compiler:"gc", Platform:"linux/amd64"}
```
</details>
### Host cluster Kubernetes distribution
❯ minikube version
minikube version: v1.25.2
### vlcuster version
from git.
### Vcluster Kubernetes distribution(k3s(default)), k8s, k0s)
default
### OS and Arch
<details>
```
❯ cat /etc/os-release
PRETTY_NAME="Ubuntu 21.10"
NAME="Ubuntu"
VERSION_ID="21.10"
VERSION="21.10 (Impish Indri)"
VERSION_CODENAME=impish
ID=ubuntu
ID_LIKE=debian
HOME_URL="https://www.ubuntu.com/"
SUPPORT_URL="https://help.ubuntu.com/"
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/"
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy"
UBUNTU_CODENAME=impish
```
</details>
|
non_process
|
error from server forbidden namespaces is forbidden user system serviceaccount vcluster default cannot list resource namespaces in api group at the cluster scope what happened i am following the docs from contributing md listing the namespaces as suggested in contributing fails vcluster on main via 🐹 ❯ devspace run dev starts fine vcluster vcluster dev kubectl get ns error from server forbidden namespaces is forbidden user system serviceaccount vcluster default cannot list resource namespaces in api group at the cluster scope what did you expect to happen according to contributing md getting the namespaces should work from the docs root vcluster vcluster kubectl get ns name status age default active kube system active kube public active kube node lease active how can we reproduce it as minimally and precisely as possible i am unsure if it if only on my machine but maybe following the instructions reproduces this on other machines too anything else we need to know no response host cluster kubernetes version console kubectl version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux kustomize version server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux host cluster kubernetes distribution ❯ minikube version minikube version vlcuster version from git vcluster kubernetes distribution default default os and arch ❯ cat etc os release pretty name ubuntu name ubuntu version id version impish indri version codename impish id ubuntu id like debian home url support url bug report url privacy policy url ubuntu codename impish
| 0
|
456,909
| 13,151,029,497
|
IssuesEvent
|
2020-08-09 14:44:43
|
chrisjsewell/docutils
|
https://api.github.com/repos/chrisjsewell/docutils
|
closed
|
polish-translation [SF:patches:43]
|
closed-accepted patches priority-5
|
author: robwolfe99
created: 2008-03-02 15:45:39
assigned: None
SF_url: https://sourceforge.net/p/docutils/patches/43
Tested with test/test\_language.py in release 0.5.
---
commenter: robwolfe99
posted: 2008-03-02 15:45:39
title: #43 polish-translation
attachments:
- https://sourceforge.net/p/docutils/patches/_discuss/thread/b8d8776f/9c0f/attachment/pl.py
docutils/languages/pl.py
---
commenter: robwolfe99
posted: 2008-03-02 15:46:55
title: #43 polish-translation
attachments:
- https://sourceforge.net/p/docutils/patches/_discuss/thread/b8d8776f/5643/attachment/pl.py
docutils/parsers/rst/languages/pl.py
---
commenter: robwolfe99
posted: 2008-03-02 15:46:55
title: #43 polish-translation
Logged In: YES
user\_id=2025005
Originator: YES
File Added: pl.py
---
commenter: grubert
posted: 2008-08-25 13:08:15
title: #43 polish-translation
- **status**: open --> closed-accepted
---
commenter: grubert
posted: 2008-08-25 13:08:15
title: #43 polish-translation
Logged In: YES
user\_id=147070
Originator: NO
Thank you for your contribution\! It has been checked in to the
Docutils repository.
You can download the most current snapshot from:
http://docutils.sourceforge.net/docutils-snapshot.tgz
|
1.0
|
polish-translation [SF:patches:43] -
author: robwolfe99
created: 2008-03-02 15:45:39
assigned: None
SF_url: https://sourceforge.net/p/docutils/patches/43
Tested with test/test\_language.py in release 0.5.
---
commenter: robwolfe99
posted: 2008-03-02 15:45:39
title: #43 polish-translation
attachments:
- https://sourceforge.net/p/docutils/patches/_discuss/thread/b8d8776f/9c0f/attachment/pl.py
docutils/languages/pl.py
---
commenter: robwolfe99
posted: 2008-03-02 15:46:55
title: #43 polish-translation
attachments:
- https://sourceforge.net/p/docutils/patches/_discuss/thread/b8d8776f/5643/attachment/pl.py
docutils/parsers/rst/languages/pl.py
---
commenter: robwolfe99
posted: 2008-03-02 15:46:55
title: #43 polish-translation
Logged In: YES
user\_id=2025005
Originator: YES
File Added: pl.py
---
commenter: grubert
posted: 2008-08-25 13:08:15
title: #43 polish-translation
- **status**: open --> closed-accepted
---
commenter: grubert
posted: 2008-08-25 13:08:15
title: #43 polish-translation
Logged In: YES
user\_id=147070
Originator: NO
Thank you for your contribution\! It has been checked in to the
Docutils repository.
You can download the most current snapshot from:
http://docutils.sourceforge.net/docutils-snapshot.tgz
|
non_process
|
polish translation author created assigned none sf url tested with test test language py in release commenter posted title polish translation attachments docutils languages pl py commenter posted title polish translation attachments docutils parsers rst languages pl py commenter posted title polish translation logged in yes user id originator yes file added pl py commenter grubert posted title polish translation status open closed accepted commenter grubert posted title polish translation logged in yes user id originator no thank you for your contribution it has been checked in to the docutils repository you can download the most current snapshot from
| 0
|
22,374
| 31,142,281,169
|
IssuesEvent
|
2023-08-16 01:43:51
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Flaky test: post-install.sh script on cypress-example-kitchensink
|
process: flaky test topic: flake ❄️ stage: fire watch stale
|
### Link to dashboard or CircleCI failure
- https://app.circleci.com/pipelines/github/cypress-io/cypress/42278/workflows/3e16eb19-8b95-4ac7-9d6c-e417606d2da4/jobs/1755693
- https://app.circleci.com/pipelines/github/cypress-io/cypress/42270/workflows/03848f21-f9fa-455a-a1ff-2540b183f110/jobs/1755135
### Link to failing test in GitHub
File in question: https://github.com/cypress-io/cypress/blob/develop/cli/scripts/post-install.js
### Analysis
This CI failure happens while installing dependencies for Cypress kitchensink. Some logs of interest:
1. Failed integrity check:
<img width="600" alt="Screen Shot 2022-08-20 at 8 59 52 PM" src="https://user-images.githubusercontent.com/26726429/185774845-ea38f147-6fce-4e16-bc45-5502aab32854.png">
2. Lots of unmet / incorrect peer dependencies:
<img width="600" alt="Screen Shot 2022-08-20 at 9 01 56 PM" src="https://user-images.githubusercontent.com/26726429/185774906-76c1abe2-5750-48e4-a5a5-efa1d2f93382.png">
3. Error thrown in post-install script for Cypress:
<img width="600" alt="Screen Shot 2022-08-20 at 9 00 31 PM" src="https://user-images.githubusercontent.com/26726429/185774858-21eb4fe6-3a9a-462b-9acf-6c5c6e094146.png">
### Cypress Version
10.5.0
### Other
|
1.0
|
Flaky test: post-install.sh script on cypress-example-kitchensink - ### Link to dashboard or CircleCI failure
- https://app.circleci.com/pipelines/github/cypress-io/cypress/42278/workflows/3e16eb19-8b95-4ac7-9d6c-e417606d2da4/jobs/1755693
- https://app.circleci.com/pipelines/github/cypress-io/cypress/42270/workflows/03848f21-f9fa-455a-a1ff-2540b183f110/jobs/1755135
### Link to failing test in GitHub
File in question: https://github.com/cypress-io/cypress/blob/develop/cli/scripts/post-install.js
### Analysis
This CI failure happens while installing dependencies for Cypress kitchensink. Some logs of interest:
1. Failed integrity check:
<img width="600" alt="Screen Shot 2022-08-20 at 8 59 52 PM" src="https://user-images.githubusercontent.com/26726429/185774845-ea38f147-6fce-4e16-bc45-5502aab32854.png">
2. Lots of unmet / incorrect peer dependencies:
<img width="600" alt="Screen Shot 2022-08-20 at 9 01 56 PM" src="https://user-images.githubusercontent.com/26726429/185774906-76c1abe2-5750-48e4-a5a5-efa1d2f93382.png">
3. Error thrown in post-install script for Cypress:
<img width="600" alt="Screen Shot 2022-08-20 at 9 00 31 PM" src="https://user-images.githubusercontent.com/26726429/185774858-21eb4fe6-3a9a-462b-9acf-6c5c6e094146.png">
### Cypress Version
10.5.0
### Other
|
process
|
flaky test post install sh script on cypress example kitchensink link to dashboard or circleci failure link to failing test in github file in question analysis this ci failure happens while installing dependencies for cypress kitchensink some logs of interest failed integrity check img width alt screen shot at pm src lots of unmet incorrect peer dependencies img width alt screen shot at pm src error thrown in post install script for cypress img width alt screen shot at pm src cypress version other
| 1
|
2,507
| 5,283,004,970
|
IssuesEvent
|
2017-02-07 20:16:35
|
opentrials/opentrials
|
https://api.github.com/repos/opentrials/opentrials
|
reopened
|
Normalize organisations names
|
Data data cleaning Processors
|
Organisation names are tricky: you can go as far as looking for the ownership trees to find companies with different names that have the same owner. This issue isn't about that, but just dealing with the simplest case: organisations with the same name but different wordings (e.g. `ACME` and `ACME Ltda`). They should all have the same name in our database. In general, prefer the longer wording (e.g. `Johnson and Johnson` instead of `J&J`).
This will probably involve changes on https://github.com/opentrials/processors/blob/master/processors/base/writers/organisation.py, so all processors using this writer will automatically have its name cleaned.
|
1.0
|
Normalize organisations names - Organisation names are tricky: you can go as far as looking for the ownership trees to find companies with different names that have the same owner. This issue isn't about that, but just dealing with the simplest case: organisations with the same name but different wordings (e.g. `ACME` and `ACME Ltda`). They should all have the same name in our database. In general, prefer the longer wording (e.g. `Johnson and Johnson` instead of `J&J`).
This will probably involve changes on https://github.com/opentrials/processors/blob/master/processors/base/writers/organisation.py, so all processors using this writer will automatically have its name cleaned.
|
process
|
normalize organisations names organisation names are tricky you can go as far as looking for the ownership trees to find companies with different names that have the same owner this issue isn t about that but just dealing with the simplest case organisations with the same name but different wordings e g acme and acme ltda they should all have the same name in our database in general prefer the longer wording e g johnson and johnson instead of j j this will probably involve changes on so all processors using this writer will automatically have its name cleaned
| 1
|
37,127
| 9,962,884,617
|
IssuesEvent
|
2019-07-07 18:22:08
|
raphw/byte-buddy
|
https://api.github.com/repos/raphw/byte-buddy
|
closed
|
Publish Gradle plugin to official repository
|
build
|
The Gradle plugin should be published in the official repository upon release:
See: https://discuss.gradle.org/t/upload-artifact-to-plugin-portal-without-plugin/19344/5
|
1.0
|
Publish Gradle plugin to official repository - The Gradle plugin should be published in the official repository upon release:
See: https://discuss.gradle.org/t/upload-artifact-to-plugin-portal-without-plugin/19344/5
|
non_process
|
publish gradle plugin to official repository the gradle plugin should be published in the official repository upon release see
| 0
|
554,547
| 16,432,731,905
|
IssuesEvent
|
2021-05-20 05:18:27
|
Sequel-Ace/Sequel-Ace
|
https://api.github.com/repos/Sequel-Ace/Sequel-Ace
|
closed
|
Quickly losing its connection to the DB host
|
Bug Highest Priority
|
Version of Sequel-Ace? v3.1.0 (build 3013)
Version of macOS? Catalina, 10.15.7
Device(s): Macbook pro
Installed via homebrew (cask)
Issue: Sequel Ace is very quickly losing its connection to the DB host.
I connect using SSH method. If I open a database table to view its contents and leave the connection idle for about 20 to 30 seconds I get a dialog that pops up that indicates the following: "Connection Lost - Sequel Ace appear to have lost the connection to the server, or the server has stopped responding", with a choice to close the connection or reconnect.
If I choose reconnect, it attempts to connect and reload the database table I was looking at. But the dialog pops up again with the same message and choices. The only way to move forward is to close the connection and open a new one.
This behavior began after updating to v3.1.0 from v2.3.2. On the 2-series release this wasn't an issue.
I've toggled the Keep Alive option under Network preferences and also adjusted the connection timeout. Nothing seems to resolve the issue (even setting the timeout value to zero "0" with keep-alive enabled.)
|
1.0
|
Quickly losing its connection to the DB host - Version of Sequel-Ace? v3.1.0 (build 3013)
Version of macOS? Catalina, 10.15.7
Device(s): Macbook pro
Installed via homebrew (cask)
Issue: Sequel Ace is very quickly losing its connection to the DB host.
I connect using SSH method. If I open a database table to view its contents and leave the connection idle for about 20 to 30 seconds I get a dialog that pops up that indicates the following: "Connection Lost - Sequel Ace appear to have lost the connection to the server, or the server has stopped responding", with a choice to close the connection or reconnect.
If I choose reconnect, it attempts to connect and reload the database table I was looking at. But the dialog pops up again with the same message and choices. The only way to move forward is to close the connection and open a new one.
This behavior began after updating to v3.1.0 from v2.3.2. On the 2-series release this wasn't an issue.
I've toggled the Keep Alive option under Network preferences and also adjusted the connection timeout. Nothing seems to resolve the issue (even setting the timeout value to zero "0" with keep-alive enabled.)
|
non_process
|
quickly losing its connection to the db host version of sequel ace build version of macos catalina device s macbook pro installed via homebrew cask issue sequel ace is very quickly losing its connection to the db host i connect using ssh method if i open a database table to view its contents and leave the connection idle for about to seconds i get a dialog that pops up that indicates the following connection lost sequel ace appear to have lost the connection to the server or the server has stopped responding with a choice to close the connection or reconnect if i choose reconnect it attempts to connect and reload the database table i was looking at but the dialog pops up again with the same message and choices the only way to move forward is to close the connection and open a new one this behavior began after updating to from on the series release this wasn t an issue i ve toggled the keep alive option under network preferences and also adjusted the connection timeout nothing seems to resolve the issue even setting the timeout value to zero with keep alive enabled
| 0
|
11,274
| 14,073,917,352
|
IssuesEvent
|
2020-11-04 06:11:31
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
Add document for algorithm "gdal:contour_polygon"
|
3.14 Processing Alg
|
## Description
There's a "gdal:contour_polygon" tool in QGIS's toolbox, and the help button links to [gdalcontour_polygon](https://docs.qgis.org/3.14/zh/docs/user_manual/processing_algs/gdal/rasterextraction.html#gdalcontour_polygon). But the document pages only have the specification of the gdal:contour tool. The built-in python help tool works, `processing.algorithmHelp("gdal:contour_polygon") `. I guess that paragraph might be forgotten, so please add it back to the document.
Page URL:
https://docs.qgis.org/3.14/zh/docs/user_manual/processing_algs/gdal/rasterextraction.html#gdalcontour_polygon
|
1.0
|
Add document for algorithm "gdal:contour_polygon" - ## Description
There's a "gdal:contour_polygon" tool in QGIS's toolbox, and the help button links to [gdalcontour_polygon](https://docs.qgis.org/3.14/zh/docs/user_manual/processing_algs/gdal/rasterextraction.html#gdalcontour_polygon). But the document pages only have the specification of the gdal:contour tool. The built-in python help tool works, `processing.algorithmHelp("gdal:contour_polygon") `. I guess that paragraph might be forgotten, so please add it back to the document.
Page URL:
https://docs.qgis.org/3.14/zh/docs/user_manual/processing_algs/gdal/rasterextraction.html#gdalcontour_polygon
|
process
|
add document for algorithm gdal contour polygon description there s a gdal contour polygon tool in qgis s toolbox and the help button links to but the document pages only have the specification of the gdal contour tool the built in python help tool works processing algorithmhelp gdal contour polygon i guess that paragraph might be forgotten so please add it back to the document page url
| 1
|
126
| 2,563,744,686
|
IssuesEvent
|
2015-02-06 15:20:08
|
tinkerpop/tinkerpop3
|
https://api.github.com/repos/tinkerpop/tinkerpop3
|
closed
|
Consider adding dedicated PropertyTraversal interface
|
enhancement process
|
Reasoning behind this is detailed in https://github.com/tinkerpop/tinkerpop3/issues/497
Three of us got this wrong with the other folks being much better at gremlin than I. Therefore users will also stumble at this.
If you're traversing a property then instead of using:
```
g.v().properties("foo").has(T.value, "prop")....
```
if would be nicer if you could use
```
g.v().properties("foo").hasValue("prop").... //or any other has signature
```
If you are using an IDE then the code completion will give you the answer without having to resort to the docs.
|
1.0
|
Consider adding dedicated PropertyTraversal interface - Reasoning behind this is detailed in https://github.com/tinkerpop/tinkerpop3/issues/497
Three of us got this wrong with the other folks being much better at gremlin than I. Therefore users will also stumble at this.
If you're traversing a property then instead of using:
```
g.v().properties("foo").has(T.value, "prop")....
```
if would be nicer if you could use
```
g.v().properties("foo").hasValue("prop").... //or any other has signature
```
If you are using an IDE then the code completion will give you the answer without having to resort to the docs.
|
process
|
consider adding dedicated propertytraversal interface reasoning behind this is detailed in three of us got this wrong with the other folks being much better at gremlin than i therefore users will also stumble at this if you re traversing a property then instead of using g v properties foo has t value prop if would be nicer if you could use g v properties foo hasvalue prop or any other has signature if you are using an ide then the code completion will give you the answer without having to resort to the docs
| 1
|
4,006
| 6,934,970,473
|
IssuesEvent
|
2017-12-03 01:36:35
|
bartop/tpl
|
https://api.github.com/repos/bartop/tpl
|
opened
|
Add gcc 4.9 in Travis CI
|
new feature process upgrade
|
gcc4.9 has theoretically all the features this library needs to work. We should try to add it to the matrix.
|
1.0
|
Add gcc 4.9 in Travis CI - gcc4.9 has theoretically all the features this library needs to work. We should try to add it to the matrix.
|
process
|
add gcc in travis ci has theoretically all the features this library needs to work we should try to add it to the matrix
| 1
|
20,024
| 26,503,762,147
|
IssuesEvent
|
2023-01-18 12:20:24
|
inmanta/inmanta-core
|
https://api.github.com/repos/inmanta/inmanta-core
|
closed
|
PluginModuleFinder uses deprecated base class
|
compiler agent process
|
`inmanta.loader.PluginModuleFinder` inherits from `importlib.abc.Finder` which has been deprecated in favor of `importlib.abc.MetaPathFinder` and `importlib.abc.PathEntryFinder`.
|
1.0
|
PluginModuleFinder uses deprecated base class - `inmanta.loader.PluginModuleFinder` inherits from `importlib.abc.Finder` which has been deprecated in favor of `importlib.abc.MetaPathFinder` and `importlib.abc.PathEntryFinder`.
|
process
|
pluginmodulefinder uses deprecated base class inmanta loader pluginmodulefinder inherits from importlib abc finder which has been deprecated in favor of importlib abc metapathfinder and importlib abc pathentryfinder
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.