Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
20,236
3,800,035,845
IssuesEvent
2016-03-23 17:45:36
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
circleci: failed tests (15094): TestCommitWaitState
Robot test-failure
The following test appears to have failed: [#15094](https://circleci.com/gh/cockroachdb/cockroach/15094): ``` E160323 17:44:51.857301 client/txn.go:358 failure aborting transaction: kv/txn_coord_sender.go:344: writing transaction timed out, was aborted, or ran on multiple coordinators; abort caused by: node unavailable; try another peer W160323 17:44:51.857409 kv/dist_sender.go:676 failed to invoke AdminSplit [/Table/50/0,/Min): storage/replica_command.go:1664: split at key /Table/50 failed: node unavailable; try another peer I160323 17:44:51.857452 client/db.go:422 failed batch: storage/replica_command.go:1664: split at key /Table/50 failed: node unavailable; try another peer E160323 17:44:51.857510 storage/queue.go:444 [split] range=5 [/Table/14-/Max): error: storage/split_queue.go:100: unable to split range=5 [/Table/14-/Max) at key "/Table/50/0": storage/replica_command.go:1664: split at key /Table/50 failed: node unavailable; try another peer I160323 17:44:52.857210 /go/src/google.golang.org/grpc/clientconn.go:463 grpc: Conn.resetTransport failed to create client transport: connection error: desc = "transport: dial tcp 127.0.0.1:42530: getsockopt: connection refused"; Reconnecting to "127.0.0.1:42530" panic: test timed out after 1m10s goroutine 26124 [running]: panic(0x14a0e00, 0xc821015be0) /usr/local/go/src/runtime/panic.go:464 +0x3e6 testing.startAlarm.func1() /usr/local/go/src/testing/testing.go:725 +0x14b created by time.goFunc /usr/local/go/src/time/sleep.go:129 +0x3a goroutine 1 [chan receive]: testing.RunTests(0x1beda10, 0x2376a40, 0x48, 0x48, 0xf8317da5b7241101) /usr/local/go/src/testing/testing.go:583 +0x8d2 testing.(*M).Run(0xc820764f08, 0x11) /usr/local/go/src/testing/testing.go:515 +0x81 github.com/cockroachdb/cockroach/sql_test.TestMain(0xc820764f08) /go/src/github.com/cockroachdb/cockroach/sql/main_test.go:234 +0x26 main.main() github.com/cockroachdb/cockroach/sql/_test/_testmain.go:382 +0x114 goroutine 17 [syscall, 1 minutes, locked to thread]: runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:1998 +0x1 goroutine 22 [chan receive]: github.com/cockroachdb/cockroach/util/log.(*loggingT).flushDaemon(0x264b580) /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:1003 +0x64 created by github.com/cockroachdb/cockroach/util/log.init.1 /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:604 +0x8a goroutine 25032 [IO wait]: net.runtime_pollWait(0x7fb5b3e5d5f8, 0x72, 0xc8201bc800) /usr/local/go/src/runtime/netpoll.go:160 +0x60 net.(*pollDesc).Wait(0xc820d0b560, 0x72, 0x0, 0x0) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x3a net.(*pollDesc).WaitRead(0xc820d0b560, 0x0, 0x0) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x36 net.(*netFD).Read(0xc820d0b500, 0xc8201bc800, 0x400, 0x400, 0x0, 0x7fb5b3e45028, 0xc82007e0b0) /usr/local/go/src/net/fd_unix.go:250 +0x23a net.(*conn).Read(0xc8200262e0, 0xc8201bc800, 0x400, 0x400, 0x0, 0x0, 0x0) /usr/local/go/src/net/net.go:172 +0xe4 -- google.golang.org/grpc.(*Server).handleRawConn(0xc82116ae00, 0x7fb5b3e5fd98, 0xc82136a580) /go/src/google.golang.org/grpc/server.go:289 +0x4ee created by google.golang.org/grpc.(*Server).Serve /go/src/google.golang.org/grpc/server.go:261 +0x372 goroutine 24985 [semacquire]: sync.runtime_Semacquire(0xc821546024) /usr/local/go/src/runtime/sema.go:47 +0x26 sync.(*WaitGroup).Wait(0xc821546018) /usr/local/go/src/sync/waitgroup.go:127 +0xb4 github.com/cockroachdb/cockroach/util/stop.(*Stopper).Stop(0xc821546000) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:302 +0x1dd github.com/cockroachdb/cockroach/server.(*Server).Stop(0xc8206780d0) /go/src/github.com/cockroachdb/cockroach/server/server.go:359 +0x28 github.com/cockroachdb/cockroach/server.(*TestServer).Stop(0xc82123a030) /go/src/github.com/cockroachdb/cockroach/server/testserver.go:318 +0x5d -- testing.tRunner(0xc820090000, 0x23770a0) /usr/local/go/src/testing/testing.go:473 +0x98 created by testing.RunTests /usr/local/go/src/testing/testing.go:582 +0x892 goroutine 25065 [sleep]: time.Sleep(0x6e51796d) /usr/local/go/src/runtime/time.go:59 +0xf9 google.golang.org/grpc.(*Conn).resetTransport(0xc8209abb20, 0xad5200, 0x0, 0x0) /go/src/google.golang.org/grpc/clientconn.go:461 +0x79f google.golang.org/grpc.NewConn.func1(0xc8209abb20) /go/src/google.golang.org/grpc/clientconn.go:317 +0x2a created by google.golang.org/grpc.NewConn /go/src/google.golang.org/grpc/clientconn.go:323 +0x509 goroutine 25066 [select]: google.golang.org/grpc.(*Conn).Wait(0xc8209abb20, 0x7fb5b3e5f9a0, 0xc8209ce000, 0x0, 0x0, 0x0, 0x0) /go/src/google.golang.org/grpc/clientconn.go:552 +0x360 google.golang.org/grpc.(*unicastPicker).Pick(0xc8209a4420, 0x7fb5b3e5f9a0, 0xc8209ce000, 0x0, 0x0, 0x0, 0x0) /go/src/google.golang.org/grpc/picker.go:84 +0x51 google.golang.org/grpc.Invoke(0x7fb5b3e5f9a0, 0xc8209ce000, 0x1968c40, 0x1d, 0x17da680, 0xc820875e40, 0x17cc9e0, 0xc821012640, 0xc820958a00, 0x0, ...) /go/src/google.golang.org/grpc/call.go:154 +0x6d0 github.com/cockroachdb/cockroach/rpc.(*heartbeatClient).Ping(0xc8200266d8, 0x7fb5b3e5f9a0, 0xc8209ce000, 0xc820875e40, 0x0, 0x0, 0x0, 0x401000000000000, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/rpc/heartbeat.pb.go:115 +0xec github.com/cockroachdb/cockroach/rpc.(*Context).runHeartbeat(0xc820a3ab40, 0xc820958a00, 0xc821014cd0, 0xf, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/rpc/context.go:187 +0x2b5 -- github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821546000, 0xc820b36870) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 24992 [chan receive]: github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open.func1(0xc82084ea20) /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:135 +0x3a created by github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:136 +0x7c9 goroutine 25002 [chan receive]: github.com/cockroachdb/cockroach/server.RegisterAdminHandlerFromEndpoint.func1.1(0x7fb5b3e93a80, 0xc8207ba6c0, 0xc8209dc000, 0xc82095bc90, 0xf) /go/src/github.com/cockroachdb/cockroach/server/admin.pb.gw.go:177 +0x51 created by github.com/cockroachdb/cockroach/server.RegisterAdminHandlerFromEndpoint.func1 /go/src/github.com/cockroachdb/cockroach/server/admin.pb.gw.go:181 +0x205 goroutine 25046 [select]: google.golang.org/grpc/transport.(*http2Server).controller(0xc8202c61b0) /go/src/google.golang.org/grpc/transport/http2_server.go:610 +0x5da created by google.golang.org/grpc/transport.newHTTP2Server /go/src/google.golang.org/grpc/transport/http2_server.go:134 +0x853 goroutine 25048 [select]: google.golang.org/grpc/transport.(*http2Client).controller(0xc820b43770) /go/src/google.golang.org/grpc/transport/http2_client.go:831 +0x5da created by google.golang.org/grpc/transport.newHTTP2Client /go/src/google.golang.org/grpc/transport/http2_client.go:194 +0x153b goroutine 25049 [IO wait]: net.runtime_pollWait(0x7fb5b3e5d178, 0x72, 0xc8210ef000) /usr/local/go/src/runtime/netpoll.go:160 +0x60 net.(*pollDesc).Wait(0xc820d0a140, 0x72, 0x0, 0x0) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x3a net.(*pollDesc).WaitRead(0xc820d0a140, 0x0, 0x0) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x36 net.(*netFD).Read(0xc820d0a0e0, 0xc8210ef000, 0x1000, 0x1000, 0x0, 0x7fb5b3e45028, 0xc82007e0b0) /usr/local/go/src/net/fd_unix.go:250 +0x23a net.(*conn).Read(0xc820026000, 0xc8210ef000, 0x1000, 0x1000, 0x0, 0x0, 0x0) /usr/local/go/src/net/net.go:172 +0xe4 -- google.golang.org/grpc/transport.(*http2Client).reader(0xc820b43770) /go/src/google.golang.org/grpc/transport/http2_client.go:754 +0xef created by google.golang.org/grpc/transport.newHTTP2Client /go/src/google.golang.org/grpc/transport/http2_client.go:200 +0x159a goroutine 25001 [select]: google.golang.org/grpc.(*Conn).transportMonitor(0xc8212f6380) /go/src/google.golang.org/grpc/clientconn.go:511 +0x1d3 google.golang.org/grpc.NewConn.func1(0xc8212f6380) /go/src/google.golang.org/grpc/clientconn.go:322 +0x1b5 created by google.golang.org/grpc.NewConn /go/src/google.golang.org/grpc/clientconn.go:323 +0x509 FAIL github.com/cockroachdb/cockroach/sql 70.044s === RUN TestEval --- PASS: TestEval (0.01s) === RUN TestEvalError --- PASS: TestEvalError (0.00s) === RUN TestEvalComparisonExprCaching --- PASS: TestEvalComparisonExprCaching (0.00s) === RUN TestSimilarEscape --- PASS: TestSimilarEscape (0.00s) === RUN TestQualifiedNameString --- PASS: TestQualifiedNameString (0.00s) ``` Please assign, take a look and update the issue accordingly.
1.0
circleci: failed tests (15094): TestCommitWaitState - The following test appears to have failed: [#15094](https://circleci.com/gh/cockroachdb/cockroach/15094): ``` E160323 17:44:51.857301 client/txn.go:358 failure aborting transaction: kv/txn_coord_sender.go:344: writing transaction timed out, was aborted, or ran on multiple coordinators; abort caused by: node unavailable; try another peer W160323 17:44:51.857409 kv/dist_sender.go:676 failed to invoke AdminSplit [/Table/50/0,/Min): storage/replica_command.go:1664: split at key /Table/50 failed: node unavailable; try another peer I160323 17:44:51.857452 client/db.go:422 failed batch: storage/replica_command.go:1664: split at key /Table/50 failed: node unavailable; try another peer E160323 17:44:51.857510 storage/queue.go:444 [split] range=5 [/Table/14-/Max): error: storage/split_queue.go:100: unable to split range=5 [/Table/14-/Max) at key "/Table/50/0": storage/replica_command.go:1664: split at key /Table/50 failed: node unavailable; try another peer I160323 17:44:52.857210 /go/src/google.golang.org/grpc/clientconn.go:463 grpc: Conn.resetTransport failed to create client transport: connection error: desc = "transport: dial tcp 127.0.0.1:42530: getsockopt: connection refused"; Reconnecting to "127.0.0.1:42530" panic: test timed out after 1m10s goroutine 26124 [running]: panic(0x14a0e00, 0xc821015be0) /usr/local/go/src/runtime/panic.go:464 +0x3e6 testing.startAlarm.func1() /usr/local/go/src/testing/testing.go:725 +0x14b created by time.goFunc /usr/local/go/src/time/sleep.go:129 +0x3a goroutine 1 [chan receive]: testing.RunTests(0x1beda10, 0x2376a40, 0x48, 0x48, 0xf8317da5b7241101) /usr/local/go/src/testing/testing.go:583 +0x8d2 testing.(*M).Run(0xc820764f08, 0x11) /usr/local/go/src/testing/testing.go:515 +0x81 github.com/cockroachdb/cockroach/sql_test.TestMain(0xc820764f08) /go/src/github.com/cockroachdb/cockroach/sql/main_test.go:234 +0x26 main.main() github.com/cockroachdb/cockroach/sql/_test/_testmain.go:382 +0x114 goroutine 17 [syscall, 1 minutes, locked to thread]: runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:1998 +0x1 goroutine 22 [chan receive]: github.com/cockroachdb/cockroach/util/log.(*loggingT).flushDaemon(0x264b580) /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:1003 +0x64 created by github.com/cockroachdb/cockroach/util/log.init.1 /go/src/github.com/cockroachdb/cockroach/util/log/clog.go:604 +0x8a goroutine 25032 [IO wait]: net.runtime_pollWait(0x7fb5b3e5d5f8, 0x72, 0xc8201bc800) /usr/local/go/src/runtime/netpoll.go:160 +0x60 net.(*pollDesc).Wait(0xc820d0b560, 0x72, 0x0, 0x0) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x3a net.(*pollDesc).WaitRead(0xc820d0b560, 0x0, 0x0) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x36 net.(*netFD).Read(0xc820d0b500, 0xc8201bc800, 0x400, 0x400, 0x0, 0x7fb5b3e45028, 0xc82007e0b0) /usr/local/go/src/net/fd_unix.go:250 +0x23a net.(*conn).Read(0xc8200262e0, 0xc8201bc800, 0x400, 0x400, 0x0, 0x0, 0x0) /usr/local/go/src/net/net.go:172 +0xe4 -- google.golang.org/grpc.(*Server).handleRawConn(0xc82116ae00, 0x7fb5b3e5fd98, 0xc82136a580) /go/src/google.golang.org/grpc/server.go:289 +0x4ee created by google.golang.org/grpc.(*Server).Serve /go/src/google.golang.org/grpc/server.go:261 +0x372 goroutine 24985 [semacquire]: sync.runtime_Semacquire(0xc821546024) /usr/local/go/src/runtime/sema.go:47 +0x26 sync.(*WaitGroup).Wait(0xc821546018) /usr/local/go/src/sync/waitgroup.go:127 +0xb4 github.com/cockroachdb/cockroach/util/stop.(*Stopper).Stop(0xc821546000) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:302 +0x1dd github.com/cockroachdb/cockroach/server.(*Server).Stop(0xc8206780d0) /go/src/github.com/cockroachdb/cockroach/server/server.go:359 +0x28 github.com/cockroachdb/cockroach/server.(*TestServer).Stop(0xc82123a030) /go/src/github.com/cockroachdb/cockroach/server/testserver.go:318 +0x5d -- testing.tRunner(0xc820090000, 0x23770a0) /usr/local/go/src/testing/testing.go:473 +0x98 created by testing.RunTests /usr/local/go/src/testing/testing.go:582 +0x892 goroutine 25065 [sleep]: time.Sleep(0x6e51796d) /usr/local/go/src/runtime/time.go:59 +0xf9 google.golang.org/grpc.(*Conn).resetTransport(0xc8209abb20, 0xad5200, 0x0, 0x0) /go/src/google.golang.org/grpc/clientconn.go:461 +0x79f google.golang.org/grpc.NewConn.func1(0xc8209abb20) /go/src/google.golang.org/grpc/clientconn.go:317 +0x2a created by google.golang.org/grpc.NewConn /go/src/google.golang.org/grpc/clientconn.go:323 +0x509 goroutine 25066 [select]: google.golang.org/grpc.(*Conn).Wait(0xc8209abb20, 0x7fb5b3e5f9a0, 0xc8209ce000, 0x0, 0x0, 0x0, 0x0) /go/src/google.golang.org/grpc/clientconn.go:552 +0x360 google.golang.org/grpc.(*unicastPicker).Pick(0xc8209a4420, 0x7fb5b3e5f9a0, 0xc8209ce000, 0x0, 0x0, 0x0, 0x0) /go/src/google.golang.org/grpc/picker.go:84 +0x51 google.golang.org/grpc.Invoke(0x7fb5b3e5f9a0, 0xc8209ce000, 0x1968c40, 0x1d, 0x17da680, 0xc820875e40, 0x17cc9e0, 0xc821012640, 0xc820958a00, 0x0, ...) /go/src/google.golang.org/grpc/call.go:154 +0x6d0 github.com/cockroachdb/cockroach/rpc.(*heartbeatClient).Ping(0xc8200266d8, 0x7fb5b3e5f9a0, 0xc8209ce000, 0xc820875e40, 0x0, 0x0, 0x0, 0x401000000000000, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/rpc/heartbeat.pb.go:115 +0xec github.com/cockroachdb/cockroach/rpc.(*Context).runHeartbeat(0xc820a3ab40, 0xc820958a00, 0xc821014cd0, 0xf, 0x0, 0x0) /go/src/github.com/cockroachdb/cockroach/rpc/context.go:187 +0x2b5 -- github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker.func1(0xc821546000, 0xc820b36870) /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:139 +0x52 created by github.com/cockroachdb/cockroach/util/stop.(*Stopper).RunWorker /go/src/github.com/cockroachdb/cockroach/util/stop/stopper.go:140 +0x62 goroutine 24992 [chan receive]: github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open.func1(0xc82084ea20) /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:135 +0x3a created by github.com/cockroachdb/cockroach/storage/engine.(*RocksDB).Open /go/src/github.com/cockroachdb/cockroach/storage/engine/rocksdb.go:136 +0x7c9 goroutine 25002 [chan receive]: github.com/cockroachdb/cockroach/server.RegisterAdminHandlerFromEndpoint.func1.1(0x7fb5b3e93a80, 0xc8207ba6c0, 0xc8209dc000, 0xc82095bc90, 0xf) /go/src/github.com/cockroachdb/cockroach/server/admin.pb.gw.go:177 +0x51 created by github.com/cockroachdb/cockroach/server.RegisterAdminHandlerFromEndpoint.func1 /go/src/github.com/cockroachdb/cockroach/server/admin.pb.gw.go:181 +0x205 goroutine 25046 [select]: google.golang.org/grpc/transport.(*http2Server).controller(0xc8202c61b0) /go/src/google.golang.org/grpc/transport/http2_server.go:610 +0x5da created by google.golang.org/grpc/transport.newHTTP2Server /go/src/google.golang.org/grpc/transport/http2_server.go:134 +0x853 goroutine 25048 [select]: google.golang.org/grpc/transport.(*http2Client).controller(0xc820b43770) /go/src/google.golang.org/grpc/transport/http2_client.go:831 +0x5da created by google.golang.org/grpc/transport.newHTTP2Client /go/src/google.golang.org/grpc/transport/http2_client.go:194 +0x153b goroutine 25049 [IO wait]: net.runtime_pollWait(0x7fb5b3e5d178, 0x72, 0xc8210ef000) /usr/local/go/src/runtime/netpoll.go:160 +0x60 net.(*pollDesc).Wait(0xc820d0a140, 0x72, 0x0, 0x0) /usr/local/go/src/net/fd_poll_runtime.go:73 +0x3a net.(*pollDesc).WaitRead(0xc820d0a140, 0x0, 0x0) /usr/local/go/src/net/fd_poll_runtime.go:78 +0x36 net.(*netFD).Read(0xc820d0a0e0, 0xc8210ef000, 0x1000, 0x1000, 0x0, 0x7fb5b3e45028, 0xc82007e0b0) /usr/local/go/src/net/fd_unix.go:250 +0x23a net.(*conn).Read(0xc820026000, 0xc8210ef000, 0x1000, 0x1000, 0x0, 0x0, 0x0) /usr/local/go/src/net/net.go:172 +0xe4 -- google.golang.org/grpc/transport.(*http2Client).reader(0xc820b43770) /go/src/google.golang.org/grpc/transport/http2_client.go:754 +0xef created by google.golang.org/grpc/transport.newHTTP2Client /go/src/google.golang.org/grpc/transport/http2_client.go:200 +0x159a goroutine 25001 [select]: google.golang.org/grpc.(*Conn).transportMonitor(0xc8212f6380) /go/src/google.golang.org/grpc/clientconn.go:511 +0x1d3 google.golang.org/grpc.NewConn.func1(0xc8212f6380) /go/src/google.golang.org/grpc/clientconn.go:322 +0x1b5 created by google.golang.org/grpc.NewConn /go/src/google.golang.org/grpc/clientconn.go:323 +0x509 FAIL github.com/cockroachdb/cockroach/sql 70.044s === RUN TestEval --- PASS: TestEval (0.01s) === RUN TestEvalError --- PASS: TestEvalError (0.00s) === RUN TestEvalComparisonExprCaching --- PASS: TestEvalComparisonExprCaching (0.00s) === RUN TestSimilarEscape --- PASS: TestSimilarEscape (0.00s) === RUN TestQualifiedNameString --- PASS: TestQualifiedNameString (0.00s) ``` Please assign, take a look and update the issue accordingly.
non_main
circleci failed tests testcommitwaitstate the following test appears to have failed client txn go failure aborting transaction kv txn coord sender go writing transaction timed out was aborted or ran on multiple coordinators abort caused by node unavailable try another peer kv dist sender go failed to invoke adminsplit table min storage replica command go split at key table failed node unavailable try another peer client db go failed batch storage replica command go split at key table failed node unavailable try another peer storage queue go range table max error storage split queue go unable to split range table max at key table storage replica command go split at key table failed node unavailable try another peer go src google golang org grpc clientconn go grpc conn resettransport failed to create client transport connection error desc transport dial tcp getsockopt connection refused reconnecting to panic test timed out after goroutine panic usr local go src runtime panic go testing startalarm usr local go src testing testing go created by time gofunc usr local go src time sleep go goroutine testing runtests usr local go src testing testing go testing m run usr local go src testing testing go github com cockroachdb cockroach sql test testmain go src github com cockroachdb cockroach sql main test go main main github com cockroachdb cockroach sql test testmain go goroutine runtime goexit usr local go src runtime asm s goroutine github com cockroachdb cockroach util log loggingt flushdaemon go src github com cockroachdb cockroach util log clog go created by github com cockroachdb cockroach util log init go src github com cockroachdb cockroach util log clog go goroutine net runtime pollwait usr local go src runtime netpoll go net polldesc wait usr local go src net fd poll runtime go net polldesc waitread usr local go src net fd poll runtime go net netfd read usr local go src net fd unix go net conn read usr local go src net net go google golang org grpc server handlerawconn go src google golang org grpc server go created by google golang org grpc server serve go src google golang org grpc server go goroutine sync runtime semacquire usr local go src runtime sema go sync waitgroup wait usr local go src sync waitgroup go github com cockroachdb cockroach util stop stopper stop go src github com cockroachdb cockroach util stop stopper go github com cockroachdb cockroach server server stop go src github com cockroachdb cockroach server server go github com cockroachdb cockroach server testserver stop go src github com cockroachdb cockroach server testserver go testing trunner usr local go src testing testing go created by testing runtests usr local go src testing testing go goroutine time sleep usr local go src runtime time go google golang org grpc conn resettransport go src google golang org grpc clientconn go google golang org grpc newconn go src google golang org grpc clientconn go created by google golang org grpc newconn go src google golang org grpc clientconn go goroutine google golang org grpc conn wait go src google golang org grpc clientconn go google golang org grpc unicastpicker pick go src google golang org grpc picker go google golang org grpc invoke go src google golang org grpc call go github com cockroachdb cockroach rpc heartbeatclient ping go src github com cockroachdb cockroach rpc heartbeat pb go github com cockroachdb cockroach rpc context runheartbeat go src github com cockroachdb cockroach rpc context go github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go created by github com cockroachdb cockroach util stop stopper runworker go src github com cockroachdb cockroach util stop stopper go goroutine github com cockroachdb cockroach storage engine rocksdb open go src github com cockroachdb cockroach storage engine rocksdb go created by github com cockroachdb cockroach storage engine rocksdb open go src github com cockroachdb cockroach storage engine rocksdb go goroutine github com cockroachdb cockroach server registeradminhandlerfromendpoint go src github com cockroachdb cockroach server admin pb gw go created by github com cockroachdb cockroach server registeradminhandlerfromendpoint go src github com cockroachdb cockroach server admin pb gw go goroutine google golang org grpc transport controller go src google golang org grpc transport server go created by google golang org grpc transport go src google golang org grpc transport server go goroutine google golang org grpc transport controller go src google golang org grpc transport client go created by google golang org grpc transport go src google golang org grpc transport client go goroutine net runtime pollwait usr local go src runtime netpoll go net polldesc wait usr local go src net fd poll runtime go net polldesc waitread usr local go src net fd poll runtime go net netfd read usr local go src net fd unix go net conn read usr local go src net net go google golang org grpc transport reader go src google golang org grpc transport client go created by google golang org grpc transport go src google golang org grpc transport client go goroutine google golang org grpc conn transportmonitor go src google golang org grpc clientconn go google golang org grpc newconn go src google golang org grpc clientconn go created by google golang org grpc newconn go src google golang org grpc clientconn go fail github com cockroachdb cockroach sql run testeval pass testeval run testevalerror pass testevalerror run testevalcomparisonexprcaching pass testevalcomparisonexprcaching run testsimilarescape pass testsimilarescape run testqualifiednamestring pass testqualifiednamestring please assign take a look and update the issue accordingly
0
75,468
9,256,369,583
IssuesEvent
2019-03-16 18:27:47
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Duplicate GlobalKey when using MaterialApp(builder) on Hot Restart
f: material design framework
Suddenly, (only) hot restarting causes this issue. ```dart runApp(MaterialApp(builder: (_, child) => AnyBuilder(builder: (_) => AnyWidget(child: child)), ...)) // app contains functionality ``` ``` Multiple widgets used the same GlobalKey. The key [GlobalObjectKey<NavigatorState> _WidgetsAppState#bd478] was used by multiple widgets. The parents of those widgets were: - AnyBuilder - AnyWidget A GlobalKey can only be specified on one widget at a time in the widget tree. ``` In this case I tested `AnyBuilder` with `FutureBuilder` and `ValueListenableBuilder` and `AnyWidget` can even be a `Container`. This always happens when hot restarting the application and only then. I looked at the Android Studio `Local History` of my project from before this behavior occured and after it occured and the `builder` in the `MaterialApp` has been unchanged. Having said that, there are many other widgets included in the actual code, e.g. `InheritedWidget`'s before `MaterialApp` etc. It is just really weird that this happens because I am not using any `GlobalKey`'s that could cause this and the exception is just really weird because neither `AnyBuilder` nor `AnyWidget` was ever supplied with any `Key` by me. I tried to change the `builder` in `MaterialApp` in many different ways but this bug (?) remains. Oh, and since this issue has been around my workflow is `Hot Restart -> Hot Reload` because a hot reload will fix it (otherwise the application is unusable because this `builder` sits at the top of the widget tree). `Channel master, v1.3.10-pre.27`
1.0
Duplicate GlobalKey when using MaterialApp(builder) on Hot Restart - Suddenly, (only) hot restarting causes this issue. ```dart runApp(MaterialApp(builder: (_, child) => AnyBuilder(builder: (_) => AnyWidget(child: child)), ...)) // app contains functionality ``` ``` Multiple widgets used the same GlobalKey. The key [GlobalObjectKey<NavigatorState> _WidgetsAppState#bd478] was used by multiple widgets. The parents of those widgets were: - AnyBuilder - AnyWidget A GlobalKey can only be specified on one widget at a time in the widget tree. ``` In this case I tested `AnyBuilder` with `FutureBuilder` and `ValueListenableBuilder` and `AnyWidget` can even be a `Container`. This always happens when hot restarting the application and only then. I looked at the Android Studio `Local History` of my project from before this behavior occured and after it occured and the `builder` in the `MaterialApp` has been unchanged. Having said that, there are many other widgets included in the actual code, e.g. `InheritedWidget`'s before `MaterialApp` etc. It is just really weird that this happens because I am not using any `GlobalKey`'s that could cause this and the exception is just really weird because neither `AnyBuilder` nor `AnyWidget` was ever supplied with any `Key` by me. I tried to change the `builder` in `MaterialApp` in many different ways but this bug (?) remains. Oh, and since this issue has been around my workflow is `Hot Restart -> Hot Reload` because a hot reload will fix it (otherwise the application is unusable because this `builder` sits at the top of the widget tree). `Channel master, v1.3.10-pre.27`
non_main
duplicate globalkey when using materialapp builder on hot restart suddenly only hot restarting causes this issue dart runapp materialapp builder child anybuilder builder anywidget child child app contains functionality multiple widgets used the same globalkey the key was used by multiple widgets the parents of those widgets were anybuilder anywidget a globalkey can only be specified on one widget at a time in the widget tree in this case i tested anybuilder with futurebuilder and valuelistenablebuilder and anywidget can even be a container this always happens when hot restarting the application and only then i looked at the android studio local history of my project from before this behavior occured and after it occured and the builder in the materialapp has been unchanged having said that there are many other widgets included in the actual code e g inheritedwidget s before materialapp etc it is just really weird that this happens because i am not using any globalkey s that could cause this and the exception is just really weird because neither anybuilder nor anywidget was ever supplied with any key by me i tried to change the builder in materialapp in many different ways but this bug remains oh and since this issue has been around my workflow is hot restart hot reload because a hot reload will fix it otherwise the application is unusable because this builder sits at the top of the widget tree channel master pre
0
5,143
26,222,057,709
IssuesEvent
2023-01-04 15:33:55
NIAEFEUP/website-niaefeup-backend
https://api.github.com/repos/NIAEFEUP/website-niaefeup-backend
closed
global: refactor DTO structure
maintainability
Until a while ago, all the DTOs we had were a direct mapping between DTO and model, so we had a `Dto` class with useful methods like `create()` and `update()`. However, we quickly realized we need more than that kind of DTO (see discussion in https://github.com/NIAEFEUP/website-niaefeup-backend/pull/54#discussion_r1038903546). The result was a few DTOs residing alongside the controller classes, which is not ideal. The discussion above proposes the following solution: ``` dto/ model/ auth/ otherSemanticUseCase/ ``` This `dto` package should either reside in the global scope of the application or inside the `controller` package, I'll leave that to discussion. It'd also be nice to consider cases of model DTOs that need more fields than the respective model's fields (e.g. creating a project with a list of account IDs to be used as its team).
True
global: refactor DTO structure - Until a while ago, all the DTOs we had were a direct mapping between DTO and model, so we had a `Dto` class with useful methods like `create()` and `update()`. However, we quickly realized we need more than that kind of DTO (see discussion in https://github.com/NIAEFEUP/website-niaefeup-backend/pull/54#discussion_r1038903546). The result was a few DTOs residing alongside the controller classes, which is not ideal. The discussion above proposes the following solution: ``` dto/ model/ auth/ otherSemanticUseCase/ ``` This `dto` package should either reside in the global scope of the application or inside the `controller` package, I'll leave that to discussion. It'd also be nice to consider cases of model DTOs that need more fields than the respective model's fields (e.g. creating a project with a list of account IDs to be used as its team).
main
global refactor dto structure until a while ago all the dtos we had were a direct mapping between dto and model so we had a dto class with useful methods like create and update however we quickly realized we need more than that kind of dto see discussion in the result was a few dtos residing alongside the controller classes which is not ideal the discussion above proposes the following solution dto model auth othersemanticusecase this dto package should either reside in the global scope of the application or inside the controller package i ll leave that to discussion it d also be nice to consider cases of model dtos that need more fields than the respective model s fields e g creating a project with a list of account ids to be used as its team
1
125,515
10,345,260,024
IssuesEvent
2019-09-04 13:07:23
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Typeguard auto complete not working as expected for union with error type
Area/Tooling BetaTesting Component/LanguageServer Priority/High Type/Bug
**Description:** ![Screenshot from 2019-08-15 12-02-07](https://user-images.githubusercontent.com/8677511/63077052-af41e480-bf54-11e9-8770-dd59ccdeff48.png) The type guard generated has an additional `else` keyword ```ballerina error? respond = caller->respond(); if (respond is error) { } else else { } ``` **Steps to reproduce:** Click on the type guard option on any variable of union type which includes only two types from which one of them is error type i.e.: type guard generation on `boolean|int|error x = true;` works but fails for `boolean|error x = true;` **Affected Versions:** Ballerina 1.0.0-beta-SNAPSHOT VSCode Version: 1.36.1
1.0
Typeguard auto complete not working as expected for union with error type - **Description:** ![Screenshot from 2019-08-15 12-02-07](https://user-images.githubusercontent.com/8677511/63077052-af41e480-bf54-11e9-8770-dd59ccdeff48.png) The type guard generated has an additional `else` keyword ```ballerina error? respond = caller->respond(); if (respond is error) { } else else { } ``` **Steps to reproduce:** Click on the type guard option on any variable of union type which includes only two types from which one of them is error type i.e.: type guard generation on `boolean|int|error x = true;` works but fails for `boolean|error x = true;` **Affected Versions:** Ballerina 1.0.0-beta-SNAPSHOT VSCode Version: 1.36.1
non_main
typeguard auto complete not working as expected for union with error type description the type guard generated has an additional else keyword ballerina error respond caller respond if respond is error else else steps to reproduce click on the type guard option on any variable of union type which includes only two types from which one of them is error type i e type guard generation on boolean int error x true works but fails for boolean error x true affected versions ballerina beta snapshot vscode version
0
1,903
6,577,556,021
IssuesEvent
2017-09-12 01:44:15
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
mount is not idempotent for bind mount (Ubuntu 14.04, ansible 2.0.1.0)
affects_2.0 bug_report waiting_on_maintainer
##### Issue Type: - Bug Report ##### Plugin Name: mount ##### Ansible Version: ``` ansible 2.0.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### Ansible Configuration: no changes from default ##### Environment: Ubuntu 14.04 control host and target host, but see below for exact steps to reproduce ##### Summary: Using the mount module to set up and mount a bind mount is not idempotent. ##### Steps To Reproduce: The following playbook illustrates the problem: ``` --- - hosts: all tasks: - file: path=/tmp/1 state=directory - file: path=/tmp/2 state=directory - mount: src=/tmp/1 name=/tmp/2 fstype=- opts=bind state=mounted ``` Here's a detailed recipe for reproducing this if you run into trouble. - Launch an instance in ec2 off the latest ubuntu trusty AMI. A t2.nano is sufficient - On the AMI: ``` sudo -s apt-get -y install software-properties-common apt-add-repository -y ppa:ansible/ansible apt-get update apt-get -y install ansible cd /tmp cat >a.yml <<EOF --- - hosts: all tasks: - file: path=/tmp/1 state=directory - file: path=/tmp/2 state=directory - mount: src=/tmp/1 name=/tmp/2 fstype=- opts=bind state=mounted EOF cat >local <<EOF [local] 127.0.0.1 ansible_connection=local EOF ansible-playbook -i local a.yml ansible-playbook -i local a.yml ``` ##### Expected Results: The second time, nothing should have changed since /tmp/1 is already mounted on /tmp/2 ##### Actual Results: The directory was mounted again. The line: ``` /tmp/1 on /tmp/2 type none (rw,bind) ``` appears twice in the output of `mount`.
True
mount is not idempotent for bind mount (Ubuntu 14.04, ansible 2.0.1.0) - ##### Issue Type: - Bug Report ##### Plugin Name: mount ##### Ansible Version: ``` ansible 2.0.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### Ansible Configuration: no changes from default ##### Environment: Ubuntu 14.04 control host and target host, but see below for exact steps to reproduce ##### Summary: Using the mount module to set up and mount a bind mount is not idempotent. ##### Steps To Reproduce: The following playbook illustrates the problem: ``` --- - hosts: all tasks: - file: path=/tmp/1 state=directory - file: path=/tmp/2 state=directory - mount: src=/tmp/1 name=/tmp/2 fstype=- opts=bind state=mounted ``` Here's a detailed recipe for reproducing this if you run into trouble. - Launch an instance in ec2 off the latest ubuntu trusty AMI. A t2.nano is sufficient - On the AMI: ``` sudo -s apt-get -y install software-properties-common apt-add-repository -y ppa:ansible/ansible apt-get update apt-get -y install ansible cd /tmp cat >a.yml <<EOF --- - hosts: all tasks: - file: path=/tmp/1 state=directory - file: path=/tmp/2 state=directory - mount: src=/tmp/1 name=/tmp/2 fstype=- opts=bind state=mounted EOF cat >local <<EOF [local] 127.0.0.1 ansible_connection=local EOF ansible-playbook -i local a.yml ansible-playbook -i local a.yml ``` ##### Expected Results: The second time, nothing should have changed since /tmp/1 is already mounted on /tmp/2 ##### Actual Results: The directory was mounted again. The line: ``` /tmp/1 on /tmp/2 type none (rw,bind) ``` appears twice in the output of `mount`.
main
mount is not idempotent for bind mount ubuntu ansible issue type bug report plugin name mount ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides ansible configuration no changes from default environment ubuntu control host and target host but see below for exact steps to reproduce summary using the mount module to set up and mount a bind mount is not idempotent steps to reproduce the following playbook illustrates the problem hosts all tasks file path tmp state directory file path tmp state directory mount src tmp name tmp fstype opts bind state mounted here s a detailed recipe for reproducing this if you run into trouble launch an instance in off the latest ubuntu trusty ami a nano is sufficient on the ami sudo s apt get y install software properties common apt add repository y ppa ansible ansible apt get update apt get y install ansible cd tmp cat a yml eof hosts all tasks file path tmp state directory file path tmp state directory mount src tmp name tmp fstype opts bind state mounted eof cat local eof ansible connection local eof ansible playbook i local a yml ansible playbook i local a yml expected results the second time nothing should have changed since tmp is already mounted on tmp actual results the directory was mounted again the line tmp on tmp type none rw bind appears twice in the output of mount
1
5,541
27,743,765,161
IssuesEvent
2023-03-15 15:46:52
cosmos/ibc-rs
https://api.github.com/repos/cosmos/ibc-rs
opened
[ICS02] Split `check_misbehaviour_and_update_state` and `check_header_and_update_state` into two sets of validation/execution functions
A: breaking I: logic O: maintainability O: modularity
<!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Problem Statement Part of #173, in continuation of implementing [IBC-go/ADR006](https://github.com/cosmos/ibc-go/blob/main/docs/architecture/adr-006-02-client-refactor.md) and getting `check_misbehaviour_and_update_state()` and `check_misbehaviour_and_update_state()` compatible with the Validation/Execution design ([ADR005](https://github.com/cosmos/ibc-rs/blob/main/docs/architecture/adr-005-handlers-redesign.md))
True
[ICS02] Split `check_misbehaviour_and_update_state` and `check_header_and_update_state` into two sets of validation/execution functions - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Problem Statement Part of #173, in continuation of implementing [IBC-go/ADR006](https://github.com/cosmos/ibc-go/blob/main/docs/architecture/adr-006-02-client-refactor.md) and getting `check_misbehaviour_and_update_state()` and `check_misbehaviour_and_update_state()` compatible with the Validation/Execution design ([ADR005](https://github.com/cosmos/ibc-rs/blob/main/docs/architecture/adr-005-handlers-redesign.md))
main
split check misbehaviour and update state and check header and update state into two sets of validation execution functions ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v please also ensure that this is not a duplicate issue ☺ problem statement part of in continuation of implementing and getting check misbehaviour and update state and check misbehaviour and update state compatible with the validation execution design
1
62,434
17,023,922,311
IssuesEvent
2021-07-03 04:34:26
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
OSM website Map Key is missing "cliff"
Component: website Priority: minor Resolution: wontfix Type: defect
**[Submitted to the original trac issue database at 10.41pm, Sunday, 21st June 2015]** The Map Key on www.openstreetmap.org/search is missing the "cliff" line type item. See screenshot for example: [[Image(http://i.imgur.com/aOUEkQz.png)]] You can view this map position [https://www.openstreetmap.org/search?query=san%20miguel%20de%20allende#map=16/20.8959/-100.7382 here].
1.0
OSM website Map Key is missing "cliff" - **[Submitted to the original trac issue database at 10.41pm, Sunday, 21st June 2015]** The Map Key on www.openstreetmap.org/search is missing the "cliff" line type item. See screenshot for example: [[Image(http://i.imgur.com/aOUEkQz.png)]] You can view this map position [https://www.openstreetmap.org/search?query=san%20miguel%20de%20allende#map=16/20.8959/-100.7382 here].
non_main
osm website map key is missing cliff the map key on is missing the cliff line type item see screenshot for example you can view this map position
0
744
4,349,990,219
IssuesEvent
2016-07-30 23:29:48
christoff-buerger/racr
https://api.github.com/repos/christoff-buerger/racr
closed
RACR-meta: separation and testing of annotation facilities
enhancement low maintainability user interface
To simplify the core implementation of _RACR_ and, in particular, clearly identify and elaborate the essential facilities required for meta-instantiation and reasoning (cf. issues #62 and #63 ), the existing abstract syntax tree annotation functionalities should become a separate _Scheme_ library, that can be loaded additionally to `(racr-meta core)` if required. Since unit tests for the existing annotation facilities are missing, they have to be added to ensure correct refactoring and improve code coverage.
True
RACR-meta: separation and testing of annotation facilities - To simplify the core implementation of _RACR_ and, in particular, clearly identify and elaborate the essential facilities required for meta-instantiation and reasoning (cf. issues #62 and #63 ), the existing abstract syntax tree annotation functionalities should become a separate _Scheme_ library, that can be loaded additionally to `(racr-meta core)` if required. Since unit tests for the existing annotation facilities are missing, they have to be added to ensure correct refactoring and improve code coverage.
main
racr meta separation and testing of annotation facilities to simplify the core implementation of racr and in particular clearly identify and elaborate the essential facilities required for meta instantiation and reasoning cf issues and the existing abstract syntax tree annotation functionalities should become a separate scheme library that can be loaded additionally to racr meta core if required since unit tests for the existing annotation facilities are missing they have to be added to ensure correct refactoring and improve code coverage
1
2,495
8,655,457,949
IssuesEvent
2018-11-27 16:00:18
codestation/qcma
https://api.github.com/repos/codestation/qcma
closed
[Issues] Disconnections
unmaintained
I'm using QCMA on Win 8.1 Pro with my Vita running on the latest official firmware Vita is connecting to PC via WiFi I'm getting disconnections in the following situation - Under Content Manager, I open Photos or Videos - Under Content Manager, I leave the Vita and QCMA to run overnight to back up all my apps (not a full backup, just the Vita games) any idea what's wrong? I can also provide a log - just need to know where to get that
True
[Issues] Disconnections - I'm using QCMA on Win 8.1 Pro with my Vita running on the latest official firmware Vita is connecting to PC via WiFi I'm getting disconnections in the following situation - Under Content Manager, I open Photos or Videos - Under Content Manager, I leave the Vita and QCMA to run overnight to back up all my apps (not a full backup, just the Vita games) any idea what's wrong? I can also provide a log - just need to know where to get that
main
disconnections i m using qcma on win pro with my vita running on the latest official firmware vita is connecting to pc via wifi i m getting disconnections in the following situation under content manager i open photos or videos under content manager i leave the vita and qcma to run overnight to back up all my apps not a full backup just the vita games any idea what s wrong i can also provide a log just need to know where to get that
1
622,413
19,634,682,162
IssuesEvent
2022-01-08 03:48:41
TASVideos/tasvideos
https://api.github.com/repos/TASVideos/tasvideos
closed
Editing the post there's 2 previews
enhancement Priority: Low
One preview is at the top and doesn't change/update, and another one appears at the bottom when you click the Preview button. I like that it appears at the bottom so you keep seeing your actual text above it, but I don't think the static preview at the top is needed.
1.0
Editing the post there's 2 previews - One preview is at the top and doesn't change/update, and another one appears at the bottom when you click the Preview button. I like that it appears at the bottom so you keep seeing your actual text above it, but I don't think the static preview at the top is needed.
non_main
editing the post there s previews one preview is at the top and doesn t change update and another one appears at the bottom when you click the preview button i like that it appears at the bottom so you keep seeing your actual text above it but i don t think the static preview at the top is needed
0
3,909
17,381,248,951
IssuesEvent
2021-07-31 19:07:19
dotnet/Silk.NET
https://api.github.com/repos/dotnet/Silk.NET
opened
Use ClangSharp's autogenerated bindings tests (i.e. for blittability, size, etc)
area-SilkTouch help wanted tenet-Maintainability
This is something that is left out of the initial pass of the SilkTouch Scraper because I (knowingly and idiotically) did not add support for tests in single-file XML dumps upstream in ClangSharp. The work is: - [ ] Investigate adding support upstream in ClangSharp for outputting test code where the XML output mode is used with a single file - currently, ClangSharp's OutputBuilderFactory is hardcoded to create a CSharpOutputBuilder. options are: - return a CSharpOutputBuilder that has C# code for stuff like `CreateTests` within the XML, so it outputs something like `<bindings><tests name="DXGI_OUTPUT_DESCTests"><code></code></tests></bindings>`. This would work nicely, but there may not be enough XML markers to allow us to change the name of the class under test, for example, - add to the XML the test "goals" i.e. `<bindings><test type="DXGI_OUTPUT_DESC" isSize32="32" isSize64="64" /></bindings>` - [ ] Recognise the new XML elements in the SilkTouch Scraper - [ ] Add required configuration structures, so that this functionality can be configured for each project - [ ] Output into the test project (the path of which indicated by the configuration structure) all the test code. cc @tannergooding
True
Use ClangSharp's autogenerated bindings tests (i.e. for blittability, size, etc) - This is something that is left out of the initial pass of the SilkTouch Scraper because I (knowingly and idiotically) did not add support for tests in single-file XML dumps upstream in ClangSharp. The work is: - [ ] Investigate adding support upstream in ClangSharp for outputting test code where the XML output mode is used with a single file - currently, ClangSharp's OutputBuilderFactory is hardcoded to create a CSharpOutputBuilder. options are: - return a CSharpOutputBuilder that has C# code for stuff like `CreateTests` within the XML, so it outputs something like `<bindings><tests name="DXGI_OUTPUT_DESCTests"><code></code></tests></bindings>`. This would work nicely, but there may not be enough XML markers to allow us to change the name of the class under test, for example, - add to the XML the test "goals" i.e. `<bindings><test type="DXGI_OUTPUT_DESC" isSize32="32" isSize64="64" /></bindings>` - [ ] Recognise the new XML elements in the SilkTouch Scraper - [ ] Add required configuration structures, so that this functionality can be configured for each project - [ ] Output into the test project (the path of which indicated by the configuration structure) all the test code. cc @tannergooding
main
use clangsharp s autogenerated bindings tests i e for blittability size etc this is something that is left out of the initial pass of the silktouch scraper because i knowingly and idiotically did not add support for tests in single file xml dumps upstream in clangsharp the work is investigate adding support upstream in clangsharp for outputting test code where the xml output mode is used with a single file currently clangsharp s outputbuilderfactory is hardcoded to create a csharpoutputbuilder options are return a csharpoutputbuilder that has c code for stuff like createtests within the xml so it outputs something like this would work nicely but there may not be enough xml markers to allow us to change the name of the class under test for example add to the xml the test goals i e recognise the new xml elements in the silktouch scraper add required configuration structures so that this functionality can be configured for each project output into the test project the path of which indicated by the configuration structure all the test code cc tannergooding
1
305,908
23,136,785,299
IssuesEvent
2022-07-28 14:54:17
eCapitalAdvisors/Incorta_templates
https://api.github.com/repos/eCapitalAdvisors/Incorta_templates
closed
Design Document
documentation
Identified data sources: https://www.chicagobooth.edu/research/kilts/datasets/dominicks - [x] Examine Dominick's Dataset New Direction: I will be using the data in Incorta from the PAC Dev Instance (OnlineStoreData) - [x] Understanding the schema structure of the dataset - [x] Identified Analytics: Using R Studio to write sparklyr syntax - [x] Identified software solution: Incorta
1.0
Design Document - Identified data sources: https://www.chicagobooth.edu/research/kilts/datasets/dominicks - [x] Examine Dominick's Dataset New Direction: I will be using the data in Incorta from the PAC Dev Instance (OnlineStoreData) - [x] Understanding the schema structure of the dataset - [x] Identified Analytics: Using R Studio to write sparklyr syntax - [x] Identified software solution: Incorta
non_main
design document identified data sources examine dominick s dataset new direction i will be using the data in incorta from the pac dev instance onlinestoredata understanding the schema structure of the dataset identified analytics using r studio to write sparklyr syntax identified software solution incorta
0
2,191
7,735,704,514
IssuesEvent
2018-05-27 18:02:30
Chromeroni/Hera-Chatbot
https://api.github.com/repos/Chromeroni/Hera-Chatbot
closed
Store log-files in a cloud
maintainability
Currently, log files are stored localy on the execution environment. This prevents easy access to the log files for all developers. **Change** Implement an interface to the google drive API, so that the daily created log files can automatically be stored in the developer cloud. **Prerequisite** Issue #16 & #27
True
Store log-files in a cloud - Currently, log files are stored localy on the execution environment. This prevents easy access to the log files for all developers. **Change** Implement an interface to the google drive API, so that the daily created log files can automatically be stored in the developer cloud. **Prerequisite** Issue #16 & #27
main
store log files in a cloud currently log files are stored localy on the execution environment this prevents easy access to the log files for all developers change implement an interface to the google drive api so that the daily created log files can automatically be stored in the developer cloud prerequisite issue
1
1,157
5,047,424,283
IssuesEvent
2016-12-20 09:23:18
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker_container with empty links list always restarts
affects_2.2 bug_report cloud docker waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.2.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION *none* ##### OS / ENVIRONMENT Ubuntu 16.04 ##### SUMMARY When using `links: []`, the container is always restarted. ##### STEPS TO REPRODUCE Run this playbook twice: ``` - hosts: localhost tasks: - docker_container: name: test image: alpine links: [] command: sleep 10000 ``` ##### EXPECTED RESULTS The container is started once. It is not restarted when running the second time. ##### ACTUAL RESULTS The container is restarted on the second run. When running with `-vvv --diff`: ``` TASK [docker_container] ******************************************************** [...] changed: [localhost] => { "ansible_facts": {}, "changed": true, "diff": { "differences": [ { "expected_links": { "container": null, "parameter": [] } } ] }, "invocation": { [...] ``` The problem seems to be that the modules compares an empty list with `None`. From `docker.log` (when enabled): ``` check differences expected_links [] vs None primitive compare: expected_links [...] differences [ { "expected_links": { "container": null, "parameter": [] } } ] ```
True
docker_container with empty links list always restarts - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.2.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION *none* ##### OS / ENVIRONMENT Ubuntu 16.04 ##### SUMMARY When using `links: []`, the container is always restarted. ##### STEPS TO REPRODUCE Run this playbook twice: ``` - hosts: localhost tasks: - docker_container: name: test image: alpine links: [] command: sleep 10000 ``` ##### EXPECTED RESULTS The container is started once. It is not restarted when running the second time. ##### ACTUAL RESULTS The container is restarted on the second run. When running with `-vvv --diff`: ``` TASK [docker_container] ******************************************************** [...] changed: [localhost] => { "ansible_facts": {}, "changed": true, "diff": { "differences": [ { "expected_links": { "container": null, "parameter": [] } } ] }, "invocation": { [...] ``` The problem seems to be that the modules compares an empty list with `None`. From `docker.log` (when enabled): ``` check differences expected_links [] vs None primitive compare: expected_links [...] differences [ { "expected_links": { "container": null, "parameter": [] } } ] ```
main
docker container with empty links list always restarts issue type bug report component name docker container ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration none os environment ubuntu summary when using links the container is always restarted steps to reproduce run this playbook twice hosts localhost tasks docker container name test image alpine links command sleep expected results the container is started once it is not restarted when running the second time actual results the container is restarted on the second run when running with vvv diff task changed ansible facts changed true diff differences expected links container null parameter invocation the problem seems to be that the modules compares an empty list with none from docker log when enabled check differences expected links vs none primitive compare expected links differences expected links container null parameter
1
5,160
26,274,240,629
IssuesEvent
2023-01-06 20:11:20
cosmos/ibc-rs
https://api.github.com/repos/cosmos/ibc-rs
closed
Add `cargo doc` check to CI
O: maintainability I: CI/CD
<!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Acceptance Criteria - CI checks to ensure that the `cargo doc` command has been successfully executed ____ ## For Admin Use - [ ] Not duplicate issue - [ ] Appropriate labels applied - [ ] Appropriate milestone (priority) applied - [ ] Appropriate contributors tagged - [ ] Contributor assigned/self-assigned
True
Add `cargo doc` check to CI - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Acceptance Criteria - CI checks to ensure that the `cargo doc` command has been successfully executed ____ ## For Admin Use - [ ] Not duplicate issue - [ ] Appropriate labels applied - [ ] Appropriate milestone (priority) applied - [ ] Appropriate contributors tagged - [ ] Contributor assigned/self-assigned
main
add cargo doc check to ci ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v please also ensure that this is not a duplicate issue ☺ acceptance criteria ci checks to ensure that the cargo doc command has been successfully executed for admin use not duplicate issue appropriate labels applied appropriate milestone priority applied appropriate contributors tagged contributor assigned self assigned
1
5,633
28,292,276,901
IssuesEvent
2023-04-09 11:09:40
WiredForWar/machines
https://api.github.com/repos/WiredForWar/machines
opened
Machines do self-destruct right on the start of Battle / Fortress / Siege in multiplayer
Bug Maintainance UI & UX Network
The scenario https://github.com/WiredForWar/machines/blob/dev/data/seige.scn#L27 works only if the start positions are fixed (rather than random). 1. Add `FIXED_POSITIONS` property to the scenarios format 2. Adjust the scenario file to indicate that it requires fixed positions. 3. Disable the positions combobox if only fixed positions are supported by the scn. 4. It seems that there is a scenarios cache file which is not invalidated on a scenario changed. Fix/implement the invalidation.
True
Machines do self-destruct right on the start of Battle / Fortress / Siege in multiplayer - The scenario https://github.com/WiredForWar/machines/blob/dev/data/seige.scn#L27 works only if the start positions are fixed (rather than random). 1. Add `FIXED_POSITIONS` property to the scenarios format 2. Adjust the scenario file to indicate that it requires fixed positions. 3. Disable the positions combobox if only fixed positions are supported by the scn. 4. It seems that there is a scenarios cache file which is not invalidated on a scenario changed. Fix/implement the invalidation.
main
machines do self destruct right on the start of battle fortress siege in multiplayer the scenario works only if the start positions are fixed rather than random add fixed positions property to the scenarios format adjust the scenario file to indicate that it requires fixed positions disable the positions combobox if only fixed positions are supported by the scn it seems that there is a scenarios cache file which is not invalidated on a scenario changed fix implement the invalidation
1
5,396
27,115,666,274
IssuesEvent
2023-02-15 18:21:50
VA-Explorer/va_explorer
https://api.github.com/repos/VA-Explorer/va_explorer
closed
Ensure that edited VAs are subjected to re-coding
Type: Maintainance good first issue Language: Python Domain: API/ Databases Status: Inactive
**What is the expected state?** If I change a VA, the CoD assigned or not assigned to it may also change. Since this is the case, I expect edited VAs to be considered by any code responsible for assigning CoD. **What is the actual state?** Edited VAs currently do not appear to be subject to CoD assignment unless they previously did not have a CoD to begin with (either due to errors or having never been considered by the CoD assignment system). **Relevant context** `va_explorer/va_data_management/utils/coding.py::run_coding_algorithms()`
True
Ensure that edited VAs are subjected to re-coding - **What is the expected state?** If I change a VA, the CoD assigned or not assigned to it may also change. Since this is the case, I expect edited VAs to be considered by any code responsible for assigning CoD. **What is the actual state?** Edited VAs currently do not appear to be subject to CoD assignment unless they previously did not have a CoD to begin with (either due to errors or having never been considered by the CoD assignment system). **Relevant context** `va_explorer/va_data_management/utils/coding.py::run_coding_algorithms()`
main
ensure that edited vas are subjected to re coding what is the expected state if i change a va the cod assigned or not assigned to it may also change since this is the case i expect edited vas to be considered by any code responsible for assigning cod what is the actual state edited vas currently do not appear to be subject to cod assignment unless they previously did not have a cod to begin with either due to errors or having never been considered by the cod assignment system relevant context va explorer va data management utils coding py run coding algorithms
1
644
4,158,721,236
IssuesEvent
2016-06-17 05:00:19
Microsoft/DirectXTK
https://api.github.com/repos/Microsoft/DirectXTK
opened
Raise minimum supported Feature Level to 9.3
maintainence
The DirectX Tool Kit has supported all feature levels 9.1 and up, primarily to support Surface RT and Surface RT 2 devices. Since these devices will not be getting upgraded to Windows 10, and given the fact that the vast majority of PCs have 10.0 or better hardware, we will be retiring support for FL 9.1. There are almost no 9.2 devices of relevance today, but Windows phone 8.1 support requires FL 9.3.
True
Raise minimum supported Feature Level to 9.3 - The DirectX Tool Kit has supported all feature levels 9.1 and up, primarily to support Surface RT and Surface RT 2 devices. Since these devices will not be getting upgraded to Windows 10, and given the fact that the vast majority of PCs have 10.0 or better hardware, we will be retiring support for FL 9.1. There are almost no 9.2 devices of relevance today, but Windows phone 8.1 support requires FL 9.3.
main
raise minimum supported feature level to the directx tool kit has supported all feature levels and up primarily to support surface rt and surface rt devices since these devices will not be getting upgraded to windows and given the fact that the vast majority of pcs have or better hardware we will be retiring support for fl there are almost no devices of relevance today but windows phone support requires fl
1
5,327
26,903,221,819
IssuesEvent
2023-02-06 17:03:11
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Add logged in user's name in the app header user icon dropdown
work: frontend status: ready restricted: maintainers
<img width="251" alt="Screenshot 2022-12-07 at 3 36 44 AM" src="https://user-images.githubusercontent.com/11032856/206033602-d55d998d-a2b6-4135-94ad-934be54fc588.png">
True
Add logged in user's name in the app header user icon dropdown - <img width="251" alt="Screenshot 2022-12-07 at 3 36 44 AM" src="https://user-images.githubusercontent.com/11032856/206033602-d55d998d-a2b6-4135-94ad-934be54fc588.png">
main
add logged in user s name in the app header user icon dropdown img width alt screenshot at am src
1
88,046
3,770,904,325
IssuesEvent
2016-03-16 15:58:17
Vannevelj/VSDiagnostics
https://api.github.com/repos/Vannevelj/VSDiagnostics
opened
Implement LINQUsed
priority - medium type - feature
In performance-sensitive code, teams might sometimes decide to not allow any LINQ (known for its many allocations). We can write an analyzer that basically warns anytime LINQ is used. We'll have to do some thinking about how we address this: do we just check `using` statements? Do we search for exact occurrences after we've done that? This should be disabled by default (users have to explicitly opt-in since only a very small subset will use this).
1.0
Implement LINQUsed - In performance-sensitive code, teams might sometimes decide to not allow any LINQ (known for its many allocations). We can write an analyzer that basically warns anytime LINQ is used. We'll have to do some thinking about how we address this: do we just check `using` statements? Do we search for exact occurrences after we've done that? This should be disabled by default (users have to explicitly opt-in since only a very small subset will use this).
non_main
implement linqused in performance sensitive code teams might sometimes decide to not allow any linq known for its many allocations we can write an analyzer that basically warns anytime linq is used we ll have to do some thinking about how we address this do we just check using statements do we search for exact occurrences after we ve done that this should be disabled by default users have to explicitly opt in since only a very small subset will use this
0
4,033
18,853,026,066
IssuesEvent
2021-11-12 00:09:52
exercism/python
https://api.github.com/repos/exercism/python
opened
[Templates]: Create New Concept Document Issue Template
maintainer chore 🔧
We should create an new issue template for [Concept Documents], similar to this issue https://github.com/exercism/python/issues/2456, but with `details` sections that can be expanded/collapsed.
True
[Templates]: Create New Concept Document Issue Template - We should create an new issue template for [Concept Documents], similar to this issue https://github.com/exercism/python/issues/2456, but with `details` sections that can be expanded/collapsed.
main
create new concept document issue template we should create an new issue template for similar to this issue but with details sections that can be expanded collapsed
1
2,879
10,319,547,782
IssuesEvent
2019-08-30 17:50:12
backdrop-ops/contrib
https://api.github.com/repos/backdrop-ops/contrib
closed
Port of Facebook OAuth (FBOAuth)
Maintainer application Port in progress
Hi there. I recently became maintainer of FBOAuth module, thanks to Nate. I want to port it to Backdrop.
True
Port of Facebook OAuth (FBOAuth) - Hi there. I recently became maintainer of FBOAuth module, thanks to Nate. I want to port it to Backdrop.
main
port of facebook oauth fboauth hi there i recently became maintainer of fboauth module thanks to nate i want to port it to backdrop
1
106,480
13,303,751,987
IssuesEvent
2020-08-25 15:56:05
LibraryOfCongress/concordia
https://api.github.com/repos/LibraryOfCongress/concordia
closed
Create user flows on reviewing transcriptions for specific personas
design
**User story/persona** - High touch user, particularly intrigued by reviewing or wanting to participate in a challenge - A student given an assignment - Transcribe-a-thon participant **Is your feature request related to a problem? Please describe.** At this point, the team has some ideas on how a user comes to BTP and the paths they take. Mapping of potential user's journey can help the product team gain more understanding of users motives and make better design decisions. **Acceptance Criteria** - [ ] Create user flows for each persona listed above
1.0
Create user flows on reviewing transcriptions for specific personas - **User story/persona** - High touch user, particularly intrigued by reviewing or wanting to participate in a challenge - A student given an assignment - Transcribe-a-thon participant **Is your feature request related to a problem? Please describe.** At this point, the team has some ideas on how a user comes to BTP and the paths they take. Mapping of potential user's journey can help the product team gain more understanding of users motives and make better design decisions. **Acceptance Criteria** - [ ] Create user flows for each persona listed above
non_main
create user flows on reviewing transcriptions for specific personas user story persona high touch user particularly intrigued by reviewing or wanting to participate in a challenge a student given an assignment transcribe a thon participant is your feature request related to a problem please describe at this point the team has some ideas on how a user comes to btp and the paths they take mapping of potential user s journey can help the product team gain more understanding of users motives and make better design decisions acceptance criteria create user flows for each persona listed above
0
701,804
24,109,100,070
IssuesEvent
2022-09-20 09:51:08
younginnovations/iatipublisher
https://api.github.com/repos/younginnovations/iatipublisher
closed
Bug : Drop Down Issue
type: bug priority: high Frontend Backend
Context - Desktop - Chrome 103.0.5060.114 Preconditions - stage.iatipublisher.yipl.com.np - Username: Ram - password: 12345678 Steps - Select the drop-down options Actual Result - remove the icon is missing from the drop-down options Expected Result - remove the icon should be in the drop-down options https://user-images.githubusercontent.com/78422663/188442470-e3bb2ba7-67a8-4a45-bf68-27d76ebcc1b7.mp4
1.0
Bug : Drop Down Issue - Context - Desktop - Chrome 103.0.5060.114 Preconditions - stage.iatipublisher.yipl.com.np - Username: Ram - password: 12345678 Steps - Select the drop-down options Actual Result - remove the icon is missing from the drop-down options Expected Result - remove the icon should be in the drop-down options https://user-images.githubusercontent.com/78422663/188442470-e3bb2ba7-67a8-4a45-bf68-27d76ebcc1b7.mp4
non_main
bug drop down issue context desktop chrome preconditions stage iatipublisher yipl com np username ram password steps select the drop down options actual result remove the icon is missing from the drop down options expected result remove the icon should be in the drop down options
0
884
4,543,548,386
IssuesEvent
2016-09-10 06:13:57
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker_container module always recreate container if combine hostname and network_mode: host
affects_2.1 bug_report cloud docker waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT N/A ##### SUMMARY docker_container module always recreate container if use network_mode: host and specify hostname. If you remove network_mode or hostname it will work fine. ##### STEPS TO REPRODUCE ``` - name: Simple start container hosts: localhost connection: local gather_facts: no tasks: - name: Create simple container with network host docker_container: image: alpine name: test_container hostname: test_container network_mode: host command: sleep 9999 state: started ``` ##### EXPECTED RESULTS When you rerun playbook. The container shouldn't be removed ##### ACTUAL RESULTS ``` [WARNING]: Host file not found: /etc/ansible/hosts [WARNING]: provided hosts list is empty, only localhost is available Loaded callback default of type stdout, v2.0 PLAYBOOK: tmp.yml ************************************************************** 1 plays in tmp.yml PLAY [Simple start container] ************************************************** TASK [Create simple container with network host] ******************************* task path: /home/username/ansible/tmp.yml:18 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: username <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1473347657.57-15375112345113 `" && echo ansible-tmp-1473347657.57-15375112345113="` echo $HOME/.ansible/tmp/ansible-tmp-1473347657.57-15375112345113 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpUnye2A TO /home/username/.ansible/tmp/ansible-tmp-1473347657.57-15375112345113/docker_container <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /home/username/.virtualenvs/ansible/bin/python /home/username/.ansible/tmp/ansible-tmp-1473347657.57-15375112345113/docker_container; rm -rf "/home/username/.ansible/tmp/ansible-tmp-1473347657.57-15375112345113/" > /dev/null 2>&1 && sleep 0' changed: [localhost] => {"ansible_facts": {"ansible_docker_container": {"AppArmorProfile": "", "Args": ["9999"], "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["sleep", "9999"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "WingE7240", "Image": "alpine", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-09-08T15:14:27.822665992Z", "Driver": "overlay2", "ExecIDs": null, "GraphDriver": {"Data": {"LowerDir": "/var/lib/docker/overlay2/68f64c2a47e62dab0d198b83d2da78b39bbc2486c33a1bda9702ea7a5a21b34f-init/diff:/var/lib/docker/overlay2/608bee2a4daf3cf045e7a9005e647763a6b9f57abdd4d06051a2cf5a892b046e/diff", "MergedDir": "/var/lib/docker/overlay2/68f64c2a47e62dab0d198b83d2da78b39bbc2486c33a1bda9702ea7a5a21b34f/merged", "UpperDir": "/var/lib/docker/overlay2/68f64c2a47e62dab0d198b83d2da78b39bbc2486c33a1bda9702ea7a5a21b34f/diff", "WorkDir": "/var/lib/docker/overlay2/68f64c2a47e62dab0d198b83d2da78b39bbc2486c33a1bda9702ea7a5a21b34f/work"}, "Name": "overlay2"}, "HostConfig": {"AutoRemove": false, "Binds": [], "BlkioDeviceReadBps": null, "BlkioDeviceReadIOps": null, "BlkioDeviceWriteBps": null, "BlkioDeviceWriteIOps": null, "BlkioWeight": 0, "BlkioWeightDevice": null, "CapAdd": null, "CapDrop": null, "Cgroup": "", "CgroupParent": "", "ConsoleSize": [0, 0], "ContainerIDFile": "", "CpuCount": 0, "CpuPercent": 0, "CpuPeriod": 0, "CpuQuota": 0, "CpuShares": 0, "CpusetCpus": "", "CpusetMems": "", "Devices": null, "DiskQuota": 0, "Dns": null, "DnsOptions": null, "DnsSearch": null, "ExtraHosts": null, "GroupAdd": null, "IOMaximumBandwidth": 0, "IOMaximumIOps": 0, "IpcMode": "", "Isolation": "", "KernelMemory": 0, "Links": null, "LogConfig": {"Config": {}, "Type": "json-file"}, "Memory": 0, "MemoryReservation": 0, "MemorySwap": 0, "MemorySwappiness": -1, "NetworkMode": "host", "OomKillDisable": false, "OomScoreAdj": 0, "PidMode": "", "PidsLimit": 0, "PortBindings": null, "Privileged": false, "PublishAllPorts": false, "ReadonlyRootfs": false, "RestartPolicy": {"MaximumRetryCount": 0, "Name": ""}, "Runtime": "runc", "SecurityOpt": null, "ShmSize": 67108864, "UTSMode": "", "Ulimits": null, "UsernsMode": "", "VolumeDriver": "", "VolumesFrom": null}, "HostnamePath": "/var/lib/docker/containers/34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976/hostname", "HostsPath": "/var/lib/docker/containers/34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976/hosts", "Id": "34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976", "Image": "sha256:4e38e38c8ce0b8d9041a9c4fefe786631d1416225e13b0bfe8cfa2321aec4bba", "LogPath": "/var/lib/docker/containers/34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976/34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976-json.log", "MountLabel": "", "Mounts": [], "Name": "/test_container", "NetworkSettings": {"Bridge": "", "EndpointID": "", "Gateway": "", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "HairpinMode": false, "IPAddress": "", "IPPrefixLen": 0, "IPv6Gateway": "", "LinkLocalIPv6Address": "", "LinkLocalIPv6PrefixLen": 0, "MacAddress": "", "Networks": {"host": {"Aliases": null, "EndpointID": "310addbf23897a7abd3a39bdd294b76b885b632a7bc245f6c08eeaee7ef62025", "Gateway": "", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "IPAMConfig": null, "IPAddress": "", "IPPrefixLen": 0, "IPv6Gateway": "", "Links": null, "MacAddress": "", "NetworkID": "b300f4a912b723de6d372aa11f5671dff3c6b59386ae0dcc5f14b0f719fa34f2"}}, "Ports": {}, "SandboxID": "f9306b73778b3651e5eef39cf3829b927425b3ad4baf60249404fc17af1eb52b", "SandboxKey": "/var/run/docker/netns/default", "SecondaryIPAddresses": null, "SecondaryIPv6Addresses": null}, "Path": "sleep", "ProcessLabel": "", "ResolvConfPath": "/var/lib/docker/containers/34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976/resolv.conf", "RestartCount": 0, "State": {"Dead": false, "Error": "", "ExitCode": 0, "FinishedAt": "0001-01-01T00:00:00Z", "OOMKilled": false, "Paused": false, "Pid": 23171, "Restarting": false, "Running": true, "StartedAt": "2016-09-08T15:14:28.030424162Z", "Status": "running"}}}, "changed": true, "invocation": {"module_args": {"api_version": null, "blkio_weight": null, "cacert_path": null, "capabilities": null, "cert_path": null, "cleanup": false, "command": "sleep 9999", "cpu_period": null, "cpu_quota": null, "cpu_shares": null, "cpuset_cpus": null, "cpuset_mems": null, "debug": false, "detach": true, "devices": null, "dns_opts": null, "dns_search_domains": null, "dns_servers": null, "docker_host": null, "entrypoint": null, "env": null, "env_file": null, "etc_hosts": null, "exposed_ports": null, "filter_logger": false, "force_kill": false, "groups": null, "hostname": "test_container", "image": "alpine", "interactive": false, "ipc_mode": null, "keep_volumes": true, "kernel_memory": null, "key_path": null, "kill_signal": null, "labels": null, "links": null, "log_driver": "json-file", "log_options": null, "mac_address": null, "memory": "0", "memory_reservation": null, "memory_swap": null, "memory_swappiness": null, "name": "test_container", "network_mode": "host", "networks": null, "oom_killer": null, "paused": false, "pid_mode": null, "privileged": false, "published_ports": null, "pull": false, "purge_networks": null, "read_only": false, "recreate": false, "restart": false, "restart_policy": null, "restart_retries": 0, "security_opts": null, "shm_size": null, "ssl_version": null, "state": "started", "stop_signal": null, "stop_timeout": null, "timeout": null, "tls": null, "tls_hostname": null, "tls_verify": null, "trust_image_content": false, "tty": false, "ulimits": null, "user": null, "uts": null, "volume_driver": null, "volumes": null, "volumes_from": null}, "module_name": "docker_container"}} PLAY RECAP ********************************************************************* localhost : ok=1 changed=1 unreachable=0 failed=0 TASK: Create simple container with network host ------------------------ 10.65s ```
True
docker_container module always recreate container if combine hostname and network_mode: host - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT N/A ##### SUMMARY docker_container module always recreate container if use network_mode: host and specify hostname. If you remove network_mode or hostname it will work fine. ##### STEPS TO REPRODUCE ``` - name: Simple start container hosts: localhost connection: local gather_facts: no tasks: - name: Create simple container with network host docker_container: image: alpine name: test_container hostname: test_container network_mode: host command: sleep 9999 state: started ``` ##### EXPECTED RESULTS When you rerun playbook. The container shouldn't be removed ##### ACTUAL RESULTS ``` [WARNING]: Host file not found: /etc/ansible/hosts [WARNING]: provided hosts list is empty, only localhost is available Loaded callback default of type stdout, v2.0 PLAYBOOK: tmp.yml ************************************************************** 1 plays in tmp.yml PLAY [Simple start container] ************************************************** TASK [Create simple container with network host] ******************************* task path: /home/username/ansible/tmp.yml:18 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: username <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1473347657.57-15375112345113 `" && echo ansible-tmp-1473347657.57-15375112345113="` echo $HOME/.ansible/tmp/ansible-tmp-1473347657.57-15375112345113 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpUnye2A TO /home/username/.ansible/tmp/ansible-tmp-1473347657.57-15375112345113/docker_container <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /home/username/.virtualenvs/ansible/bin/python /home/username/.ansible/tmp/ansible-tmp-1473347657.57-15375112345113/docker_container; rm -rf "/home/username/.ansible/tmp/ansible-tmp-1473347657.57-15375112345113/" > /dev/null 2>&1 && sleep 0' changed: [localhost] => {"ansible_facts": {"ansible_docker_container": {"AppArmorProfile": "", "Args": ["9999"], "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["sleep", "9999"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "WingE7240", "Image": "alpine", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-09-08T15:14:27.822665992Z", "Driver": "overlay2", "ExecIDs": null, "GraphDriver": {"Data": {"LowerDir": "/var/lib/docker/overlay2/68f64c2a47e62dab0d198b83d2da78b39bbc2486c33a1bda9702ea7a5a21b34f-init/diff:/var/lib/docker/overlay2/608bee2a4daf3cf045e7a9005e647763a6b9f57abdd4d06051a2cf5a892b046e/diff", "MergedDir": "/var/lib/docker/overlay2/68f64c2a47e62dab0d198b83d2da78b39bbc2486c33a1bda9702ea7a5a21b34f/merged", "UpperDir": "/var/lib/docker/overlay2/68f64c2a47e62dab0d198b83d2da78b39bbc2486c33a1bda9702ea7a5a21b34f/diff", "WorkDir": "/var/lib/docker/overlay2/68f64c2a47e62dab0d198b83d2da78b39bbc2486c33a1bda9702ea7a5a21b34f/work"}, "Name": "overlay2"}, "HostConfig": {"AutoRemove": false, "Binds": [], "BlkioDeviceReadBps": null, "BlkioDeviceReadIOps": null, "BlkioDeviceWriteBps": null, "BlkioDeviceWriteIOps": null, "BlkioWeight": 0, "BlkioWeightDevice": null, "CapAdd": null, "CapDrop": null, "Cgroup": "", "CgroupParent": "", "ConsoleSize": [0, 0], "ContainerIDFile": "", "CpuCount": 0, "CpuPercent": 0, "CpuPeriod": 0, "CpuQuota": 0, "CpuShares": 0, "CpusetCpus": "", "CpusetMems": "", "Devices": null, "DiskQuota": 0, "Dns": null, "DnsOptions": null, "DnsSearch": null, "ExtraHosts": null, "GroupAdd": null, "IOMaximumBandwidth": 0, "IOMaximumIOps": 0, "IpcMode": "", "Isolation": "", "KernelMemory": 0, "Links": null, "LogConfig": {"Config": {}, "Type": "json-file"}, "Memory": 0, "MemoryReservation": 0, "MemorySwap": 0, "MemorySwappiness": -1, "NetworkMode": "host", "OomKillDisable": false, "OomScoreAdj": 0, "PidMode": "", "PidsLimit": 0, "PortBindings": null, "Privileged": false, "PublishAllPorts": false, "ReadonlyRootfs": false, "RestartPolicy": {"MaximumRetryCount": 0, "Name": ""}, "Runtime": "runc", "SecurityOpt": null, "ShmSize": 67108864, "UTSMode": "", "Ulimits": null, "UsernsMode": "", "VolumeDriver": "", "VolumesFrom": null}, "HostnamePath": "/var/lib/docker/containers/34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976/hostname", "HostsPath": "/var/lib/docker/containers/34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976/hosts", "Id": "34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976", "Image": "sha256:4e38e38c8ce0b8d9041a9c4fefe786631d1416225e13b0bfe8cfa2321aec4bba", "LogPath": "/var/lib/docker/containers/34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976/34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976-json.log", "MountLabel": "", "Mounts": [], "Name": "/test_container", "NetworkSettings": {"Bridge": "", "EndpointID": "", "Gateway": "", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "HairpinMode": false, "IPAddress": "", "IPPrefixLen": 0, "IPv6Gateway": "", "LinkLocalIPv6Address": "", "LinkLocalIPv6PrefixLen": 0, "MacAddress": "", "Networks": {"host": {"Aliases": null, "EndpointID": "310addbf23897a7abd3a39bdd294b76b885b632a7bc245f6c08eeaee7ef62025", "Gateway": "", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "IPAMConfig": null, "IPAddress": "", "IPPrefixLen": 0, "IPv6Gateway": "", "Links": null, "MacAddress": "", "NetworkID": "b300f4a912b723de6d372aa11f5671dff3c6b59386ae0dcc5f14b0f719fa34f2"}}, "Ports": {}, "SandboxID": "f9306b73778b3651e5eef39cf3829b927425b3ad4baf60249404fc17af1eb52b", "SandboxKey": "/var/run/docker/netns/default", "SecondaryIPAddresses": null, "SecondaryIPv6Addresses": null}, "Path": "sleep", "ProcessLabel": "", "ResolvConfPath": "/var/lib/docker/containers/34f613253024563f630cba84c94a8d5f7e8b06a4bfc2e76ade99497afbe2c976/resolv.conf", "RestartCount": 0, "State": {"Dead": false, "Error": "", "ExitCode": 0, "FinishedAt": "0001-01-01T00:00:00Z", "OOMKilled": false, "Paused": false, "Pid": 23171, "Restarting": false, "Running": true, "StartedAt": "2016-09-08T15:14:28.030424162Z", "Status": "running"}}}, "changed": true, "invocation": {"module_args": {"api_version": null, "blkio_weight": null, "cacert_path": null, "capabilities": null, "cert_path": null, "cleanup": false, "command": "sleep 9999", "cpu_period": null, "cpu_quota": null, "cpu_shares": null, "cpuset_cpus": null, "cpuset_mems": null, "debug": false, "detach": true, "devices": null, "dns_opts": null, "dns_search_domains": null, "dns_servers": null, "docker_host": null, "entrypoint": null, "env": null, "env_file": null, "etc_hosts": null, "exposed_ports": null, "filter_logger": false, "force_kill": false, "groups": null, "hostname": "test_container", "image": "alpine", "interactive": false, "ipc_mode": null, "keep_volumes": true, "kernel_memory": null, "key_path": null, "kill_signal": null, "labels": null, "links": null, "log_driver": "json-file", "log_options": null, "mac_address": null, "memory": "0", "memory_reservation": null, "memory_swap": null, "memory_swappiness": null, "name": "test_container", "network_mode": "host", "networks": null, "oom_killer": null, "paused": false, "pid_mode": null, "privileged": false, "published_ports": null, "pull": false, "purge_networks": null, "read_only": false, "recreate": false, "restart": false, "restart_policy": null, "restart_retries": 0, "security_opts": null, "shm_size": null, "ssl_version": null, "state": "started", "stop_signal": null, "stop_timeout": null, "timeout": null, "tls": null, "tls_hostname": null, "tls_verify": null, "trust_image_content": false, "tty": false, "ulimits": null, "user": null, "uts": null, "volume_driver": null, "volumes": null, "volumes_from": null}, "module_name": "docker_container"}} PLAY RECAP ********************************************************************* localhost : ok=1 changed=1 unreachable=0 failed=0 TASK: Create simple container with network host ------------------------ 10.65s ```
main
docker container module always recreate container if combine hostname and network mode host issue type bug report component name docker container ansible version ansible configuration n a os environment n a summary docker container module always recreate container if use network mode host and specify hostname if you remove network mode or hostname it will work fine steps to reproduce name simple start container hosts localhost connection local gather facts no tasks name create simple container with network host docker container image alpine name test container hostname test container network mode host command sleep state started expected results when you rerun playbook the container shouldn t be removed actual results host file not found etc ansible hosts provided hosts list is empty only localhost is available loaded callback default of type stdout playbook tmp yml plays in tmp yml play task task path home username ansible tmp yml establish local connection for user username exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home username ansible tmp ansible tmp docker container exec bin sh c lang en us utf lc all en us utf lc messages en us utf home username virtualenvs ansible bin python home username ansible tmp ansible tmp docker container rm rf home username ansible tmp ansible tmp dev null sleep changed ansible facts ansible docker container apparmorprofile args config attachstderr false attachstdin false attachstdout false cmd domainname entrypoint null env hostname image alpine labels onbuild null openstdin false stdinonce false tty false user volumes null workingdir created driver execids null graphdriver data lowerdir var lib docker init diff var lib docker diff mergeddir var lib docker merged upperdir var lib docker diff workdir var lib docker work name hostconfig autoremove false binds blkiodevicereadbps null blkiodevicereadiops null blkiodevicewritebps null blkiodevicewriteiops null blkioweight blkioweightdevice null capadd null capdrop null cgroup cgroupparent consolesize containeridfile cpucount cpupercent cpuperiod cpuquota cpushares cpusetcpus cpusetmems devices null diskquota dns null dnsoptions null dnssearch null extrahosts null groupadd null iomaximumbandwidth iomaximumiops ipcmode isolation kernelmemory links null logconfig config type json file memory memoryreservation memoryswap memoryswappiness networkmode host oomkilldisable false oomscoreadj pidmode pidslimit portbindings null privileged false publishallports false readonlyrootfs false restartpolicy maximumretrycount name runtime runc securityopt null shmsize utsmode ulimits null usernsmode volumedriver volumesfrom null hostnamepath var lib docker containers hostname hostspath var lib docker containers hosts id image logpath var lib docker containers json log mountlabel mounts name test container networksettings bridge endpointid gateway hairpinmode false ipaddress ipprefixlen macaddress networks host aliases null endpointid gateway ipamconfig null ipaddress ipprefixlen links null macaddress networkid ports sandboxid sandboxkey var run docker netns default secondaryipaddresses null null path sleep processlabel resolvconfpath var lib docker containers resolv conf restartcount state dead false error exitcode finishedat oomkilled false paused false pid restarting false running true startedat status running changed true invocation module args api version null blkio weight null cacert path null capabilities null cert path null cleanup false command sleep cpu period null cpu quota null cpu shares null cpuset cpus null cpuset mems null debug false detach true devices null dns opts null dns search domains null dns servers null docker host null entrypoint null env null env file null etc hosts null exposed ports null filter logger false force kill false groups null hostname test container image alpine interactive false ipc mode null keep volumes true kernel memory null key path null kill signal null labels null links null log driver json file log options null mac address null memory memory reservation null memory swap null memory swappiness null name test container network mode host networks null oom killer null paused false pid mode null privileged false published ports null pull false purge networks null read only false recreate false restart false restart policy null restart retries security opts null shm size null ssl version null state started stop signal null stop timeout null timeout null tls null tls hostname null tls verify null trust image content false tty false ulimits null user null uts null volume driver null volumes null volumes from null module name docker container play recap localhost ok changed unreachable failed task create simple container with network host
1
280,624
21,313,844,988
IssuesEvent
2022-04-16 01:08:22
dankamongmen/notcurses
https://api.github.com/repos/dankamongmen/notcurses
opened
Support "passthrough" printing
documentation enhancement
Sometimes I have pre-colored strings I want to print. It would be nice if notcurses supported parsing and converting those strings so I don't have to. Today, the first one vanishes, and the second one print out in blue: ```c ncplane_putstr(n, "\x1b[34ma ═ nccolor\x1b[0m\n"); //... printf("\x1b[34ma ═ directcolor\x1b[0m\n"); ```
1.0
Support "passthrough" printing - Sometimes I have pre-colored strings I want to print. It would be nice if notcurses supported parsing and converting those strings so I don't have to. Today, the first one vanishes, and the second one print out in blue: ```c ncplane_putstr(n, "\x1b[34ma ═ nccolor\x1b[0m\n"); //... printf("\x1b[34ma ═ directcolor\x1b[0m\n"); ```
non_main
support passthrough printing sometimes i have pre colored strings i want to print it would be nice if notcurses supported parsing and converting those strings so i don t have to today the first one vanishes and the second one print out in blue c ncplane putstr n ═ nccolor n printf ═ directcolor n
0
21,466
11,660,239,703
IssuesEvent
2020-03-03 02:39:50
cityofaustin/atd-mobility-project-database
https://api.github.com/repos/cityofaustin/atd-mobility-project-database
closed
Discuss GIS <---> Knack integration
Project: Mobility Project Database Service: PM Type: Meeting Workgroup: AMD Workgroup: ATSD
@amenity commented on [Fri Jun 07 2019](https://github.com/cityofaustin/atd-data-tech/issues/185) ### Objective For MoPed, we'll need to be able to edit features in a GIS application. ### Participants Jaime, John, Nathan, Cole, Amenity ### Agenda <!--- Add agenda inline here or link to an online document. ---> ------ - [ ] Schedule meeting - [ ] Meet - [ ] Attach meeting notes to this issue - [ ] Create resulting issues
1.0
Discuss GIS <---> Knack integration - @amenity commented on [Fri Jun 07 2019](https://github.com/cityofaustin/atd-data-tech/issues/185) ### Objective For MoPed, we'll need to be able to edit features in a GIS application. ### Participants Jaime, John, Nathan, Cole, Amenity ### Agenda <!--- Add agenda inline here or link to an online document. ---> ------ - [ ] Schedule meeting - [ ] Meet - [ ] Attach meeting notes to this issue - [ ] Create resulting issues
non_main
discuss gis knack integration amenity commented on objective for moped we ll need to be able to edit features in a gis application participants jaime john nathan cole amenity agenda schedule meeting meet attach meeting notes to this issue create resulting issues
0
74,546
15,353,864,029
IssuesEvent
2021-03-01 09:06:38
AlexRogalskiy/github-action-open-jscharts
https://api.github.com/repos/AlexRogalskiy/github-action-open-jscharts
opened
CVE-2020-28499 (High) detected in merge-1.2.1.tgz
security vulnerability
## CVE-2020-28499 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>merge-1.2.1.tgz</b></p></summary> <p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p> <p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p> <p>Path to dependency file: github-action-open-jscharts/package.json</p> <p>Path to vulnerable library: github-action-open-jscharts/node_modules/merge/package.json</p> <p> Dependency Hierarchy: - cz-conventional-changelog-3.3.0.tgz (Root Library) - commitizen-4.2.3.tgz - find-node-modules-2.0.0.tgz - :x: **merge-1.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-open-jscharts/commit/5de582648dd07b6e4340db81e5c597889ca61c68">5de582648dd07b6e4340db81e5c597889ca61c68</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge . <p>Publish Date: 2021-02-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-28499 (High) detected in merge-1.2.1.tgz - ## CVE-2020-28499 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>merge-1.2.1.tgz</b></p></summary> <p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p> <p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.1.tgz">https://registry.npmjs.org/merge/-/merge-1.2.1.tgz</a></p> <p>Path to dependency file: github-action-open-jscharts/package.json</p> <p>Path to vulnerable library: github-action-open-jscharts/node_modules/merge/package.json</p> <p> Dependency Hierarchy: - cz-conventional-changelog-3.3.0.tgz (Root Library) - commitizen-4.2.3.tgz - find-node-modules-2.0.0.tgz - :x: **merge-1.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-open-jscharts/commit/5de582648dd07b6e4340db81e5c597889ca61c68">5de582648dd07b6e4340db81e5c597889ca61c68</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge . <p>Publish Date: 2021-02-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in merge tgz cve high severity vulnerability vulnerable library merge tgz merge multiple objects into one optionally creating a new cloned object similar to the jquery extend but more flexible works in node js and the browser library home page a href path to dependency file github action open jscharts package json path to vulnerable library github action open jscharts node modules merge package json dependency hierarchy cz conventional changelog tgz root library commitizen tgz find node modules tgz x merge tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package merge are vulnerable to prototype pollution via recursivemerge publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with whitesource
0
2,031
6,793,189,634
IssuesEvent
2017-11-01 05:54:59
cannawen/metric_units_reddit_bot
https://api.github.com/repos/cannawen/metric_units_reddit_bot
closed
Move documentation to /docs folder
chore first timers only hacktoberfest in progress maintainer approved
[metric_units](https://www.reddit.com/user/metric_units) is a sassy reddit bot that finds imperial units, and replies with a metric conversion. ## First timers only This issue is reserved for anyone who has **never** made a pull request to Open Source Software. If you are not a first timer, we would still love your help! Please see our other open issues :) [Read our New to OSS guide](https://github.com/cannawen/metric_units_reddit_bot/blob/master/NEW-TO-OSS.md) for an overview of what to expect ### IMPORTANT: Comment below if you would like to volunteer to tackle this issue. --- ## Recommended experience - Some familiarity with basic markdown syntax - No programming required to complete this task ## Time estimate 30-60 minutes ## The problem Our project is getting cluttered with all this documentation at the top-level directory. Let's do some cleanup, and move all documentation into some subdirectories! ## Instructions Ready? Let's get started! 1. Fork this repository ([detailed instructions](https://guides.github.com/activities/forking/#fork)) 2. Clone your fork ([detailed instructions](https://guides.github.com/activities/forking/#clone)) 3. Open the source code folder, and create two new folder called "docs" and ".github" 4. Move "CONTRIBUTING.md" "ISSUE_TEMPLATE.md", and "PULL_REQUEST_TEMPLATE.md" into the `.github` directory ([see why here](https://github.com/blog/2111-issue-and-pull-request-templates)) 4. Move all other documentation into the "docs" folder, except for the README. 5. Go through all the documentation, and update any broken links (for example, the README must point to `./docs/CONTRIBUTING.md` instead of `./CONSTRIBUTING.md`) 6. Commit the changes with a descriptive git commit message 7. Push your changes back to github, and click around the documentation starting from `https://github.com/{{your-github-username}}/metric_units_reddit_bot` to make sure your links work 8. Make a pull request back to the main repository ([detailed instructions](https://guides.github.com/activities/forking/#making-a-pull-request)) OPTIONAL: Don't forget to [give yourself credit](https://github.com/cannawen/metric_units_reddit_bot/blob/master/CONTRIBUTING.md#add-yourself-to-the-contributors-list)! Thank you for contributing to metric_units bot! 9. Wait for your PR to be reviewed by a maintainer, and address any comments they may have ## Step 10: YOUR CHANGE GETS MERGED!! Party!!!
True
Move documentation to /docs folder - [metric_units](https://www.reddit.com/user/metric_units) is a sassy reddit bot that finds imperial units, and replies with a metric conversion. ## First timers only This issue is reserved for anyone who has **never** made a pull request to Open Source Software. If you are not a first timer, we would still love your help! Please see our other open issues :) [Read our New to OSS guide](https://github.com/cannawen/metric_units_reddit_bot/blob/master/NEW-TO-OSS.md) for an overview of what to expect ### IMPORTANT: Comment below if you would like to volunteer to tackle this issue. --- ## Recommended experience - Some familiarity with basic markdown syntax - No programming required to complete this task ## Time estimate 30-60 minutes ## The problem Our project is getting cluttered with all this documentation at the top-level directory. Let's do some cleanup, and move all documentation into some subdirectories! ## Instructions Ready? Let's get started! 1. Fork this repository ([detailed instructions](https://guides.github.com/activities/forking/#fork)) 2. Clone your fork ([detailed instructions](https://guides.github.com/activities/forking/#clone)) 3. Open the source code folder, and create two new folder called "docs" and ".github" 4. Move "CONTRIBUTING.md" "ISSUE_TEMPLATE.md", and "PULL_REQUEST_TEMPLATE.md" into the `.github` directory ([see why here](https://github.com/blog/2111-issue-and-pull-request-templates)) 4. Move all other documentation into the "docs" folder, except for the README. 5. Go through all the documentation, and update any broken links (for example, the README must point to `./docs/CONTRIBUTING.md` instead of `./CONSTRIBUTING.md`) 6. Commit the changes with a descriptive git commit message 7. Push your changes back to github, and click around the documentation starting from `https://github.com/{{your-github-username}}/metric_units_reddit_bot` to make sure your links work 8. Make a pull request back to the main repository ([detailed instructions](https://guides.github.com/activities/forking/#making-a-pull-request)) OPTIONAL: Don't forget to [give yourself credit](https://github.com/cannawen/metric_units_reddit_bot/blob/master/CONTRIBUTING.md#add-yourself-to-the-contributors-list)! Thank you for contributing to metric_units bot! 9. Wait for your PR to be reviewed by a maintainer, and address any comments they may have ## Step 10: YOUR CHANGE GETS MERGED!! Party!!!
main
move documentation to docs folder is a sassy reddit bot that finds imperial units and replies with a metric conversion first timers only this issue is reserved for anyone who has never made a pull request to open source software if you are not a first timer we would still love your help please see our other open issues for an overview of what to expect important comment below if you would like to volunteer to tackle this issue recommended experience some familiarity with basic markdown syntax no programming required to complete this task time estimate minutes the problem our project is getting cluttered with all this documentation at the top level directory let s do some cleanup and move all documentation into some subdirectories instructions ready let s get started fork this repository clone your fork open the source code folder and create two new folder called docs and github move contributing md issue template md and pull request template md into the github directory move all other documentation into the docs folder except for the readme go through all the documentation and update any broken links for example the readme must point to docs contributing md instead of constributing md commit the changes with a descriptive git commit message push your changes back to github and click around the documentation starting from to make sure your links work make a pull request back to the main repository optional don t forget to thank you for contributing to metric units bot wait for your pr to be reviewed by a maintainer and address any comments they may have step your change gets merged party
1
344,610
30,751,873,225
IssuesEvent
2023-07-28 20:06:16
saltstack/salt
https://api.github.com/repos/saltstack/salt
opened
[Increase Test Coverage] Batch 35
Tests
Increase the code coverage percent on the following files to at least 80%. Please be aware that currently the percentage might be inaccurate if the module uses salt due to #64696 File | Percent salt/modules/ps.py 51 salt/modules/slsutil.py 71 salt/modules/textfsm_mod.py 13 salt/states/http.py 70 salt/beacons/aix_account.py 25
1.0
[Increase Test Coverage] Batch 35 - Increase the code coverage percent on the following files to at least 80%. Please be aware that currently the percentage might be inaccurate if the module uses salt due to #64696 File | Percent salt/modules/ps.py 51 salt/modules/slsutil.py 71 salt/modules/textfsm_mod.py 13 salt/states/http.py 70 salt/beacons/aix_account.py 25
non_main
batch increase the code coverage percent on the following files to at least please be aware that currently the percentage might be inaccurate if the module uses salt due to file percent salt modules ps py salt modules slsutil py salt modules textfsm mod py salt states http py salt beacons aix account py
0
1,678
6,574,117,695
IssuesEvent
2017-09-11 11:33:59
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Support docker --sysctls options in docker_container
affects_2.3 bug_report cloud docker waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME cloud/docker_container ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.3.0 config file = /etc/ansible/ansible.cfg configured module search path = ['/home/kassian/bundle/ansible/core', '/home/kassian/bundle/ansible/extras'] ``` ##### SUMMARY docker has new '--sysctl=' options with 1.12 release, docker-py also came with sysctls options in release 1.10, we should add this feature in docker module
True
Support docker --sysctls options in docker_container - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME cloud/docker_container ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.3.0 config file = /etc/ansible/ansible.cfg configured module search path = ['/home/kassian/bundle/ansible/core', '/home/kassian/bundle/ansible/extras'] ``` ##### SUMMARY docker has new '--sysctl=' options with 1.12 release, docker-py also came with sysctls options in release 1.10, we should add this feature in docker module
main
support docker sysctls options in docker container issue type bug report component name cloud docker container ansible version ansible config file etc ansible ansible cfg configured module search path summary docker has new sysctl options with release docker py also came with sysctls options in release we should add this feature in docker module
1
758,993
26,576,383,845
IssuesEvent
2023-01-21 21:46:57
ApplETS/Notre-Dame
https://api.github.com/repos/ApplETS/Notre-Dame
closed
Recreate CI CD pipelines
enhancement CI priority: medium ready to develop
**Is your feature request related to a problem? Please describe.** As a developer, the CI and CD pipelines contain too much conditions depending on the execution context. - Creating pre-release vs release builds - Lint execution and version increment **Describe the solution you'd like** - [ ] Create development workflow - [ ] Create master workflow **Describe alternatives you've considered** N/A **Additional context** N/A
1.0
Recreate CI CD pipelines - **Is your feature request related to a problem? Please describe.** As a developer, the CI and CD pipelines contain too much conditions depending on the execution context. - Creating pre-release vs release builds - Lint execution and version increment **Describe the solution you'd like** - [ ] Create development workflow - [ ] Create master workflow **Describe alternatives you've considered** N/A **Additional context** N/A
non_main
recreate ci cd pipelines is your feature request related to a problem please describe as a developer the ci and cd pipelines contain too much conditions depending on the execution context creating pre release vs release builds lint execution and version increment describe the solution you d like create development workflow create master workflow describe alternatives you ve considered n a additional context n a
0
683,072
23,367,631,212
IssuesEvent
2022-08-10 16:43:53
ASSETS-Conference/assets2022
https://api.github.com/repos/ASSETS-Conference/assets2022
closed
Adding table of content for long pages
medium priority
Potentially using https://tscanlin.github.io/tocbot/ for the long pages. Look into it and check if it is accessible.
1.0
Adding table of content for long pages - Potentially using https://tscanlin.github.io/tocbot/ for the long pages. Look into it and check if it is accessible.
non_main
adding table of content for long pages potentially using for the long pages look into it and check if it is accessible
0
1,072
4,890,830,236
IssuesEvent
2016-11-18 15:03:12
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
systemd fails on masking nonexistent service
affects_2.2 bug_report in progress waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME systemd ##### ANSIBLE VERSION ``` ansible 2.2.0.0 ``` ##### CONFIGURATION Default ##### OS / ENVIRONMENT Debian Jessie 64 GNU/Linux ##### SUMMARY Module systemd fails to mask nonexistent service. Shell command `systemctl mask hello` works fine with nonexistent services. It is useful to mask service before package installation to avoid autostart, e.g.: ``` systemctl mask redis-server apt-get install redis-server systemctl unmask redis-server ``` ##### STEPS TO REPRODUCE ``` ansible -m systemd -a 'name=hello masked=True' localhost -s ``` ##### EXPECTED RESULTS As in systemctl command: ``` $ systemctl mask hello Created symlink from /etc/systemd/system/hello.service to /dev/null. ``` ##### ACTUAL RESULTS ``` localhost | FAILED! => { "changed": false, "failed": true, "msg": "Could not find the requested service \"'hello'\": " } ```
True
systemd fails on masking nonexistent service - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME systemd ##### ANSIBLE VERSION ``` ansible 2.2.0.0 ``` ##### CONFIGURATION Default ##### OS / ENVIRONMENT Debian Jessie 64 GNU/Linux ##### SUMMARY Module systemd fails to mask nonexistent service. Shell command `systemctl mask hello` works fine with nonexistent services. It is useful to mask service before package installation to avoid autostart, e.g.: ``` systemctl mask redis-server apt-get install redis-server systemctl unmask redis-server ``` ##### STEPS TO REPRODUCE ``` ansible -m systemd -a 'name=hello masked=True' localhost -s ``` ##### EXPECTED RESULTS As in systemctl command: ``` $ systemctl mask hello Created symlink from /etc/systemd/system/hello.service to /dev/null. ``` ##### ACTUAL RESULTS ``` localhost | FAILED! => { "changed": false, "failed": true, "msg": "Could not find the requested service \"'hello'\": " } ```
main
systemd fails on masking nonexistent service issue type bug report component name systemd ansible version ansible configuration default os environment debian jessie gnu linux summary module systemd fails to mask nonexistent service shell command systemctl mask hello works fine with nonexistent services it is useful to mask service before package installation to avoid autostart e g systemctl mask redis server apt get install redis server systemctl unmask redis server steps to reproduce ansible m systemd a name hello masked true localhost s expected results as in systemctl command systemctl mask hello created symlink from etc systemd system hello service to dev null actual results localhost failed changed false failed true msg could not find the requested service hello
1
313,255
26,913,466,038
IssuesEvent
2023-02-07 03:06:37
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
[wasi] `System.Collections.Tests` failing with `System.BadImageFormatException : Bad method token 0x1b00008f`
arch-wasm test-failure wasi
[Build](https://dev.azure.com/dnceng-public/public/_build/results?buildId=160649&view=logs&j=f02b8cf9-dd4d-54fc-c292-2bb1d305b019&t=fd9015e5-cd94-57d7-8286-4fe749355863&l=55), and [log](https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-76143faa4dbf4bd38b/System.Collections.Tests/1/console.fcb9b94b.log?helixlogtype=result): ``` [08:32:36] info: Starting: System.Collections.Tests.dll [08:34:40] fail: [FAIL] System.Collections.Generic.Tests.NullableUInt16EnumComparersTests.Comparer_IComparerCompareWithObjectsNotOfMatchingTypeShouldThrow [08:34:40] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:40] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:40] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:40] fail: [FAIL] System.Collections.Generic.Tests.NullableUInt16EnumComparersTests.EqualityComparer_IEqualityComparerGetHashCodeWithObjectNotOfMatchingType [08:34:40] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:40] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:40] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:40] fail: [FAIL] System.Collections.Generic.Tests.NullableUInt16EnumComparersTests.EqualityComparer_IEqualityComparerEqualsWithObjectsNotOfMatchingType [08:34:40] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:40] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:40] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:43] fail: [FAIL] System.Collections.Generic.Tests.NullableSByteEnumComparersTests.Comparer_IComparerCompareWithObjectsNotOfMatchingTypeShouldThrow [08:34:43] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:43] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:43] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:43] fail: [FAIL] System.Collections.Generic.Tests.NullableSByteEnumComparersTests.EqualityComparer_IEqualityComparerGetHashCodeWithObjectNotOfMatchingType [08:34:43] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:43] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:43] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:44] fail: [FAIL] System.Collections.Generic.Tests.NullableSByteEnumComparersTests.EqualityComparer_IEqualityComparerEqualsWithObjectsNotOfMatchingType [08:34:44] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:44] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:44] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:44] info: Finished: System.Collections.Tests.dll ```
1.0
[wasi] `System.Collections.Tests` failing with `System.BadImageFormatException : Bad method token 0x1b00008f` - [Build](https://dev.azure.com/dnceng-public/public/_build/results?buildId=160649&view=logs&j=f02b8cf9-dd4d-54fc-c292-2bb1d305b019&t=fd9015e5-cd94-57d7-8286-4fe749355863&l=55), and [log](https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-76143faa4dbf4bd38b/System.Collections.Tests/1/console.fcb9b94b.log?helixlogtype=result): ``` [08:32:36] info: Starting: System.Collections.Tests.dll [08:34:40] fail: [FAIL] System.Collections.Generic.Tests.NullableUInt16EnumComparersTests.Comparer_IComparerCompareWithObjectsNotOfMatchingTypeShouldThrow [08:34:40] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:40] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:40] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:40] fail: [FAIL] System.Collections.Generic.Tests.NullableUInt16EnumComparersTests.EqualityComparer_IEqualityComparerGetHashCodeWithObjectNotOfMatchingType [08:34:40] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:40] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:40] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:40] fail: [FAIL] System.Collections.Generic.Tests.NullableUInt16EnumComparersTests.EqualityComparer_IEqualityComparerEqualsWithObjectsNotOfMatchingType [08:34:40] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:40] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:40] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:43] fail: [FAIL] System.Collections.Generic.Tests.NullableSByteEnumComparersTests.Comparer_IComparerCompareWithObjectsNotOfMatchingTypeShouldThrow [08:34:43] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:43] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:43] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:43] fail: [FAIL] System.Collections.Generic.Tests.NullableSByteEnumComparersTests.EqualityComparer_IEqualityComparerGetHashCodeWithObjectNotOfMatchingType [08:34:43] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:43] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:43] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:44] fail: [FAIL] System.Collections.Generic.Tests.NullableSByteEnumComparersTests.EqualityComparer_IEqualityComparerEqualsWithObjectsNotOfMatchingType [08:34:44] info: System.BadImageFormatException : Bad method token 0x1b00008f. [08:34:44] info: at System.Reflection.MethodInvoker.InterpretedInvoke(Object , IntPtr* ) [08:34:44] info: at System.Reflection.MethodInvoker.Invoke(Object , IntPtr* , BindingFlags ) [08:34:44] info: Finished: System.Collections.Tests.dll ```
non_main
system collections tests failing with system badimageformatexception bad method token and info starting system collections tests dll fail system collections generic tests comparer icomparercomparewithobjectsnotofmatchingtypeshouldthrow info system badimageformatexception bad method token info at system reflection methodinvoker interpretedinvoke object intptr info at system reflection methodinvoker invoke object intptr bindingflags fail system collections generic tests equalitycomparer iequalitycomparergethashcodewithobjectnotofmatchingtype info system badimageformatexception bad method token info at system reflection methodinvoker interpretedinvoke object intptr info at system reflection methodinvoker invoke object intptr bindingflags fail system collections generic tests equalitycomparer iequalitycomparerequalswithobjectsnotofmatchingtype info system badimageformatexception bad method token info at system reflection methodinvoker interpretedinvoke object intptr info at system reflection methodinvoker invoke object intptr bindingflags fail system collections generic tests nullablesbyteenumcomparerstests comparer icomparercomparewithobjectsnotofmatchingtypeshouldthrow info system badimageformatexception bad method token info at system reflection methodinvoker interpretedinvoke object intptr info at system reflection methodinvoker invoke object intptr bindingflags fail system collections generic tests nullablesbyteenumcomparerstests equalitycomparer iequalitycomparergethashcodewithobjectnotofmatchingtype info system badimageformatexception bad method token info at system reflection methodinvoker interpretedinvoke object intptr info at system reflection methodinvoker invoke object intptr bindingflags fail system collections generic tests nullablesbyteenumcomparerstests equalitycomparer iequalitycomparerequalswithobjectsnotofmatchingtype info system badimageformatexception bad method token info at system reflection methodinvoker interpretedinvoke object intptr info at system reflection methodinvoker invoke object intptr bindingflags info finished system collections tests dll
0
269,428
20,381,933,731
IssuesEvent
2022-02-21 23:26:50
s-merritt/NotWordle
https://api.github.com/repos/s-merritt/NotWordle
opened
Add Code Documentation and Doxygen
documentation
I'd like to become more familiar with doxygen and how to use it for future projects, so let's see if we can get that added to this project. This will also force me to write up the code comments for everything.
1.0
Add Code Documentation and Doxygen - I'd like to become more familiar with doxygen and how to use it for future projects, so let's see if we can get that added to this project. This will also force me to write up the code comments for everything.
non_main
add code documentation and doxygen i d like to become more familiar with doxygen and how to use it for future projects so let s see if we can get that added to this project this will also force me to write up the code comments for everything
0
259,184
8,189,355,411
IssuesEvent
2018-08-30 07:09:33
kubeflow/tf-operator
https://api.github.com/repos/kubeflow/tf-operator
closed
Backwards compatibility support "Master" as chief
api/v1alpha2 priority/p1 release/0.3.0
For backwards compatibility with older versions of TF.estimator we need to treat replicas with name "Master" as "Chief". One place where this doesn't seem to be happening is the exit policy. If there is a replica of type, "Master" than TFJob should be marked completed when the "Master" exits. It would be good to add an E2E test for this case. We should probably add validation to ensure only "Master" or "Chief" is provided. @ScorpioCPH Would you be willing to pick this up?
1.0
Backwards compatibility support "Master" as chief - For backwards compatibility with older versions of TF.estimator we need to treat replicas with name "Master" as "Chief". One place where this doesn't seem to be happening is the exit policy. If there is a replica of type, "Master" than TFJob should be marked completed when the "Master" exits. It would be good to add an E2E test for this case. We should probably add validation to ensure only "Master" or "Chief" is provided. @ScorpioCPH Would you be willing to pick this up?
non_main
backwards compatibility support master as chief for backwards compatibility with older versions of tf estimator we need to treat replicas with name master as chief one place where this doesn t seem to be happening is the exit policy if there is a replica of type master than tfjob should be marked completed when the master exits it would be good to add an test for this case we should probably add validation to ensure only master or chief is provided scorpiocph would you be willing to pick this up
0
34,559
9,412,174,293
IssuesEvent
2019-04-10 02:49:52
habitat-sh/builder
https://api.github.com/repos/habitat-sh/builder
closed
How to make an existing origin private
A-builder C-feature V-bldr
Does Builder support the ability to switch origins to private visibility after it has been created as public visibility? If an origin goes private, does that mean any user/client has to be part of that origin to gain at least read access to the packages?
1.0
How to make an existing origin private - Does Builder support the ability to switch origins to private visibility after it has been created as public visibility? If an origin goes private, does that mean any user/client has to be part of that origin to gain at least read access to the packages?
non_main
how to make an existing origin private does builder support the ability to switch origins to private visibility after it has been created as public visibility if an origin goes private does that mean any user client has to be part of that origin to gain at least read access to the packages
0
206,651
7,114,554,672
IssuesEvent
2018-01-18 01:23:04
tgockel/zookeeper-cpp
https://api.github.com/repos/tgockel/zookeeper-cpp
closed
Calling server::shutdown will always crash
bug lib/server priority/high
If `server::shutdown` is called with `true`, the `std::thread` will be joined again in `~server`, which will always throw an `std::invalid_argument` leading to an abort, since `~server` is `noexcept`. ``` #0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51 #1 0x00007ffff36993aa in __GI_abort () at abort.c:89 #2 0x00007ffff3fd9095 in __gnu_cxx::__verbose_terminate_handler() () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #3 0x00007ffff3fd6c86 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #4 0x00007ffff3fd5b89 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #5 0x00007ffff3fd6578 in __gxx_personality_v0 () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #6 0x00007ffff4b9a9bf in __libunwind_Unwind_RaiseException () from /usr/lib/x86_64-linux-gnu/libunwind.so.8 #7 0x00007ffff3fd6f07 in __cxa_throw () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #8 0x00007ffff40030be in std::__throw_system_error(int) () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #9 0x00007ffff40033cc in std::thread::join() () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #10 0x00007ffff6e7ee09 in zk::server::server::shutdown (this=<optimized out>, wait_for_stop=<optimized out>) at //root/zookeeper-cpp/src/zk/server/server.cpp:32 #11 0x00007ffff6e7ee1e in zk::server::server::~server (this=0x1002307c0, __in_chrg=<optimized out>) at //root/zookeeper-cpp/src/zk/server/server.cpp:19 ... ```
1.0
Calling server::shutdown will always crash - If `server::shutdown` is called with `true`, the `std::thread` will be joined again in `~server`, which will always throw an `std::invalid_argument` leading to an abort, since `~server` is `noexcept`. ``` #0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51 #1 0x00007ffff36993aa in __GI_abort () at abort.c:89 #2 0x00007ffff3fd9095 in __gnu_cxx::__verbose_terminate_handler() () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #3 0x00007ffff3fd6c86 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #4 0x00007ffff3fd5b89 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #5 0x00007ffff3fd6578 in __gxx_personality_v0 () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #6 0x00007ffff4b9a9bf in __libunwind_Unwind_RaiseException () from /usr/lib/x86_64-linux-gnu/libunwind.so.8 #7 0x00007ffff3fd6f07 in __cxa_throw () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #8 0x00007ffff40030be in std::__throw_system_error(int) () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #9 0x00007ffff40033cc in std::thread::join() () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6 #10 0x00007ffff6e7ee09 in zk::server::server::shutdown (this=<optimized out>, wait_for_stop=<optimized out>) at //root/zookeeper-cpp/src/zk/server/server.cpp:32 #11 0x00007ffff6e7ee1e in zk::server::server::~server (this=0x1002307c0, __in_chrg=<optimized out>) at //root/zookeeper-cpp/src/zk/server/server.cpp:19 ... ```
non_main
calling server shutdown will always crash if server shutdown is called with true the std thread will be joined again in server which will always throw an std invalid argument leading to an abort since server is noexcept gi raise sig sig entry at sysdeps unix sysv linux raise c in gi abort at abort c in gnu cxx verbose terminate handler from usr lib linux gnu libstdc so in from usr lib linux gnu libstdc so in from usr lib linux gnu libstdc so in gxx personality from usr lib linux gnu libstdc so in libunwind unwind raiseexception from usr lib linux gnu libunwind so in cxa throw from usr lib linux gnu libstdc so in std throw system error int from usr lib linux gnu libstdc so in std thread join from usr lib linux gnu libstdc so in zk server server shutdown this wait for stop at root zookeeper cpp src zk server server cpp in zk server server server this in chrg at root zookeeper cpp src zk server server cpp
0
154,253
5,916,657,908
IssuesEvent
2017-05-22 11:07:45
DOAJ/doaj
https://api.github.com/repos/DOAJ/doaj
closed
Libraries IP information
feedback high priority
Dear CL Do we have this info? where can I get it? Hello. I am the Director at the Tom Green County Library and we have had a change in IP address. Please provide the information you currently have on file for our IPs. Thank you.
1.0
Libraries IP information - Dear CL Do we have this info? where can I get it? Hello. I am the Director at the Tom Green County Library and we have had a change in IP address. Please provide the information you currently have on file for our IPs. Thank you.
non_main
libraries ip information dear cl do we have this info where can i get it hello i am the director at the tom green county library and we have had a change in ip address please provide the information you currently have on file for our ips thank you
0
1,932
6,609,874,302
IssuesEvent
2017-09-19 15:49:27
Kristinita/Erics-Green-Room
https://api.github.com/repos/Kristinita/Erics-Green-Room
closed
[Feature request] Пауза
need-maintainer
### 1. Запрос Было бы неплохо, если б отыгрыш пакета в комнатах возможно было бы ставить на паузу. ### 2. Аргументация Игрока могут отвлекать различные факторы. От просьбы что-то сделать по дому до срочных сообщений в соцсетях. Сейчас игрок, который хочет узнать свой объективный результат, но которого отвлекли во время отыгрыша пакета, вынужден заново начинать тренировка. ### 3. Пример реализации Игрок запустил какой-либо пакет. Вводит команду `!пауза` → игра приостанавливается. Вводит команду `!продолжим` → игра возобновляется. Можно сделать и следующее: если после того, как команда `!пауза` введена прошло, например, 15 минут, игра останавливается и комната закрывается. ### 4. Примеры в других приложениях Данная функция реализована в [**DragTrainer**](https://dvorkin.by/dragtrainer/). Спасибо.
True
[Feature request] Пауза - ### 1. Запрос Было бы неплохо, если б отыгрыш пакета в комнатах возможно было бы ставить на паузу. ### 2. Аргументация Игрока могут отвлекать различные факторы. От просьбы что-то сделать по дому до срочных сообщений в соцсетях. Сейчас игрок, который хочет узнать свой объективный результат, но которого отвлекли во время отыгрыша пакета, вынужден заново начинать тренировка. ### 3. Пример реализации Игрок запустил какой-либо пакет. Вводит команду `!пауза` → игра приостанавливается. Вводит команду `!продолжим` → игра возобновляется. Можно сделать и следующее: если после того, как команда `!пауза` введена прошло, например, 15 минут, игра останавливается и комната закрывается. ### 4. Примеры в других приложениях Данная функция реализована в [**DragTrainer**](https://dvorkin.by/dragtrainer/). Спасибо.
main
пауза запрос было бы неплохо если б отыгрыш пакета в комнатах возможно было бы ставить на паузу аргументация игрока могут отвлекать различные факторы от просьбы что то сделать по дому до срочных сообщений в соцсетях сейчас игрок который хочет узнать свой объективный результат но которого отвлекли во время отыгрыша пакета вынужден заново начинать тренировка пример реализации игрок запустил какой либо пакет вводит команду пауза → игра приостанавливается вводит команду продолжим → игра возобновляется можно сделать и следующее если после того как команда пауза введена прошло например минут игра останавливается и комната закрывается примеры в других приложениях данная функция реализована в спасибо
1
337,691
30,257,124,995
IssuesEvent
2023-07-07 04:22:07
rusefi/hellen154hyundai
https://api.github.com/repos/rusefi/hellen154hyundai
closed
Smiley face "Daniel" board now "3.6-1"
rev-C-hall-drama waiting-for-test
Some traces are FUBAR we would have to try the other fix on this board? ![WhatsApp Image 2023-06-30 at 14 53 03](https://github.com/rusefi/hellen154hyundai/assets/48498823/070a1353-a40a-4393-a039-e246d0ba4ddf)
1.0
Smiley face "Daniel" board now "3.6-1" - Some traces are FUBAR we would have to try the other fix on this board? ![WhatsApp Image 2023-06-30 at 14 53 03](https://github.com/rusefi/hellen154hyundai/assets/48498823/070a1353-a40a-4393-a039-e246d0ba4ddf)
non_main
smiley face daniel board now some traces are fubar we would have to try the other fix on this board
0
193,038
15,365,617,510
IssuesEvent
2021-03-01 23:57:27
spree/spree
https://api.github.com/repos/spree/spree
closed
Active Storage - CloudFlare - Image Caching & CDN
Documentation :book:
This is not really a bug with Spree, more of an issue with image caching and using a CDN for images served through Rails 5.2 Active Storage, but I figured it might be useful to have a documented solution on this issue for other Spree users. ## Issue Active Storage is simple to set up and get working, but offers no out of the box solution for serving your public facing images through a content delivery network or browser caching, with images having a signed link that expires after 5 minutes, this is great for secure files and download paid for content links, but not ideal for real world usage when serving website product images. ## Solution This solution uses CloudFlare for the CDN and caching. ## Work Around With Spree using active_storage/representations to display images you can simply monkey patch the following file in Rails 5.2.1: ```rails-5.2.1/activestorage/app/controllers/active_storage/representations_controller.rb``` With the following code ``` # frozen_string_literal: true # Take a signed permanent reference for a blob representation and turn it into an expiring service URL for download. # Note: These URLs are publicly accessible. If you need to enforce access protection beyond the # security-through-obscurity factor of the signed blob and variation reference, you'll need to implement your own # authenticated redirection controller. class ActiveStorage::RepresentationsController < ActiveStorage::BaseController include ActiveStorage::SetBlob def show expires_in 5.year, public: true variant = @blob.representation(params[:variation_key]).processed send_data @blob.service.download(variant.key), type: @blob.content_type || DEFAULT_SEND_FILE_TYPE, disposition: 'inline' end end ``` This will then give you Active Storage images served with a public accessible file with a 5 year expire date, if you are running your new Spree app through CloudFlare you can now change your active_storage/* page rule from bypass cache to: <img width="799" alt="cf" src="https://user-images.githubusercontent.com/1240766/46920404-a2c3c700-cfe5-11e8-84c7-2b569fa59f98.png"> And that should do it, your files will be served through the CloudFlare CDN with browser caching. Please be aware that this fix may be outdated as Rails updates past version 5.2.1
1.0
Active Storage - CloudFlare - Image Caching & CDN - This is not really a bug with Spree, more of an issue with image caching and using a CDN for images served through Rails 5.2 Active Storage, but I figured it might be useful to have a documented solution on this issue for other Spree users. ## Issue Active Storage is simple to set up and get working, but offers no out of the box solution for serving your public facing images through a content delivery network or browser caching, with images having a signed link that expires after 5 minutes, this is great for secure files and download paid for content links, but not ideal for real world usage when serving website product images. ## Solution This solution uses CloudFlare for the CDN and caching. ## Work Around With Spree using active_storage/representations to display images you can simply monkey patch the following file in Rails 5.2.1: ```rails-5.2.1/activestorage/app/controllers/active_storage/representations_controller.rb``` With the following code ``` # frozen_string_literal: true # Take a signed permanent reference for a blob representation and turn it into an expiring service URL for download. # Note: These URLs are publicly accessible. If you need to enforce access protection beyond the # security-through-obscurity factor of the signed blob and variation reference, you'll need to implement your own # authenticated redirection controller. class ActiveStorage::RepresentationsController < ActiveStorage::BaseController include ActiveStorage::SetBlob def show expires_in 5.year, public: true variant = @blob.representation(params[:variation_key]).processed send_data @blob.service.download(variant.key), type: @blob.content_type || DEFAULT_SEND_FILE_TYPE, disposition: 'inline' end end ``` This will then give you Active Storage images served with a public accessible file with a 5 year expire date, if you are running your new Spree app through CloudFlare you can now change your active_storage/* page rule from bypass cache to: <img width="799" alt="cf" src="https://user-images.githubusercontent.com/1240766/46920404-a2c3c700-cfe5-11e8-84c7-2b569fa59f98.png"> And that should do it, your files will be served through the CloudFlare CDN with browser caching. Please be aware that this fix may be outdated as Rails updates past version 5.2.1
non_main
active storage cloudflare image caching cdn this is not really a bug with spree more of an issue with image caching and using a cdn for images served through rails active storage but i figured it might be useful to have a documented solution on this issue for other spree users issue active storage is simple to set up and get working but offers no out of the box solution for serving your public facing images through a content delivery network or browser caching with images having a signed link that expires after minutes this is great for secure files and download paid for content links but not ideal for real world usage when serving website product images solution this solution uses cloudflare for the cdn and caching work around with spree using active storage representations to display images you can simply monkey patch the following file in rails rails activestorage app controllers active storage representations controller rb with the following code frozen string literal true take a signed permanent reference for a blob representation and turn it into an expiring service url for download note these urls are publicly accessible if you need to enforce access protection beyond the security through obscurity factor of the signed blob and variation reference you ll need to implement your own authenticated redirection controller class activestorage representationscontroller activestorage basecontroller include activestorage setblob def show expires in year public true variant blob representation params processed send data blob service download variant key type blob content type default send file type disposition inline end end this will then give you active storage images served with a public accessible file with a year expire date if you are running your new spree app through cloudflare you can now change your active storage page rule from bypass cache to img width alt cf src and that should do it your files will be served through the cloudflare cdn with browser caching please be aware that this fix may be outdated as rails updates past version
0
1,659
6,574,047,879
IssuesEvent
2017-09-11 11:14:45
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
azure resource group creation fails
affects_2.2 azure bug_report cloud waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME azure_rm_resourcegroup ##### ANSIBLE VERSION ansible 2.2.0.0 also tried with github pip install at this time ##### CONFIGURATION ##### OS / ENVIRONMENT Linux ``` pip freeze | grep azure azure==2.0.0rc5 azure-batch==0.30.0rc5 azure-cli==0.1.0b10 azure-cli-acr==0.1.0b10 azure-cli-acs==0.1.0b10 azure-cli-appservice==0.1.0b10 azure-cli-cloud==0.1.0b10 azure-cli-component==0.1.0b10 azure-cli-configure==0.1.0b10 azure-cli-container==0.1.0b10 azure-cli-context==0.1.0b10 azure-cli-core==0.1.0b10 azure-cli-feedback==0.1.0b10 azure-cli-network==0.1.0b10 azure-cli-profile==0.1.0b10 azure-cli-resource==0.1.0b10 azure-cli-role==0.1.0b10 azure-cli-storage==0.1.0b10 azure-cli-vm==0.1.0b10 azure-common==1.1.4 azure-graphrbac==0.30.0rc6 azure-mgmt==0.30.0rc5 azure-mgmt-authorization==0.30.0rc6 azure-mgmt-batch==0.30.0rc5 azure-mgmt-cdn==0.30.0rc5 azure-mgmt-cognitiveservices==0.30.0rc5 azure-mgmt-commerce==0.30.0rc5 azure-mgmt-compute==0.32.1 azure-mgmt-containerregistry==0.1.0 azure-mgmt-dns==0.30.0rc6 azure-mgmt-keyvault==0.30.0rc5 azure-mgmt-logic==0.30.0rc5 azure-mgmt-network==0.30.0rc6 azure-mgmt-notificationhubs==0.30.0rc5 azure-mgmt-nspkg==1.0.0 azure-mgmt-powerbiembedded==0.30.0rc5 azure-mgmt-redis==0.30.0rc5 azure-mgmt-resource==0.30.2 azure-mgmt-scheduler==0.30.0rc5 azure-mgmt-storage==0.30.0rc6 azure-mgmt-trafficmanager==0.30.0rc6 azure-mgmt-web==0.30.1 azure-nspkg==1.0.0 azure-servicebus==0.20.2 azure-servicemanagement-legacy==0.20.3 azure-storage==0.33.0 msrestazure==0.4.5 ``` ##### SUMMARY Resource Group Creation fails due to 403 on existence check url ##### STEPS TO REPRODUCE ``` - name: Create a resource group azure_rm_resourcegroup: name: devtesting location: westus tags: testing: testing delete: never ``` ##### EXPECTED RESULTS Expected resource group to be created ##### ACTUAL RESULTS ``` fatal: [development-tools]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_args": { "ad_user": null, "append_tags": true, "client_id": null, "force": false, "location": "westus", "name": "devtesting", "password": null, "profile": null, "secret": null, "state": "present", "subscription_id": null, "tags": { "delete": "never", "testing": "testing" }, "tenant": null }, "module_name": "azure_rm_resourcegroup" }, "msg": "Error checking for existence of name devtesting - 403 Client Error: Forbidden for url: https://management.azure.com/subscriptions/MYSUBID/resourcegroups/devtesting?api-version=2016-09-01" } ``` ```
True
azure resource group creation fails - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME azure_rm_resourcegroup ##### ANSIBLE VERSION ansible 2.2.0.0 also tried with github pip install at this time ##### CONFIGURATION ##### OS / ENVIRONMENT Linux ``` pip freeze | grep azure azure==2.0.0rc5 azure-batch==0.30.0rc5 azure-cli==0.1.0b10 azure-cli-acr==0.1.0b10 azure-cli-acs==0.1.0b10 azure-cli-appservice==0.1.0b10 azure-cli-cloud==0.1.0b10 azure-cli-component==0.1.0b10 azure-cli-configure==0.1.0b10 azure-cli-container==0.1.0b10 azure-cli-context==0.1.0b10 azure-cli-core==0.1.0b10 azure-cli-feedback==0.1.0b10 azure-cli-network==0.1.0b10 azure-cli-profile==0.1.0b10 azure-cli-resource==0.1.0b10 azure-cli-role==0.1.0b10 azure-cli-storage==0.1.0b10 azure-cli-vm==0.1.0b10 azure-common==1.1.4 azure-graphrbac==0.30.0rc6 azure-mgmt==0.30.0rc5 azure-mgmt-authorization==0.30.0rc6 azure-mgmt-batch==0.30.0rc5 azure-mgmt-cdn==0.30.0rc5 azure-mgmt-cognitiveservices==0.30.0rc5 azure-mgmt-commerce==0.30.0rc5 azure-mgmt-compute==0.32.1 azure-mgmt-containerregistry==0.1.0 azure-mgmt-dns==0.30.0rc6 azure-mgmt-keyvault==0.30.0rc5 azure-mgmt-logic==0.30.0rc5 azure-mgmt-network==0.30.0rc6 azure-mgmt-notificationhubs==0.30.0rc5 azure-mgmt-nspkg==1.0.0 azure-mgmt-powerbiembedded==0.30.0rc5 azure-mgmt-redis==0.30.0rc5 azure-mgmt-resource==0.30.2 azure-mgmt-scheduler==0.30.0rc5 azure-mgmt-storage==0.30.0rc6 azure-mgmt-trafficmanager==0.30.0rc6 azure-mgmt-web==0.30.1 azure-nspkg==1.0.0 azure-servicebus==0.20.2 azure-servicemanagement-legacy==0.20.3 azure-storage==0.33.0 msrestazure==0.4.5 ``` ##### SUMMARY Resource Group Creation fails due to 403 on existence check url ##### STEPS TO REPRODUCE ``` - name: Create a resource group azure_rm_resourcegroup: name: devtesting location: westus tags: testing: testing delete: never ``` ##### EXPECTED RESULTS Expected resource group to be created ##### ACTUAL RESULTS ``` fatal: [development-tools]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_args": { "ad_user": null, "append_tags": true, "client_id": null, "force": false, "location": "westus", "name": "devtesting", "password": null, "profile": null, "secret": null, "state": "present", "subscription_id": null, "tags": { "delete": "never", "testing": "testing" }, "tenant": null }, "module_name": "azure_rm_resourcegroup" }, "msg": "Error checking for existence of name devtesting - 403 Client Error: Forbidden for url: https://management.azure.com/subscriptions/MYSUBID/resourcegroups/devtesting?api-version=2016-09-01" } ``` ```
main
azure resource group creation fails issue type bug report component name azure rm resourcegroup ansible version ansible also tried with github pip install at this time configuration os environment linux pip freeze grep azure azure azure batch azure cli azure cli acr azure cli acs azure cli appservice azure cli cloud azure cli component azure cli configure azure cli container azure cli context azure cli core azure cli feedback azure cli network azure cli profile azure cli resource azure cli role azure cli storage azure cli vm azure common azure graphrbac azure mgmt azure mgmt authorization azure mgmt batch azure mgmt cdn azure mgmt cognitiveservices azure mgmt commerce azure mgmt compute azure mgmt containerregistry azure mgmt dns azure mgmt keyvault azure mgmt logic azure mgmt network azure mgmt notificationhubs azure mgmt nspkg azure mgmt powerbiembedded azure mgmt redis azure mgmt resource azure mgmt scheduler azure mgmt storage azure mgmt trafficmanager azure mgmt web azure nspkg azure servicebus azure servicemanagement legacy azure storage msrestazure summary resource group creation fails due to on existence check url steps to reproduce name create a resource group azure rm resourcegroup name devtesting location westus tags testing testing delete never expected results expected resource group to be created actual results fatal failed changed false failed true invocation module args ad user null append tags true client id null force false location westus name devtesting password null profile null secret null state present subscription id null tags delete never testing testing tenant null module name azure rm resourcegroup msg error checking for existence of name devtesting client error forbidden for url
1
293,759
25,321,101,259
IssuesEvent
2022-11-18 04:02:33
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
closed
[chip-test] chip_sw_clkmgr_extended_range
Component:ChipLevelTest
### Test point name [chip_sw_clkmgr_extended_range](https://github.com/lowRISC/opentitan/blob/c1686dfd538165041d727fe1427e139bc20e848c/hw/top_earlgrey/data/chip_testplan.hjson#L867) Description: Check that the system can run at a reduced, calibrated clock frequency (70MHz) with jitter enabled (which can momentarily bring it down to ~55 MHz). This option is intended as a fall-back in case there are issues running the system at 100MHz (calibrated). This testpoint can likely be covered by extending the DV environment to support the extended range clock option via a flag, and running several existing chip-level tests with that option, e.g. chip_sw_clkmgr_jitter chip_sw_flash_ctrl_ops_jitter_en chip_sw_flash_ctrl_access_jitter_en chip_sw_otbn_ecdsa_op_irq_jitter_en chip_sw_aes_enc_jitter_en chip_sw_hmac_enc_jitter_en chip_sw_keymgr_key_derivation_jitter_en chip_sw_kmac_mode_kmac_jitter_en chip_sw_sram_ctrl_main_scrambled_access_jitter_en ### Host side component None ### OpenTitanTool infrastructure implemented Unknown ### Contact person @Jacob-Levy @arnonsha @tjaychen ### Checklist Please fill out this checklist as items are completed. Link to PRs and issues as appropriate. - [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes) - [ ] Device-side (C) component developed - [ ] Bazel build rules developed - [ ] Host-side component developed - [ ] HJSON test plan updated with test name (so it shows up in the dashboard) - [ ] Test added to dvsim nightly regression (and passing at time of checking)
1.0
[chip-test] chip_sw_clkmgr_extended_range - ### Test point name [chip_sw_clkmgr_extended_range](https://github.com/lowRISC/opentitan/blob/c1686dfd538165041d727fe1427e139bc20e848c/hw/top_earlgrey/data/chip_testplan.hjson#L867) Description: Check that the system can run at a reduced, calibrated clock frequency (70MHz) with jitter enabled (which can momentarily bring it down to ~55 MHz). This option is intended as a fall-back in case there are issues running the system at 100MHz (calibrated). This testpoint can likely be covered by extending the DV environment to support the extended range clock option via a flag, and running several existing chip-level tests with that option, e.g. chip_sw_clkmgr_jitter chip_sw_flash_ctrl_ops_jitter_en chip_sw_flash_ctrl_access_jitter_en chip_sw_otbn_ecdsa_op_irq_jitter_en chip_sw_aes_enc_jitter_en chip_sw_hmac_enc_jitter_en chip_sw_keymgr_key_derivation_jitter_en chip_sw_kmac_mode_kmac_jitter_en chip_sw_sram_ctrl_main_scrambled_access_jitter_en ### Host side component None ### OpenTitanTool infrastructure implemented Unknown ### Contact person @Jacob-Levy @arnonsha @tjaychen ### Checklist Please fill out this checklist as items are completed. Link to PRs and issues as appropriate. - [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes) - [ ] Device-side (C) component developed - [ ] Bazel build rules developed - [ ] Host-side component developed - [ ] HJSON test plan updated with test name (so it shows up in the dashboard) - [ ] Test added to dvsim nightly regression (and passing at time of checking)
non_main
chip sw clkmgr extended range test point name description check that the system can run at a reduced calibrated clock frequency with jitter enabled which can momentarily bring it down to mhz this option is intended as a fall back in case there are issues running the system at calibrated this testpoint can likely be covered by extending the dv environment to support the extended range clock option via a flag and running several existing chip level tests with that option e g chip sw clkmgr jitter chip sw flash ctrl ops jitter en chip sw flash ctrl access jitter en chip sw otbn ecdsa op irq jitter en chip sw aes enc jitter en chip sw hmac enc jitter en chip sw keymgr key derivation jitter en chip sw kmac mode kmac jitter en chip sw sram ctrl main scrambled access jitter en host side component none opentitantool infrastructure implemented unknown contact person jacob levy arnonsha tjaychen checklist please fill out this checklist as items are completed link to prs and issues as appropriate check if existing test covers most or all of this testpoint if so either extend said test to cover all points or skip the next checkboxes device side c component developed bazel build rules developed host side component developed hjson test plan updated with test name so it shows up in the dashboard test added to dvsim nightly regression and passing at time of checking
0
3,076
11,645,010,136
IssuesEvent
2020-02-29 22:09:16
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
AppVeyor configuration file still uses JDK1.8
maintainability
AppVeyor configuration file still uses JDK1.8. Since we have recommended use of JDK8 the config must be updated too. cc: @thadguidry @wetneb your views ?
True
AppVeyor configuration file still uses JDK1.8 - AppVeyor configuration file still uses JDK1.8. Since we have recommended use of JDK8 the config must be updated too. cc: @thadguidry @wetneb your views ?
main
appveyor configuration file still uses appveyor configuration file still uses since we have recommended use of the config must be updated too cc thadguidry wetneb your views
1
795
4,397,162,824
IssuesEvent
2016-08-10 06:56:39
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
allow to specify version number when installing with zypper
feature_idea waiting_on_maintainer
##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME packaging/os/zypper.py ##### ANSIBLE VERSION 2.2, devel ##### CONFIGURATION N/A ##### OS / ENVIRONMENT any Suse distro ##### SUMMARY Allow to put version information in the package name and handle that correctly. ##### STEPS TO REPRODUCE Things that should work: ``` - zypper: name=openssh>=6.6 state=present - zypper: name=openssh=6.6 state=present ``` Allowed operators in zypper: `<`, `>`, `<=`, `>=`, `=`. Version should be `[0-9.-]*` Tests for those are also needed. ##### EXPECTED RESULTS The package should be installed under the given version restrictions. Output/diffs should handle the version info correctly and not put it as part of the package name. ##### ACTUAL RESULTS Depending on the case this is broken. `state=absent` does not work with versions, the rest is not fully tested.
True
allow to specify version number when installing with zypper - ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME packaging/os/zypper.py ##### ANSIBLE VERSION 2.2, devel ##### CONFIGURATION N/A ##### OS / ENVIRONMENT any Suse distro ##### SUMMARY Allow to put version information in the package name and handle that correctly. ##### STEPS TO REPRODUCE Things that should work: ``` - zypper: name=openssh>=6.6 state=present - zypper: name=openssh=6.6 state=present ``` Allowed operators in zypper: `<`, `>`, `<=`, `>=`, `=`. Version should be `[0-9.-]*` Tests for those are also needed. ##### EXPECTED RESULTS The package should be installed under the given version restrictions. Output/diffs should handle the version info correctly and not put it as part of the package name. ##### ACTUAL RESULTS Depending on the case this is broken. `state=absent` does not work with versions, the rest is not fully tested.
main
allow to specify version number when installing with zypper issue type feature idea component name packaging os zypper py ansible version devel configuration n a os environment any suse distro summary allow to put version information in the package name and handle that correctly steps to reproduce things that should work zypper name openssh state present zypper name openssh state present allowed operators in zypper version should be tests for those are also needed expected results the package should be installed under the given version restrictions output diffs should handle the version info correctly and not put it as part of the package name actual results depending on the case this is broken state absent does not work with versions the rest is not fully tested
1
890
4,553,165,577
IssuesEvent
2016-09-13 03:00:35
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Support starting stopped EC2 instances by tag
affects_1.7 aws cloud feature_idea waiting_on_maintainer
##### Issue Type: Feature Idea ##### Ansible Version: ansible 1.7.2 ##### Environment: N/A ##### Summary: Allow user to start all stopped EC2 instances associated with a particular tag. ##### Steps To Reproduce: Extend the ec2 module to enable, say, starting all stopped instances tagged with key=Name, Value=ExtraPower ``` - local_action: module: ec2 instance_tags: Name: ExtraPower state: running ``` ##### Expected Results: Any stopped instance with the associated tag will be started (any instance already running with the tag would be unaffected). ##### Actual Results: N/A
True
Support starting stopped EC2 instances by tag - ##### Issue Type: Feature Idea ##### Ansible Version: ansible 1.7.2 ##### Environment: N/A ##### Summary: Allow user to start all stopped EC2 instances associated with a particular tag. ##### Steps To Reproduce: Extend the ec2 module to enable, say, starting all stopped instances tagged with key=Name, Value=ExtraPower ``` - local_action: module: ec2 instance_tags: Name: ExtraPower state: running ``` ##### Expected Results: Any stopped instance with the associated tag will be started (any instance already running with the tag would be unaffected). ##### Actual Results: N/A
main
support starting stopped instances by tag issue type feature idea ansible version ansible environment n a summary allow user to start all stopped instances associated with a particular tag steps to reproduce extend the module to enable say starting all stopped instances tagged with key name value extrapower local action module instance tags name extrapower state running expected results any stopped instance with the associated tag will be started any instance already running with the tag would be unaffected actual results n a
1
105,692
4,240,542,202
IssuesEvent
2016-07-06 13:46:40
azurestandard/api-spec
https://api.github.com/repos/azurestandard/api-spec
closed
Country of Origin
Low Priority
Allow the country of origin to be returned via the API. See Website Issue [#478](https://github.com/azurestandard/website/issues/478)
1.0
Country of Origin - Allow the country of origin to be returned via the API. See Website Issue [#478](https://github.com/azurestandard/website/issues/478)
non_main
country of origin allow the country of origin to be returned via the api see website issue
0
3,542
13,942,237,310
IssuesEvent
2020-10-22 20:41:49
polyfacet/ArasDeveloperTool
https://api.github.com/repos/polyfacet/ArasDeveloperTool
closed
General Refactoring
maintainability
Do a general refactor before replacing IOM with InnovatorClient and splitting the VS-project
True
General Refactoring - Do a general refactor before replacing IOM with InnovatorClient and splitting the VS-project
main
general refactoring do a general refactor before replacing iom with innovatorclient and splitting the vs project
1
2,089
7,106,605,165
IssuesEvent
2018-01-16 17:06:31
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
STDOUT logging on network failures
affects_2.3 feature_idea module needs_maintainer networking support:community
<!--- Verify first that your issue/request is not already reported on GitHub. Also test if the latest release, and master branch are affected too. --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the module/plugin/task/feature --> network modules ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.3.0 config file = /etc/ansible/ansible.cfg configured module search path = [u'/etc/ntc/ansible/library', u'/etc/ntc/ntc-ansible-plugins/library'] python version = 2.7.6 (default, Oct 26 2016, 20:32:47) [GCC 4.8.4] ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. Also mention the specific version of what you are trying to control, e.g. if this is a network bug the version of firmware on the network device. --> ##### SUMMARY <!--- Explain the problem briefly --> Prior to 2.3, when using network modules (ios, eos, nxos, etc...) that required ssh support, connection failures produced usable output to indicate the connection issue. With the use of persistence that feature has gone away and replaced with a multi-step process to log, then view the logs. While it is understood that you can have access to the logs as described here: https://docs.ansible.com/ansible/network_debug_troubleshooting.html#unable-to-open-shell There is no immediate feedback to the user. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem, using a minimal test-case. For new features, show how the feature would be used. --> Run any playbook with bad credentials or non-existent IP. <!--- Paste example playbooks or commands between quotes below --> ```yaml ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Same as 2.2 ``` (ansible2.2) ken@kcelenza-ubuntu:/github/cookbook$ ansible-playbook -i inventory ios.yml -vvv Using /etc/ansible/ansible.cfg as config file PLAYBOOK: ios.yml ************************************************************** 1 plays in ios.yml PLAY [RUN COMMAND AND SAVE] **************************************************** TASK [RUN COMMANDS] ************************************************************ task path: /github/cookbook/ios.yml:23 Using module file /home/ken/.virtualenvs/ansible2.2/local/lib/python2.7/site-packages/ansible/modules/core/network/ios/ios_command.py <10.1.100.46> ESTABLISH LOCAL CONNECTION FOR USER: ken <10.1.100.46> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo ~/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661 `" && echo ansible-tmp-1500042841.83-229486189401661="` echo ~/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661 `" ) && sleep 0' Using module file /home/ken/.virtualenvs/ansible2.2/local/lib/python2.7/site-packages/ansible/modules/core/network/ios/ios_command.py <10.1.100.1> ESTABLISH LOCAL CONNECTION FOR USER: ken <10.1.100.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo ~/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487 `" && echo ansible-tmp-1500042841.83-15461642620487="` echo ~/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487 `" ) && sleep 0' <10.1.100.46> PUT /tmp/tmpnNfZ50 TO /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661/ios_command.py <10.1.100.46> EXEC /bin/sh -c 'chmod u+x /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661/ /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661/ios_command.py && sleep 0' <10.1.100.1> PUT /tmp/tmpOgUCo4 TO /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487/ios_command.py <10.1.100.1> EXEC /bin/sh -c 'chmod u+x /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487/ /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487/ios_command.py && sleep 0' <10.1.100.46> EXEC /bin/sh -c 'python /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661/ios_command.py; rm -rf "/home/ken/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661/" > /dev/null 2>&1 && sleep 0' <10.1.100.1> EXEC /bin/sh -c 'python /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487/ios_command.py; rm -rf "/home/ken/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487/" > /dev/null 2>&1 && sleep 0' fatal: [ios]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_args": { "auth_pass": null, "authorize": false, "commands": [ "show inventory", "show version" ], "host": "10.1.100.1", "interval": 1, "match": "all", "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": { "host": "10.1.100.1", "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "username": "ntc" }, "retries": 10, "ssh_keyfile": null, "timeout": 10, "transport": null, "use_ssl": true, "username": "ntc", "validate_certs": true, "wait_for": null }, "module_name": "ios_command" }, "msg": "failed to connect to 10.1.100.1:22" } An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_sPVrBz/ansible_module_ios_command.py", line 237, in <module> main() File "/tmp/ansible_sPVrBz/ansible_module_ios_command.py", line 200, in main runner.add_command(**cmd) File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/netcli.py", line 147, in add_command File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/network.py", line 117, in cli File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/network.py", line 148, in connect File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/ios.py", line 180, in connect File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/shell.py", line 226, in connect File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/shell.py", line 96, in open File "/home/ken/.virtualenvs/ansible2.2/local/lib/python2.7/site-packages/paramiko/client.py", line 329, in connect raise NoValidConnectionsError(errors) paramiko.ssh_exception.NoValidConnectionsError: [Errno None] Unable to connect to port 22 on 10.1.100.46 fatal: [ios1]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_name": "ios_command" }, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_sPVrBz/ansible_module_ios_command.py\", line 237, in <module>\n main()\n File \"/tmp/ansible_sPVrBz/ansible_module_ios_command.py\", line 200, in main\n runner.add_command(**cmd)\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 147, in add_command\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/network.py\", line 117, in cli\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/network.py\", line 148, in connect\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/ios.py\", line 180, in connect\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/shell.py\", line 226, in connect\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/shell.py\", line 96, in open\n File \"/home/ken/.virtualenvs/ansible2.2/local/lib/python2.7/site-packages/paramiko/client.py\", line 329, in connect\n raise NoValidConnectionsError(errors)\nparamiko.ssh_exception.NoValidConnectionsError: [Errno None] Unable to connect to port 22 on 10.1.100.46\n", "module_stdout": "", "msg": "MODULE FAILURE" } PLAY RECAP ********************************************************************* ios : ok=0 changed=0 unreachable=0 failed=1 ios1 : ok=0 changed=0 unreachable=0 failed=1 (ansible2.2) ken@kcelenza-ubuntu:/github/cookbook$ ``` ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` (ansible2.3) ken@kcelenza-ubuntu:/github/cookbook$ ansible-playbook -i inventory ios.yml -vvv Using /etc/ansible/ansible.cfg as config file PLAYBOOK: ios.yml ************************************************************************************************************************************************************************************************************************************************************** 1 plays in ios.yml PLAY [RUN COMMAND AND SAVE] **************************************************************************************************************************************************************************************************************************************************** META: ran handlers TASK [RUN COMMANDS] ************************************************************************************************************************************************************************************************************************************************************ task path: /github/cookbook/ios.yml:23 fatal: [ios1]: FAILED! => { "changed": false, "failed": true, "msg": "unable to open shell", "rc": 255 } fatal: [ios]: FAILED! => { "changed": false, "failed": true, "msg": "unable to open shell", "rc": 255 } PLAY RECAP ********************************************************************************************************************************************************************************************************************************************************************* ios : ok=0 changed=0 unreachable=0 failed=1 ios1 : ok=0 changed=0 unreachable=0 failed=1 (ansible2.3) ken@kcelenza-ubuntu:/github/cookbook$ ansible --version ansible 2.3.0 config file = /etc/ansible/ansible.cfg configured module search path = [u'/etc/ntc/ansible/library', u'/etc/ntc/ntc-ansible-plugins/library'] python version = 2.7.6 (default, Oct 26 2016, 20:32:47) [GCC 4.8.4] (ansible2.3) ken@kcelenza-ubuntu:/github/cookbook$ workon ansible2.4 (ansible2.4) ken@kcelenza-ubuntu:/github/cookbook$ ansible-playbook -i inventory ios.yml -vvv ansible-playbook 2.4.0 (kc_pan_sec_rule_idemp 7ee2685edb) last updated 2017/07/09 23:25:10 (GMT -400) config file = /etc/ansible/ansible.cfg configured module search path = [u'/etc/ntc/ansible/library', u'/etc/ntc/ntc-ansible-plugins/library'] ansible python module location = /home/ken/ansible/lib/ansible executable location = /home/ken/.virtualenvs/ansible2.4/bin/ansible-playbook python version = 2.7.6 (default, Oct 26 2016, 20:32:47) [GCC 4.8.4] Using /etc/ansible/ansible.cfg as config file Parsed /github/cookbook/inventory inventory source with ini plugin PLAYBOOK: ios.yml ************************************************************************************************************************************************************************************************************************************************************** 1 plays in ios.yml PLAY [RUN COMMAND AND SAVE] **************************************************************************************************************************************************************************************************************************************************** META: ran handlers TASK [RUN COMMANDS] ************************************************************************************************************************************************************************************************************************************************************ task path: /github/cookbook/ios.yml:23 <10.1.100.1> using connection plugin network_cli <10.1.100.46> using connection plugin network_cli fatal: [ios1]: FAILED! => { "changed": false, "failed": true, "msg": "unable to open shell. Please see: https://docs.ansible.com/ansible/network_debug_troubleshooting.html#unable-to-open-shell" } fatal: [ios]: FAILED! => { "changed": false, "failed": true, "msg": "unable to open shell. Please see: https://docs.ansible.com/ansible/network_debug_troubleshooting.html#unable-to-open-shell" } PLAY RECAP ********************************************************************************************************************************************************************************************************************************************************************* ios : ok=0 changed=0 unreachable=0 failed=1 ios1 : ok=0 changed=0 unreachable=0 failed=1 ``` pinging @privateip @gundalow @rcarrillocruz
True
STDOUT logging on network failures - <!--- Verify first that your issue/request is not already reported on GitHub. Also test if the latest release, and master branch are affected too. --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the module/plugin/task/feature --> network modules ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.3.0 config file = /etc/ansible/ansible.cfg configured module search path = [u'/etc/ntc/ansible/library', u'/etc/ntc/ntc-ansible-plugins/library'] python version = 2.7.6 (default, Oct 26 2016, 20:32:47) [GCC 4.8.4] ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. Also mention the specific version of what you are trying to control, e.g. if this is a network bug the version of firmware on the network device. --> ##### SUMMARY <!--- Explain the problem briefly --> Prior to 2.3, when using network modules (ios, eos, nxos, etc...) that required ssh support, connection failures produced usable output to indicate the connection issue. With the use of persistence that feature has gone away and replaced with a multi-step process to log, then view the logs. While it is understood that you can have access to the logs as described here: https://docs.ansible.com/ansible/network_debug_troubleshooting.html#unable-to-open-shell There is no immediate feedback to the user. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem, using a minimal test-case. For new features, show how the feature would be used. --> Run any playbook with bad credentials or non-existent IP. <!--- Paste example playbooks or commands between quotes below --> ```yaml ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Same as 2.2 ``` (ansible2.2) ken@kcelenza-ubuntu:/github/cookbook$ ansible-playbook -i inventory ios.yml -vvv Using /etc/ansible/ansible.cfg as config file PLAYBOOK: ios.yml ************************************************************** 1 plays in ios.yml PLAY [RUN COMMAND AND SAVE] **************************************************** TASK [RUN COMMANDS] ************************************************************ task path: /github/cookbook/ios.yml:23 Using module file /home/ken/.virtualenvs/ansible2.2/local/lib/python2.7/site-packages/ansible/modules/core/network/ios/ios_command.py <10.1.100.46> ESTABLISH LOCAL CONNECTION FOR USER: ken <10.1.100.46> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo ~/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661 `" && echo ansible-tmp-1500042841.83-229486189401661="` echo ~/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661 `" ) && sleep 0' Using module file /home/ken/.virtualenvs/ansible2.2/local/lib/python2.7/site-packages/ansible/modules/core/network/ios/ios_command.py <10.1.100.1> ESTABLISH LOCAL CONNECTION FOR USER: ken <10.1.100.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo ~/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487 `" && echo ansible-tmp-1500042841.83-15461642620487="` echo ~/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487 `" ) && sleep 0' <10.1.100.46> PUT /tmp/tmpnNfZ50 TO /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661/ios_command.py <10.1.100.46> EXEC /bin/sh -c 'chmod u+x /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661/ /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661/ios_command.py && sleep 0' <10.1.100.1> PUT /tmp/tmpOgUCo4 TO /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487/ios_command.py <10.1.100.1> EXEC /bin/sh -c 'chmod u+x /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487/ /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487/ios_command.py && sleep 0' <10.1.100.46> EXEC /bin/sh -c 'python /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661/ios_command.py; rm -rf "/home/ken/.ansible/tmp/ansible-tmp-1500042841.83-229486189401661/" > /dev/null 2>&1 && sleep 0' <10.1.100.1> EXEC /bin/sh -c 'python /home/ken/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487/ios_command.py; rm -rf "/home/ken/.ansible/tmp/ansible-tmp-1500042841.83-15461642620487/" > /dev/null 2>&1 && sleep 0' fatal: [ios]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_args": { "auth_pass": null, "authorize": false, "commands": [ "show inventory", "show version" ], "host": "10.1.100.1", "interval": 1, "match": "all", "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": { "host": "10.1.100.1", "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "username": "ntc" }, "retries": 10, "ssh_keyfile": null, "timeout": 10, "transport": null, "use_ssl": true, "username": "ntc", "validate_certs": true, "wait_for": null }, "module_name": "ios_command" }, "msg": "failed to connect to 10.1.100.1:22" } An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_sPVrBz/ansible_module_ios_command.py", line 237, in <module> main() File "/tmp/ansible_sPVrBz/ansible_module_ios_command.py", line 200, in main runner.add_command(**cmd) File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/netcli.py", line 147, in add_command File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/network.py", line 117, in cli File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/network.py", line 148, in connect File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/ios.py", line 180, in connect File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/shell.py", line 226, in connect File "/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/shell.py", line 96, in open File "/home/ken/.virtualenvs/ansible2.2/local/lib/python2.7/site-packages/paramiko/client.py", line 329, in connect raise NoValidConnectionsError(errors) paramiko.ssh_exception.NoValidConnectionsError: [Errno None] Unable to connect to port 22 on 10.1.100.46 fatal: [ios1]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_name": "ios_command" }, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_sPVrBz/ansible_module_ios_command.py\", line 237, in <module>\n main()\n File \"/tmp/ansible_sPVrBz/ansible_module_ios_command.py\", line 200, in main\n runner.add_command(**cmd)\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 147, in add_command\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/network.py\", line 117, in cli\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/network.py\", line 148, in connect\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/ios.py\", line 180, in connect\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/shell.py\", line 226, in connect\n File \"/tmp/ansible_sPVrBz/ansible_modlib.zip/ansible/module_utils/shell.py\", line 96, in open\n File \"/home/ken/.virtualenvs/ansible2.2/local/lib/python2.7/site-packages/paramiko/client.py\", line 329, in connect\n raise NoValidConnectionsError(errors)\nparamiko.ssh_exception.NoValidConnectionsError: [Errno None] Unable to connect to port 22 on 10.1.100.46\n", "module_stdout": "", "msg": "MODULE FAILURE" } PLAY RECAP ********************************************************************* ios : ok=0 changed=0 unreachable=0 failed=1 ios1 : ok=0 changed=0 unreachable=0 failed=1 (ansible2.2) ken@kcelenza-ubuntu:/github/cookbook$ ``` ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` (ansible2.3) ken@kcelenza-ubuntu:/github/cookbook$ ansible-playbook -i inventory ios.yml -vvv Using /etc/ansible/ansible.cfg as config file PLAYBOOK: ios.yml ************************************************************************************************************************************************************************************************************************************************************** 1 plays in ios.yml PLAY [RUN COMMAND AND SAVE] **************************************************************************************************************************************************************************************************************************************************** META: ran handlers TASK [RUN COMMANDS] ************************************************************************************************************************************************************************************************************************************************************ task path: /github/cookbook/ios.yml:23 fatal: [ios1]: FAILED! => { "changed": false, "failed": true, "msg": "unable to open shell", "rc": 255 } fatal: [ios]: FAILED! => { "changed": false, "failed": true, "msg": "unable to open shell", "rc": 255 } PLAY RECAP ********************************************************************************************************************************************************************************************************************************************************************* ios : ok=0 changed=0 unreachable=0 failed=1 ios1 : ok=0 changed=0 unreachable=0 failed=1 (ansible2.3) ken@kcelenza-ubuntu:/github/cookbook$ ansible --version ansible 2.3.0 config file = /etc/ansible/ansible.cfg configured module search path = [u'/etc/ntc/ansible/library', u'/etc/ntc/ntc-ansible-plugins/library'] python version = 2.7.6 (default, Oct 26 2016, 20:32:47) [GCC 4.8.4] (ansible2.3) ken@kcelenza-ubuntu:/github/cookbook$ workon ansible2.4 (ansible2.4) ken@kcelenza-ubuntu:/github/cookbook$ ansible-playbook -i inventory ios.yml -vvv ansible-playbook 2.4.0 (kc_pan_sec_rule_idemp 7ee2685edb) last updated 2017/07/09 23:25:10 (GMT -400) config file = /etc/ansible/ansible.cfg configured module search path = [u'/etc/ntc/ansible/library', u'/etc/ntc/ntc-ansible-plugins/library'] ansible python module location = /home/ken/ansible/lib/ansible executable location = /home/ken/.virtualenvs/ansible2.4/bin/ansible-playbook python version = 2.7.6 (default, Oct 26 2016, 20:32:47) [GCC 4.8.4] Using /etc/ansible/ansible.cfg as config file Parsed /github/cookbook/inventory inventory source with ini plugin PLAYBOOK: ios.yml ************************************************************************************************************************************************************************************************************************************************************** 1 plays in ios.yml PLAY [RUN COMMAND AND SAVE] **************************************************************************************************************************************************************************************************************************************************** META: ran handlers TASK [RUN COMMANDS] ************************************************************************************************************************************************************************************************************************************************************ task path: /github/cookbook/ios.yml:23 <10.1.100.1> using connection plugin network_cli <10.1.100.46> using connection plugin network_cli fatal: [ios1]: FAILED! => { "changed": false, "failed": true, "msg": "unable to open shell. Please see: https://docs.ansible.com/ansible/network_debug_troubleshooting.html#unable-to-open-shell" } fatal: [ios]: FAILED! => { "changed": false, "failed": true, "msg": "unable to open shell. Please see: https://docs.ansible.com/ansible/network_debug_troubleshooting.html#unable-to-open-shell" } PLAY RECAP ********************************************************************************************************************************************************************************************************************************************************************* ios : ok=0 changed=0 unreachable=0 failed=1 ios1 : ok=0 changed=0 unreachable=0 failed=1 ``` pinging @privateip @gundalow @rcarrillocruz
main
stdout logging on network failures verify first that your issue request is not already reported on github also test if the latest release and master branch are affected too issue type feature idea component name network modules ansible version ansible config file etc ansible ansible cfg configured module search path python version default oct configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific also mention the specific version of what you are trying to control e g if this is a network bug the version of firmware on the network device summary prior to when using network modules ios eos nxos etc that required ssh support connection failures produced usable output to indicate the connection issue with the use of persistence that feature has gone away and replaced with a multi step process to log then view the logs while it is understood that you can have access to the logs as described here there is no immediate feedback to the user steps to reproduce for bugs show exactly how to reproduce the problem using a minimal test case for new features show how the feature would be used run any playbook with bad credentials or non existent ip yaml expected results same as ken kcelenza ubuntu github cookbook ansible playbook i inventory ios yml vvv using etc ansible ansible cfg as config file playbook ios yml plays in ios yml play task task path github cookbook ios yml using module file home ken virtualenvs local lib site packages ansible modules core network ios ios command py establish local connection for user ken exec bin sh c umask mkdir p echo ansible tmp ansible tmp echo ansible tmp echo ansible tmp ansible tmp sleep using module file home ken virtualenvs local lib site packages ansible modules core network ios ios command py establish local connection for user ken exec bin sh c umask mkdir p echo ansible tmp ansible tmp echo ansible tmp echo ansible tmp ansible tmp sleep put tmp to home ken ansible tmp ansible tmp ios command py exec bin sh c chmod u x home ken ansible tmp ansible tmp home ken ansible tmp ansible tmp ios command py sleep put tmp to home ken ansible tmp ansible tmp ios command py exec bin sh c chmod u x home ken ansible tmp ansible tmp home ken ansible tmp ansible tmp ios command py sleep exec bin sh c python home ken ansible tmp ansible tmp ios command py rm rf home ken ansible tmp ansible tmp dev null sleep exec bin sh c python home ken ansible tmp ansible tmp ios command py rm rf home ken ansible tmp ansible tmp dev null sleep fatal failed changed false failed true invocation module args auth pass null authorize false commands show inventory show version host interval match all password value specified in no log parameter port null provider host password value specified in no log parameter username ntc retries ssh keyfile null timeout transport null use ssl true username ntc validate certs true wait for null module name ios command msg failed to connect to an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible spvrbz ansible module ios command py line in main file tmp ansible spvrbz ansible module ios command py line in main runner add command cmd file tmp ansible spvrbz ansible modlib zip ansible module utils netcli py line in add command file tmp ansible spvrbz ansible modlib zip ansible module utils network py line in cli file tmp ansible spvrbz ansible modlib zip ansible module utils network py line in connect file tmp ansible spvrbz ansible modlib zip ansible module utils ios py line in connect file tmp ansible spvrbz ansible modlib zip ansible module utils shell py line in connect file tmp ansible spvrbz ansible modlib zip ansible module utils shell py line in open file home ken virtualenvs local lib site packages paramiko client py line in connect raise novalidconnectionserror errors paramiko ssh exception novalidconnectionserror unable to connect to port on fatal failed changed false failed true invocation module name ios command module stderr traceback most recent call last n file tmp ansible spvrbz ansible module ios command py line in n main n file tmp ansible spvrbz ansible module ios command py line in main n runner add command cmd n file tmp ansible spvrbz ansible modlib zip ansible module utils netcli py line in add command n file tmp ansible spvrbz ansible modlib zip ansible module utils network py line in cli n file tmp ansible spvrbz ansible modlib zip ansible module utils network py line in connect n file tmp ansible spvrbz ansible modlib zip ansible module utils ios py line in connect n file tmp ansible spvrbz ansible modlib zip ansible module utils shell py line in connect n file tmp ansible spvrbz ansible modlib zip ansible module utils shell py line in open n file home ken virtualenvs local lib site packages paramiko client py line in connect n raise novalidconnectionserror errors nparamiko ssh exception novalidconnectionserror unable to connect to port on n module stdout msg module failure play recap ios ok changed unreachable failed ok changed unreachable failed ken kcelenza ubuntu github cookbook actual results ken kcelenza ubuntu github cookbook ansible playbook i inventory ios yml vvv using etc ansible ansible cfg as config file playbook ios yml plays in ios yml play meta ran handlers task task path github cookbook ios yml fatal failed changed false failed true msg unable to open shell rc fatal failed changed false failed true msg unable to open shell rc play recap ios ok changed unreachable failed ok changed unreachable failed ken kcelenza ubuntu github cookbook ansible version ansible config file etc ansible ansible cfg configured module search path python version default oct ken kcelenza ubuntu github cookbook workon ken kcelenza ubuntu github cookbook ansible playbook i inventory ios yml vvv ansible playbook kc pan sec rule idemp last updated gmt config file etc ansible ansible cfg configured module search path ansible python module location home ken ansible lib ansible executable location home ken virtualenvs bin ansible playbook python version default oct using etc ansible ansible cfg as config file parsed github cookbook inventory inventory source with ini plugin playbook ios yml plays in ios yml play meta ran handlers task task path github cookbook ios yml using connection plugin network cli using connection plugin network cli fatal failed changed false failed true msg unable to open shell please see fatal failed changed false failed true msg unable to open shell please see play recap ios ok changed unreachable failed ok changed unreachable failed pinging privateip gundalow rcarrillocruz
1
891
4,553,420,776
IssuesEvent
2016-09-13 04:41:27
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
ec2_elb_facts module can't use
affects_2.0 aws bug_report cloud waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2_elb_facts.py ##### ANSIBLE VERSION ``` ansible 2.0.1.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT CentOS 6.5 ##### SUMMARY ec2_elb_facts module can't get elb information ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: test hosts: 127.0.0.1 gather_facts: False become: yes tasks: - name: get elb facts local_action: ec2_elb_facts region='eu-west-1' name='test' aws_access_key='xxxxx' aws_secret_key='yyyy' register: elb_facts - name: debug local_action: debug msg="{{ elb_facts }}" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` An exception occurred during task execution. To see the full traceback, use -vvv. The error was: ValueError: invalid literal for int() with base 10: '80/api?url=http%3A%2F%2Fwww.shencan.net' fatal: [localhost -> localhost]: FAILED! => {"changed": false, "failed": true, "parsed": false} ```
True
ec2_elb_facts module can't use - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2_elb_facts.py ##### ANSIBLE VERSION ``` ansible 2.0.1.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT CentOS 6.5 ##### SUMMARY ec2_elb_facts module can't get elb information ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: test hosts: 127.0.0.1 gather_facts: False become: yes tasks: - name: get elb facts local_action: ec2_elb_facts region='eu-west-1' name='test' aws_access_key='xxxxx' aws_secret_key='yyyy' register: elb_facts - name: debug local_action: debug msg="{{ elb_facts }}" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` An exception occurred during task execution. To see the full traceback, use -vvv. The error was: ValueError: invalid literal for int() with base 10: '80/api?url=http%3A%2F%2Fwww.shencan.net' fatal: [localhost -> localhost]: FAILED! => {"changed": false, "failed": true, "parsed": false} ```
main
elb facts module can t use issue type bug report component name elb facts py ansible version ansible configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment centos summary elb facts module can t get elb information steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name test hosts gather facts false become yes tasks name get elb facts local action elb facts region eu west name test aws access key xxxxx aws secret key yyyy register elb facts name debug local action debug msg elb facts expected results actual results an exception occurred during task execution to see the full traceback use vvv the error was valueerror invalid literal for int with base api url http fatal failed changed false failed true parsed false
1
3,271
12,490,012,844
IssuesEvent
2020-05-31 21:40:26
short-d/short
https://api.github.com/repos/short-d/short
opened
[Refactor] Migrate switch statement for error type checking to user errors.As instead
maintainability
**What is frustrating you?** When short was first created, error types were checked using switch statement. Since then, golang has added two new constructs to work with errors: `Is` and `As`. Can be read about here: https://blog.golang.org/go1.13-errors Error type checking is still done the old way in the following three instances: - `authmutation.go` [1](https://github.com/short-d/short/blob/a900268c57539d040500f31b8f01ec406e002cc4/backend/app/adapter/gqlapi/resolver/authmutation.go#L66-L77) - `changelog.go` [1](https://github.com/short-d/short/blob/a900268c57539d040500f31b8f01ec406e002cc4/backend/app/usecase/changelog/changelog.go#L58-L61) [2](https://github.com/short-d/short/blob/a900268c57539d040500f31b8f01ec406e002cc4/backend/app/usecase/changelog/changelog.go#L74-L82) **Your solution** Since we're checking for error types, `errors.As` should be used to check the error type against each of the different error types. **Alternatives considered** Can also use `errors.Is` to check for sentinel error values, but like [this article](https://dave.cheney.net/2016/04/27/dont-just-check-errors-handle-them-gracefully) points out, sentinel errors are not very flexible as even a change of context for the error will fail the check. The same article also suggests opaque errors, which allow errors to be checked by its behaviour instead of its type. This can also be considered as an alternative over using `errors.As` but will require a greater refactor effort.
True
[Refactor] Migrate switch statement for error type checking to user errors.As instead - **What is frustrating you?** When short was first created, error types were checked using switch statement. Since then, golang has added two new constructs to work with errors: `Is` and `As`. Can be read about here: https://blog.golang.org/go1.13-errors Error type checking is still done the old way in the following three instances: - `authmutation.go` [1](https://github.com/short-d/short/blob/a900268c57539d040500f31b8f01ec406e002cc4/backend/app/adapter/gqlapi/resolver/authmutation.go#L66-L77) - `changelog.go` [1](https://github.com/short-d/short/blob/a900268c57539d040500f31b8f01ec406e002cc4/backend/app/usecase/changelog/changelog.go#L58-L61) [2](https://github.com/short-d/short/blob/a900268c57539d040500f31b8f01ec406e002cc4/backend/app/usecase/changelog/changelog.go#L74-L82) **Your solution** Since we're checking for error types, `errors.As` should be used to check the error type against each of the different error types. **Alternatives considered** Can also use `errors.Is` to check for sentinel error values, but like [this article](https://dave.cheney.net/2016/04/27/dont-just-check-errors-handle-them-gracefully) points out, sentinel errors are not very flexible as even a change of context for the error will fail the check. The same article also suggests opaque errors, which allow errors to be checked by its behaviour instead of its type. This can also be considered as an alternative over using `errors.As` but will require a greater refactor effort.
main
migrate switch statement for error type checking to user errors as instead what is frustrating you when short was first created error types were checked using switch statement since then golang has added two new constructs to work with errors is and as can be read about here error type checking is still done the old way in the following three instances authmutation go changelog go your solution since we re checking for error types errors as should be used to check the error type against each of the different error types alternatives considered can also use errors is to check for sentinel error values but like points out sentinel errors are not very flexible as even a change of context for the error will fail the check the same article also suggests opaque errors which allow errors to be checked by its behaviour instead of its type this can also be considered as an alternative over using errors as but will require a greater refactor effort
1
130,219
18,155,332,307
IssuesEvent
2021-09-27 00:05:24
ghc-dev/Haley-Sexton
https://api.github.com/repos/ghc-dev/Haley-Sexton
opened
CVE-2020-14332 (Medium) detected in ansible-2.9.9.tar.gz
security vulnerability
## CVE-2020-14332 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.9.9.tar.gz</b></p></summary> <p>Radically simple IT automation</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz">https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz</a></p> <p>Path to dependency file: Haley-Sexton/requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **ansible-2.9.9.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Haley-Sexton/commit/0cf1b8b91bc2e7172bcdb1d9eba98d5cf37e05a3">0cf1b8b91bc2e7172bcdb1d9eba98d5cf37e05a3</a></p> <p>Found in base branch: <b>feature_branch</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the Ansible Engine when using module_args. Tasks executed with check mode (--check-mode) do not properly neutralize sensitive data exposed in the event data. This flaw allows unauthorized users to read this data. The highest threat from this vulnerability is to confidentiality. <p>Publish Date: 2020-09-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14332>CVE-2020-14332</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-14332">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-14332</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 2.8.14,2.9.12</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"ansible","packageVersion":"2.9.9","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"ansible:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.8.14,2.9.12"}],"baseBranches":["feature_branch"],"vulnerabilityIdentifier":"CVE-2020-14332","vulnerabilityDetails":"A flaw was found in the Ansible Engine when using module_args. Tasks executed with check mode (--check-mode) do not properly neutralize sensitive data exposed in the event data. This flaw allows unauthorized users to read this data. The highest threat from this vulnerability is to confidentiality.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14332","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-14332 (Medium) detected in ansible-2.9.9.tar.gz - ## CVE-2020-14332 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.9.9.tar.gz</b></p></summary> <p>Radically simple IT automation</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz">https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz</a></p> <p>Path to dependency file: Haley-Sexton/requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **ansible-2.9.9.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Haley-Sexton/commit/0cf1b8b91bc2e7172bcdb1d9eba98d5cf37e05a3">0cf1b8b91bc2e7172bcdb1d9eba98d5cf37e05a3</a></p> <p>Found in base branch: <b>feature_branch</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the Ansible Engine when using module_args. Tasks executed with check mode (--check-mode) do not properly neutralize sensitive data exposed in the event data. This flaw allows unauthorized users to read this data. The highest threat from this vulnerability is to confidentiality. <p>Publish Date: 2020-09-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14332>CVE-2020-14332</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-14332">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-14332</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 2.8.14,2.9.12</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"ansible","packageVersion":"2.9.9","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"ansible:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.8.14,2.9.12"}],"baseBranches":["feature_branch"],"vulnerabilityIdentifier":"CVE-2020-14332","vulnerabilityDetails":"A flaw was found in the Ansible Engine when using module_args. Tasks executed with check mode (--check-mode) do not properly neutralize sensitive data exposed in the event data. This flaw allows unauthorized users to read this data. The highest threat from this vulnerability is to confidentiality.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14332","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> -->
non_main
cve medium detected in ansible tar gz cve medium severity vulnerability vulnerable library ansible tar gz radically simple it automation library home page a href path to dependency file haley sexton requirements txt path to vulnerable library requirements txt dependency hierarchy x ansible tar gz vulnerable library found in head commit a href found in base branch feature branch vulnerability details a flaw was found in the ansible engine when using module args tasks executed with check mode check mode do not properly neutralize sensitive data exposed in the event data this flaw allows unauthorized users to read this data the highest threat from this vulnerability is to confidentiality publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree ansible isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails a flaw was found in the ansible engine when using module args tasks executed with check mode check mode do not properly neutralize sensitive data exposed in the event data this flaw allows unauthorized users to read this data the highest threat from this vulnerability is to confidentiality vulnerabilityurl
0
4,124
19,547,178,924
IssuesEvent
2022-01-02 03:52:49
go-playground/validator
https://api.github.com/repos/go-playground/validator
opened
⚠️Validator - Looking for Additional Maintainers
help wanted maintainers
The Validator package is looking for additional maintainers to help with the growing number inquiries, Pull Requests and documentation needs. I can no longer fully dedicate the majority on my time to this one package 📦 I will continue to maintain and support this package but I don't want the community or contributors to suffer waiting for help or support due to my time constraints. This request is currently for the `validator` package but may extend to other packages within [go-playground](https://github.com/go-playground) in the future. Maintainer Core Asks: - A demonstrated history of OSS contributions. This is important not only to be trustworthy but also good with people. - Ideally contribute, review and help with issues for 1-3 months before being promoted to an admin. This will also provide the time to understand the projects current state, it's goals and where it's going. - Strong sense of responsibility. Many people will be counting on your thorough reviews and vetting of changed or additional code. Please join the go-playground discord server and request in the `#validator-maintainers-requests` channel [here](https://discord.gg/WdMCY3bwGw) for anyone interested :) Thank You!
True
⚠️Validator - Looking for Additional Maintainers - The Validator package is looking for additional maintainers to help with the growing number inquiries, Pull Requests and documentation needs. I can no longer fully dedicate the majority on my time to this one package 📦 I will continue to maintain and support this package but I don't want the community or contributors to suffer waiting for help or support due to my time constraints. This request is currently for the `validator` package but may extend to other packages within [go-playground](https://github.com/go-playground) in the future. Maintainer Core Asks: - A demonstrated history of OSS contributions. This is important not only to be trustworthy but also good with people. - Ideally contribute, review and help with issues for 1-3 months before being promoted to an admin. This will also provide the time to understand the projects current state, it's goals and where it's going. - Strong sense of responsibility. Many people will be counting on your thorough reviews and vetting of changed or additional code. Please join the go-playground discord server and request in the `#validator-maintainers-requests` channel [here](https://discord.gg/WdMCY3bwGw) for anyone interested :) Thank You!
main
⚠️validator looking for additional maintainers the validator package is looking for additional maintainers to help with the growing number inquiries pull requests and documentation needs i can no longer fully dedicate the majority on my time to this one package 📦 i will continue to maintain and support this package but i don t want the community or contributors to suffer waiting for help or support due to my time constraints this request is currently for the validator package but may extend to other packages within in the future maintainer core asks a demonstrated history of oss contributions this is important not only to be trustworthy but also good with people ideally contribute review and help with issues for months before being promoted to an admin this will also provide the time to understand the projects current state it s goals and where it s going strong sense of responsibility many people will be counting on your thorough reviews and vetting of changed or additional code please join the go playground discord server and request in the validator maintainers requests channel for anyone interested thank you
1
1,017
4,804,581,604
IssuesEvent
2016-11-02 13:59:18
duckduckgo/zeroclickinfo-fathead
https://api.github.com/repos/duckduckgo/zeroclickinfo-fathead
closed
C++ Reference Docs: Need a new maintainer
Maintainer Submitted Mission: Programming
For now I can't anymore be the maintainer of the C++ Reference Docs fathead, so we need a new maintainer ! @duckduckgo/duckduckhack-contributors would anyone out there be interested ? Here are the pending issues : #301 #300 and #203. Good luck ! https://duck.co/ia/view/cppreference_doc
True
C++ Reference Docs: Need a new maintainer - For now I can't anymore be the maintainer of the C++ Reference Docs fathead, so we need a new maintainer ! @duckduckgo/duckduckhack-contributors would anyone out there be interested ? Here are the pending issues : #301 #300 and #203. Good luck ! https://duck.co/ia/view/cppreference_doc
main
c reference docs need a new maintainer for now i can t anymore be the maintainer of the c reference docs fathead so we need a new maintainer duckduckgo duckduckhack contributors would anyone out there be interested here are the pending issues and good luck
1
2,890
10,319,629,859
IssuesEvent
2019-08-30 18:04:10
backdrop-ops/contrib
https://api.github.com/repos/backdrop-ops/contrib
closed
Backdrop geo kit
Maintainer application Port in progress
I've made an initial port of a collection of drupal mapping projects : geophp, geocoder, geofield, and leaflet. All were straightforward, except geocoder, which is marked with the infamous “dependencies[] = ctools”. I followed the @jenlampton method in the addressfield project, and it works... That's located at https://github.com/gifad/backdrop-geokit
True
Backdrop geo kit - I've made an initial port of a collection of drupal mapping projects : geophp, geocoder, geofield, and leaflet. All were straightforward, except geocoder, which is marked with the infamous “dependencies[] = ctools”. I followed the @jenlampton method in the addressfield project, and it works... That's located at https://github.com/gifad/backdrop-geokit
main
backdrop geo kit i ve made an initial port of a collection of drupal mapping projects geophp geocoder geofield and leaflet all were straightforward except geocoder which is marked with the infamous “dependencies ctools” i followed the jenlampton method in the addressfield project and it works that s located at
1
4,218
20,869,490,092
IssuesEvent
2022-03-22 10:33:41
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
opened
Enhance MiKo_3111 to only underline 'EqualTo(0)'
enhancement Area: analyzer Area: maintainability
Enhance MiKo_3111 to only underline `EqualTo(0)` as that's where the issue is.
True
Enhance MiKo_3111 to only underline 'EqualTo(0)' - Enhance MiKo_3111 to only underline `EqualTo(0)` as that's where the issue is.
main
enhance miko to only underline equalto enhance miko to only underline equalto as that s where the issue is
1
132,847
5,194,216,607
IssuesEvent
2017-01-23 02:09:15
UPBGE/blender
https://api.github.com/repos/UPBGE/blender
closed
Feature Request: Logic Brick to Interact with Python Component
feature request low priority
If you have a python component, it would be nice to be able to present a logic-brick interface to it. **Use cases:** - A FPS movement script, that exposes functions like "duck" to the logic brick interface - A vehicle that expose functions like "drive" and "toot horn" - A "damage" component that provided a logic brick interface for "onDeath" and "onHit" as sensors. - A "keymapper" component that abstracts from "W" on the component to the "Forwards" action. - A "hover-car" component that allows the user to control the size of the flames coming out the back. - A weapon component that exposes a "reload" and a "shoot" interface. **Suggested interface from python:** Ideally two dictionaries of: ``` { component_function_name: {parameter_name:default_value, parameter_name:default_value} component_function_name: {parameter_name:default_value, parameter_name:default_value} } ``` But could be a 3-deep tuple instead: ``` ( component_function_name, ((parameter_name, default_value), (parameter_name, default_value)) ) ``` It needs two of them, one to allow the python component to act as a sensor, one to allow it to act as an actuator. **Suggested interface from logic bricks:** - A drop down box to select from python components attached to the body - The parameters appear in the brick
1.0
Feature Request: Logic Brick to Interact with Python Component - If you have a python component, it would be nice to be able to present a logic-brick interface to it. **Use cases:** - A FPS movement script, that exposes functions like "duck" to the logic brick interface - A vehicle that expose functions like "drive" and "toot horn" - A "damage" component that provided a logic brick interface for "onDeath" and "onHit" as sensors. - A "keymapper" component that abstracts from "W" on the component to the "Forwards" action. - A "hover-car" component that allows the user to control the size of the flames coming out the back. - A weapon component that exposes a "reload" and a "shoot" interface. **Suggested interface from python:** Ideally two dictionaries of: ``` { component_function_name: {parameter_name:default_value, parameter_name:default_value} component_function_name: {parameter_name:default_value, parameter_name:default_value} } ``` But could be a 3-deep tuple instead: ``` ( component_function_name, ((parameter_name, default_value), (parameter_name, default_value)) ) ``` It needs two of them, one to allow the python component to act as a sensor, one to allow it to act as an actuator. **Suggested interface from logic bricks:** - A drop down box to select from python components attached to the body - The parameters appear in the brick
non_main
feature request logic brick to interact with python component if you have a python component it would be nice to be able to present a logic brick interface to it use cases a fps movement script that exposes functions like duck to the logic brick interface a vehicle that expose functions like drive and toot horn a damage component that provided a logic brick interface for ondeath and onhit as sensors a keymapper component that abstracts from w on the component to the forwards action a hover car component that allows the user to control the size of the flames coming out the back a weapon component that exposes a reload and a shoot interface suggested interface from python ideally two dictionaries of component function name parameter name default value parameter name default value component function name parameter name default value parameter name default value but could be a deep tuple instead component function name parameter name default value parameter name default value it needs two of them one to allow the python component to act as a sensor one to allow it to act as an actuator suggested interface from logic bricks a drop down box to select from python components attached to the body the parameters appear in the brick
0
56,936
23,963,376,700
IssuesEvent
2022-09-12 21:24:26
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
closed
dms-vpc-role is not configured properly when creating aws_dms_replication_instance
bug service/iam service/dms
This is a similar (or the same) issue as [terraform-providers/terraform-provider-aws#7748](https://github.com/terraform-providers/terraform-provider-aws/issues/7748) which was closed. <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ``` Terraform v0.12.16 + provider.aws v2.39.0 ``` ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * aws_dms_replication_subnet_group * aws_dms_replication_instance ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl # Roles defined as per official documentation: # https://www.terraform.io/docs/providers/aws/r/dms_replication_instance.html # Database Migration Service requires the below IAM Roles to be created before # replication instances can be created. See the DMS Documentation for # additional information: https://docs.aws.amazon.com/dms/latest/userguide/CHAP_Security.APIRole.html # * dms-vpc-role # * dms-cloudwatch-logs-role # * dms-access-for-endpoint data "aws_iam_policy_document" "dms_assume_role" { statement { actions = ["sts:AssumeRole"] principals { identifiers = ["dms.amazonaws.com"] type = "Service" } } } resource "aws_iam_role" "dms-access-for-endpoint" { assume_role_policy = "${data.aws_iam_policy_document.dms_assume_role.json}" name = "dms-access-for-endpoint" } resource "aws_iam_role_policy_attachment" "dms-access-for-endpoint-AmazonDMSRedshiftS3Role" { policy_arn = "arn:aws:iam::aws:policy/service-role/AmazonDMSRedshiftS3Role" role = "${aws_iam_role.dms-access-for-endpoint.name}" } resource "aws_iam_role" "dms-cloudwatch-logs-role" { assume_role_policy = "${data.aws_iam_policy_document.dms_assume_role.json}" name = "dms-cloudwatch-logs-role" } resource "aws_iam_role_policy_attachment" "dms-cloudwatch-logs-role-AmazonDMSCloudWatchLogsRole" { policy_arn = "arn:aws:iam::aws:policy/service-role/AmazonDMSCloudWatchLogsRole" role = "${aws_iam_role.dms-cloudwatch-logs-role.name}" } resource "aws_iam_role" "dms-vpc-role" { assume_role_policy = "${data.aws_iam_policy_document.dms_assume_role.json}" name = "dms-vpc-role" } resource "aws_iam_role_policy_attachment" "dms-vpc-role-AmazonDMSVPCManagementRole" { policy_arn = "arn:aws:iam::aws:policy/service-role/AmazonDMSVPCManagementRole" role = "${aws_iam_role.dms-vpc-role.name}" } # Issue when creating aws_dms_replication_subnet_group # (required for aws_dms_replication_instance) resource "aws_dms_replication_subnet_group" "replication_subnet" { replication_subnet_group_description = "Test replication subnet group" replication_subnet_group_id = "test-dms-replication-subnet-group-tf" subnet_ids = "${aws_subnet.database_subnet.*.id}" # Explicit depends_on for required role depends_on = ["aws_iam_role.dms-vpc-role"] } ``` ### Debug Output Error applying plan: * Error: AccessDeniedFault: The IAM Role arn:aws:iam::xxxxxxxx:role/dms-vpc-role is not configured properly. status code: 400, request id: xxxxxxxx * on dms.tf line xxx, in resource "aws_dms_replication_subnet_group" "replication_subnet": xxx: resource "aws_dms_replication_subnet_group" "replication_subnet" { ### Expected Behavior On first `terraform apply`: * * Apply complete! Resources: X added, 0 changed, 0 destroyed. ### Actual Behavior On first `terraform apply`: Error applying plan: * Error: AccessDeniedFault: The IAM Role arn:aws:iam::xxxxxxxx:role/dms-vpc-role is not configured properly. status code: 400, request id: xxxxxxxx * on dms.tf line xxx, in resource "aws_dms_replication_subnet_group" "replication_subnet": xxx: resource "aws_dms_replication_subnet_group" "replication_subnet" { On second `terraform apply`: * Apply complete! Resources: X added, 0 changed, 0 destroyed. ### Steps to Reproduce 1. `terraform apply`
2.0
dms-vpc-role is not configured properly when creating aws_dms_replication_instance - This is a similar (or the same) issue as [terraform-providers/terraform-provider-aws#7748](https://github.com/terraform-providers/terraform-provider-aws/issues/7748) which was closed. <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ``` Terraform v0.12.16 + provider.aws v2.39.0 ``` ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * aws_dms_replication_subnet_group * aws_dms_replication_instance ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl # Roles defined as per official documentation: # https://www.terraform.io/docs/providers/aws/r/dms_replication_instance.html # Database Migration Service requires the below IAM Roles to be created before # replication instances can be created. See the DMS Documentation for # additional information: https://docs.aws.amazon.com/dms/latest/userguide/CHAP_Security.APIRole.html # * dms-vpc-role # * dms-cloudwatch-logs-role # * dms-access-for-endpoint data "aws_iam_policy_document" "dms_assume_role" { statement { actions = ["sts:AssumeRole"] principals { identifiers = ["dms.amazonaws.com"] type = "Service" } } } resource "aws_iam_role" "dms-access-for-endpoint" { assume_role_policy = "${data.aws_iam_policy_document.dms_assume_role.json}" name = "dms-access-for-endpoint" } resource "aws_iam_role_policy_attachment" "dms-access-for-endpoint-AmazonDMSRedshiftS3Role" { policy_arn = "arn:aws:iam::aws:policy/service-role/AmazonDMSRedshiftS3Role" role = "${aws_iam_role.dms-access-for-endpoint.name}" } resource "aws_iam_role" "dms-cloudwatch-logs-role" { assume_role_policy = "${data.aws_iam_policy_document.dms_assume_role.json}" name = "dms-cloudwatch-logs-role" } resource "aws_iam_role_policy_attachment" "dms-cloudwatch-logs-role-AmazonDMSCloudWatchLogsRole" { policy_arn = "arn:aws:iam::aws:policy/service-role/AmazonDMSCloudWatchLogsRole" role = "${aws_iam_role.dms-cloudwatch-logs-role.name}" } resource "aws_iam_role" "dms-vpc-role" { assume_role_policy = "${data.aws_iam_policy_document.dms_assume_role.json}" name = "dms-vpc-role" } resource "aws_iam_role_policy_attachment" "dms-vpc-role-AmazonDMSVPCManagementRole" { policy_arn = "arn:aws:iam::aws:policy/service-role/AmazonDMSVPCManagementRole" role = "${aws_iam_role.dms-vpc-role.name}" } # Issue when creating aws_dms_replication_subnet_group # (required for aws_dms_replication_instance) resource "aws_dms_replication_subnet_group" "replication_subnet" { replication_subnet_group_description = "Test replication subnet group" replication_subnet_group_id = "test-dms-replication-subnet-group-tf" subnet_ids = "${aws_subnet.database_subnet.*.id}" # Explicit depends_on for required role depends_on = ["aws_iam_role.dms-vpc-role"] } ``` ### Debug Output Error applying plan: * Error: AccessDeniedFault: The IAM Role arn:aws:iam::xxxxxxxx:role/dms-vpc-role is not configured properly. status code: 400, request id: xxxxxxxx * on dms.tf line xxx, in resource "aws_dms_replication_subnet_group" "replication_subnet": xxx: resource "aws_dms_replication_subnet_group" "replication_subnet" { ### Expected Behavior On first `terraform apply`: * * Apply complete! Resources: X added, 0 changed, 0 destroyed. ### Actual Behavior On first `terraform apply`: Error applying plan: * Error: AccessDeniedFault: The IAM Role arn:aws:iam::xxxxxxxx:role/dms-vpc-role is not configured properly. status code: 400, request id: xxxxxxxx * on dms.tf line xxx, in resource "aws_dms_replication_subnet_group" "replication_subnet": xxx: resource "aws_dms_replication_subnet_group" "replication_subnet" { On second `terraform apply`: * Apply complete! Resources: X added, 0 changed, 0 destroyed. ### Steps to Reproduce 1. `terraform apply`
non_main
dms vpc role is not configured properly when creating aws dms replication instance this is a similar or the same issue as which was closed community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version terraform provider aws affected resource s aws dms replication subnet group aws dms replication instance terraform configuration files hcl  roles defined as per official documentation   database migration service requires the below iam roles to be created before replication instances can be created see the dms documentation for additional information dms vpc role dms cloudwatch logs role dms access for endpoint data aws iam policy document dms assume role statement actions principals identifiers type service resource aws iam role dms access for endpoint assume role policy data aws iam policy document dms assume role json name dms access for endpoint resource aws iam role policy attachment dms access for endpoint policy arn arn aws iam aws policy service role role aws iam role dms access for endpoint name resource aws iam role dms cloudwatch logs role assume role policy data aws iam policy document dms assume role json name dms cloudwatch logs role resource aws iam role policy attachment dms cloudwatch logs role amazondmscloudwatchlogsrole policy arn arn aws iam aws policy service role amazondmscloudwatchlogsrole role aws iam role dms cloudwatch logs role name resource aws iam role dms vpc role assume role policy data aws iam policy document dms assume role json name dms vpc role resource aws iam role policy attachment dms vpc role amazondmsvpcmanagementrole policy arn arn aws iam aws policy service role amazondmsvpcmanagementrole role aws iam role dms vpc role name issue when creating aws dms replication subnet group required for aws dms replication instance resource aws dms replication subnet group replication subnet replication subnet group description test replication subnet group replication subnet group id test dms replication subnet group tf subnet ids aws subnet database subnet id explicit depends on for required role depends on debug output error applying plan error accessdeniedfault the iam role arn aws iam xxxxxxxx role dms vpc role is not configured properly status code request id xxxxxxxx on dms tf line xxx in resource aws dms replication subnet group replication subnet xxx resource aws dms replication subnet group replication subnet expected behavior on first terraform apply apply complete resources x added changed destroyed actual behavior on first terraform apply error applying plan error accessdeniedfault the iam role arn aws iam xxxxxxxx role dms vpc role is not configured properly status code request id xxxxxxxx on dms tf line xxx in resource aws dms replication subnet group replication subnet xxx resource aws dms replication subnet group replication subnet on second terraform apply apply complete resources x added changed destroyed steps to reproduce terraform apply
0
730,825
25,190,570,690
IssuesEvent
2022-11-12 00:01:48
simonbaird/tiddlyhost
https://api.github.com/repos/simonbaird/tiddlyhost
closed
Feature request: Clone wiki
priority
Shortly after the noob stage, a TW user identifies "favourite plugins, settings and customizations" that he wants for all his wikis. I dare say this happens to *all* TW users. And those that go deeper into tiddlyverse probably develop fine tuned recurring setups (e.g *public* vs *private* wikis, *work* vs *non-work* etc). But to manually drag'n drop plugins + modified shadow tids and other tidbits is a rather annoying task. Therefore, I wonder if TH could feature a simple "Clone wiki" feature in the *Your sites* page. It could appear as a menu option in the "Actions" button and it can lead to the same page as the "Create site" button, i.e to register a new TH site, but instead of an empty wiki it is a clone. Thoughts?
1.0
Feature request: Clone wiki - Shortly after the noob stage, a TW user identifies "favourite plugins, settings and customizations" that he wants for all his wikis. I dare say this happens to *all* TW users. And those that go deeper into tiddlyverse probably develop fine tuned recurring setups (e.g *public* vs *private* wikis, *work* vs *non-work* etc). But to manually drag'n drop plugins + modified shadow tids and other tidbits is a rather annoying task. Therefore, I wonder if TH could feature a simple "Clone wiki" feature in the *Your sites* page. It could appear as a menu option in the "Actions" button and it can lead to the same page as the "Create site" button, i.e to register a new TH site, but instead of an empty wiki it is a clone. Thoughts?
non_main
feature request clone wiki shortly after the noob stage a tw user identifies favourite plugins settings and customizations that he wants for all his wikis i dare say this happens to all tw users and those that go deeper into tiddlyverse probably develop fine tuned recurring setups e g public vs private wikis work vs non work etc but to manually drag n drop plugins modified shadow tids and other tidbits is a rather annoying task therefore i wonder if th could feature a simple clone wiki feature in the your sites page it could appear as a menu option in the actions button and it can lead to the same page as the create site button i e to register a new th site but instead of an empty wiki it is a clone thoughts
0
5,394
27,096,282,619
IssuesEvent
2023-02-15 03:18:51
OpenLightingProject/ola
https://api.github.com/repos/OpenLightingProject/ola
closed
Python version check bug
question Maintainability API-Python Language-Python
Hi Your python version check in OlaClient.py fails on Python 3.10 line 957 if sys.version >= '3.2': request.data = data.tobytes() else: request.data = data.tostring() The best solution I can see, others may have a better solution, is pv = platform.python_version_tuple() if int(pv[0]) >= 3 and int(pv[1]) > 2: request.data = data.tobytes() else: request.data = data.tostring()
True
Python version check bug - Hi Your python version check in OlaClient.py fails on Python 3.10 line 957 if sys.version >= '3.2': request.data = data.tobytes() else: request.data = data.tostring() The best solution I can see, others may have a better solution, is pv = platform.python_version_tuple() if int(pv[0]) >= 3 and int(pv[1]) > 2: request.data = data.tobytes() else: request.data = data.tostring()
main
python version check bug hi your python version check in olaclient py fails on python line if sys version request data data tobytes else request data data tostring the best solution i can see others may have a better solution is pv platform python version tuple if int pv and int pv request data data tobytes else request data data tostring
1
711
4,288,799,410
IssuesEvent
2016-07-17 18:00:33
topogram/topogram
https://api.github.com/repos/topogram/topogram
closed
update to Meteor 1.3 / ES2015 design patterns
enhancement maintaining
This is a trillion time better, as we can control imports, require files, etc. Instead of having a big autoload, you just have clean importing / dependencies patterns. It will make code stronger, easy to test and debug, plus easier to move out Meteor if needed one day. check the whole story here http://guide.meteor.com/structure.html this is a big change, but will totally push topogram from alpha to beta !
True
update to Meteor 1.3 / ES2015 design patterns - This is a trillion time better, as we can control imports, require files, etc. Instead of having a big autoload, you just have clean importing / dependencies patterns. It will make code stronger, easy to test and debug, plus easier to move out Meteor if needed one day. check the whole story here http://guide.meteor.com/structure.html this is a big change, but will totally push topogram from alpha to beta !
main
update to meteor design patterns this is a trillion time better as we can control imports require files etc instead of having a big autoload you just have clean importing dependencies patterns it will make code stronger easy to test and debug plus easier to move out meteor if needed one day check the whole story here this is a big change but will totally push topogram from alpha to beta
1
1,208
5,162,389,391
IssuesEvent
2017-01-17 00:24:59
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
iam_cert doesn't return the ARN value after creating a certificate
affects_2.0 aws bug_report cloud feature_idea waiting_on_maintainer
##### Issue Type: <!--- Please pick one and delete the rest: --> - Bug Report ##### Ansible Version: ``` ansible 2.0.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### Environment: Ubuntu 14.04.4 LTS ##### Summary: I'm using the `iam_cert` module to create a certificate to use with ELB. I've registered a variable with the results of running that module and dumped it. There's no ARN field that I need to feed to the `ec2_elb_lb` module. ##### Steps To Reproduce: ``` - name: Upload SSL certificate iam_cert: name: ssl state: present cert: cert.pem key: cert.key register: ssl_certificate - debug: var=ssl_certificate ``` ##### Expected Results: I expect the ARN field in the returned result. AWS CLI gives that information: ``` $ aws iam list-server-certificates { "ServerCertificateMetadataList": [ { "ServerCertificateId": "AS...SU", "ServerCertificateName": "ssl", "Expiration": "2025-11-16T12:52:40Z", "Path": "/", "Arn": "arn:aws:iam::4...9:server-certificate/ssl", "UploadDate": "2016-03-03T21:46:09Z" } ] } ``` So the info is there, but Ansible doesn't use it. ##### Actual Results: ``` ok: [localhost] => { "ssl_certificate": { "cert_body": "-----BEGIN CERTIFICATE-----\n...\n-----END CERTIFICATE-----", "cert_path": "/", "changed": false, "expiration_date": "2025-11-16T12:52:40Z", "msg": "No new path or name specified. No changes made", "name": "ssl", "upload_date": "2016-03-03T21:46:09Z" } } ```
True
iam_cert doesn't return the ARN value after creating a certificate - ##### Issue Type: <!--- Please pick one and delete the rest: --> - Bug Report ##### Ansible Version: ``` ansible 2.0.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### Environment: Ubuntu 14.04.4 LTS ##### Summary: I'm using the `iam_cert` module to create a certificate to use with ELB. I've registered a variable with the results of running that module and dumped it. There's no ARN field that I need to feed to the `ec2_elb_lb` module. ##### Steps To Reproduce: ``` - name: Upload SSL certificate iam_cert: name: ssl state: present cert: cert.pem key: cert.key register: ssl_certificate - debug: var=ssl_certificate ``` ##### Expected Results: I expect the ARN field in the returned result. AWS CLI gives that information: ``` $ aws iam list-server-certificates { "ServerCertificateMetadataList": [ { "ServerCertificateId": "AS...SU", "ServerCertificateName": "ssl", "Expiration": "2025-11-16T12:52:40Z", "Path": "/", "Arn": "arn:aws:iam::4...9:server-certificate/ssl", "UploadDate": "2016-03-03T21:46:09Z" } ] } ``` So the info is there, but Ansible doesn't use it. ##### Actual Results: ``` ok: [localhost] => { "ssl_certificate": { "cert_body": "-----BEGIN CERTIFICATE-----\n...\n-----END CERTIFICATE-----", "cert_path": "/", "changed": false, "expiration_date": "2025-11-16T12:52:40Z", "msg": "No new path or name specified. No changes made", "name": "ssl", "upload_date": "2016-03-03T21:46:09Z" } } ```
main
iam cert doesn t return the arn value after creating a certificate issue type bug report ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides environment ubuntu lts summary i m using the iam cert module to create a certificate to use with elb i ve registered a variable with the results of running that module and dumped it there s no arn field that i need to feed to the elb lb module steps to reproduce name upload ssl certificate iam cert name ssl state present cert cert pem key cert key register ssl certificate debug var ssl certificate expected results i expect the arn field in the returned result aws cli gives that information aws iam list server certificates servercertificatemetadatalist servercertificateid as su servercertificatename ssl expiration path arn arn aws iam server certificate ssl uploaddate so the info is there but ansible doesn t use it actual results ok ssl certificate cert body begin certificate n n end certificate cert path changed false expiration date msg no new path or name specified no changes made name ssl upload date
1
673,993
23,036,238,715
IssuesEvent
2022-07-22 19:06:21
GoogleCloudPlatform/golang-samples
https://api.github.com/repos/GoogleCloudPlatform/golang-samples
closed
compute/custom-hostname-instance: TestInstanceHostnameSnippets failed
type: bug priority: p1 api: compute samples flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: a9c8603f3a8d007af2f7d0258fa9777254f421b3 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f6f05963-eece-411b-b423-5a4d2aa943c6), [Sponge](http://sponge2/f6f05963-eece-411b-b423-5a4d2aa943c6) status: failed <details><summary>Test output</summary><br><pre> instance_hostname_test.go:51: createInstanceWithCustomHostname got err: unable to create instance: googleapi: Error 503: Internal error. Please try again or contact Google Support. (Code: '5E453FC690AA2.6E64DD7.3E034DC3') instance_hostname_test.go:56: createInstanceWithCustomHostname got "", want "Instance created" instance_hostname_test.go:67: unable to get instance: googleapi: Error 404: The resource 'projects/golang-samples-tests-2/zones/europe-central2-b/instances/test-3783530460895391398' was not found instance_hostname_test.go:71: Instance has incorrect hostname got "", want "host.domain.com" instance_hostname_test.go:77: getInstanceHostname got err: unable to get instance: googleapi: Error 404: The resource 'projects/golang-samples-tests-2/zones/europe-central2-b/instances/test-3783530460895391398' was not found instance_hostname_test.go:82: getInstanceHostname got "", want "Instance test-3783530460895391398 has hostname: " instance_hostname_test.go:93: unable to delete instance: googleapi: Error 404: The resource 'projects/golang-samples-tests-2/zones/europe-central2-b/instances/test-3783530460895391398' was not found panic: runtime error: invalid memory address or nil pointer dereference /usr/local/go/src/testing/testing.go:1209 +0x24e /usr/local/go/src/testing/testing.go:1212 +0x218 /usr/local/go/src/runtime/panic.go:1038 +0x215 /go/pkg/mod/cloud.google.com/go/compute@v1.6.1/apiv1/operations.go:51 +0x92 /tmpfs/src/github/golang-samples/compute/custom-hostname-instance/instance_hostname_test.go:96 +0xae8 /usr/local/go/src/testing/testing.go:1259 +0x102 /usr/local/go/src/testing/testing.go:1306 +0x35a</pre></details>
1.0
compute/custom-hostname-instance: TestInstanceHostnameSnippets failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: a9c8603f3a8d007af2f7d0258fa9777254f421b3 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f6f05963-eece-411b-b423-5a4d2aa943c6), [Sponge](http://sponge2/f6f05963-eece-411b-b423-5a4d2aa943c6) status: failed <details><summary>Test output</summary><br><pre> instance_hostname_test.go:51: createInstanceWithCustomHostname got err: unable to create instance: googleapi: Error 503: Internal error. Please try again or contact Google Support. (Code: '5E453FC690AA2.6E64DD7.3E034DC3') instance_hostname_test.go:56: createInstanceWithCustomHostname got "", want "Instance created" instance_hostname_test.go:67: unable to get instance: googleapi: Error 404: The resource 'projects/golang-samples-tests-2/zones/europe-central2-b/instances/test-3783530460895391398' was not found instance_hostname_test.go:71: Instance has incorrect hostname got "", want "host.domain.com" instance_hostname_test.go:77: getInstanceHostname got err: unable to get instance: googleapi: Error 404: The resource 'projects/golang-samples-tests-2/zones/europe-central2-b/instances/test-3783530460895391398' was not found instance_hostname_test.go:82: getInstanceHostname got "", want "Instance test-3783530460895391398 has hostname: " instance_hostname_test.go:93: unable to delete instance: googleapi: Error 404: The resource 'projects/golang-samples-tests-2/zones/europe-central2-b/instances/test-3783530460895391398' was not found panic: runtime error: invalid memory address or nil pointer dereference /usr/local/go/src/testing/testing.go:1209 +0x24e /usr/local/go/src/testing/testing.go:1212 +0x218 /usr/local/go/src/runtime/panic.go:1038 +0x215 /go/pkg/mod/cloud.google.com/go/compute@v1.6.1/apiv1/operations.go:51 +0x92 /tmpfs/src/github/golang-samples/compute/custom-hostname-instance/instance_hostname_test.go:96 +0xae8 /usr/local/go/src/testing/testing.go:1259 +0x102 /usr/local/go/src/testing/testing.go:1306 +0x35a</pre></details>
non_main
compute custom hostname instance testinstancehostnamesnippets failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output instance hostname test go createinstancewithcustomhostname got err unable to create instance googleapi error internal error please try again or contact google support code instance hostname test go createinstancewithcustomhostname got want instance created instance hostname test go unable to get instance googleapi error the resource projects golang samples tests zones europe b instances test was not found instance hostname test go instance has incorrect hostname got want host domain com instance hostname test go getinstancehostname got err unable to get instance googleapi error the resource projects golang samples tests zones europe b instances test was not found instance hostname test go getinstancehostname got want instance test has hostname instance hostname test go unable to delete instance googleapi error the resource projects golang samples tests zones europe b instances test was not found panic runtime error invalid memory address or nil pointer dereference usr local go src testing testing go usr local go src testing testing go usr local go src runtime panic go go pkg mod cloud google com go compute operations go tmpfs src github golang samples compute custom hostname instance instance hostname test go usr local go src testing testing go usr local go src testing testing go
0
301,416
26,045,733,774
IssuesEvent
2022-12-22 14:15:04
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
closed
[UI test automation] Improve coverage of the full functional UI test suite
eng:ui-test
### Summary Improve coverage of the [Full functional UI test suite](https://testrail.stage.mozaws.net/index.php?/suites/view/3192&group_by=cases:section_id&group_order=asc&group_id=45740): • Create new UI tests • Add missing steps ┆Issue is synchronized with this [Jira Task](https://mozilla-hub.atlassian.net/browse/FNXV2-21180)
1.0
[UI test automation] Improve coverage of the full functional UI test suite - ### Summary Improve coverage of the [Full functional UI test suite](https://testrail.stage.mozaws.net/index.php?/suites/view/3192&group_by=cases:section_id&group_order=asc&group_id=45740): • Create new UI tests • Add missing steps ┆Issue is synchronized with this [Jira Task](https://mozilla-hub.atlassian.net/browse/FNXV2-21180)
non_main
improve coverage of the full functional ui test suite summary improve coverage of the • create new ui tests • add missing steps ┆issue is synchronized with this
0
251,366
21,475,014,185
IssuesEvent
2022-04-26 12:59:45
kubernetes-sigs/cluster-api
https://api.github.com/repos/kubernetes-sigs/cluster-api
reopened
Integrate internal/matchers into Komega style test abstraction in Controller Runtime
kind/feature lifecycle/rotten area/testing
**User Story** As a developer I would like to have a consistent abstract toolset to quickly write unit tests in Cluster API. **Detailed Description** The matchers package added in #5259 brings generic matching to Cluster API objects and makes it easier to match only on required fields at test time. In controller runtime a similar approach with some differences - called Komega in its current iteration at https://github.com/kubernetes-sigs/controller-runtime/pull/1364 These approaches could be integrated inside controller runtime to bring generic json based matching with the Komega abstraction over the controller runtime client to reduce the amount of boiler plate code involved in writing these tests. /kind feature /area testing
1.0
Integrate internal/matchers into Komega style test abstraction in Controller Runtime - **User Story** As a developer I would like to have a consistent abstract toolset to quickly write unit tests in Cluster API. **Detailed Description** The matchers package added in #5259 brings generic matching to Cluster API objects and makes it easier to match only on required fields at test time. In controller runtime a similar approach with some differences - called Komega in its current iteration at https://github.com/kubernetes-sigs/controller-runtime/pull/1364 These approaches could be integrated inside controller runtime to bring generic json based matching with the Komega abstraction over the controller runtime client to reduce the amount of boiler plate code involved in writing these tests. /kind feature /area testing
non_main
integrate internal matchers into komega style test abstraction in controller runtime user story as a developer i would like to have a consistent abstract toolset to quickly write unit tests in cluster api detailed description the matchers package added in brings generic matching to cluster api objects and makes it easier to match only on required fields at test time in controller runtime a similar approach with some differences called komega in its current iteration at these approaches could be integrated inside controller runtime to bring generic json based matching with the komega abstraction over the controller runtime client to reduce the amount of boiler plate code involved in writing these tests kind feature area testing
0
805,078
29,507,282,137
IssuesEvent
2023-06-03 13:18:27
exellian/rex
https://api.github.com/repos/exellian/rex
opened
DRAFT: Add template children syntax
priority: low
I should add a syntax for embedding children into a template like contentChildren or props.children.
1.0
DRAFT: Add template children syntax - I should add a syntax for embedding children into a template like contentChildren or props.children.
non_main
draft add template children syntax i should add a syntax for embedding children into a template like contentchildren or props children
0
181,482
14,040,118,963
IssuesEvent
2020-11-01 00:25:29
trailheadapps/lwc-recipes
https://api.github.com/repos/trailheadapps/lwc-recipes
closed
Implement exception handling in c/mortgage component
new-test no-issue-activity
This component doesn't implement any exception handling, or handle false inputs that are not number-based. It should implement some, and add tests that exercise the logic. Also, look into implementing the static picklist values as a getter instead of a function. For reference see this [comment](https://github.com/trailheadapps/lwc-recipes/pull/326#discussion_r453829591) on #326 .
1.0
Implement exception handling in c/mortgage component - This component doesn't implement any exception handling, or handle false inputs that are not number-based. It should implement some, and add tests that exercise the logic. Also, look into implementing the static picklist values as a getter instead of a function. For reference see this [comment](https://github.com/trailheadapps/lwc-recipes/pull/326#discussion_r453829591) on #326 .
non_main
implement exception handling in c mortgage component this component doesn t implement any exception handling or handle false inputs that are not number based it should implement some and add tests that exercise the logic also look into implementing the static picklist values as a getter instead of a function for reference see this on
0
670
4,212,800,213
IssuesEvent
2016-06-29 17:14:26
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
closed
Tips IA maintainer request
Maintainer Input Requested
There has been several maintainer timeouts for the Tips IA. If @mattlehning does not have any objections I would like to takeover as maintainer. Thanks! https://duck.co/ia/view/tips cc @moollaza @edgesince84 @zekiel
True
Tips IA maintainer request - There has been several maintainer timeouts for the Tips IA. If @mattlehning does not have any objections I would like to takeover as maintainer. Thanks! https://duck.co/ia/view/tips cc @moollaza @edgesince84 @zekiel
main
tips ia maintainer request there has been several maintainer timeouts for the tips ia if mattlehning does not have any objections i would like to takeover as maintainer thanks cc moollaza zekiel
1
81,055
30,688,674,619
IssuesEvent
2023-07-26 13:55:29
DependencyTrack/dependency-track
https://api.github.com/repos/DependencyTrack/dependency-track
opened
Vulnerability Project API
defect in triage
### Current Behavior Using utilising api endpoint `/api/v1/vulnerability/project/{project_id}` `offset` query parameter doesn't work. Changing `offset` parameter doesn't change output from endpoint. Tested as well with query parameters `limit` & `page` and `pageNumber` & `pageSize` When using `offset` query parameter in endpoint `/api/v1/component/project/{project_id}` endpoint works as expected. ### Steps to Reproduce 1. Create project 2. Upload SBOM file containing components having vulnerabilities 3. Call `/api/v1/vulnerability/project/{project_id}` with `offset=0` 4. Call `/api/v1/vulnerability/project/{project_id}` with `offset=1` 5. Compare response ### Expected Behavior If project contains vulnerabilities then response from third and fourth step should be different ### Dependency-Track Version 4.8.2 ### Dependency-Track Distribution Container Image ### Database Server N/A ### Database Server Version _No response_ ### Browser N/A ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
1.0
Vulnerability Project API - ### Current Behavior Using utilising api endpoint `/api/v1/vulnerability/project/{project_id}` `offset` query parameter doesn't work. Changing `offset` parameter doesn't change output from endpoint. Tested as well with query parameters `limit` & `page` and `pageNumber` & `pageSize` When using `offset` query parameter in endpoint `/api/v1/component/project/{project_id}` endpoint works as expected. ### Steps to Reproduce 1. Create project 2. Upload SBOM file containing components having vulnerabilities 3. Call `/api/v1/vulnerability/project/{project_id}` with `offset=0` 4. Call `/api/v1/vulnerability/project/{project_id}` with `offset=1` 5. Compare response ### Expected Behavior If project contains vulnerabilities then response from third and fourth step should be different ### Dependency-Track Version 4.8.2 ### Dependency-Track Distribution Container Image ### Database Server N/A ### Database Server Version _No response_ ### Browser N/A ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
non_main
vulnerability project api current behavior using utilising api endpoint api vulnerability project project id offset query parameter doesn t work changing offset parameter doesn t change output from endpoint tested as well with query parameters limit page and pagenumber pagesize when using offset query parameter in endpoint api component project project id endpoint works as expected steps to reproduce create project upload sbom file containing components having vulnerabilities call api vulnerability project project id with offset call api vulnerability project project id with offset compare response expected behavior if project contains vulnerabilities then response from third and fourth step should be different dependency track version dependency track distribution container image database server n a database server version no response browser n a checklist i have read and understand the i have checked the for whether this defect was already reported
0
1,148
5,005,855,851
IssuesEvent
2016-12-12 12:10:23
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
Feature request: Allow casks to be searchable by their names specified by `name`
awaiting maintainer feedback
### Description of feature/enhancement Sometimes, a cask has multiple names, all of which are used in the `name` stanza in its cask file. However, only the file name is searchable. Additional names in the `name` stanza should also be searchable. ### Justification This would certainly be helpful for programs with multiple names, or even multiple version, since the cask file names tend to be all abbreviated versions of each one's full name. ### Example use case Sometimes, a program's file name is its well-known acronym, but that means `cask search` for its full name returns no results, even if the full name is listed as an additional `name`. For example, the BOINC cask would not be found in `cask search` if you searched for "Berkeley" or any other of the full words of the name. Similarly, for the many different versions of PyCharm, `cask search pycharm` returns abbreviated suffixes for all 4 versions. I know I could `cask info` each, but it would probably be helpful if i could just search for part of the full name and get the right one immediately.
True
Feature request: Allow casks to be searchable by their names specified by `name` - ### Description of feature/enhancement Sometimes, a cask has multiple names, all of which are used in the `name` stanza in its cask file. However, only the file name is searchable. Additional names in the `name` stanza should also be searchable. ### Justification This would certainly be helpful for programs with multiple names, or even multiple version, since the cask file names tend to be all abbreviated versions of each one's full name. ### Example use case Sometimes, a program's file name is its well-known acronym, but that means `cask search` for its full name returns no results, even if the full name is listed as an additional `name`. For example, the BOINC cask would not be found in `cask search` if you searched for "Berkeley" or any other of the full words of the name. Similarly, for the many different versions of PyCharm, `cask search pycharm` returns abbreviated suffixes for all 4 versions. I know I could `cask info` each, but it would probably be helpful if i could just search for part of the full name and get the right one immediately.
main
feature request allow casks to be searchable by their names specified by name description of feature enhancement sometimes a cask has multiple names all of which are used in the name stanza in its cask file however only the file name is searchable additional names in the name stanza should also be searchable justification this would certainly be helpful for programs with multiple names or even multiple version since the cask file names tend to be all abbreviated versions of each one s full name example use case sometimes a program s file name is its well known acronym but that means cask search for its full name returns no results even if the full name is listed as an additional name for example the boinc cask would not be found in cask search if you searched for berkeley or any other of the full words of the name similarly for the many different versions of pycharm cask search pycharm returns abbreviated suffixes for all versions i know i could cask info each but it would probably be helpful if i could just search for part of the full name and get the right one immediately
1
1,200
5,133,075,979
IssuesEvent
2017-01-11 01:41:13
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Deleting elasticache cluster on redis throws TypeError
affects_2.0 aws bug_report cloud P2 waiting_on_maintainer
##### Issue Type: - Bug Report ##### Component Name: elasticache module ##### Ansible Version: ansible 2.0.0.2 ##### Ansible Configuration: `any_errors_fatal = True` ##### Environment: redis on aws elasticache ##### Summary: Errors on deleting cluster. ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 2727, in <module> main() File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 562, in main module.exit_json(**facts_result) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 2099, in exit_json kwargs = remove_values(kwargs, self.no_log_values) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in remove_values return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in <genexpr> return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in remove_values return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in <genexpr> return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in remove_values return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in <genexpr> return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 980, in remove_values return [remove_values(elem, no_log_strings) for elem in value] File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 991, in remove_values raise TypeError('Value of unknown type: %s, %s' % (type(value), value)) TypeError: Value of unknown type: <type 'object'>, <object object at 0x10bf01100> fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "elasticache"}, "parsed": false} ``` ##### Steps To Reproduce: Create cluster with elasticache module. Attempt to delete cluster with `state: absent`. If it helps, the cluster name has a hyphen. e.g. `abc-redis` ``` - name: launch elasticache instance elasticache: cache_engine_version: 2.8.24 cache_port: 6379 cache_security_groups: [] cache_subnet_group: --- engine: redis hard_modify: false name: abc-redis node_type: cache.t2.micro num_nodes: 1 region: ---- security_group_ids: ---- state: present wait: true - name: terminate cluster elasticache: name: abc-redis region: --- state: absent ``` ##### Expected Results: Expect cluster to be deleted without errors. ##### Actual Results: As far as I can tell the cluster is gone from the AWS console, however there shouldn't be any errors thrown.
True
Deleting elasticache cluster on redis throws TypeError - ##### Issue Type: - Bug Report ##### Component Name: elasticache module ##### Ansible Version: ansible 2.0.0.2 ##### Ansible Configuration: `any_errors_fatal = True` ##### Environment: redis on aws elasticache ##### Summary: Errors on deleting cluster. ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 2727, in <module> main() File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 562, in main module.exit_json(**facts_result) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 2099, in exit_json kwargs = remove_values(kwargs, self.no_log_values) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in remove_values return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in <genexpr> return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in remove_values return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in <genexpr> return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in remove_values return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 982, in <genexpr> return dict((k, remove_values(v, no_log_strings)) for k, v in value.items()) File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 980, in remove_values return [remove_values(elem, no_log_strings) for elem in value] File "/Users/james/.ansible/tmp/ansible-tmp-1455652399.45-51638225824520/elasticache", line 991, in remove_values raise TypeError('Value of unknown type: %s, %s' % (type(value), value)) TypeError: Value of unknown type: <type 'object'>, <object object at 0x10bf01100> fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "elasticache"}, "parsed": false} ``` ##### Steps To Reproduce: Create cluster with elasticache module. Attempt to delete cluster with `state: absent`. If it helps, the cluster name has a hyphen. e.g. `abc-redis` ``` - name: launch elasticache instance elasticache: cache_engine_version: 2.8.24 cache_port: 6379 cache_security_groups: [] cache_subnet_group: --- engine: redis hard_modify: false name: abc-redis node_type: cache.t2.micro num_nodes: 1 region: ---- security_group_ids: ---- state: present wait: true - name: terminate cluster elasticache: name: abc-redis region: --- state: absent ``` ##### Expected Results: Expect cluster to be deleted without errors. ##### Actual Results: As far as I can tell the cluster is gone from the AWS console, however there shouldn't be any errors thrown.
main
deleting elasticache cluster on redis throws typeerror issue type bug report component name elasticache module ansible version ansible ansible configuration any errors fatal true environment redis on aws elasticache summary errors on deleting cluster an exception occurred during task execution the full traceback is traceback most recent call last file users james ansible tmp ansible tmp elasticache line in main file users james ansible tmp ansible tmp elasticache line in main module exit json facts result file users james ansible tmp ansible tmp elasticache line in exit json kwargs remove values kwargs self no log values file users james ansible tmp ansible tmp elasticache line in remove values return dict k remove values v no log strings for k v in value items file users james ansible tmp ansible tmp elasticache line in return dict k remove values v no log strings for k v in value items file users james ansible tmp ansible tmp elasticache line in remove values return dict k remove values v no log strings for k v in value items file users james ansible tmp ansible tmp elasticache line in return dict k remove values v no log strings for k v in value items file users james ansible tmp ansible tmp elasticache line in remove values return dict k remove values v no log strings for k v in value items file users james ansible tmp ansible tmp elasticache line in return dict k remove values v no log strings for k v in value items file users james ansible tmp ansible tmp elasticache line in remove values return file users james ansible tmp ansible tmp elasticache line in remove values raise typeerror value of unknown type s s type value value typeerror value of unknown type fatal failed changed false failed true invocation module name elasticache parsed false steps to reproduce create cluster with elasticache module attempt to delete cluster with state absent if it helps the cluster name has a hyphen e g abc redis name launch elasticache instance elasticache cache engine version cache port cache security groups cache subnet group engine redis hard modify false name abc redis node type cache micro num nodes region security group ids state present wait true name terminate cluster elasticache name abc redis region state absent expected results expect cluster to be deleted without errors actual results as far as i can tell the cluster is gone from the aws console however there shouldn t be any errors thrown
1
3,362
13,033,007,647
IssuesEvent
2020-07-28 05:53:15
arcticicestudio/nord
https://api.github.com/repos/arcticicestudio/nord
closed
Repo label system
context-workflow resolution-answered scope-maintainability type-question
This has nothing to do with the wonderful Nord theme (that I've been using everywhere I can), but more about the label system you have in place for the repo. I find it more useful than the usual (default) label system that has Github and I want to implement this system (and colors) for my own repos. Is there any kind of documentation about your workflow that I can read? Also, is there any way to easily configure the labels in new repos? Doing this by hand is long and tedious!
True
Repo label system - This has nothing to do with the wonderful Nord theme (that I've been using everywhere I can), but more about the label system you have in place for the repo. I find it more useful than the usual (default) label system that has Github and I want to implement this system (and colors) for my own repos. Is there any kind of documentation about your workflow that I can read? Also, is there any way to easily configure the labels in new repos? Doing this by hand is long and tedious!
main
repo label system this has nothing to do with the wonderful nord theme that i ve been using everywhere i can but more about the label system you have in place for the repo i find it more useful than the usual default label system that has github and i want to implement this system and colors for my own repos is there any kind of documentation about your workflow that i can read also is there any way to easily configure the labels in new repos doing this by hand is long and tedious
1
2,727
9,634,177,613
IssuesEvent
2019-05-15 20:33:28
Homebrew/homebrew-core
https://api.github.com/repos/Homebrew/homebrew-core
closed
Tracking issue for migrating to OpenBLAS
help wanted maintainer feedback
## Formulae using Accelerate The following formula still appear to depend on Apple's Accelerate framework: - [x] arpack (uses `veclibfort`) - [x] dynare (uses `veclibfort`) - [x] hypre (uses `veclibfort`) - [x] octave (uses `veclibfort`) - [x] qrupdate (uses `veclibfort`) - [x] sundials (uses `veclibfort`) - [x] suite-sparse (explicit `-framework Accelerate`) - [x] superlu (uses `veclibfort`) This list was compiled via `brew uses veclibfort` and also `ag -w 'accelerate'` There are additional recursive dependencies, and `shogun` may need modifying because it pulls in the `lapack` formula under certain circumstances, but `OpenBLAS` provides `lapack` functionality. See also brewsci/homebrew-num which defaults to OpenBLAS for all formulae requiring BLAS: https://github.com/brewsci/homebrew-num/tree/master/Formula ## Additional recursive dependencies: ### `veclibfort` - armadillo - libbi - shogun ### suite-sparse - ceres-solver - petsc - petsc-complex - rawtoaces
True
Tracking issue for migrating to OpenBLAS - ## Formulae using Accelerate The following formula still appear to depend on Apple's Accelerate framework: - [x] arpack (uses `veclibfort`) - [x] dynare (uses `veclibfort`) - [x] hypre (uses `veclibfort`) - [x] octave (uses `veclibfort`) - [x] qrupdate (uses `veclibfort`) - [x] sundials (uses `veclibfort`) - [x] suite-sparse (explicit `-framework Accelerate`) - [x] superlu (uses `veclibfort`) This list was compiled via `brew uses veclibfort` and also `ag -w 'accelerate'` There are additional recursive dependencies, and `shogun` may need modifying because it pulls in the `lapack` formula under certain circumstances, but `OpenBLAS` provides `lapack` functionality. See also brewsci/homebrew-num which defaults to OpenBLAS for all formulae requiring BLAS: https://github.com/brewsci/homebrew-num/tree/master/Formula ## Additional recursive dependencies: ### `veclibfort` - armadillo - libbi - shogun ### suite-sparse - ceres-solver - petsc - petsc-complex - rawtoaces
main
tracking issue for migrating to openblas formulae using accelerate the following formula still appear to depend on apple s accelerate framework arpack uses veclibfort dynare uses veclibfort hypre uses veclibfort octave uses veclibfort qrupdate uses veclibfort sundials uses veclibfort suite sparse explicit framework accelerate superlu uses veclibfort this list was compiled via brew uses veclibfort and also ag w accelerate there are additional recursive dependencies and shogun may need modifying because it pulls in the lapack formula under certain circumstances but openblas provides lapack functionality see also brewsci homebrew num which defaults to openblas for all formulae requiring blas additional recursive dependencies veclibfort armadillo libbi shogun suite sparse ceres solver petsc petsc complex rawtoaces
1
4,540
23,638,389,951
IssuesEvent
2022-08-25 14:58:11
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Error:
type: bug work: backend status: ready restricted: maintainers
I get this error any time I try to load the records for the Checkouts table, when that table is created using our original method (in Cycle 2) of import tsv files and running the my SQL transform. I can do more work to try providing better reproduction steps if needed. <details> <summary><h3>Traceback</h3></summary> ``` Environment: Request Method: GET Request URL: http://localhost:8000/api/db/v0/tables/16/records/ Django Version: 3.1.14 Python Version: 3.9.8 Installed Applications: ['django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'django_filters', 'django_property_filter', 'mathesar'] Installed Middleware: ['django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware'] Traceback (most recent call last): File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner response = get_response(request) File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view return view_func(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view return self.dispatch(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch response = self.handle_exception(exc) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception response = exception_handler(exc, context) File "/code/mathesar/exception_handlers.py", line 53, in mathesar_exception_handler raise exc File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch response = handler(request, *args, **kwargs) File "/code/mathesar/api/db/viewsets/records.py", line 65, in list records = paginator.paginate_queryset( File "/code/mathesar/api/pagination.py", line 89, in paginate_queryset records = query.get_records( File "/code/mathesar/models/query.py", line 121, in get_records return self.db_query.get_records(**kwargs) File "/code/mathesar/models/query.py", line 149, in db_query initial_columns=self._db_initial_columns, File "/code/mathesar/models/query.py", line 157, in _db_initial_columns return tuple( File "/code/mathesar/models/query.py", line 158, in <genexpr> _db_initial_column_from_json(json_col) File "/code/mathesar/models/query.py", line 208, in _db_initial_column_from_json column_pair = _get_column_pair_from_id(col_json["id"]) File "/code/mathesar/models/query.py", line 203, in _get_column_pair_from_id col = Column.objects.get(id=col_id) File "/usr/local/lib/python3.9/site-packages/django/db/models/manager.py", line 85, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/django/db/models/query.py", line 429, in get raise self.model.DoesNotExist( Exception Type: DoesNotExist at /api/db/v0/tables/16/records/ Exception Value: Column matching query does not exist. ``` </details>
True
Error: - I get this error any time I try to load the records for the Checkouts table, when that table is created using our original method (in Cycle 2) of import tsv files and running the my SQL transform. I can do more work to try providing better reproduction steps if needed. <details> <summary><h3>Traceback</h3></summary> ``` Environment: Request Method: GET Request URL: http://localhost:8000/api/db/v0/tables/16/records/ Django Version: 3.1.14 Python Version: 3.9.8 Installed Applications: ['django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'django_filters', 'django_property_filter', 'mathesar'] Installed Middleware: ['django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware'] Traceback (most recent call last): File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner response = get_response(request) File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view return view_func(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view return self.dispatch(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch response = self.handle_exception(exc) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception response = exception_handler(exc, context) File "/code/mathesar/exception_handlers.py", line 53, in mathesar_exception_handler raise exc File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch response = handler(request, *args, **kwargs) File "/code/mathesar/api/db/viewsets/records.py", line 65, in list records = paginator.paginate_queryset( File "/code/mathesar/api/pagination.py", line 89, in paginate_queryset records = query.get_records( File "/code/mathesar/models/query.py", line 121, in get_records return self.db_query.get_records(**kwargs) File "/code/mathesar/models/query.py", line 149, in db_query initial_columns=self._db_initial_columns, File "/code/mathesar/models/query.py", line 157, in _db_initial_columns return tuple( File "/code/mathesar/models/query.py", line 158, in <genexpr> _db_initial_column_from_json(json_col) File "/code/mathesar/models/query.py", line 208, in _db_initial_column_from_json column_pair = _get_column_pair_from_id(col_json["id"]) File "/code/mathesar/models/query.py", line 203, in _get_column_pair_from_id col = Column.objects.get(id=col_id) File "/usr/local/lib/python3.9/site-packages/django/db/models/manager.py", line 85, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/django/db/models/query.py", line 429, in get raise self.model.DoesNotExist( Exception Type: DoesNotExist at /api/db/v0/tables/16/records/ Exception Value: Column matching query does not exist. ``` </details>
main
error i get this error any time i try to load the records for the checkouts table when that table is created using our original method in cycle of import tsv files and running the my sql transform i can do more work to try providing better reproduction steps if needed traceback environment request method get request url django version python version installed applications django contrib admin django contrib auth django contrib contenttypes django contrib sessions django contrib messages django contrib staticfiles rest framework django filters django property filter mathesar installed middleware django middleware security securitymiddleware django contrib sessions middleware sessionmiddleware django middleware common commonmiddleware django middleware csrf csrfviewmiddleware django contrib auth middleware authenticationmiddleware django contrib messages middleware messagemiddleware django middleware clickjacking xframeoptionsmiddleware traceback most recent call last file usr local lib site packages django core handlers exception py line in inner response get response request file usr local lib site packages django core handlers base py line in get response response wrapped callback request callback args callback kwargs file usr local lib site packages django views decorators csrf py line in wrapped view return view func args kwargs file usr local lib site packages rest framework viewsets py line in view return self dispatch request args kwargs file usr local lib site packages rest framework views py line in dispatch response self handle exception exc file usr local lib site packages rest framework views py line in handle exception response exception handler exc context file code mathesar exception handlers py line in mathesar exception handler raise exc file usr local lib site packages rest framework views py line in dispatch response handler request args kwargs file code mathesar api db viewsets records py line in list records paginator paginate queryset file code mathesar api pagination py line in paginate queryset records query get records file code mathesar models query py line in get records return self db query get records kwargs file code mathesar models query py line in db query initial columns self db initial columns file code mathesar models query py line in db initial columns return tuple file code mathesar models query py line in db initial column from json json col file code mathesar models query py line in db initial column from json column pair get column pair from id col json file code mathesar models query py line in get column pair from id col column objects get id col id file usr local lib site packages django db models manager py line in manager method return getattr self get queryset name args kwargs file usr local lib site packages django db models query py line in get raise self model doesnotexist exception type doesnotexist at api db tables records exception value column matching query does not exist
1
3,477
13,394,055,274
IssuesEvent
2020-09-03 05:55:45
Homebrew/homebrew-cask
https://api.github.com/repos/Homebrew/homebrew-cask
closed
Error using --force to install/update some already installed casks
awaiting maintainer feedback
#### General troubleshooting steps - [x] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/Homebrew/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). - [x] I have retried my command with `--force`. - [x] I ran `brew update-reset && brew update` and retried my command. - [x] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [x] I have checked the instructions for [reporting bugs](https://github.com/Homebrew/homebrew-cask#reporting-bugs). - [x] I made doubly sure this is not a [checksum does not match](https://github.com/Homebrew/homebrew-cask/blob/master/doc/reporting_bugs/checksum_does_not_match_error.md) error. #### Description of issue Some casks do not get installed correctly when using `--force` if app is already installed, e.g. nextcloud, github... <!-- Please DO NOT delete the backticks. Only change the “{{replace this}}” text. --> #### Command that failed nextcloud is just an example, fails on multiple casks if cask is already installed. ``` brew cask install --force nextcloud && brew cask install --force nextcloud ``` #### Output of command with `--force --verbose --debug` ``` brew cask install --force nextcloud --verbose --debug ==> Cask::Installer#install ==> Printing caveats ==> Cask::Installer#fetch ==> Downloading ==> Downloading https://download.nextcloud.com/desktop/releases/Mac/Installer/Nextcloud-3.0.1.pkg /usr/bin/curl --disable --globoff --show-error --user-agent Homebrew/2.4.14-27-ga39aae8\ \(Macintosh\;\ Intel\ Mac\ OS\ X\ 10.15.6\)\ curl/7.64.1 --retry 3 --location --silent --head --request GET https://download.nextcloud.com/desktop/releases/Mac/Installer/Nextcloud-3.0.1.pkg Already downloaded: /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg ==> Checking quarantine support /usr/bin/xattr /usr/bin/swift /usr/local/Homebrew/Library/Homebrew/cask/utils/quarantine.swift ==> Quarantine is available. ==> Verifying Gatekeeper status of /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg /usr/bin/xattr -p com.apple.quarantine /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg ==> /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg is quarantined ==> Downloaded to -> /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg ==> Verifying SHA-256 checksum for Cask 'nextcloud'. ==> Installing Cask nextcloud ==> Cask::Installer#stage ==> Extracting primary container ==> Using container class UnpackStrategy::Pkg for /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg cp -p /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg /usr/local/Caskroom/nextcloud/3.0.1/Nextcloud-3.0.1.pkg ==> Verifying Gatekeeper status of /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg /usr/bin/xattr -p com.apple.quarantine /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg ==> /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg is quarantined ==> Propagating quarantine from /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg to /usr/local/Caskroom/nextcloud/3.0.1 /usr/bin/xattr -p com.apple.quarantine /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg /usr/bin/xargs -0 -- /bin/chmod -h u\+w /usr/bin/xargs -0 -- /usr/bin/xattr -w com.apple.quarantine 0181\;5f4abfa7\;Homebrew\\x20Cask\;7D556E3E-732F-42A0-9226-7C600A6A1D18 ==> Creating metadata directory /usr/local/Caskroom/nextcloud/.metadata/3.0.1/20200829205456.344. ==> Creating metadata subdirectory /usr/local/Caskroom/nextcloud/.metadata/3.0.1/20200829205456.344/Casks. ==> Installing artifacts ==> 3 artifact/s defined #<SortedSet: {#<Cask::Artifact::Uninstall:0x00007f90c02248b0 @cask=#<Cask::Cask:0x00007f90681a4aa0 @token="nextcloud", @sourcefile_path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb>, @tap=#<Tap:0x00007f90c01cea50 @user="Homebrew", @repo="cask", @name="homebrew/cask", @full_name="Homebrew/homebrew-cask", @path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask>, @alias_table=nil, @alias_reverse_table=nil, @cask_dir=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks>>, @block=#<Proc:0x00007f90681a4a00@/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb:1>, @caskroom_path=#<Pathname:/usr/local/Caskroom/nextcloud>, @metadata_master_container_path=#<Pathname:/usr/local/Caskroom/nextcloud/.metadata>, @config=#<Cask::Config:0x00007f90c0206ab8 @explicit={}, @binarydir=#<Pathname:/usr/local/bin>, @default={:languages=>["de-DE"], :appdir=>#<Pathname:/Applications>, :colorpickerdir=>#<Pathname:/Users/tom/Library/ColorPickers>, :prefpanedir=>#<Pathname:/Users/tom/Library/PreferencePanes>, :qlplugindir=>#<Pathname:/Users/tom/Library/QuickLook>, :mdimporterdir=>#<Pathname:/Users/tom/Library/Spotlight>, :dictionarydir=>#<Pathname:/Users/tom/Library/Dictionaries>, :fontdir=>#<Pathname:/Users/tom/Library/Fonts>, :servicedir=>#<Pathname:/Users/tom/Library/Services>, :input_methoddir=>#<Pathname:/Users/tom/Library/Input Methods>, :internet_plugindir=>#<Pathname:/Users/tom/Library/Internet Plug-Ins>, :audio_unit_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/Components>, :vst_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST>, :vst3_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST3>, :screen_saverdir=>#<Pathname:/Users/tom/Library/Screen Savers>}, @env={}>, @dsl=#<Cask::DSL:0x00007f90c0205e60 @cask=#<Cask::Cask:0x00007f90681a4aa0 ...>, @token="nextcloud", @version="3.0.1", @sha256="aea710cf0fae1f42fdffe6ca94b46929ef4a288ac51809de513a04a6f2a074b4", @url=#<URL:0x00007f90c0204d80 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/Nextcloud-3.0.1.pkg>, @user_agent=:default, @specs={}>, @appcast=#<Cask::DSL::Appcast:0x00007f90c020e9c0 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/>, @parameters={}>, @name=["Nextcloud"], @homepage="https://nextcloud.com/", @depends_on={:macos=>#<MacOSRequirement: [] version>=10.10>}, @artifacts=#<SortedSet: {...}>, @staged_path=#<Pathname:/usr/local/Caskroom/nextcloud/3.0.1>, @language=nil>>, @directives={:pkgutil=>"com.nextcloud.desktopclient", :signal=>[]}>, #<Cask::Artifact::Pkg:0x00007f90c0214de8 @cask=#<Cask::Cask:0x00007f90681a4aa0 @token="nextcloud", @sourcefile_path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb>, @tap=#<Tap:0x00007f90c01cea50 @user="Homebrew", @repo="cask", @name="homebrew/cask", @full_name="Homebrew/homebrew-cask", @path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask>, @alias_table=nil, @alias_reverse_table=nil, @cask_dir=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks>>, @block=#<Proc:0x00007f90681a4a00@/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb:1>, @caskroom_path=#<Pathname:/usr/local/Caskroom/nextcloud>, @metadata_master_container_path=#<Pathname:/usr/local/Caskroom/nextcloud/.metadata>, @config=#<Cask::Config:0x00007f90c0206ab8 @explicit={}, @binarydir=#<Pathname:/usr/local/bin>, @default={:languages=>["de-DE"], :appdir=>#<Pathname:/Applications>, :colorpickerdir=>#<Pathname:/Users/tom/Library/ColorPickers>, :prefpanedir=>#<Pathname:/Users/tom/Library/PreferencePanes>, :qlplugindir=>#<Pathname:/Users/tom/Library/QuickLook>, :mdimporterdir=>#<Pathname:/Users/tom/Library/Spotlight>, :dictionarydir=>#<Pathname:/Users/tom/Library/Dictionaries>, :fontdir=>#<Pathname:/Users/tom/Library/Fonts>, :servicedir=>#<Pathname:/Users/tom/Library/Services>, :input_methoddir=>#<Pathname:/Users/tom/Library/Input Methods>, :internet_plugindir=>#<Pathname:/Users/tom/Library/Internet Plug-Ins>, :audio_unit_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/Components>, :vst_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST>, :vst3_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST3>, :screen_saverdir=>#<Pathname:/Users/tom/Library/Screen Savers>}, @env={}>, @dsl=#<Cask::DSL:0x00007f90c0205e60 @cask=#<Cask::Cask:0x00007f90681a4aa0 ...>, @token="nextcloud", @version="3.0.1", @sha256="aea710cf0fae1f42fdffe6ca94b46929ef4a288ac51809de513a04a6f2a074b4", @url=#<URL:0x00007f90c0204d80 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/Nextcloud-3.0.1.pkg>, @user_agent=:default, @specs={}>, @appcast=#<Cask::DSL::Appcast:0x00007f90c020e9c0 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/>, @parameters={}>, @name=["Nextcloud"], @homepage="https://nextcloud.com/", @depends_on={:macos=>#<MacOSRequirement: [] version>=10.10>}, @artifacts=#<SortedSet: {...}>, @staged_path=#<Pathname:/usr/local/Caskroom/nextcloud/3.0.1>, @language=nil>>, @path=#<Pathname:/usr/local/Caskroom/nextcloud/3.0.1/Nextcloud-3.0.1.pkg>, @stanza_options={}>, #<Cask::Artifact::Binary:0x00007f90c021d448 @cask=#<Cask::Cask:0x00007f90681a4aa0 @token="nextcloud", @sourcefile_path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb>, @tap=#<Tap:0x00007f90c01cea50 @user="Homebrew", @repo="cask", @name="homebrew/cask", @full_name="Homebrew/homebrew-cask", @path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask>, @alias_table=nil, @alias_reverse_table=nil, @cask_dir=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks>>, @block=#<Proc:0x00007f90681a4a00@/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb:1>, @caskroom_path=#<Pathname:/usr/local/Caskroom/nextcloud>, @metadata_master_container_path=#<Pathname:/usr/local/Caskroom/nextcloud/.metadata>, @config=#<Cask::Config:0x00007f90c0206ab8 @explicit={}, @binarydir=#<Pathname:/usr/local/bin>, @default={:languages=>["de-DE"], :appdir=>#<Pathname:/Applications>, :colorpickerdir=>#<Pathname:/Users/tom/Library/ColorPickers>, :prefpanedir=>#<Pathname:/Users/tom/Library/PreferencePanes>, :qlplugindir=>#<Pathname:/Users/tom/Library/QuickLook>, :mdimporterdir=>#<Pathname:/Users/tom/Library/Spotlight>, :dictionarydir=>#<Pathname:/Users/tom/Library/Dictionaries>, :fontdir=>#<Pathname:/Users/tom/Library/Fonts>, :servicedir=>#<Pathname:/Users/tom/Library/Services>, :input_methoddir=>#<Pathname:/Users/tom/Library/Input Methods>, :internet_plugindir=>#<Pathname:/Users/tom/Library/Internet Plug-Ins>, :audio_unit_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/Components>, :vst_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST>, :vst3_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST3>, :screen_saverdir=>#<Pathname:/Users/tom/Library/Screen Savers>}, @env={}>, @dsl=#<Cask::DSL:0x00007f90c0205e60 @cask=#<Cask::Cask:0x00007f90681a4aa0 ...>, @token="nextcloud", @version="3.0.1", @sha256="aea710cf0fae1f42fdffe6ca94b46929ef4a288ac51809de513a04a6f2a074b4", @url=#<URL:0x00007f90c0204d80 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/Nextcloud-3.0.1.pkg>, @user_agent=:default, @specs={}>, @appcast=#<Cask::DSL::Appcast:0x00007f90c020e9c0 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/>, @parameters={}>, @name=["Nextcloud"], @homepage="https://nextcloud.com/", @depends_on={:macos=>#<MacOSRequirement: [] version>=10.10>}, @artifacts=#<SortedSet: {...}>, @staged_path=#<Pathname:/usr/local/Caskroom/nextcloud/3.0.1>, @language=nil>>, @source_string="/Applications/nextcloud.app/Contents/MacOS/nextcloudcmd", @target_string="", @source=#<Pathname:/Applications/nextcloud.app/Contents/MacOS/nextcloudcmd>, @target=#<Pathname:/usr/local/bin/nextcloudcmd>>}> ==> Installing artifact of class Cask::Artifact::Pkg ==> Running installer for nextcloud; your password may be necessary. ==> Package installers may write to any location; options such as --appdir are ignored. /usr/bin/sudo -E -- /usr/bin/env LOGNAME=tom USER=tom USERNAME=tom /usr/sbin/installer -pkg /usr/local/Caskroom/nextcloud/3.0.1/Nextcloud-3.0.1.pkg -target / -verboseR Password: installer: Package name is Nextcloud Client installer: Upgrading at base path / installer:PHASE:Installation vorbereiten … installer:PHASE:Volume vorbereiten … installer:PHASE:„Nextcloud Client“ vorbereiten … installer:PHASE:Warten, bis andere Installationen abgeschlossen werden … installer:PHASE:Installation konfigurieren … installer:STATUS: installer:%17.584231 installer:PHASE:Dateien schreiben … installer:%43.442713 installer:PHASE:Dateien schreiben … installer:%69.648134 installer:PHASE:Objekte an Zielort bewegen … installer:%89.891156 installer:PHASE:Paketskripte ausführen … installer:%90.887821 installer:PHASE:Paketskripte ausführen … installer:%91.181448 installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Pakete überprüfen … installer:%97.750000 installer:STATUS:Installationsaktionen ausführen … installer:STATUS: installer:PHASE:Installation abschließen … installer:STATUS: installer:%100.000000 installer:PHASE:Die Software wurde erfolgreich installiert. installer: The upgrade was successful. ==> Installing artifact of class Cask::Artifact::Binary ==> Purging files for version 3.0.1 of Cask nextcloud Error: It seems there already exists a Binary at '/usr/local/bin/nextcloudcmd'; not overwriting. /usr/local/Homebrew/Library/Homebrew/cask/artifact/symlinked.rb:51:in `link' /usr/local/Homebrew/Library/Homebrew/cask/artifact/binary.rb:12:in `link' /usr/local/Homebrew/Library/Homebrew/cask/artifact/symlinked.rb:20:in `install_phase' /usr/local/Homebrew/Library/Homebrew/cask/installer.rb:221:in `block in install_artifacts' /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/2.6.0/set.rb:777:in `each' /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/2.6.0/set.rb:777:in `each' /usr/local/Homebrew/Library/Homebrew/cask/installer.rb:212:in `install_artifacts' /usr/local/Homebrew/Library/Homebrew/cask/installer.rb:104:in `install' /usr/local/Homebrew/Library/Homebrew/cask/cmd/install.rb:44:in `block in run' /usr/local/Homebrew/Library/Homebrew/cask/cmd/install.rb:43:in `each' /usr/local/Homebrew/Library/Homebrew/cask/cmd/install.rb:43:in `run' /usr/local/Homebrew/Library/Homebrew/cask/cmd/abstract_command.rb:89:in `run' /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:238:in `run' /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:165:in `run' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:14:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:113:in `<main>' Error: Kernel.exit /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:243:in `exit' /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:243:in `rescue in run' /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:213:in `run' /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:165:in `run' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:14:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:113:in `<main>' ``` #### Output of `brew cask doctor` ``` brew cask doctor ==> Homebrew Version 2.4.14-27-ga39aae8 ==> macOS 10.15.6 ==> SIP Unknown (custom configuration) configuration: apple internal: disabled kext signing: enabled filesystem protections: disabled debugging restrictions: disabled dtrace restrictions: enabled nvram protections: enabled basesystem verification: enabled ==> Java 14.0.1 ==> Homebrew Cask Staging Location /usr/local/Caskroom ==> Homebrew Cask Taps: /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask (3642 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask-versions (160 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask-fonts (1558 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask-drivers (157 casks) ==> $LOAD_PATHS /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/thor-1.0.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/ruby-macho-2.2.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rubocop-rspec-1.43.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rubocop-performance-1.7.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rubocop-0.89.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/unicode-display_width-1.7.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/ruby-progressbar-1.10.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rubocop-ast-0.3.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-wait-0.0.9/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-retry-0.6.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-its-1.3.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-3.9.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-mocks-3.9.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-expectations-3.9.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-core-3.9.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-support-3.9.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/ronn-0.7.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/regexp_parser-1.7.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rdiscount-2.2.0.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/rdiscount-2.2.0.2 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/pry-0.13.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/plist-3.5.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/patchelf-1.3.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/parlour-4.0.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rainbow-3.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/parser-2.7.1.4/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/parallel_tests-3.1.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/parallel-1.19.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/mustache-1.1.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/method_source-1.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/mechanize-2.7.6/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/webrobots-0.1.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/ntlm-http-0.1.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/nokogiri-1.10.10/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/nokogiri-1.10.10 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/mini_portile2-2.4.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/net-http-persistent-4.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/net-http-digest_auth-1.4.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/mime-types-3.3.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/mime-types-data-3.2020.0512/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/http-cookie-1.0.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/hpricot-0.8.6/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/hpricot-0.8.6 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/elftools-1.1.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/domain_name-0.5.20190701/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/unf-0.1.4/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/unf_ext-0.0.7.7/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/unf_ext-0.0.7.7 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/diff-lcs-1.4.4/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/connection_pool-2.2.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/commander-4.5.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/highline-2.0.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/coderay-1.1.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/codecov-0.2.8/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/simplecov-0.19.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/simplecov-html-0.12.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/docile-1.3.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/json-2.3.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/json-2.3.1 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/byebug-11.1.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/byebug-11.1.3 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/ /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/bindata-2.4.8/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/ast-2.4.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/activesupport-6.0.3.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/zeitwerk-2.4.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/tzinfo-1.2.7/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/thread_safe-0.3.6/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/minitest-5.14.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/i18n-1.8.5/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/concurrent-ruby-1.1.7/lib/concurrent-ruby /Library/Ruby/Site/2.6.0 /Library/Ruby/Site/2.6.0/universal-darwin19 /Library/Ruby/Site /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/vendor_ruby/2.6.0 /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/vendor_ruby/2.6.0/universal-darwin19 /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/vendor_ruby /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/2.6.0 /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/2.6.0/universal-darwin19 /usr/local/Homebrew/Library/Homebrew ==> Cask Environment Variables: BUNDLE_PATH CHRUBY_VERSION GEM_HOME GEM_PATH HOMEBREW_CASK_OPTS LC_ALL PATH RBENV_VERSION RUBYLIB RUBYOPT RUBYPATH SHELL ``` #### Output of `brew tap` ``` homebrew/cask homebrew/cask-drivers homebrew/cask-fonts homebrew/cask-versions homebrew/core vitorgalvao/tiny-scripts ```
True
Error using --force to install/update some already installed casks - #### General troubleshooting steps - [x] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/Homebrew/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). - [x] I have retried my command with `--force`. - [x] I ran `brew update-reset && brew update` and retried my command. - [x] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [x] I have checked the instructions for [reporting bugs](https://github.com/Homebrew/homebrew-cask#reporting-bugs). - [x] I made doubly sure this is not a [checksum does not match](https://github.com/Homebrew/homebrew-cask/blob/master/doc/reporting_bugs/checksum_does_not_match_error.md) error. #### Description of issue Some casks do not get installed correctly when using `--force` if app is already installed, e.g. nextcloud, github... <!-- Please DO NOT delete the backticks. Only change the “{{replace this}}” text. --> #### Command that failed nextcloud is just an example, fails on multiple casks if cask is already installed. ``` brew cask install --force nextcloud && brew cask install --force nextcloud ``` #### Output of command with `--force --verbose --debug` ``` brew cask install --force nextcloud --verbose --debug ==> Cask::Installer#install ==> Printing caveats ==> Cask::Installer#fetch ==> Downloading ==> Downloading https://download.nextcloud.com/desktop/releases/Mac/Installer/Nextcloud-3.0.1.pkg /usr/bin/curl --disable --globoff --show-error --user-agent Homebrew/2.4.14-27-ga39aae8\ \(Macintosh\;\ Intel\ Mac\ OS\ X\ 10.15.6\)\ curl/7.64.1 --retry 3 --location --silent --head --request GET https://download.nextcloud.com/desktop/releases/Mac/Installer/Nextcloud-3.0.1.pkg Already downloaded: /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg ==> Checking quarantine support /usr/bin/xattr /usr/bin/swift /usr/local/Homebrew/Library/Homebrew/cask/utils/quarantine.swift ==> Quarantine is available. ==> Verifying Gatekeeper status of /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg /usr/bin/xattr -p com.apple.quarantine /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg ==> /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg is quarantined ==> Downloaded to -> /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg ==> Verifying SHA-256 checksum for Cask 'nextcloud'. ==> Installing Cask nextcloud ==> Cask::Installer#stage ==> Extracting primary container ==> Using container class UnpackStrategy::Pkg for /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg cp -p /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg /usr/local/Caskroom/nextcloud/3.0.1/Nextcloud-3.0.1.pkg ==> Verifying Gatekeeper status of /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg /usr/bin/xattr -p com.apple.quarantine /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg ==> /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg is quarantined ==> Propagating quarantine from /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg to /usr/local/Caskroom/nextcloud/3.0.1 /usr/bin/xattr -p com.apple.quarantine /Users/tom/Library/Caches/Homebrew/downloads/d5ec41b33f3d569bfdba5da811cafddfdd6286cddde5f209848e7c2cd50aaea3--Nextcloud-3.0.1.pkg /usr/bin/xargs -0 -- /bin/chmod -h u\+w /usr/bin/xargs -0 -- /usr/bin/xattr -w com.apple.quarantine 0181\;5f4abfa7\;Homebrew\\x20Cask\;7D556E3E-732F-42A0-9226-7C600A6A1D18 ==> Creating metadata directory /usr/local/Caskroom/nextcloud/.metadata/3.0.1/20200829205456.344. ==> Creating metadata subdirectory /usr/local/Caskroom/nextcloud/.metadata/3.0.1/20200829205456.344/Casks. ==> Installing artifacts ==> 3 artifact/s defined #<SortedSet: {#<Cask::Artifact::Uninstall:0x00007f90c02248b0 @cask=#<Cask::Cask:0x00007f90681a4aa0 @token="nextcloud", @sourcefile_path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb>, @tap=#<Tap:0x00007f90c01cea50 @user="Homebrew", @repo="cask", @name="homebrew/cask", @full_name="Homebrew/homebrew-cask", @path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask>, @alias_table=nil, @alias_reverse_table=nil, @cask_dir=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks>>, @block=#<Proc:0x00007f90681a4a00@/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb:1>, @caskroom_path=#<Pathname:/usr/local/Caskroom/nextcloud>, @metadata_master_container_path=#<Pathname:/usr/local/Caskroom/nextcloud/.metadata>, @config=#<Cask::Config:0x00007f90c0206ab8 @explicit={}, @binarydir=#<Pathname:/usr/local/bin>, @default={:languages=>["de-DE"], :appdir=>#<Pathname:/Applications>, :colorpickerdir=>#<Pathname:/Users/tom/Library/ColorPickers>, :prefpanedir=>#<Pathname:/Users/tom/Library/PreferencePanes>, :qlplugindir=>#<Pathname:/Users/tom/Library/QuickLook>, :mdimporterdir=>#<Pathname:/Users/tom/Library/Spotlight>, :dictionarydir=>#<Pathname:/Users/tom/Library/Dictionaries>, :fontdir=>#<Pathname:/Users/tom/Library/Fonts>, :servicedir=>#<Pathname:/Users/tom/Library/Services>, :input_methoddir=>#<Pathname:/Users/tom/Library/Input Methods>, :internet_plugindir=>#<Pathname:/Users/tom/Library/Internet Plug-Ins>, :audio_unit_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/Components>, :vst_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST>, :vst3_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST3>, :screen_saverdir=>#<Pathname:/Users/tom/Library/Screen Savers>}, @env={}>, @dsl=#<Cask::DSL:0x00007f90c0205e60 @cask=#<Cask::Cask:0x00007f90681a4aa0 ...>, @token="nextcloud", @version="3.0.1", @sha256="aea710cf0fae1f42fdffe6ca94b46929ef4a288ac51809de513a04a6f2a074b4", @url=#<URL:0x00007f90c0204d80 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/Nextcloud-3.0.1.pkg>, @user_agent=:default, @specs={}>, @appcast=#<Cask::DSL::Appcast:0x00007f90c020e9c0 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/>, @parameters={}>, @name=["Nextcloud"], @homepage="https://nextcloud.com/", @depends_on={:macos=>#<MacOSRequirement: [] version>=10.10>}, @artifacts=#<SortedSet: {...}>, @staged_path=#<Pathname:/usr/local/Caskroom/nextcloud/3.0.1>, @language=nil>>, @directives={:pkgutil=>"com.nextcloud.desktopclient", :signal=>[]}>, #<Cask::Artifact::Pkg:0x00007f90c0214de8 @cask=#<Cask::Cask:0x00007f90681a4aa0 @token="nextcloud", @sourcefile_path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb>, @tap=#<Tap:0x00007f90c01cea50 @user="Homebrew", @repo="cask", @name="homebrew/cask", @full_name="Homebrew/homebrew-cask", @path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask>, @alias_table=nil, @alias_reverse_table=nil, @cask_dir=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks>>, @block=#<Proc:0x00007f90681a4a00@/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb:1>, @caskroom_path=#<Pathname:/usr/local/Caskroom/nextcloud>, @metadata_master_container_path=#<Pathname:/usr/local/Caskroom/nextcloud/.metadata>, @config=#<Cask::Config:0x00007f90c0206ab8 @explicit={}, @binarydir=#<Pathname:/usr/local/bin>, @default={:languages=>["de-DE"], :appdir=>#<Pathname:/Applications>, :colorpickerdir=>#<Pathname:/Users/tom/Library/ColorPickers>, :prefpanedir=>#<Pathname:/Users/tom/Library/PreferencePanes>, :qlplugindir=>#<Pathname:/Users/tom/Library/QuickLook>, :mdimporterdir=>#<Pathname:/Users/tom/Library/Spotlight>, :dictionarydir=>#<Pathname:/Users/tom/Library/Dictionaries>, :fontdir=>#<Pathname:/Users/tom/Library/Fonts>, :servicedir=>#<Pathname:/Users/tom/Library/Services>, :input_methoddir=>#<Pathname:/Users/tom/Library/Input Methods>, :internet_plugindir=>#<Pathname:/Users/tom/Library/Internet Plug-Ins>, :audio_unit_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/Components>, :vst_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST>, :vst3_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST3>, :screen_saverdir=>#<Pathname:/Users/tom/Library/Screen Savers>}, @env={}>, @dsl=#<Cask::DSL:0x00007f90c0205e60 @cask=#<Cask::Cask:0x00007f90681a4aa0 ...>, @token="nextcloud", @version="3.0.1", @sha256="aea710cf0fae1f42fdffe6ca94b46929ef4a288ac51809de513a04a6f2a074b4", @url=#<URL:0x00007f90c0204d80 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/Nextcloud-3.0.1.pkg>, @user_agent=:default, @specs={}>, @appcast=#<Cask::DSL::Appcast:0x00007f90c020e9c0 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/>, @parameters={}>, @name=["Nextcloud"], @homepage="https://nextcloud.com/", @depends_on={:macos=>#<MacOSRequirement: [] version>=10.10>}, @artifacts=#<SortedSet: {...}>, @staged_path=#<Pathname:/usr/local/Caskroom/nextcloud/3.0.1>, @language=nil>>, @path=#<Pathname:/usr/local/Caskroom/nextcloud/3.0.1/Nextcloud-3.0.1.pkg>, @stanza_options={}>, #<Cask::Artifact::Binary:0x00007f90c021d448 @cask=#<Cask::Cask:0x00007f90681a4aa0 @token="nextcloud", @sourcefile_path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb>, @tap=#<Tap:0x00007f90c01cea50 @user="Homebrew", @repo="cask", @name="homebrew/cask", @full_name="Homebrew/homebrew-cask", @path=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask>, @alias_table=nil, @alias_reverse_table=nil, @cask_dir=#<Pathname:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks>>, @block=#<Proc:0x00007f90681a4a00@/usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask/Casks/nextcloud.rb:1>, @caskroom_path=#<Pathname:/usr/local/Caskroom/nextcloud>, @metadata_master_container_path=#<Pathname:/usr/local/Caskroom/nextcloud/.metadata>, @config=#<Cask::Config:0x00007f90c0206ab8 @explicit={}, @binarydir=#<Pathname:/usr/local/bin>, @default={:languages=>["de-DE"], :appdir=>#<Pathname:/Applications>, :colorpickerdir=>#<Pathname:/Users/tom/Library/ColorPickers>, :prefpanedir=>#<Pathname:/Users/tom/Library/PreferencePanes>, :qlplugindir=>#<Pathname:/Users/tom/Library/QuickLook>, :mdimporterdir=>#<Pathname:/Users/tom/Library/Spotlight>, :dictionarydir=>#<Pathname:/Users/tom/Library/Dictionaries>, :fontdir=>#<Pathname:/Users/tom/Library/Fonts>, :servicedir=>#<Pathname:/Users/tom/Library/Services>, :input_methoddir=>#<Pathname:/Users/tom/Library/Input Methods>, :internet_plugindir=>#<Pathname:/Users/tom/Library/Internet Plug-Ins>, :audio_unit_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/Components>, :vst_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST>, :vst3_plugindir=>#<Pathname:/Users/tom/Library/Audio/Plug-Ins/VST3>, :screen_saverdir=>#<Pathname:/Users/tom/Library/Screen Savers>}, @env={}>, @dsl=#<Cask::DSL:0x00007f90c0205e60 @cask=#<Cask::Cask:0x00007f90681a4aa0 ...>, @token="nextcloud", @version="3.0.1", @sha256="aea710cf0fae1f42fdffe6ca94b46929ef4a288ac51809de513a04a6f2a074b4", @url=#<URL:0x00007f90c0204d80 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/Nextcloud-3.0.1.pkg>, @user_agent=:default, @specs={}>, @appcast=#<Cask::DSL::Appcast:0x00007f90c020e9c0 @uri=#<URI::HTTPS https://download.nextcloud.com/desktop/releases/Mac/Installer/>, @parameters={}>, @name=["Nextcloud"], @homepage="https://nextcloud.com/", @depends_on={:macos=>#<MacOSRequirement: [] version>=10.10>}, @artifacts=#<SortedSet: {...}>, @staged_path=#<Pathname:/usr/local/Caskroom/nextcloud/3.0.1>, @language=nil>>, @source_string="/Applications/nextcloud.app/Contents/MacOS/nextcloudcmd", @target_string="", @source=#<Pathname:/Applications/nextcloud.app/Contents/MacOS/nextcloudcmd>, @target=#<Pathname:/usr/local/bin/nextcloudcmd>>}> ==> Installing artifact of class Cask::Artifact::Pkg ==> Running installer for nextcloud; your password may be necessary. ==> Package installers may write to any location; options such as --appdir are ignored. /usr/bin/sudo -E -- /usr/bin/env LOGNAME=tom USER=tom USERNAME=tom /usr/sbin/installer -pkg /usr/local/Caskroom/nextcloud/3.0.1/Nextcloud-3.0.1.pkg -target / -verboseR Password: installer: Package name is Nextcloud Client installer: Upgrading at base path / installer:PHASE:Installation vorbereiten … installer:PHASE:Volume vorbereiten … installer:PHASE:„Nextcloud Client“ vorbereiten … installer:PHASE:Warten, bis andere Installationen abgeschlossen werden … installer:PHASE:Installation konfigurieren … installer:STATUS: installer:%17.584231 installer:PHASE:Dateien schreiben … installer:%43.442713 installer:PHASE:Dateien schreiben … installer:%69.648134 installer:PHASE:Objekte an Zielort bewegen … installer:%89.891156 installer:PHASE:Paketskripte ausführen … installer:%90.887821 installer:PHASE:Paketskripte ausführen … installer:%91.181448 installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Paketskripte ausführen … installer:PHASE:Pakete überprüfen … installer:%97.750000 installer:STATUS:Installationsaktionen ausführen … installer:STATUS: installer:PHASE:Installation abschließen … installer:STATUS: installer:%100.000000 installer:PHASE:Die Software wurde erfolgreich installiert. installer: The upgrade was successful. ==> Installing artifact of class Cask::Artifact::Binary ==> Purging files for version 3.0.1 of Cask nextcloud Error: It seems there already exists a Binary at '/usr/local/bin/nextcloudcmd'; not overwriting. /usr/local/Homebrew/Library/Homebrew/cask/artifact/symlinked.rb:51:in `link' /usr/local/Homebrew/Library/Homebrew/cask/artifact/binary.rb:12:in `link' /usr/local/Homebrew/Library/Homebrew/cask/artifact/symlinked.rb:20:in `install_phase' /usr/local/Homebrew/Library/Homebrew/cask/installer.rb:221:in `block in install_artifacts' /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/2.6.0/set.rb:777:in `each' /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/2.6.0/set.rb:777:in `each' /usr/local/Homebrew/Library/Homebrew/cask/installer.rb:212:in `install_artifacts' /usr/local/Homebrew/Library/Homebrew/cask/installer.rb:104:in `install' /usr/local/Homebrew/Library/Homebrew/cask/cmd/install.rb:44:in `block in run' /usr/local/Homebrew/Library/Homebrew/cask/cmd/install.rb:43:in `each' /usr/local/Homebrew/Library/Homebrew/cask/cmd/install.rb:43:in `run' /usr/local/Homebrew/Library/Homebrew/cask/cmd/abstract_command.rb:89:in `run' /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:238:in `run' /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:165:in `run' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:14:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:113:in `<main>' Error: Kernel.exit /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:243:in `exit' /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:243:in `rescue in run' /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:213:in `run' /usr/local/Homebrew/Library/Homebrew/cask/cmd.rb:165:in `run' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:14:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:113:in `<main>' ``` #### Output of `brew cask doctor` ``` brew cask doctor ==> Homebrew Version 2.4.14-27-ga39aae8 ==> macOS 10.15.6 ==> SIP Unknown (custom configuration) configuration: apple internal: disabled kext signing: enabled filesystem protections: disabled debugging restrictions: disabled dtrace restrictions: enabled nvram protections: enabled basesystem verification: enabled ==> Java 14.0.1 ==> Homebrew Cask Staging Location /usr/local/Caskroom ==> Homebrew Cask Taps: /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask (3642 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask-versions (160 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask-fonts (1558 casks) /usr/local/Homebrew/Library/Taps/homebrew/homebrew-cask-drivers (157 casks) ==> $LOAD_PATHS /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/thor-1.0.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/ruby-macho-2.2.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rubocop-rspec-1.43.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rubocop-performance-1.7.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rubocop-0.89.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/unicode-display_width-1.7.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/ruby-progressbar-1.10.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rubocop-ast-0.3.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-wait-0.0.9/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-retry-0.6.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-its-1.3.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-3.9.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-mocks-3.9.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-expectations-3.9.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-core-3.9.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rspec-support-3.9.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/ronn-0.7.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/regexp_parser-1.7.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rdiscount-2.2.0.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/rdiscount-2.2.0.2 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/pry-0.13.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/plist-3.5.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/patchelf-1.3.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/parlour-4.0.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/rainbow-3.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/parser-2.7.1.4/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/parallel_tests-3.1.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/parallel-1.19.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/mustache-1.1.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/method_source-1.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/mechanize-2.7.6/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/webrobots-0.1.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/ntlm-http-0.1.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/nokogiri-1.10.10/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/nokogiri-1.10.10 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/mini_portile2-2.4.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/net-http-persistent-4.0.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/net-http-digest_auth-1.4.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/mime-types-3.3.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/mime-types-data-3.2020.0512/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/http-cookie-1.0.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/hpricot-0.8.6/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/hpricot-0.8.6 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/elftools-1.1.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/domain_name-0.5.20190701/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/unf-0.1.4/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/unf_ext-0.0.7.7/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/unf_ext-0.0.7.7 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/diff-lcs-1.4.4/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/connection_pool-2.2.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/commander-4.5.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/highline-2.0.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/coderay-1.1.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/codecov-0.2.8/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/simplecov-0.19.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/simplecov-html-0.12.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/docile-1.3.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/json-2.3.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/json-2.3.1 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/byebug-11.1.3/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/extensions/universal-darwin-19/2.6.0/byebug-11.1.3 /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/ /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/bindata-2.4.8/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/ast-2.4.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/activesupport-6.0.3.2/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/zeitwerk-2.4.0/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/tzinfo-1.2.7/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/thread_safe-0.3.6/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/minitest-5.14.1/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/i18n-1.8.5/lib /usr/local/Homebrew/Library/Homebrew/vendor/bundle/bundler/../ruby/2.6.0/gems/concurrent-ruby-1.1.7/lib/concurrent-ruby /Library/Ruby/Site/2.6.0 /Library/Ruby/Site/2.6.0/universal-darwin19 /Library/Ruby/Site /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/vendor_ruby/2.6.0 /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/vendor_ruby/2.6.0/universal-darwin19 /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/vendor_ruby /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/2.6.0 /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib/ruby/2.6.0/universal-darwin19 /usr/local/Homebrew/Library/Homebrew ==> Cask Environment Variables: BUNDLE_PATH CHRUBY_VERSION GEM_HOME GEM_PATH HOMEBREW_CASK_OPTS LC_ALL PATH RBENV_VERSION RUBYLIB RUBYOPT RUBYPATH SHELL ``` #### Output of `brew tap` ``` homebrew/cask homebrew/cask-drivers homebrew/cask-fonts homebrew/cask-versions homebrew/core vitorgalvao/tiny-scripts ```
main
error using force to install update some already installed casks general troubleshooting steps i understand that i have retried my command with force i ran brew update reset brew update and retried my command i ran brew doctor fixed as many issues as possible and retried my command i have checked the instructions for i made doubly sure this is not a error description of issue some casks do not get installed correctly when using force if app is already installed e g nextcloud github command that failed nextcloud is just an example fails on multiple casks if cask is already installed brew cask install force nextcloud brew cask install force nextcloud output of command with force verbose debug brew cask install force nextcloud verbose debug cask installer install printing caveats cask installer fetch downloading downloading usr bin curl disable globoff show error user agent homebrew macintosh intel mac os x curl retry location silent head request get already downloaded users tom library caches homebrew downloads nextcloud pkg checking quarantine support usr bin xattr usr bin swift usr local homebrew library homebrew cask utils quarantine swift quarantine is available verifying gatekeeper status of users tom library caches homebrew downloads nextcloud pkg usr bin xattr p com apple quarantine users tom library caches homebrew downloads nextcloud pkg users tom library caches homebrew downloads nextcloud pkg is quarantined downloaded to users tom library caches homebrew downloads nextcloud pkg verifying sha checksum for cask nextcloud installing cask nextcloud cask installer stage extracting primary container using container class unpackstrategy pkg for users tom library caches homebrew downloads nextcloud pkg cp p users tom library caches homebrew downloads nextcloud pkg usr local caskroom nextcloud nextcloud pkg verifying gatekeeper status of users tom library caches homebrew downloads nextcloud pkg usr bin xattr p com apple quarantine users tom library caches homebrew downloads nextcloud pkg users tom library caches homebrew downloads nextcloud pkg is quarantined propagating quarantine from users tom library caches homebrew downloads nextcloud pkg to usr local caskroom nextcloud usr bin xattr p com apple quarantine users tom library caches homebrew downloads nextcloud pkg usr bin xargs bin chmod h u w usr bin xargs usr bin xattr w com apple quarantine homebrew creating metadata directory usr local caskroom nextcloud metadata creating metadata subdirectory usr local caskroom nextcloud metadata casks installing artifacts artifact s defined tap alias table nil alias reverse table nil cask dir block caskroom path metadata master container path config default languages appdir colorpickerdir prefpanedir qlplugindir mdimporterdir dictionarydir fontdir servicedir input methoddir internet plugindir audio unit plugindir vst plugindir plugindir screen saverdir env dsl token nextcloud version url appcast name homepage depends on macos artifacts staged path language nil directives pkgutil com nextcloud desktopclient signal tap alias table nil alias reverse table nil cask dir block caskroom path metadata master container path config default languages appdir colorpickerdir prefpanedir qlplugindir mdimporterdir dictionarydir fontdir servicedir input methoddir internet plugindir audio unit plugindir vst plugindir plugindir screen saverdir env dsl token nextcloud version url appcast name homepage depends on macos artifacts staged path language nil path stanza options tap alias table nil alias reverse table nil cask dir block caskroom path metadata master container path config default languages appdir colorpickerdir prefpanedir qlplugindir mdimporterdir dictionarydir fontdir servicedir input methoddir internet plugindir audio unit plugindir vst plugindir plugindir screen saverdir env dsl token nextcloud version url appcast name homepage depends on macos artifacts staged path language nil source string applications nextcloud app contents macos nextcloudcmd target string source target installing artifact of class cask artifact pkg running installer for nextcloud your password may be necessary package installers may write to any location options such as appdir are ignored usr bin sudo e usr bin env logname tom user tom username tom usr sbin installer pkg usr local caskroom nextcloud nextcloud pkg target verboser password installer package name is nextcloud client installer upgrading at base path installer phase installation vorbereiten … installer phase volume vorbereiten … installer phase „nextcloud client“ vorbereiten … installer phase warten bis andere installationen abgeschlossen werden … installer phase installation konfigurieren … installer status installer installer phase dateien schreiben … installer installer phase dateien schreiben … installer installer phase objekte an zielort bewegen … installer installer phase paketskripte ausführen … installer installer phase paketskripte ausführen … installer installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase paketskripte ausführen … installer phase pakete überprüfen … installer installer status installationsaktionen ausführen … installer status installer phase installation abschließen … installer status installer installer phase die software wurde erfolgreich installiert installer the upgrade was successful installing artifact of class cask artifact binary purging files for version of cask nextcloud error it seems there already exists a binary at usr local bin nextcloudcmd not overwriting usr local homebrew library homebrew cask artifact symlinked rb in link usr local homebrew library homebrew cask artifact binary rb in link usr local homebrew library homebrew cask artifact symlinked rb in install phase usr local homebrew library homebrew cask installer rb in block in install artifacts system library frameworks ruby framework versions usr lib ruby set rb in each system library frameworks ruby framework versions usr lib ruby set rb in each usr local homebrew library homebrew cask installer rb in install artifacts usr local homebrew library homebrew cask installer rb in install usr local homebrew library homebrew cask cmd install rb in block in run usr local homebrew library homebrew cask cmd install rb in each usr local homebrew library homebrew cask cmd install rb in run usr local homebrew library homebrew cask cmd abstract command rb in run usr local homebrew library homebrew cask cmd rb in run usr local homebrew library homebrew cask cmd rb in run usr local homebrew library homebrew cmd cask rb in cask usr local homebrew library homebrew brew rb in error kernel exit usr local homebrew library homebrew cask cmd rb in exit usr local homebrew library homebrew cask cmd rb in rescue in run usr local homebrew library homebrew cask cmd rb in run usr local homebrew library homebrew cask cmd rb in run usr local homebrew library homebrew cmd cask rb in cask usr local homebrew library homebrew brew rb in output of brew cask doctor brew cask doctor homebrew version macos sip unknown custom configuration configuration apple internal disabled kext signing enabled filesystem protections disabled debugging restrictions disabled dtrace restrictions enabled nvram protections enabled basesystem verification enabled java homebrew cask staging location usr local caskroom homebrew cask taps usr local homebrew library taps homebrew homebrew cask casks usr local homebrew library taps homebrew homebrew cask versions casks usr local homebrew library taps homebrew homebrew cask fonts casks usr local homebrew library taps homebrew homebrew cask drivers casks load paths usr local homebrew library homebrew vendor bundle bundler ruby gems thor lib usr local homebrew library homebrew vendor bundle bundler ruby gems ruby macho lib usr local homebrew library homebrew vendor bundle bundler ruby gems rubocop rspec lib usr local homebrew library homebrew vendor bundle bundler ruby gems rubocop performance lib usr local homebrew library homebrew vendor bundle bundler ruby gems rubocop lib usr local homebrew library homebrew vendor bundle bundler ruby gems unicode display width lib usr local homebrew library homebrew vendor bundle bundler ruby gems ruby progressbar lib usr local homebrew library homebrew vendor bundle bundler ruby gems rubocop ast lib usr local homebrew library homebrew vendor bundle bundler ruby gems rspec wait lib usr local homebrew library homebrew vendor bundle bundler ruby gems rspec retry lib usr local homebrew library homebrew vendor bundle bundler ruby gems rspec its lib usr local homebrew library homebrew vendor bundle bundler ruby gems rspec lib usr local homebrew library homebrew vendor bundle bundler ruby gems rspec mocks lib usr local homebrew library homebrew vendor bundle bundler ruby gems rspec expectations lib usr local homebrew library homebrew vendor bundle bundler ruby gems rspec core lib usr local homebrew library homebrew vendor bundle bundler ruby gems rspec support lib usr local homebrew library homebrew vendor bundle bundler ruby gems ronn lib usr local homebrew library homebrew vendor bundle bundler ruby gems regexp parser lib usr local homebrew library homebrew vendor bundle bundler ruby gems rdiscount lib usr local homebrew library homebrew vendor bundle bundler ruby extensions universal darwin rdiscount usr local homebrew library homebrew vendor bundle bundler ruby gems pry lib usr local homebrew library homebrew vendor bundle bundler ruby gems plist lib usr local homebrew library homebrew vendor bundle bundler ruby gems patchelf lib usr local homebrew library homebrew vendor bundle bundler ruby gems parlour lib usr local homebrew library homebrew vendor bundle bundler ruby gems rainbow lib usr local homebrew library homebrew vendor bundle bundler ruby gems parser lib usr local homebrew library homebrew vendor bundle bundler ruby gems parallel tests lib usr local homebrew library homebrew vendor bundle bundler ruby gems parallel lib usr local homebrew library homebrew vendor bundle bundler ruby gems mustache lib usr local homebrew library homebrew vendor bundle bundler ruby gems method source lib usr local homebrew library homebrew vendor bundle bundler ruby gems mechanize lib usr local homebrew library homebrew vendor bundle bundler ruby gems webrobots lib usr local homebrew library homebrew vendor bundle bundler ruby gems ntlm http lib usr local homebrew library homebrew vendor bundle bundler ruby gems nokogiri lib usr local homebrew library homebrew vendor bundle bundler ruby extensions universal darwin nokogiri usr local homebrew library homebrew vendor bundle bundler ruby gems mini lib usr local homebrew library homebrew vendor bundle bundler ruby gems net http persistent lib usr local homebrew library homebrew vendor bundle bundler ruby gems net http digest auth lib usr local homebrew library homebrew vendor bundle bundler ruby gems mime types lib usr local homebrew library homebrew vendor bundle bundler ruby gems mime types data lib usr local homebrew library homebrew vendor bundle bundler ruby gems http cookie lib usr local homebrew library homebrew vendor bundle bundler ruby gems hpricot lib usr local homebrew library homebrew vendor bundle bundler ruby extensions universal darwin hpricot usr local homebrew library homebrew vendor bundle bundler ruby gems elftools lib usr local homebrew library homebrew vendor bundle bundler ruby gems domain name lib usr local homebrew library homebrew vendor bundle bundler ruby gems unf lib usr local homebrew library homebrew vendor bundle bundler ruby gems unf ext lib usr local homebrew library homebrew vendor bundle bundler ruby extensions universal darwin unf ext usr local homebrew library homebrew vendor bundle bundler ruby gems diff lcs lib usr local homebrew library homebrew vendor bundle bundler ruby gems connection pool lib usr local homebrew library homebrew vendor bundle bundler ruby gems commander lib usr local homebrew library homebrew vendor bundle bundler ruby gems highline lib usr local homebrew library homebrew vendor bundle bundler ruby gems coderay lib usr local homebrew library homebrew vendor bundle bundler ruby gems codecov lib usr local homebrew library homebrew vendor bundle bundler ruby gems simplecov lib usr local homebrew library homebrew vendor bundle bundler ruby gems simplecov html lib usr local homebrew library homebrew vendor bundle bundler ruby gems docile lib usr local homebrew library homebrew vendor bundle bundler ruby gems json lib usr local homebrew library homebrew vendor bundle bundler ruby extensions universal darwin json usr local homebrew library homebrew vendor bundle bundler ruby gems byebug lib usr local homebrew library homebrew vendor bundle bundler ruby extensions universal darwin byebug usr local homebrew library homebrew vendor bundle bundler usr local homebrew library homebrew vendor bundle bundler ruby gems bindata lib usr local homebrew library homebrew vendor bundle bundler ruby gems ast lib usr local homebrew library homebrew vendor bundle bundler ruby gems activesupport lib usr local homebrew library homebrew vendor bundle bundler ruby gems zeitwerk lib usr local homebrew library homebrew vendor bundle bundler ruby gems tzinfo lib usr local homebrew library homebrew vendor bundle bundler ruby gems thread safe lib usr local homebrew library homebrew vendor bundle bundler ruby gems minitest lib usr local homebrew library homebrew vendor bundle bundler ruby gems lib usr local homebrew library homebrew vendor bundle bundler ruby gems concurrent ruby lib concurrent ruby library ruby site library ruby site universal library ruby site system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby universal system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby universal usr local homebrew library homebrew cask environment variables bundle path chruby version gem home gem path homebrew cask opts lc all path rbenv version rubylib rubyopt rubypath shell output of brew tap homebrew cask homebrew cask drivers homebrew cask fonts homebrew cask versions homebrew core vitorgalvao tiny scripts
1
155,170
13,613,583,835
IssuesEvent
2020-09-23 12:06:26
liferay/clay
https://api.github.com/repos/liferay/clay
opened
Expand documentation of Layout
comp: clayui.com comp: documentation
Now that the set of Layout components has been brewing for some time in DXP we could document them a bit better. We could describe some of the more common strategies and document the markup of the component so it's more consistent with all other components. @bryceosterhaus You worked on this, perhaps you can do this, or if you don't have time, provide me with some bullet points of what to go over.
1.0
Expand documentation of Layout - Now that the set of Layout components has been brewing for some time in DXP we could document them a bit better. We could describe some of the more common strategies and document the markup of the component so it's more consistent with all other components. @bryceosterhaus You worked on this, perhaps you can do this, or if you don't have time, provide me with some bullet points of what to go over.
non_main
expand documentation of layout now that the set of layout components has been brewing for some time in dxp we could document them a bit better we could describe some of the more common strategies and document the markup of the component so it s more consistent with all other components bryceosterhaus you worked on this perhaps you can do this or if you don t have time provide me with some bullet points of what to go over
0
3,276
2,832,369,702
IssuesEvent
2015-05-25 07:24:45
HGustavs/LenaSYS
https://api.github.com/repos/HGustavs/LenaSYS
closed
Preview option dissapeared after merge.
CodeViewer highPriority
There is no longer a preview option to choose from in the "Kind". The css for the preview also seems to be missing...
1.0
Preview option dissapeared after merge. - There is no longer a preview option to choose from in the "Kind". The css for the preview also seems to be missing...
non_main
preview option dissapeared after merge there is no longer a preview option to choose from in the kind the css for the preview also seems to be missing
0
167,908
20,730,041,414
IssuesEvent
2022-03-14 08:33:06
zettatips/django-first-blog
https://api.github.com/repos/zettatips/django-first-blog
opened
CVE-2019-14234 (High) detected in django2.2.3, django2.2.3
security vulnerability
## CVE-2019-14234 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>django2.2.3</b>, <b>django2.2.3</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Django 1.11.x before 1.11.23, 2.1.x before 2.1.11, and 2.2.x before 2.2.4. Due to an error in shallow key transformation, key and index lookups for django.contrib.postgres.fields.JSONField, and key lookups for django.contrib.postgres.fields.HStoreField, were subject to SQL injection. This could, for example, be exploited via crafted use of "OR 1=1" in a key or index name to return all records, using a suitably crafted dictionary, with dictionary expansion, as the **kwargs passed to the QuerySet.filter() function. <p>Publish Date: 2019-08-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14234>CVE-2019-14234</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.djangoproject.com/weblog/2019/aug/01/security-releases/">https://www.djangoproject.com/weblog/2019/aug/01/security-releases/</a></p> <p>Release Date: 2019-08-28</p> <p>Fix Resolution: 2.2.4, 2.1.11, 1.11.23</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-14234 (High) detected in django2.2.3, django2.2.3 - ## CVE-2019-14234 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>django2.2.3</b>, <b>django2.2.3</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Django 1.11.x before 1.11.23, 2.1.x before 2.1.11, and 2.2.x before 2.2.4. Due to an error in shallow key transformation, key and index lookups for django.contrib.postgres.fields.JSONField, and key lookups for django.contrib.postgres.fields.HStoreField, were subject to SQL injection. This could, for example, be exploited via crafted use of "OR 1=1" in a key or index name to return all records, using a suitably crafted dictionary, with dictionary expansion, as the **kwargs passed to the QuerySet.filter() function. <p>Publish Date: 2019-08-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14234>CVE-2019-14234</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.djangoproject.com/weblog/2019/aug/01/security-releases/">https://www.djangoproject.com/weblog/2019/aug/01/security-releases/</a></p> <p>Release Date: 2019-08-28</p> <p>Fix Resolution: 2.2.4, 2.1.11, 1.11.23</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in cve high severity vulnerability vulnerable libraries vulnerability details an issue was discovered in django x before x before and x before due to an error in shallow key transformation key and index lookups for django contrib postgres fields jsonfield and key lookups for django contrib postgres fields hstorefield were subject to sql injection this could for example be exploited via crafted use of or in a key or index name to return all records using a suitably crafted dictionary with dictionary expansion as the kwargs passed to the queryset filter function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
233,795
17,892,780,678
IssuesEvent
2021-09-08 03:09:57
CoinAlpha/hummingbot
https://api.github.com/repos/CoinAlpha/hummingbot
opened
Documentation for `APIThrottler` class
documentation
The `APIThrottler` class is one of the two clear components (the other is `HangingOrdersTracker`) that we have individualized even in the code that are going to be used as they are and something that we want to make it usable by the community. Add to developer documentation and include as much information as possible about this class.
1.0
Documentation for `APIThrottler` class - The `APIThrottler` class is one of the two clear components (the other is `HangingOrdersTracker`) that we have individualized even in the code that are going to be used as they are and something that we want to make it usable by the community. Add to developer documentation and include as much information as possible about this class.
non_main
documentation for apithrottler class the apithrottler class is one of the two clear components the other is hangingorderstracker that we have individualized even in the code that are going to be used as they are and something that we want to make it usable by the community add to developer documentation and include as much information as possible about this class
0
541,584
15,829,677,972
IssuesEvent
2021-04-06 11:30:59
kubernetes-sigs/cluster-api
https://api.github.com/repos/kubernetes-sigs/cluster-api
closed
Set CAPI status to paused or running
area/clusterctl help wanted kind/feature lifecycle/stale priority/backlog
<!-- NOTE: ⚠️ For larger proposals, we follow the CAEP process as outlined in https://sigs.k8s.io/cluster-api/CONTRIBUTING.md. --> **User Story** As a operator I would like to set my CAPI management cluster to a state of Paused or Running for the reason of needing to safely backup or restore my CAPI cluster. **Detailed Description** In order to safely backup, restore, duplicate a cluster and leave it in 'stopped' state there is a need for the following clusterctl commands. ` clusterctl state stop clusterctl state start clusterctl status > CAPI processes are stopped > CAPI processes are running ` **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.] /kind feature
1.0
Set CAPI status to paused or running - <!-- NOTE: ⚠️ For larger proposals, we follow the CAEP process as outlined in https://sigs.k8s.io/cluster-api/CONTRIBUTING.md. --> **User Story** As a operator I would like to set my CAPI management cluster to a state of Paused or Running for the reason of needing to safely backup or restore my CAPI cluster. **Detailed Description** In order to safely backup, restore, duplicate a cluster and leave it in 'stopped' state there is a need for the following clusterctl commands. ` clusterctl state stop clusterctl state start clusterctl status > CAPI processes are stopped > CAPI processes are running ` **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.] /kind feature
non_main
set capi status to paused or running user story as a operator i would like to set my capi management cluster to a state of paused or running for the reason of needing to safely backup or restore my capi cluster detailed description in order to safely backup restore duplicate a cluster and leave it in stopped state there is a need for the following clusterctl commands clusterctl state stop clusterctl state start clusterctl status capi processes are stopped capi processes are running anything else you would like to add kind feature
0
563,941
16,706,597,895
IssuesEvent
2021-06-09 10:43:16
googleapis/google-api-ruby-client
https://api.github.com/repos/googleapis/google-api-ruby-client
closed
Synthesis failed for gmailpostmastertools-v1
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate gmailpostmastertools-v1. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the gmailpostmastertools-v1 API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` 2021-06-08 03:14:11,787 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client 2021-06-08 03:14:12,575 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2021-06-08 03:14:12,578 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2021-06-08 03:14:12,580 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2021-06-08 03:14:12,583 autosynth [DEBUG] > Running: git config push.default simple 2021-06-08 03:14:12,586 autosynth [DEBUG] > Running: git branch -f autosynth-gmailpostmastertools-v1 2021-06-08 03:14:12,589 autosynth [DEBUG] > Running: git checkout autosynth-gmailpostmastertools-v1 Switched to branch 'autosynth-gmailpostmastertools-v1' 2021-06-08 03:14:12,823 autosynth [INFO] > Running synthtool 2021-06-08 03:14:12,823 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gmailpostmastertools_v1/synth.metadata', 'synth.py', '--'] 2021-06-08 03:14:12,823 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/gmailpostmastertools/v1/sponge_log.log 2021-06-08 03:14:12,825 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-gmailpostmastertools_v1/synth.metadata synth.py -- gmailpostmastertools v1 2021-06-08 03:14:13,028 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py. On branch autosynth-gmailpostmastertools-v1 nothing to commit, working tree clean 2021-06-08 03:14:13,092 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gmailpostmastertools v1 DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gmailpostmastertools v1 git clean -df bundle install Don't run Bundler as root. Bundler can ask for sudo if it is needed, and installing your bundle as root will break this application for all non-root users on this machine. Fetching source index from https://rubygems.org/ Retrying fetcher due to error (2/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (3/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (4/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Net::HTTPBadGateway: connection failure chown -R 1000:1000 /workspace/generated 2021-06-08 03:15:04,085 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gmailpostmastertools v1: None ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gmailpostmastertools v1: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke return __callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module> shell.run(command, hide_output=False) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'gmailpostmastertools', 'v1']' returned non-zero exit status 1. 2021-06-08 03:15:04,111 autosynth [ERROR] > Synthesis failed 2021-06-08 03:15:04,111 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main ).synthesize(synth_log_path / "sponge_log.log") File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gmailpostmastertools_v1/synth.metadata', 'synth.py', '--', 'gmailpostmastertools', 'v1']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
1.0
Synthesis failed for gmailpostmastertools-v1 - Hello! Autosynth couldn't regenerate gmailpostmastertools-v1. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the gmailpostmastertools-v1 API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` 2021-06-08 03:14:11,787 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client 2021-06-08 03:14:12,575 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2021-06-08 03:14:12,578 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2021-06-08 03:14:12,580 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2021-06-08 03:14:12,583 autosynth [DEBUG] > Running: git config push.default simple 2021-06-08 03:14:12,586 autosynth [DEBUG] > Running: git branch -f autosynth-gmailpostmastertools-v1 2021-06-08 03:14:12,589 autosynth [DEBUG] > Running: git checkout autosynth-gmailpostmastertools-v1 Switched to branch 'autosynth-gmailpostmastertools-v1' 2021-06-08 03:14:12,823 autosynth [INFO] > Running synthtool 2021-06-08 03:14:12,823 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gmailpostmastertools_v1/synth.metadata', 'synth.py', '--'] 2021-06-08 03:14:12,823 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/gmailpostmastertools/v1/sponge_log.log 2021-06-08 03:14:12,825 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-gmailpostmastertools_v1/synth.metadata synth.py -- gmailpostmastertools v1 2021-06-08 03:14:13,028 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py. On branch autosynth-gmailpostmastertools-v1 nothing to commit, working tree clean 2021-06-08 03:14:13,092 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gmailpostmastertools v1 DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gmailpostmastertools v1 git clean -df bundle install Don't run Bundler as root. Bundler can ask for sudo if it is needed, and installing your bundle as root will break this application for all non-root users on this machine. Fetching source index from https://rubygems.org/ Retrying fetcher due to error (2/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (3/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (4/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Net::HTTPBadGateway: connection failure chown -R 1000:1000 /workspace/generated 2021-06-08 03:15:04,085 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gmailpostmastertools v1: None ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth gmailpostmastertools v1: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke return __callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module> shell.run(command, hide_output=False) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'gmailpostmastertools', 'v1']' returned non-zero exit status 1. 2021-06-08 03:15:04,111 autosynth [ERROR] > Synthesis failed 2021-06-08 03:15:04,111 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main ).synthesize(synth_log_path / "sponge_log.log") File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-gmailpostmastertools_v1/synth.metadata', 'synth.py', '--', 'gmailpostmastertools', 'v1']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
non_main
synthesis failed for gmailpostmastertools hello autosynth couldn t regenerate gmailpostmastertools broken heart please investigate and fix this issue within business days while it remains broken this library cannot be updated with changes to the gmailpostmastertools api and the library grows stale see for trouble shooting tips here s the output from running synth py autosynth logs will be written to tmpfs src logs google api ruby client autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth gmailpostmastertools autosynth running git checkout autosynth gmailpostmastertools switched to branch autosynth gmailpostmastertools autosynth running synthtool autosynth autosynth log file path tmpfs src logs google api ruby client gmailpostmastertools sponge log log autosynth running tmpfs src github synthtool env bin m synthtool metadata generated google apis gmailpostmastertools synth metadata synth py gmailpostmastertools synthtool executing home kbuilder cache synthtool google api ruby client synth py on branch autosynth gmailpostmastertools nothing to commit working tree clean synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gmailpostmastertools debug synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gmailpostmastertools git clean df bundle install don t run bundler as root bundler can ask for sudo if it is needed and installing your bundle as root will break this application for all non root users on this machine fetching source index from retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response service unavailable retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response service unavailable retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response service unavailable net httpbadgateway connection failure chown r workspace generated synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gmailpostmastertools none error synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth gmailpostmastertools none traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool google api ruby client synth py line in shell run command hide output false file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize synth log path sponge log log file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
599
4,097,416,693
IssuesEvent
2016-06-03 01:26:46
Particular/ServiceControl
https://api.github.com/repos/Particular/ServiceControl
closed
Password validation fails for some Windows 10 users
Tag: Installer Tag: Maintainer Prio Type: Bug
The ServiceControl Manangement UI indicates that the service account passowrd is incorrect on some Windows 10 PCs. The code throws a FileNotFoundException when initiializing the PrincipalContext for the machine. This is related to this issue with Windows 10 - http://stackoverflow.com/questions/33717673/system-directoryservices-accountmanagement-principalcontext-broken-after-windows Testing on affected Windows 10 PC both the workaround listed on stack overflow worked ( Rebuild with `prefer 32-bit` turned off and also replacing the missing reg keys)
True
Password validation fails for some Windows 10 users - The ServiceControl Manangement UI indicates that the service account passowrd is incorrect on some Windows 10 PCs. The code throws a FileNotFoundException when initiializing the PrincipalContext for the machine. This is related to this issue with Windows 10 - http://stackoverflow.com/questions/33717673/system-directoryservices-accountmanagement-principalcontext-broken-after-windows Testing on affected Windows 10 PC both the workaround listed on stack overflow worked ( Rebuild with `prefer 32-bit` turned off and also replacing the missing reg keys)
main
password validation fails for some windows users the servicecontrol manangement ui indicates that the service account passowrd is incorrect on some windows pcs the code throws a filenotfoundexception when initiializing the principalcontext for the machine this is related to this issue with windows testing on affected windows pc both the workaround listed on stack overflow worked rebuild with prefer bit turned off and also replacing the missing reg keys
1
2,941
10,563,690,223
IssuesEvent
2019-10-04 21:44:53
Unleash/unleash-client-core
https://api.github.com/repos/Unleash/unleash-client-core
opened
Add CONTRIBUTING.md guide
maintainer
Should give contributors a set of guidelines for contributing to this project
True
Add CONTRIBUTING.md guide - Should give contributors a set of guidelines for contributing to this project
main
add contributing md guide should give contributors a set of guidelines for contributing to this project
1
11,254
29,115,958,682
IssuesEvent
2023-05-17 01:00:48
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Híbrido] Desenvolvedor Full Stack Java - Softtek
CLT Sênior Python Java Git Docker Stale Híbrido Clean architecture
## Nossa empresa Fundada em 1982, a Softtek é uma fornecedora global de soluções de TI orientadas a processos com 30 escritórios na América do Norte, América Latina, Europa e Ásia. Com 15 centros de entrega global nos EUA, México, China, Brasil, Argentina, Costa Rica, Espanha, Hungria e Índia, a Softtek ajuda a melhorar e produzir resultados previsíveis para corporações de primeira linha em mais de 20 países. ## Descrição da vaga Vaga Analista Desenvolvedor Full Stack Java com modalidade híbrida 3x na semana na Avenida Nações Unidas - SP. ## Local Híbrido ## Requisitos - Java 11; - Framework Quarkus; - Docker; - Kafka; - Rest; - VPC; - Git hub; - Sonar; - Python; - Clean Architecture; - Spring boot; - Design de Software. ## Benefícios - Vale Refeição - Vale alimentação - Plano de saúde - Seguro de vida - Plano deontológico - Auxílio creche - Auxílio estudo - Auxílio mobilidade - Previdência privada. ## Contratação CLT ## Como se candidatar E-mail: julia.souza@softtek.com Telefone: (21) 99103-2805 ## Labels JAVA FRAMEWORK QUARKUS DOCKER KAFKA PYTHON SPRING BOOT REST SONAR ### Nível Sênior ### Regime CLT ### Alocação Híbrido
1.0
[Híbrido] Desenvolvedor Full Stack Java - Softtek - ## Nossa empresa Fundada em 1982, a Softtek é uma fornecedora global de soluções de TI orientadas a processos com 30 escritórios na América do Norte, América Latina, Europa e Ásia. Com 15 centros de entrega global nos EUA, México, China, Brasil, Argentina, Costa Rica, Espanha, Hungria e Índia, a Softtek ajuda a melhorar e produzir resultados previsíveis para corporações de primeira linha em mais de 20 países. ## Descrição da vaga Vaga Analista Desenvolvedor Full Stack Java com modalidade híbrida 3x na semana na Avenida Nações Unidas - SP. ## Local Híbrido ## Requisitos - Java 11; - Framework Quarkus; - Docker; - Kafka; - Rest; - VPC; - Git hub; - Sonar; - Python; - Clean Architecture; - Spring boot; - Design de Software. ## Benefícios - Vale Refeição - Vale alimentação - Plano de saúde - Seguro de vida - Plano deontológico - Auxílio creche - Auxílio estudo - Auxílio mobilidade - Previdência privada. ## Contratação CLT ## Como se candidatar E-mail: julia.souza@softtek.com Telefone: (21) 99103-2805 ## Labels JAVA FRAMEWORK QUARKUS DOCKER KAFKA PYTHON SPRING BOOT REST SONAR ### Nível Sênior ### Regime CLT ### Alocação Híbrido
non_main
desenvolvedor full stack java softtek nossa empresa fundada em a softtek é uma fornecedora global de soluções de ti orientadas a processos com escritórios na américa do norte américa latina europa e ásia com centros de entrega global nos eua méxico china brasil argentina costa rica espanha hungria e índia a softtek ajuda a melhorar e produzir resultados previsíveis para corporações de primeira linha em mais de países descrição da vaga vaga analista desenvolvedor full stack java com modalidade híbrida na semana na avenida nações unidas sp local híbrido requisitos java framework quarkus docker kafka rest vpc git hub sonar python clean architecture spring boot design de software benefícios vale refeição vale alimentação plano de saúde seguro de vida plano deontológico auxílio creche auxílio estudo auxílio mobilidade previdência privada contratação clt como se candidatar e mail julia souza softtek com telefone labels java framework quarkus docker kafka python spring boot rest sonar nível sênior regime clt alocação híbrido
0
748
4,351,169,776
IssuesEvent
2016-07-31 18:06:41
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker: _human_to_bytes broken on current devel head
bug_report cloud docker waiting_on_maintainer
##### ISSUE TYPE Bug Report ##### COMPONENT NAME _docker module ##### ANSIBLE VERSION N/A ##### SUMMARY As of 8d126bd877444c9557b1671521516447cc557d3f the _human_to_bytes() function in docker.py is failing to work at all due to throwing the exception that "Cannot convert 0 into an integer" Without digging too deep on why, it looks like the simple case of the value defaulting to 0 isn't caught and handled properly.
True
docker: _human_to_bytes broken on current devel head - ##### ISSUE TYPE Bug Report ##### COMPONENT NAME _docker module ##### ANSIBLE VERSION N/A ##### SUMMARY As of 8d126bd877444c9557b1671521516447cc557d3f the _human_to_bytes() function in docker.py is failing to work at all due to throwing the exception that "Cannot convert 0 into an integer" Without digging too deep on why, it looks like the simple case of the value defaulting to 0 isn't caught and handled properly.
main
docker human to bytes broken on current devel head issue type bug report component name docker module ansible version n a summary as of the human to bytes function in docker py is failing to work at all due to throwing the exception that cannot convert into an integer without digging too deep on why it looks like the simple case of the value defaulting to isn t caught and handled properly
1
293,432
25,291,868,070
IssuesEvent
2022-11-17 01:19:41
Hekili/hekili
https://api.github.com/repos/Hekili/hekili
closed
Convoke not recommended in M+
incomplete triage in testing retail
### Before You Begin - [X] I confirm that I have downloaded the latest version of the addon. - [X] I am not playing on a private server. - [X] I checked for an [existing, open ticket](https://github.com/Hekili/hekili/issues?q=is%3Aissue+is%3Aopen+sort%3Aupdated-desc) for this issue and was not able to find one. - [X] I edited the title of this bug report (above) so that it describes the issue I am reporting. ### WoW Version Retail (Dragonflight) ### Describe the Issue I'm not sure what caused it, but convoke was not being reccomened when running this tallent build during a dungeon. **Talent String** BcGAGX1kx6Mci9Zl2t+S+sRoPCAAAAAAIkUSDkkEJNtASkcgkAAAAAAgiAIJAQikEBJ0SoAAAAAAAgEA ### How to Reproduce 1. Spec into these tallents : BcGAGX1kx6Mci9Zl2t+S+sRoPCAAAAAAIkUSDkkEJNtASkcgkAAAAAAgiAIJAQikEBJ0SoAAAAAAAgEA 2. Run Junkyard 3. Notice convoke isn't reccomeneded after a short while in the dungeon ### Player Information (Link) [Player Information](https://pastebin.com/y76EDFHm) ### Error Messages (Link) ```Text No Error Message ``` ### Additional Information _No response_ ### Contact Information _No response_
1.0
Convoke not recommended in M+ - ### Before You Begin - [X] I confirm that I have downloaded the latest version of the addon. - [X] I am not playing on a private server. - [X] I checked for an [existing, open ticket](https://github.com/Hekili/hekili/issues?q=is%3Aissue+is%3Aopen+sort%3Aupdated-desc) for this issue and was not able to find one. - [X] I edited the title of this bug report (above) so that it describes the issue I am reporting. ### WoW Version Retail (Dragonflight) ### Describe the Issue I'm not sure what caused it, but convoke was not being reccomened when running this tallent build during a dungeon. **Talent String** BcGAGX1kx6Mci9Zl2t+S+sRoPCAAAAAAIkUSDkkEJNtASkcgkAAAAAAgiAIJAQikEBJ0SoAAAAAAAgEA ### How to Reproduce 1. Spec into these tallents : BcGAGX1kx6Mci9Zl2t+S+sRoPCAAAAAAIkUSDkkEJNtASkcgkAAAAAAgiAIJAQikEBJ0SoAAAAAAAgEA 2. Run Junkyard 3. Notice convoke isn't reccomeneded after a short while in the dungeon ### Player Information (Link) [Player Information](https://pastebin.com/y76EDFHm) ### Error Messages (Link) ```Text No Error Message ``` ### Additional Information _No response_ ### Contact Information _No response_
non_main
convoke not recommended in m before you begin i confirm that i have downloaded the latest version of the addon i am not playing on a private server i checked for an for this issue and was not able to find one i edited the title of this bug report above so that it describes the issue i am reporting wow version retail dragonflight describe the issue i m not sure what caused it but convoke was not being reccomened when running this tallent build during a dungeon talent string s how to reproduce spec into these tallents s run junkyard notice convoke isn t reccomeneded after a short while in the dungeon player information link error messages link text no error message additional information no response contact information no response
0
422,648
28,452,218,584
IssuesEvent
2023-04-17 02:27:29
timescale/docs
https://api.github.com/repos/timescale/docs
closed
[Site Bug] Issue with the page: /use-timescale/latest/continuous-aggregates/time/
bug documentation community
_Use this template for reporting bugs in the docs._ # Describe the bug Example setting chunk size does not match text. ## What do the docs say now? > To create a hypertable that uses an integer-based column as time, you need to provide the chunk time interval. In this case, each chunk is 10 minutes. ``` SELECT create_hypertable('devices', 'time', chunk_time_interval => 20); ``` ## What should the docs say? I think it should be `chunk_time_interval => 600`? # Page affected https://docs.timescale.com/use-timescale/latest/continuous-aggregates/time/ # Subject matter expert (SME) [If known, who is a good person to ask about this topic] # Screenshots [Attach images of screenshots showing the bug] # Any further info [Anything else you want to add, or further links] ## Contributing to documentation We welcome documentation contributions! * For information about how to propose a change, see the [contributing guide](https://github.com/timescale/docs/blob/latest/CONTRIBUTING.md) in our GitHub repository. * For information on style and word usage, see the [style guide](https://docs.timescale.com/about/latest/contribute-to-docs)
1.0
[Site Bug] Issue with the page: /use-timescale/latest/continuous-aggregates/time/ - _Use this template for reporting bugs in the docs._ # Describe the bug Example setting chunk size does not match text. ## What do the docs say now? > To create a hypertable that uses an integer-based column as time, you need to provide the chunk time interval. In this case, each chunk is 10 minutes. ``` SELECT create_hypertable('devices', 'time', chunk_time_interval => 20); ``` ## What should the docs say? I think it should be `chunk_time_interval => 600`? # Page affected https://docs.timescale.com/use-timescale/latest/continuous-aggregates/time/ # Subject matter expert (SME) [If known, who is a good person to ask about this topic] # Screenshots [Attach images of screenshots showing the bug] # Any further info [Anything else you want to add, or further links] ## Contributing to documentation We welcome documentation contributions! * For information about how to propose a change, see the [contributing guide](https://github.com/timescale/docs/blob/latest/CONTRIBUTING.md) in our GitHub repository. * For information on style and word usage, see the [style guide](https://docs.timescale.com/about/latest/contribute-to-docs)
non_main
issue with the page use timescale latest continuous aggregates time use this template for reporting bugs in the docs describe the bug example setting chunk size does not match text what do the docs say now to create a hypertable that uses an integer based column as time you need to provide the chunk time interval in this case each chunk is minutes select create hypertable devices time chunk time interval what should the docs say i think it should be chunk time interval page affected subject matter expert sme screenshots any further info contributing to documentation we welcome documentation contributions for information about how to propose a change see the in our github repository for information on style and word usage see the
0
2,545
8,674,894,603
IssuesEvent
2018-11-30 09:16:44
citrusframework/citrus
https://api.github.com/repos/citrusframework/citrus
opened
Change of class signature and typing of WaitConditionBuilder
Prio: High State: To discuss Type: Maintainance
**Citrus Version** >= 2.7.7 **Description** If you upgrade your Citrus version to 2.7.7 or higher, we've a breaking change in the `WaitConditionBuilder` class signature which is now abstract and of different typing. We should discuss if we build this back, because this does only affect implementations which created their own wait condition builders for their own wait conditions. **API before change** ```java public class WaitConditionBuilder<T extends Condition> ``` **API after change** ```java public abstract class WaitConditionBuilder<T extends Condition, S extends WaitConditionBuilder> { ``` **Additional information** Issue: #417 Commit: https://github.com/citrusframework/citrus/commit/515e840f9133383d19304916db197ce5fdb9ac83#diff-fc38b7c87355bad09a45d5e7589fd83eR27 BR, Sven
True
Change of class signature and typing of WaitConditionBuilder - **Citrus Version** >= 2.7.7 **Description** If you upgrade your Citrus version to 2.7.7 or higher, we've a breaking change in the `WaitConditionBuilder` class signature which is now abstract and of different typing. We should discuss if we build this back, because this does only affect implementations which created their own wait condition builders for their own wait conditions. **API before change** ```java public class WaitConditionBuilder<T extends Condition> ``` **API after change** ```java public abstract class WaitConditionBuilder<T extends Condition, S extends WaitConditionBuilder> { ``` **Additional information** Issue: #417 Commit: https://github.com/citrusframework/citrus/commit/515e840f9133383d19304916db197ce5fdb9ac83#diff-fc38b7c87355bad09a45d5e7589fd83eR27 BR, Sven
main
change of class signature and typing of waitconditionbuilder citrus version description if you upgrade your citrus version to or higher we ve a breaking change in the waitconditionbuilder class signature which is now abstract and of different typing we should discuss if we build this back because this does only affect implementations which created their own wait condition builders for their own wait conditions api before change java public class waitconditionbuilder api after change java public abstract class waitconditionbuilder additional information issue commit br sven
1
4,146
19,712,661,387
IssuesEvent
2022-01-13 07:46:47
keptn/community
https://api.github.com/repos/keptn/community
closed
REQUEST: Maintainer status for @oleg-nenashev
membership:maintainer status:approved
I would like to request the maintainer status in Keptn. I have been an active contributor since July 2021, and currently I am helping with community management and Keptn growth. I do not contribute much to the Keptn production codebase but I focus on other areas: public roadmap, issue backlog, documentation and tutorials, Keptn adoption, contributor guidelines, community marketing, etc. All of that is an important part of the Keptn community! My main use-cases for the maintainer status are: - GitHub org and repository administration so that I can help other maintainers with processing requests and improving the community - Access to the CNCF support portal o that I can handle CNCF interaction on my own for the ecosystem, e.g. community.cncf.io, Maintainer List, LFX Mentorship, LFX Insights, Slack, Core Infrastructure Initiative, etc. - Representing the Keptn project officially in the CNCF and other communities/organizations, e.g. for Keptn incubation or technical partnerships ### GitHub Username @oleg-nenashev ### Membership level maintainer ### Requirements (for member) - [x] I have reviewed the community membership guidelines (https://github.com/keptn/community/blob/master/COMMUNITY_MEMBERSHIP.md) - [x] I have enabled 2FA on my GitHub account. See https://github.com/settings/security - [x] I have subscribed to the [Keptn Slack channel](http://slack.keptn.sh/) - [x] I am actively contributing to 1 or more Keptn subprojects - [x] I have two sponsors that meet the sponsor requirements listed in the community membership guidelines. Among other requirements, sponsors must be approvers or maintainers of at least one repository in the organization and not both affiliated with the same company - [x] I have spoken to 2 sponsors (approvers or maintainers) ahead of this application, and they have agreed to sponsor my application - [x] I have filed a PR to add myself as [project member](https://github.com/keptn/keptn/blob/master/MAINTAINERS) and referenced this issue in the PR ### Requirements (for approver, if applicable) - [x] I am a reviewer of the codebase for at least 1 month - [x] I am a reviewer for/or author of substantial PRs to the codebase, with the definition of substantial subject to the maintainer's discretion (e.g., refactors/adds new functionality rather than one-line changes). - [x] I have spoken to 2 sponsors (maintainers) ahead of this application, and they have agreed to sponsor my application - [x] I have filed a PR to add myself as [project approver](https://github.com/keptn/keptn/blob/master/MAINTAINERS) and referenced this issue in the PR ### Requirements (for maintainer, if applicable) - [x] I have deep understanding of the technical goals and direction of the subproject - [x] I have deep understanding of the technical domain (specifically the language) of the subproject - [x] I did sustained contributions to design and direction by doing all of: - [x] I am authoring and reviewing proposals - [x] I am initiating, contributing, and resolving discussions (e-mails, GitHub issues, meetings) - [x] I am identifying subtle or complex issues in designs and implementation PRs - [x] I am directly contributing to the subproject through implementation and / or review - [x] I am aligning with the overall project goals, specifications, and design principles. I am bringing general questions and requests to the discussions as part of the specifications project. - [x] I have spoken to 2 sponsors (maintainers) ahead of this application, and they have agreed to sponsor my application - [x] I have filed a PR to add myself as [project maintainer](https://github.com/keptn/keptn/blob/master/MAINTAINERS) and referenced this issue in the PR ### Sponsors <!-- Replace (at) with the `@` sign --> - @grabnerandi - @AloisReitbauer - @johannes-b - @thisthat - @christian-kreuzberger-dtx Each sponsor should reply to this issue with the comment "*I support*". Please remember, it is an applicant's responsibility to get their sponsors' confirmation before submitting the request. ### List of contributions to the Keptn project - PRs reviewed / authored - Filed pull requests: https://github.com/search?q=org%3Akeptn+org%3Akeptn-sandbox+org%3Akeptn-integrations+is%3Apr+author%3Aoleg-nenashev&type=issues . Key authored ones: - https://github.com/keptn/keptn/pull/6405 - https://github.com/keptn/keptn.github.io/pull/989 - https://github.com/keptn/tutorials/pull/209 - https://github.com/keptn/community/pull/77 - Reviewed PRs: https://github.com/search?q=org%3Akeptn+org%3Akeptn-sandbox+org%3Akeptn-integrations+is%3Apr+reviewed-by%3Aoleg-nenashev&type=issues - Issues responded to - Keptn RFEs relevant to the community feedback and requests - Keptn community roadmap on https://github.com/keptn/community/issues - User support in the `#help` channel - SIG projects I am involved with - CDF: Events SIG (Keptn Events is a part of it), Interoperability SIG, TOC, Governing Board - CNCF: [TAG App Delivery](https://github.com/cncf/tag-app-delivery) - Many [Jenkins SIGs](https://www.jenkins.io/sigs)
True
REQUEST: Maintainer status for @oleg-nenashev - I would like to request the maintainer status in Keptn. I have been an active contributor since July 2021, and currently I am helping with community management and Keptn growth. I do not contribute much to the Keptn production codebase but I focus on other areas: public roadmap, issue backlog, documentation and tutorials, Keptn adoption, contributor guidelines, community marketing, etc. All of that is an important part of the Keptn community! My main use-cases for the maintainer status are: - GitHub org and repository administration so that I can help other maintainers with processing requests and improving the community - Access to the CNCF support portal o that I can handle CNCF interaction on my own for the ecosystem, e.g. community.cncf.io, Maintainer List, LFX Mentorship, LFX Insights, Slack, Core Infrastructure Initiative, etc. - Representing the Keptn project officially in the CNCF and other communities/organizations, e.g. for Keptn incubation or technical partnerships ### GitHub Username @oleg-nenashev ### Membership level maintainer ### Requirements (for member) - [x] I have reviewed the community membership guidelines (https://github.com/keptn/community/blob/master/COMMUNITY_MEMBERSHIP.md) - [x] I have enabled 2FA on my GitHub account. See https://github.com/settings/security - [x] I have subscribed to the [Keptn Slack channel](http://slack.keptn.sh/) - [x] I am actively contributing to 1 or more Keptn subprojects - [x] I have two sponsors that meet the sponsor requirements listed in the community membership guidelines. Among other requirements, sponsors must be approvers or maintainers of at least one repository in the organization and not both affiliated with the same company - [x] I have spoken to 2 sponsors (approvers or maintainers) ahead of this application, and they have agreed to sponsor my application - [x] I have filed a PR to add myself as [project member](https://github.com/keptn/keptn/blob/master/MAINTAINERS) and referenced this issue in the PR ### Requirements (for approver, if applicable) - [x] I am a reviewer of the codebase for at least 1 month - [x] I am a reviewer for/or author of substantial PRs to the codebase, with the definition of substantial subject to the maintainer's discretion (e.g., refactors/adds new functionality rather than one-line changes). - [x] I have spoken to 2 sponsors (maintainers) ahead of this application, and they have agreed to sponsor my application - [x] I have filed a PR to add myself as [project approver](https://github.com/keptn/keptn/blob/master/MAINTAINERS) and referenced this issue in the PR ### Requirements (for maintainer, if applicable) - [x] I have deep understanding of the technical goals and direction of the subproject - [x] I have deep understanding of the technical domain (specifically the language) of the subproject - [x] I did sustained contributions to design and direction by doing all of: - [x] I am authoring and reviewing proposals - [x] I am initiating, contributing, and resolving discussions (e-mails, GitHub issues, meetings) - [x] I am identifying subtle or complex issues in designs and implementation PRs - [x] I am directly contributing to the subproject through implementation and / or review - [x] I am aligning with the overall project goals, specifications, and design principles. I am bringing general questions and requests to the discussions as part of the specifications project. - [x] I have spoken to 2 sponsors (maintainers) ahead of this application, and they have agreed to sponsor my application - [x] I have filed a PR to add myself as [project maintainer](https://github.com/keptn/keptn/blob/master/MAINTAINERS) and referenced this issue in the PR ### Sponsors <!-- Replace (at) with the `@` sign --> - @grabnerandi - @AloisReitbauer - @johannes-b - @thisthat - @christian-kreuzberger-dtx Each sponsor should reply to this issue with the comment "*I support*". Please remember, it is an applicant's responsibility to get their sponsors' confirmation before submitting the request. ### List of contributions to the Keptn project - PRs reviewed / authored - Filed pull requests: https://github.com/search?q=org%3Akeptn+org%3Akeptn-sandbox+org%3Akeptn-integrations+is%3Apr+author%3Aoleg-nenashev&type=issues . Key authored ones: - https://github.com/keptn/keptn/pull/6405 - https://github.com/keptn/keptn.github.io/pull/989 - https://github.com/keptn/tutorials/pull/209 - https://github.com/keptn/community/pull/77 - Reviewed PRs: https://github.com/search?q=org%3Akeptn+org%3Akeptn-sandbox+org%3Akeptn-integrations+is%3Apr+reviewed-by%3Aoleg-nenashev&type=issues - Issues responded to - Keptn RFEs relevant to the community feedback and requests - Keptn community roadmap on https://github.com/keptn/community/issues - User support in the `#help` channel - SIG projects I am involved with - CDF: Events SIG (Keptn Events is a part of it), Interoperability SIG, TOC, Governing Board - CNCF: [TAG App Delivery](https://github.com/cncf/tag-app-delivery) - Many [Jenkins SIGs](https://www.jenkins.io/sigs)
main
request maintainer status for oleg nenashev i would like to request the maintainer status in keptn i have been an active contributor since july and currently i am helping with community management and keptn growth i do not contribute much to the keptn production codebase but i focus on other areas public roadmap issue backlog documentation and tutorials keptn adoption contributor guidelines community marketing etc all of that is an important part of the keptn community my main use cases for the maintainer status are github org and repository administration so that i can help other maintainers with processing requests and improving the community access to the cncf support portal o that i can handle cncf interaction on my own for the ecosystem e g community cncf io maintainer list lfx mentorship lfx insights slack core infrastructure initiative etc representing the keptn project officially in the cncf and other communities organizations e g for keptn incubation or technical partnerships github username oleg nenashev membership level maintainer requirements for member i have reviewed the community membership guidelines i have enabled on my github account see i have subscribed to the i am actively contributing to or more keptn subprojects i have two sponsors that meet the sponsor requirements listed in the community membership guidelines among other requirements sponsors must be approvers or maintainers of at least one repository in the organization and not both affiliated with the same company i have spoken to sponsors approvers or maintainers ahead of this application and they have agreed to sponsor my application i have filed a pr to add myself as and referenced this issue in the pr requirements for approver if applicable i am a reviewer of the codebase for at least month i am a reviewer for or author of substantial prs to the codebase with the definition of substantial subject to the maintainer s discretion e g refactors adds new functionality rather than one line changes i have spoken to sponsors maintainers ahead of this application and they have agreed to sponsor my application i have filed a pr to add myself as and referenced this issue in the pr requirements for maintainer if applicable i have deep understanding of the technical goals and direction of the subproject i have deep understanding of the technical domain specifically the language of the subproject i did sustained contributions to design and direction by doing all of i am authoring and reviewing proposals i am initiating contributing and resolving discussions e mails github issues meetings i am identifying subtle or complex issues in designs and implementation prs i am directly contributing to the subproject through implementation and or review i am aligning with the overall project goals specifications and design principles i am bringing general questions and requests to the discussions as part of the specifications project i have spoken to sponsors maintainers ahead of this application and they have agreed to sponsor my application i have filed a pr to add myself as and referenced this issue in the pr sponsors grabnerandi aloisreitbauer johannes b thisthat christian kreuzberger dtx each sponsor should reply to this issue with the comment i support please remember it is an applicant s responsibility to get their sponsors confirmation before submitting the request list of contributions to the keptn project prs reviewed authored filed pull requests key authored ones reviewed prs issues responded to keptn rfes relevant to the community feedback and requests keptn community roadmap on user support in the help channel sig projects i am involved with cdf events sig keptn events is a part of it interoperability sig toc governing board cncf many
1
995
4,759,595,993
IssuesEvent
2016-10-24 23:11:58
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
azure_rm_storageaccount throws SkuName' is not defined
affects_2.2 azure bug_report cloud waiting_on_maintainer
Moving #17949 to the correct location with my own information. ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME azure_rm_storageaccount ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT RHEL 7 ##### SUMMARY Using a similar example to the one found in the Azure guide, I receive a module failure that the SkuName is not defined ``` - name: testing storage module hosts: localhost tasks: - name: Create storage account azure_rm_storageaccount: resource_group: mperz name: testing account_type: Standard_GRS state: present ``` ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py", line 442, in <module> main() File "/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py", line 439, in main AzureRMStorageAccount() File "/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py", line 180, in __init__ for key in SkuName: NameError: global name 'SkuName' is not defined fatal: [localhost]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_name": "azure_rm_storageaccount" }, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py\", line 442, in <module>\n main()\n File \"/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py\", line 439, in main\n AzureRMStorageAccount()\n File \"/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py\", line 180, in __init__\n for key in SkuName:\nNameError: global name 'SkuName' is not defined\n", "module_stdout": "", "msg": "MODULE FAILURE" } ``` ##### Additional Info ``` pip freeze|grep azure azure==2.0.0rc5 azure-batch==0.30.0rc5 azure-common==1.1.4 azure-graphrbac==0.30.0rc5 azure-mgmt==0.30.0rc5 azure-mgmt-authorization==0.30.0rc5 azure-mgmt-batch==0.30.0rc5 azure-mgmt-cdn==0.30.0rc5 azure-mgmt-cognitiveservices==0.30.0rc5 azure-mgmt-commerce==0.30.0rc5 azure-mgmt-compute==0.30.0rc5 azure-mgmt-keyvault==0.30.0rc5 azure-mgmt-logic==0.30.0rc5 azure-mgmt-network==0.30.0rc5 azure-mgmt-notificationhubs==0.30.0rc5 azure-mgmt-nspkg==1.0.0 azure-mgmt-powerbiembedded==0.30.0rc5 azure-mgmt-redis==0.30.0rc5 azure-mgmt-resource==0.30.0rc5 azure-mgmt-scheduler==0.30.0rc5 azure-mgmt-storage==0.30.0rc5 azure-mgmt-web==0.30.0rc5 azure-nspkg==1.0.0 azure-servicebus==0.20.2 azure-servicemanagement-legacy==0.20.3 azure-storage==0.32.0 ``` This happens regardless of which account_type you select.
True
azure_rm_storageaccount throws SkuName' is not defined - Moving #17949 to the correct location with my own information. ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME azure_rm_storageaccount ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT RHEL 7 ##### SUMMARY Using a similar example to the one found in the Azure guide, I receive a module failure that the SkuName is not defined ``` - name: testing storage module hosts: localhost tasks: - name: Create storage account azure_rm_storageaccount: resource_group: mperz name: testing account_type: Standard_GRS state: present ``` ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py", line 442, in <module> main() File "/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py", line 439, in main AzureRMStorageAccount() File "/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py", line 180, in __init__ for key in SkuName: NameError: global name 'SkuName' is not defined fatal: [localhost]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_name": "azure_rm_storageaccount" }, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py\", line 442, in <module>\n main()\n File \"/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py\", line 439, in main\n AzureRMStorageAccount()\n File \"/tmp/ansible_Bb1sra/ansible_module_azure_rm_storageaccount.py\", line 180, in __init__\n for key in SkuName:\nNameError: global name 'SkuName' is not defined\n", "module_stdout": "", "msg": "MODULE FAILURE" } ``` ##### Additional Info ``` pip freeze|grep azure azure==2.0.0rc5 azure-batch==0.30.0rc5 azure-common==1.1.4 azure-graphrbac==0.30.0rc5 azure-mgmt==0.30.0rc5 azure-mgmt-authorization==0.30.0rc5 azure-mgmt-batch==0.30.0rc5 azure-mgmt-cdn==0.30.0rc5 azure-mgmt-cognitiveservices==0.30.0rc5 azure-mgmt-commerce==0.30.0rc5 azure-mgmt-compute==0.30.0rc5 azure-mgmt-keyvault==0.30.0rc5 azure-mgmt-logic==0.30.0rc5 azure-mgmt-network==0.30.0rc5 azure-mgmt-notificationhubs==0.30.0rc5 azure-mgmt-nspkg==1.0.0 azure-mgmt-powerbiembedded==0.30.0rc5 azure-mgmt-redis==0.30.0rc5 azure-mgmt-resource==0.30.0rc5 azure-mgmt-scheduler==0.30.0rc5 azure-mgmt-storage==0.30.0rc5 azure-mgmt-web==0.30.0rc5 azure-nspkg==1.0.0 azure-servicebus==0.20.2 azure-servicemanagement-legacy==0.20.3 azure-storage==0.32.0 ``` This happens regardless of which account_type you select.
main
azure rm storageaccount throws skuname is not defined moving to the correct location with my own information issue type bug report component name azure rm storageaccount ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides os environment rhel summary using a similar example to the one found in the azure guide i receive a module failure that the skuname is not defined name testing storage module hosts localhost tasks name create storage account azure rm storageaccount resource group mperz name testing account type standard grs state present an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible ansible module azure rm storageaccount py line in main file tmp ansible ansible module azure rm storageaccount py line in main azurermstorageaccount file tmp ansible ansible module azure rm storageaccount py line in init for key in skuname nameerror global name skuname is not defined fatal failed changed false failed true invocation module name azure rm storageaccount module stderr traceback most recent call last n file tmp ansible ansible module azure rm storageaccount py line in n main n file tmp ansible ansible module azure rm storageaccount py line in main n azurermstorageaccount n file tmp ansible ansible module azure rm storageaccount py line in init n for key in skuname nnameerror global name skuname is not defined n module stdout msg module failure additional info pip freeze grep azure azure azure batch azure common azure graphrbac azure mgmt azure mgmt authorization azure mgmt batch azure mgmt cdn azure mgmt cognitiveservices azure mgmt commerce azure mgmt compute azure mgmt keyvault azure mgmt logic azure mgmt network azure mgmt notificationhubs azure mgmt nspkg azure mgmt powerbiembedded azure mgmt redis azure mgmt resource azure mgmt scheduler azure mgmt storage azure mgmt web azure nspkg azure servicebus azure servicemanagement legacy azure storage this happens regardless of which account type you select
1
174
2,755,882,987
IssuesEvent
2015-04-27 01:09:20
canadainc/quran10
https://api.github.com/repos/canadainc/quran10
closed
Refactor Juz and Surah pages to move common elements into common helper
Component-UI Maintainability Type-Enhancement
Move LazyMediaPlayer, VersesListView, etc.
True
Refactor Juz and Surah pages to move common elements into common helper - Move LazyMediaPlayer, VersesListView, etc.
main
refactor juz and surah pages to move common elements into common helper move lazymediaplayer verseslistview etc
1
1,911
6,577,573,138
IssuesEvent
2017-09-12 01:51:13
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
rax module fails to create Cloud Server with multiple Cloud Images with same human_id
affects_1.9 bug_report cloud waiting_on_maintainer
##### Issue Type: - Bug Report ##### Plugin Name: `rax` ##### Ansible Version: ``` ansible 1.9.3 configured module search path = None ``` This includes vendored `pyrax==1.9.7`. ##### Ansible Configuration: N/A ##### Environment: N/A ##### Summary: The rax module supports the image key to be either `id` (eg UUID) or the `human_id`. The Cloud Images API v2 supports images to be created with the same `human_id`. However when there are greater than one `human_id` in an images list, the `rax` module fails a create. ##### Steps To Reproduce: - Create Cloud Server - Create Cloud Image of Cloud Server - Call `rax` module to create additional Cloud Server from `human_id` image of above Cloud Image. You should succeed. - Create Cloud Image of same Cloud Server reusing the same `human_id`. - Call `rax` module to create additional Cloud Server from `human_id` image of above Cloud Image. You fail with `No matching image found (${IMAGE_NAME})` - Delete Cloud Image created from the same `human_id`, leaving behind one Cloud Image of the given `human_id`. - Call `rax` module to create Cloud Server from `human_id` image of above Cloud Image. You should succeed. ##### Expected Results: You should succeed with the second Cloud Server build. ##### Actual Results: You fail with `No matching image found (${IMAGE_NAME})`.
True
rax module fails to create Cloud Server with multiple Cloud Images with same human_id - ##### Issue Type: - Bug Report ##### Plugin Name: `rax` ##### Ansible Version: ``` ansible 1.9.3 configured module search path = None ``` This includes vendored `pyrax==1.9.7`. ##### Ansible Configuration: N/A ##### Environment: N/A ##### Summary: The rax module supports the image key to be either `id` (eg UUID) or the `human_id`. The Cloud Images API v2 supports images to be created with the same `human_id`. However when there are greater than one `human_id` in an images list, the `rax` module fails a create. ##### Steps To Reproduce: - Create Cloud Server - Create Cloud Image of Cloud Server - Call `rax` module to create additional Cloud Server from `human_id` image of above Cloud Image. You should succeed. - Create Cloud Image of same Cloud Server reusing the same `human_id`. - Call `rax` module to create additional Cloud Server from `human_id` image of above Cloud Image. You fail with `No matching image found (${IMAGE_NAME})` - Delete Cloud Image created from the same `human_id`, leaving behind one Cloud Image of the given `human_id`. - Call `rax` module to create Cloud Server from `human_id` image of above Cloud Image. You should succeed. ##### Expected Results: You should succeed with the second Cloud Server build. ##### Actual Results: You fail with `No matching image found (${IMAGE_NAME})`.
main
rax module fails to create cloud server with multiple cloud images with same human id issue type bug report plugin name rax ansible version ansible configured module search path none this includes vendored pyrax ansible configuration n a environment n a summary the rax module supports the image key to be either id eg uuid or the human id the cloud images api supports images to be created with the same human id however when there are greater than one human id in an images list the rax module fails a create steps to reproduce create cloud server create cloud image of cloud server call rax module to create additional cloud server from human id image of above cloud image you should succeed create cloud image of same cloud server reusing the same human id call rax module to create additional cloud server from human id image of above cloud image you fail with no matching image found image name delete cloud image created from the same human id leaving behind one cloud image of the given human id call rax module to create cloud server from human id image of above cloud image you should succeed expected results you should succeed with the second cloud server build actual results you fail with no matching image found image name
1
2,182
7,674,568,160
IssuesEvent
2018-05-15 04:53:18
Kristinita/SashaMiscellaneous
https://api.github.com/repos/Kristinita/SashaMiscellaneous
opened
feature_request(repl.it): support Stylus
need-maintainer
### 1. Summary It would be nice, if Repl.it will support [**Stylus**](http://stylus-lang.com/) CSS preprocessor. ### 2. Argumentation I can use [**Stylus in Codepen.io**](https://blog.codepen.io/documentation/views/editor-view/#the-css-editor-1), but Codepen doesn't support [**more than one project (multiple files) for free plan**](https://codepen.io/pro/projects). It would be nice, if Repl.it users will have the opportunity use multiple Stylus files for repls. It a reason for Stylus users migrate from Codepen to Repl.it. Thanks.
True
feature_request(repl.it): support Stylus - ### 1. Summary It would be nice, if Repl.it will support [**Stylus**](http://stylus-lang.com/) CSS preprocessor. ### 2. Argumentation I can use [**Stylus in Codepen.io**](https://blog.codepen.io/documentation/views/editor-view/#the-css-editor-1), but Codepen doesn't support [**more than one project (multiple files) for free plan**](https://codepen.io/pro/projects). It would be nice, if Repl.it users will have the opportunity use multiple Stylus files for repls. It a reason for Stylus users migrate from Codepen to Repl.it. Thanks.
main
feature request repl it support stylus summary it would be nice if repl it will support css preprocessor argumentation i can use but codepen doesn t support it would be nice if repl it users will have the opportunity use multiple stylus files for repls it a reason for stylus users migrate from codepen to repl it thanks
1
212,524
16,488,012,471
IssuesEvent
2021-05-24 21:10:28
AlaskaAirlines/WC-Generator
https://api.github.com/repos/AlaskaAirlines/WC-Generator
closed
As a [user] I need help with conventions
Status: Work In Progress Type: Documentation
# General Support Request As the Auro team is growing and more people are getting into the mix, the more the difference in naming conventions and commit messages can create momentary confusion. ## Support request It is healthy and time to discuss a naming convention for things like branches and update the standard for commit messages. ## Possible Solution When it comes to branch names, clarity and reference will always be preferred. On larger teams in the past, there has always been a convention that included the author's name and a reference to the issue they are solving. Having a brief description of the issue or a reference to the title is helpful too. An example I'd like to offer is, ```shell dsande/i18nApi#110 or dsande/i18n/#110 ``` I like this example as this quickly gives me the info I need. Also, as I am pulling other author's branches, this will help group branches between authors. When it comes to commit messages, the format we already have is defined in the [Conventional Commits](https://auro.alaskaair.com/contributing) section of the Auro site. ```shell <type>[optional scope]: <description> ``` I would like to propose the following 1. The optional scope is not optional if there are more than one commits to a PR 2. The issue number is added to at least one comment in the PR ```shell feat(fixed API): update fixed feature to accept arguments #905 ``` Please offer your opinions in the comments below.
1.0
As a [user] I need help with conventions - # General Support Request As the Auro team is growing and more people are getting into the mix, the more the difference in naming conventions and commit messages can create momentary confusion. ## Support request It is healthy and time to discuss a naming convention for things like branches and update the standard for commit messages. ## Possible Solution When it comes to branch names, clarity and reference will always be preferred. On larger teams in the past, there has always been a convention that included the author's name and a reference to the issue they are solving. Having a brief description of the issue or a reference to the title is helpful too. An example I'd like to offer is, ```shell dsande/i18nApi#110 or dsande/i18n/#110 ``` I like this example as this quickly gives me the info I need. Also, as I am pulling other author's branches, this will help group branches between authors. When it comes to commit messages, the format we already have is defined in the [Conventional Commits](https://auro.alaskaair.com/contributing) section of the Auro site. ```shell <type>[optional scope]: <description> ``` I would like to propose the following 1. The optional scope is not optional if there are more than one commits to a PR 2. The issue number is added to at least one comment in the PR ```shell feat(fixed API): update fixed feature to accept arguments #905 ``` Please offer your opinions in the comments below.
non_main
as a i need help with conventions general support request as the auro team is growing and more people are getting into the mix the more the difference in naming conventions and commit messages can create momentary confusion support request it is healthy and time to discuss a naming convention for things like branches and update the standard for commit messages possible solution when it comes to branch names clarity and reference will always be preferred on larger teams in the past there has always been a convention that included the author s name and a reference to the issue they are solving having a brief description of the issue or a reference to the title is helpful too an example i d like to offer is shell dsande or dsande i like this example as this quickly gives me the info i need also as i am pulling other author s branches this will help group branches between authors when it comes to commit messages the format we already have is defined in the section of the auro site shell i would like to propose the following the optional scope is not optional if there are more than one commits to a pr the issue number is added to at least one comment in the pr shell feat fixed api update fixed feature to accept arguments please offer your opinions in the comments below
0
2,520
8,655,460,291
IssuesEvent
2018-11-27 16:00:34
codestation/qcma
https://api.github.com/repos/codestation/qcma
closed
QCMA Sometimes Requires a Restart
unmaintained
I am using Mac OS X 10.13.5 and quite frequently (especially on the PSTV) I have to restart QCMA before my PSTV will connect properly. Also I have to select quit twice for the app to fully quit and then sometimes crashes. Both PS Vita-1000 and PSTV on Enso 3.65 spoofed to 3.68.
True
QCMA Sometimes Requires a Restart - I am using Mac OS X 10.13.5 and quite frequently (especially on the PSTV) I have to restart QCMA before my PSTV will connect properly. Also I have to select quit twice for the app to fully quit and then sometimes crashes. Both PS Vita-1000 and PSTV on Enso 3.65 spoofed to 3.68.
main
qcma sometimes requires a restart i am using mac os x and quite frequently especially on the pstv i have to restart qcma before my pstv will connect properly also i have to select quit twice for the app to fully quit and then sometimes crashes both ps vita and pstv on enso spoofed to
1
5,332
26,915,122,782
IssuesEvent
2023-02-07 05:25:59
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
opened
Spack recipies are way out of date.
maintainability
## Expected behavior ## Packaging with spack is getting more and more popular, we should probably bring the versions into the modern age. It will be a bit challenging getting the versioning of deps right in the spack language, but just thought I should flag. AFAIK they only have up to 1.0 https://github.com/spack/spack/blob/develop/var/spack/repos/builtin/packages/py-mdanalysis/package.py
True
Spack recipies are way out of date. - ## Expected behavior ## Packaging with spack is getting more and more popular, we should probably bring the versions into the modern age. It will be a bit challenging getting the versioning of deps right in the spack language, but just thought I should flag. AFAIK they only have up to 1.0 https://github.com/spack/spack/blob/develop/var/spack/repos/builtin/packages/py-mdanalysis/package.py
main
spack recipies are way out of date expected behavior packaging with spack is getting more and more popular we should probably bring the versions into the modern age it will be a bit challenging getting the versioning of deps right in the spack language but just thought i should flag afaik they only have up to
1
3,272
12,491,467,536
IssuesEvent
2020-06-01 04:12:10
short-d/short
https://api.github.com/repos/short-d/short
opened
[Refactor] Integration tests give error because of null values in primary key on migrate up or migrate down
maintainability
**What is frustrating you?** Two errors: ``` panic: pq: column "email" contains null values handling 19_add_constraints_to_user_changelog_table.sql [recovered] panic: pq: column "email" contains null values handling 19_add_constraints_to_user_changelog_table.sql ``` ``` panic: pq: column "user_email" contains null values handling 16_switch_from_email_to_id.sql [recovered] panic: pq: column "user_email" contains null values handling 16_switch_from_email_to_id.sql ``` **Your solution** Remove primary key constraint from user_email column in user_url_relation table and email column in user_changelog table on migrate up and migrate down.
True
[Refactor] Integration tests give error because of null values in primary key on migrate up or migrate down - **What is frustrating you?** Two errors: ``` panic: pq: column "email" contains null values handling 19_add_constraints_to_user_changelog_table.sql [recovered] panic: pq: column "email" contains null values handling 19_add_constraints_to_user_changelog_table.sql ``` ``` panic: pq: column "user_email" contains null values handling 16_switch_from_email_to_id.sql [recovered] panic: pq: column "user_email" contains null values handling 16_switch_from_email_to_id.sql ``` **Your solution** Remove primary key constraint from user_email column in user_url_relation table and email column in user_changelog table on migrate up and migrate down.
main
integration tests give error because of null values in primary key on migrate up or migrate down what is frustrating you two errors panic pq column email contains null values handling add constraints to user changelog table sql panic pq column email contains null values handling add constraints to user changelog table sql panic pq column user email contains null values handling switch from email to id sql panic pq column user email contains null values handling switch from email to id sql your solution remove primary key constraint from user email column in user url relation table and email column in user changelog table on migrate up and migrate down
1
583,761
17,397,948,604
IssuesEvent
2021-08-02 15:34:42
thespacedoctor/sherlock
https://api.github.com/repos/thespacedoctor/sherlock
closed
SDSS name conversion error
priority: 1 type: bug
Getting some skipped annotations because the code falls over when it tries to convert SDSS objects into names. https://star.pst.qub.ac.uk/sne/atlas4/candidate/1122646551085257700/ ``` * 08:35:34 - ERROR: /usr/local/swtools/python/atls/anaconda3/envs/sherlock37/lib/python3.7/site-packages/astrocalc/coords/unit_conversion.py:dec_decimal_to_sexegesimal:443 > DEC must be between -90 - 90 degrees Traceback (most recent call last): File "/usr/local/swtools/python/atls/anaconda3/envs/sherlock37/bin/sherlock", line 8, in <module> sys.exit(main()) File "/usr/local/swtools/python/atls/anaconda3/envs/sherlock37/lib/python3.7/site-packages/sherlock/cl_utils.py", line 208, in main classifier.classify() File "/usr/local/swtools/python/atls/anaconda3/envs/sherlock37/lib/python3.7/site-packages/sherlock/transient_classifier.py", line 493, in classify self.updatePeakMags) File "/usr/local/swtools/python/atls/anaconda3/envs/sherlock37/lib/python3.7/site-packages/sherlock/transient_classifier.py", line 1545, in update_classification_annotations_and_summaries match=row, updatePeakMagnitudes=updatePeakMagnitudes) File "/usr/local/swtools/python/atls/anaconda3/envs/sherlock37/lib/python3.7/site-packages/sherlock/transient_classifier.py", line 1948, in generate_match_annotation betterName = "SDSS J" + ra[0:9] + dec[0:9] TypeError: 'int' object is not subscriptable ```
1.0
SDSS name conversion error - Getting some skipped annotations because the code falls over when it tries to convert SDSS objects into names. https://star.pst.qub.ac.uk/sne/atlas4/candidate/1122646551085257700/ ``` * 08:35:34 - ERROR: /usr/local/swtools/python/atls/anaconda3/envs/sherlock37/lib/python3.7/site-packages/astrocalc/coords/unit_conversion.py:dec_decimal_to_sexegesimal:443 > DEC must be between -90 - 90 degrees Traceback (most recent call last): File "/usr/local/swtools/python/atls/anaconda3/envs/sherlock37/bin/sherlock", line 8, in <module> sys.exit(main()) File "/usr/local/swtools/python/atls/anaconda3/envs/sherlock37/lib/python3.7/site-packages/sherlock/cl_utils.py", line 208, in main classifier.classify() File "/usr/local/swtools/python/atls/anaconda3/envs/sherlock37/lib/python3.7/site-packages/sherlock/transient_classifier.py", line 493, in classify self.updatePeakMags) File "/usr/local/swtools/python/atls/anaconda3/envs/sherlock37/lib/python3.7/site-packages/sherlock/transient_classifier.py", line 1545, in update_classification_annotations_and_summaries match=row, updatePeakMagnitudes=updatePeakMagnitudes) File "/usr/local/swtools/python/atls/anaconda3/envs/sherlock37/lib/python3.7/site-packages/sherlock/transient_classifier.py", line 1948, in generate_match_annotation betterName = "SDSS J" + ra[0:9] + dec[0:9] TypeError: 'int' object is not subscriptable ```
non_main
sdss name conversion error getting some skipped annotations because the code falls over when it tries to convert sdss objects into names error usr local swtools python atls envs lib site packages astrocalc coords unit conversion py dec decimal to sexegesimal dec must be between degrees traceback most recent call last file usr local swtools python atls envs bin sherlock line in sys exit main file usr local swtools python atls envs lib site packages sherlock cl utils py line in main classifier classify file usr local swtools python atls envs lib site packages sherlock transient classifier py line in classify self updatepeakmags file usr local swtools python atls envs lib site packages sherlock transient classifier py line in update classification annotations and summaries match row updatepeakmagnitudes updatepeakmagnitudes file usr local swtools python atls envs lib site packages sherlock transient classifier py line in generate match annotation bettername sdss j ra dec typeerror int object is not subscriptable
0