Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
10,022
13,043,926,124
IssuesEvent
2020-07-29 03:04:21
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `Locate2ArgsUTF8` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `Locate2ArgsUTF8` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `Locate2ArgsUTF8` from TiDB - ## Description Port the scalar function `Locate2ArgsUTF8` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function from tidb description port the scalar function from tidb to coprocessor score mentor s andylokandy recommended skills rust programming learning materials already implemented expressions ported from tidb
1
2,185
5,036,744,813
IssuesEvent
2016-12-17 08:03:02
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
closed
[subtitles] [eng] Revue de la semaine n°10
Language: English Process: [2] Ready for review (1) Process: [6] Approved
# Video title RDLS10 - MATTEO RENZI, EUROPE, PAUVRETÉ, ZONES MORTES DANS LES OCÉANS, SNOWDEN # URL https://www.youtube.com/watch?v=DwvM9oV3v3I # Youtube subtitle language English # Duration 25:16 # URL subtitles https://www.youtube.com/timedtext_editor?ref=player&action_mde_edit_form=1&ui=hd&v=DwvM9oV3v3I&tab=captions&bl=vmp&lang=en
2.0
[subtitles] [eng] Revue de la semaine n°10 - # Video title RDLS10 - MATTEO RENZI, EUROPE, PAUVRETÉ, ZONES MORTES DANS LES OCÉANS, SNOWDEN # URL https://www.youtube.com/watch?v=DwvM9oV3v3I # Youtube subtitle language English # Duration 25:16 # URL subtitles https://www.youtube.com/timedtext_editor?ref=player&action_mde_edit_form=1&ui=hd&v=DwvM9oV3v3I&tab=captions&bl=vmp&lang=en
process
revue de la semaine n° video title matteo renzi europe pauvreté zones mortes dans les océans snowden url youtube subtitle language english duration url subtitles
1
19,075
25,109,935,159
IssuesEvent
2022-11-08 19:37:36
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
vscode multi-root workspace: open terminal asks me to choose cwd even if both have been set to same place
bug help wanted terminal-process
Issue Type: <b>Feature Request</b> Not sure if this is intended behaviour, but I have two root folders in my workspace, I've overriden one to use the other as the cwd, yet I still get asked which one to use when I open a new terminal: ![workflowy](https://user-images.githubusercontent.com/4956427/175555109-c7032e18-b9e6-4e6c-b0c4-9c086dab4195.png) Please can you change this so that when cwds are identical there is no prompt? VS Code version: Code 1.68.1 (30d9c6cd9483b2cc586687151bcbcd635f373630, 2022-06-15T02:58:26.441Z) OS version: Linux x64 5.10.110-15808-ge5740beba59b Restricted Mode: No <!-- generated by issue reporter -->
1.0
vscode multi-root workspace: open terminal asks me to choose cwd even if both have been set to same place - Issue Type: <b>Feature Request</b> Not sure if this is intended behaviour, but I have two root folders in my workspace, I've overriden one to use the other as the cwd, yet I still get asked which one to use when I open a new terminal: ![workflowy](https://user-images.githubusercontent.com/4956427/175555109-c7032e18-b9e6-4e6c-b0c4-9c086dab4195.png) Please can you change this so that when cwds are identical there is no prompt? VS Code version: Code 1.68.1 (30d9c6cd9483b2cc586687151bcbcd635f373630, 2022-06-15T02:58:26.441Z) OS version: Linux x64 5.10.110-15808-ge5740beba59b Restricted Mode: No <!-- generated by issue reporter -->
process
vscode multi root workspace open terminal asks me to choose cwd even if both have been set to same place issue type feature request not sure if this is intended behaviour but i have two root folders in my workspace i ve overriden one to use the other as the cwd yet i still get asked which one to use when i open a new terminal please can you change this so that when cwds are identical there is no prompt vs code version code os version linux restricted mode no
1
163,348
25,797,182,907
IssuesEvent
2022-12-10 17:28:21
boostcampwm-2022/web07-zokboo.com
https://api.github.com/repos/boostcampwm-2022/web07-zokboo.com
closed
문제집 검색 후 문제집 상세보기 페이지
🧩 frontend 🎨 design
- [x] 구조잡기 완료 - [x] 제목, 설명, 작성자 등등 기본정보 보여주기 - [x] 문제 보여주기 - [ ] #207 - [ ] 시험보러가기 버튼 - 문제집 정보 - 문제 목록
1.0
문제집 검색 후 문제집 상세보기 페이지 - - [x] 구조잡기 완료 - [x] 제목, 설명, 작성자 등등 기본정보 보여주기 - [x] 문제 보여주기 - [ ] #207 - [ ] 시험보러가기 버튼 - 문제집 정보 - 문제 목록
non_process
문제집 검색 후 문제집 상세보기 페이지 구조잡기 완료 제목 설명 작성자 등등 기본정보 보여주기 문제 보여주기 시험보러가기 버튼 문제집 정보 문제 목록
0
95,481
8,559,886,029
IssuesEvent
2018-11-08 22:46:00
FreeUKGen/MyopicVicar
https://api.github.com/repos/FreeUKGen/MyopicVicar
closed
Physical File Processing - Processor Stopping unexpectedly with very little Error info
Ready for Deployment testing
The CSV processor stopped processing files for approx 12 hours or so, starting between 10am & 11am on Sept 28th. I have downloaded a copy of the passenger.log file from that day, if required. The issue has resolved itself, but there was very little feedback from the logs, and nothing shown in errbit, to notify of what the issue was, although we suspect it was caused by a problem file that was uploaded. **Action**: Determine the root cause of the issue to prevent re-occurrence, if the cause can be found in the passenger log file errors. needs #1621
1.0
Physical File Processing - Processor Stopping unexpectedly with very little Error info - The CSV processor stopped processing files for approx 12 hours or so, starting between 10am & 11am on Sept 28th. I have downloaded a copy of the passenger.log file from that day, if required. The issue has resolved itself, but there was very little feedback from the logs, and nothing shown in errbit, to notify of what the issue was, although we suspect it was caused by a problem file that was uploaded. **Action**: Determine the root cause of the issue to prevent re-occurrence, if the cause can be found in the passenger log file errors. needs #1621
non_process
physical file processing processor stopping unexpectedly with very little error info the csv processor stopped processing files for approx hours or so starting between on sept i have downloaded a copy of the passenger log file from that day if required the issue has resolved itself but there was very little feedback from the logs and nothing shown in errbit to notify of what the issue was although we suspect it was caused by a problem file that was uploaded action determine the root cause of the issue to prevent re occurrence if the cause can be found in the passenger log file errors needs
0
21,403
29,326,579,325
IssuesEvent
2023-05-26 00:05:54
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] [Bug] Metadata calculation not working correctly for aggregation that contains a `:metric` clause
Type:Bug .Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
Failing test in `metabase.query-processor-test.case-test` ```clj Can we use case in metric with temporary :metabase.models.metric/Metric with attributes {:creator_id 2, :definition {:source-table 32407, :aggregation [:sum [:case [[[:< [:field 133191 nil] 4] [:field 133191 nil]]]]]}, :description "Lookin' for a blueberry", :name "Toucans in the rainforest", :table_id 32407} legacy query = {:database 2368, :type :query, :query {:aggregation [[:metric 316]], :source-table 32407}} pMBQL = {:lib/type :mbql/query, :database 2368, :type :pipeline, :stages [{:lib/type :mbql.stage/mbql, :source-table 32407, :aggregation [[:metric {:lib/uuid "6a455487-6033-4f35-b159-91558080ac1b"} 316]]}]} Failed to calculated metadata for query {:via [{:type clojure.lang.ExceptionInfo, :message "[\"Not a valid base type: :metabase.lib.schema.expression/type.unknown, received: :metabase.lib.schema.expression/type.unknown\"]", :data {:type :malli.core/invalid-output, :data {:output :metabase.lib.schema.common/base-type, :value :metabase.lib.schema.expression/type.unknown, :args [{:lib/type :mbql/query, :lib/metadata (metabase.lib.metadata.cached-provider/cached-metadata-provider (metabase.lib.metadata.jvm/->UncachedApplicationDatabaseMetadataProvider 2368)), :database 2368, :type :pipeline, :stages [{:lib/type :mbql.stage/mbql, :source-table 32407, :aggregation [[:metric #:lib{:uuid "06174580-70bd-474a-b167-95a9ef3899cc"} 318]]}]} -1 [:case #:lib{:uuid "01480a56-78ee-4b7f-9407-23a6493cdd7d"} [[[:< #:lib{:uuid "e078b90f-de46-4dbb-9903-930c3891c375"} [:field #:lib{:uuid "3454c908-194a-447d-af44-cbedee1ba44d"} 133191] 4] [:field #:lib{:uuid "4d3ae622-03b7-40f6-af20-1ae413065640"} 133191]]]]], :schema [:=> [:cat :metabase.lib.schema/query :int :any] :metabase.lib.schema.common/base-type], :fn-name lib.metadata.calculation/type-of}, :humanized ["Not a valid base type: :metabase.lib.schema.expression/type.unknown, received: :metabase.lib.schema.expression/type.unknown"], :link "https://malli.io?schema=%3Ametabase.lib.schema.common%2Fbase-type%0A&value=%3Ametabase.lib.schema.expression%2Ftype.unknown%0A"}, :at [metabase.util.malli$explain_fn_fail_BANG_ invokeStatic "malli.cljc" 58]}], :trace [[metabase.util.malli$explain_fn_fail_BANG_ invokeStatic "malli.cljc" 58] [metabase.util.malli$explain_fn_fail_BANG_ invoke "malli.cljc" 50] [malli.instrument$_strument_BANG_$fn__30546$fn__30547 invoke "instrument.clj" 23] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2529] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.metadata.calculation$eval32823$fn__32825 invoke "calculation.cljc" 155] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32809$type_of__32810 invoke "calculation.cljc" 141] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.metric$eval75577$fn__75578 invoke "metric.cljc" 34] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32809$type_of__32810 invoke "calculation.cljc" 141] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.metric$eval75586$fn__75588 invoke "metric.cljc" 40] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32809$type_of__32810 invoke "calculation.cljc" 141] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.metadata.calculation$eval32843$fn__32844 invoke "calculation.cljc" 173] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32848$metadata__32849 invoke "calculation.cljc" 197] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.aggregation$eval73541$aggregations__73542$fn__73543 invoke "aggregation.cljc" 217] [clojure.core$map_indexed$mapi__8638$fn__8639 invoke "core.clj" 7376] [clojure.lang.LazySeq sval "LazySeq.java" 42] [clojure.lang.LazySeq seq "LazySeq.java" 51] [clojure.lang.RT seq "RT.java" 535] [clojure.core$seq__5467 invokeStatic "core.clj" 139] [clojure.core.protocols$seq_reduce invokeStatic "protocols.clj" 24] [clojure.core.protocols$fn__8236 invokeStatic "protocols.clj" 75] [clojure.core.protocols$fn__8236 invoke "protocols.clj" 75] [clojure.core.protocols$fn__8178$G__8173__8191 invoke "protocols.clj" 13] [clojure.core$reduce invokeStatic "core.clj" 6886] [clojure.core$reduce invoke "core.clj" 6868] [malli.core$_collection_schema$reify$reify__28325$fn__28326 invoke "core.cljc" 1194] [malli.core$_maybe_schema$reify$reify__28467$fn__28468 invoke "core.cljc" 1466] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2528] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 421] [metabase.lib.stage$eval75811$aggregations_columns__75812 invoke "stage.cljc" 107] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.stage$eval75861$breakout_ags_fields_columns__75862$fn__75863 invoke "stage.cljc" 139] [clojure.core$map$fn__5931$fn__5932 invoke "core.clj" 2759] [clojure.lang.PersistentVector reduce "PersistentVector.java" 343] [clojure.core$transduce invokeStatic "core.clj" 6946] [clojure.core$into invokeStatic "core.clj" 6962] [clojure.core$into invoke "core.clj" 6950] [metabase.lib.stage$eval75861$breakout_ags_fields_columns__75862 invoke "stage.cljc" 137] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.stage$eval76011$stage_metadata__76012 invoke "stage.cljc" 315] [clojure.lang.AFn applyToHelper "AFn.java" 156] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn invoke "RestFn.java" 421] [metabase.lib.stage$eval76022$fn__76023 invoke "stage.cljc" 322] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32848$metadata__32849 invoke "calculation.cljc" 197] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.query$eval33071$fn__33072 invoke "query.cljc" 66] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32848$metadata__32849 invoke "calculation.cljc" 197] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.metadata.calculation$eval32848$metadata__32849 invoke "calculation.cljc" 191] [clojure.lang.AFn applyToHelper "AFn.java" 154] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 408] [metabase.query_processor_test.test_mlv2$test_mlv2_metadata$fn__142634$fn__142635 invoke "test_mlv2.clj" 97] [metabase.query_processor_test.test_mlv2$do_with_pMBQL_query_testing_context invokeStatic "test_mlv2.clj" 24] [metabase.query_processor_test.test_mlv2$do_with_pMBQL_query_testing_context invoke "test_mlv2.clj" 22] [metabase.query_processor_test.test_mlv2$test_mlv2_metadata$fn__142634 invoke "test_mlv2.clj" 91] [metabase.query_processor_test.test_mlv2$do_with_legacy_query_testing_context invokeStatic "test_mlv2.clj" 20] [metabase.query_processor_test.test_mlv2$do_with_legacy_query_testing_context invoke "test_mlv2.clj" 18] [metabase.query_processor_test.test_mlv2$test_mlv2_metadata invokeStatic "test_mlv2.clj" 83] [metabase.query_processor_test.test_mlv2$test_mlv2_metadata invoke "test_mlv2.clj" 80] [metabase.query_processor_test.test_mlv2$post_processing_middleware$fn__142666 invoke "test_mlv2.clj" 127] [metabase.query_processor.middleware.annotate$add_column_info$add_column_info_rff_STAR___88584 invoke "annotate.clj" 711] [metabase.query_processor.middleware.mbql_to_native$mbql__GT_native$fn__98111$fn__98112 invoke "mbql_to_native.clj" 27] [metabase_enterprise.advanced_permissions.query_processor.middleware.permissions$check_download_permissions$fn__100598$fn__100599 invoke "permissions.clj" 126] [metabase.query_processor.context.default$default_reducef invokeStatic "default.clj" 55] [metabase.query_processor.context.default$default_reducef invoke "default.clj" 49] [metabase.query_processor.context$reducef invokeStatic "context.clj" 70] [metabase.query_processor.context$reducef invoke "context.clj" 63] [metabase.query_processor.context.default$default_runf$respond_STAR___72458 invoke "default.clj" 69] [metabase.driver.sql_jdbc.execute$execute_reducible_query invokeStatic "execute.clj" 513] [metabase.driver.sql_jdbc.execute$execute_reducible_query invoke "execute.clj" 489] [metabase.driver.sql_jdbc.execute$execute_reducible_query invokeStatic "execute.clj" 497] [metabase.driver.sql_jdbc.execute$execute_reducible_query invoke "execute.clj" 489] [metabase.driver.sql_jdbc$eval127004$fn__127005 invoke "sql_jdbc.clj" 63] [metabase.driver.h2$eval127896$fn__127897 invoke "h2.clj" 221] [clojure.lang.MultiFn invoke "MultiFn.java" 244] [metabase.query_processor.context$executef invokeStatic "context.clj" 60] [metabase.query_processor.context$executef invoke "context.clj" 49] [metabase.query_processor.context.default$default_runf invokeStatic "default.clj" 68] [metabase.query_processor.context.default$default_runf invoke "default.clj" 66] [metabase.query_processor.context$runf invokeStatic "context.clj" 46] [metabase.query_processor.context$runf invoke "context.clj" 40] [metabase.query_processor.reducible$identity_qp invokeStatic "reducible.clj" 15] [metabase.query_processor.reducible$identity_qp invoke "reducible.clj" 12] [metabase.query_processor.middleware.cache$maybe_return_cached_results$maybe_return_cached_results_STAR___96299 invoke "cache.clj" 224] [metabase.query_processor.middleware.permissions$check_query_permissions$fn__96639 invoke "permissions.clj" 126] [metabase_enterprise.advanced_permissions.query_processor.middleware.permissions$check_download_permissions$fn__100598 invoke "permissions.clj" 125] [metabase_enterprise.sandbox.query_processor.middleware.column_level_perms_check$maybe_apply_column_level_perms_check$fn__100845 invoke "column_level_perms_check.clj" 31] [metabase.query_processor.middleware.mbql_to_native$mbql__GT_native$fn__98111 invoke "mbql_to_native.clj" 24] [metabase.query_processor$fn__101672$combined_post_process__101677$combined_post_process_STAR___101678 invoke "query_processor.clj" 254] [metabase.query_processor$fn__101672$combined_pre_process__101673$combined_pre_process_STAR___101674 invoke "query_processor.clj" 251] [metabase.query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__99381$fn__99386 invoke "resolve_database_and_driver.clj" 36] [metabase.driver$do_with_driver invokeStatic "driver.clj" 90] [metabase.driver$do_with_driver invoke "driver.clj" 86] [metabase.query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__99381 invoke "resolve_database_and_driver.clj" 35] [metabase.query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__97914 invoke "fetch_source_query.clj" 310] [metabase.query_processor.middleware.store$initialize_store$fn__100367$fn__100368 invoke "store.clj" 12] [metabase.query_processor.store$do_with_store invokeStatic "store.clj" 46] [metabase.query_processor.store$do_with_store invoke "store.clj" 40] [metabase.query_processor.middleware.store$initialize_store$fn__100367 invoke "store.clj" 11] [metabase.query_processor_test.test_mlv2$around_middleware$fn__142669 invoke "test_mlv2.clj" 136] [metabase.query_processor.middleware.normalize_query$normalize$fn__98141 invoke "normalize_query.clj" 36] [metabase_enterprise.audit_app.query_processor.middleware.handle_audit_queries$handle_internal_queries$fn__100811 invoke "handle_audit_queries.clj" 131] [metabase.query_processor.reducible$async_qp$qp_STAR___72537$thunk__72539 invoke "reducible.clj" 103] [metabase.query_processor.reducible$async_qp$qp_STAR___72537 invoke "reducible.clj" 109] [metabase.query_processor.reducible$async_qp$qp_STAR___72537 invoke "reducible.clj" 94] [metabase.query_processor.reducible$async_qp$qp_STAR___72537 invoke "reducible.clj" 91] [clojure.lang.AFn applyToHelper "AFn.java" 154] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [metabase.query_processor$base_qp$fn__101684 doInvoke "query_processor.clj" 272] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [metabase.query_processor.reducible$sync_qp$qp_STAR___72549 doInvoke "reducible.clj" 129] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 669] [clojure.core$apply invoke "core.clj" 662] [metabase.query_processor$process_query invokeStatic "query_processor.clj" 289] [metabase.query_processor$process_query doInvoke "query_processor.clj" 283] [clojure.lang.RestFn invoke "RestFn.java" 410] [metabase.test.data$run_mbql_query_STAR_ invokeStatic "data.clj" 190] [metabase.test.data$run_mbql_query_STAR_ invoke "data.clj" 187] [metabase.query_processor_test.case_test$test_case invokeStatic "case_test.clj" 9] [metabase.query_processor_test.case_test$test_case invoke "case_test.clj" 7] [metabase.query_processor_test.case_test$fn__143175$fn__143192$fn__143193 invoke "case_test.clj" 51] [toucan2.tools.with_temp$do_with_temp_STAR__primary_method_default invokeStatic "with_temp.clj" 62] [toucan2.tools.with_temp$do_with_temp_STAR__primary_method_default invoke "with_temp.clj" 44] [clojure.lang.AFn applyToHelper "AFn.java" 165] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.lang.AFunction$1 doInvoke "AFunction.java" 31] [clojure.lang.RestFn invoke "RestFn.java" 457] [clojure.core$partial$fn__5908 invoke "core.clj" 2643] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.RestFn applyTo "RestFn.java" 132] [clojure.lang.AFunction$1 doInvoke "AFunction.java" 31] [clojure.lang.RestFn invoke "RestFn.java" 436] [methodical.impl.combo.threaded$eval2207$fn__2208$fn__2209$fn__2212 invoke "threaded.clj" 71] [methodical.impl.combo.threaded$reducer_fn$fn__2175$fn__2179 invoke "threaded.clj" 23] [clojure.lang.ArrayChunk reduce "ArrayChunk.java" 58] [clojure.core.protocols$fn__8244 invokeStatic "protocols.clj" 136] [clojure.core.protocols$fn__8244 invoke "protocols.clj" 124] [clojure.core.protocols$fn__8204$G__8199__8213 invoke "protocols.clj" 19] [clojure.core.protocols$seq_reduce invokeStatic "protocols.clj" 31] [clojure.core.protocols$fn__8236 invokeStatic "protocols.clj" 75] [clojure.core.protocols$fn__8236 invoke "protocols.clj" 75] [clojure.core.protocols$fn__8178$G__8173__8191 invoke "protocols.clj" 13] [clojure.core$reduce invokeStatic "core.clj" 6886] [clojure.core$reduce invoke "core.clj" 6868] [methodical.impl.combo.threaded$reducer_fn$fn__2175 invoke "threaded.clj" 21] [clojure.core$comp$fn__5876 invoke "core.clj" 2588] [methodical.impl.combo.threaded$combine_with_threader$fn__2185 invoke "threaded.clj" 44] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.RestFn applyTo "RestFn.java" 132] [clojure.lang.AFunction$1 doInvoke "AFunction.java" 31] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.test.redefs$do_with_temp_STAR__around_method_default invokeStatic "redefs.clj" 23] [metabase.test.redefs$do_with_temp_STAR__around_method_default invoke "redefs.clj" 10] [clojure.lang.AFn applyToHelper "AFn.java" 165] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.lang.AFunction$1 doInvoke "AFunction.java" 31] [clojure.lang.RestFn invoke "RestFn.java" 457] [clojure.core$partial$fn__5908 invoke "core.clj" 2643] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.RestFn applyTo "RestFn.java" 132] [clojure.lang.AFunction$1 doInvoke "AFunction.java" 31] [clojure.lang.RestFn invoke "RestFn.java" 436] [methodical.impl.standard$invoke_multifn invokeStatic "standard.clj" 58] [methodical.impl.standard$invoke_multifn invoke "standard.clj" 47] [methodical.impl.standard.StandardMultiFn invoke "standard.clj" 195] [toucan2.tools.with_temp$do_with_temp invokeStatic "with_temp.clj" 68] [toucan2.tools.with_temp$do_with_temp invoke "with_temp.clj" 66] [metabase.query_processor_test.case_test$fn__143175$fn__143192 invoke "case_test.clj" 46] [metabase.test.data.datasets$do_with_driver_when_testing$fn__110589$fn__110590 invoke "datasets.clj" 42] [metabase.driver$do_with_driver invokeStatic "driver.clj" 90] [metabase.driver$do_with_driver invoke "driver.clj" 86] [metabase.test.data.datasets$do_with_driver_when_testing$fn__110589 invoke "datasets.clj" 41] [metabase.test.data.datasets$do_when_testing_driver invokeStatic "datasets.clj" 30] [metabase.test.data.datasets$do_when_testing_driver invoke "datasets.clj" 22] [metabase.test.data.datasets$do_with_driver_when_testing invokeStatic "datasets.clj" 40] [metabase.test.data.datasets$do_with_driver_when_testing invoke "datasets.clj" 39] [metabase.query_processor_test.case_test$fn__143175 invokeStatic "case_test.clj" 44] [metabase.query_processor_test.case_test$fn__143175 invoke "case_test.clj" 43] [cider.nrepl.middleware.test$test_var$fn__134683 invoke "test.clj" 244] [cider.nrepl.middleware.test$test_var invokeStatic "test.clj" 244] [cider.nrepl.middleware.test$test_var invoke "test.clj" 236] [cider.nrepl.middleware.test$test_vars$fn__134687$fn__134692 invoke "test.clj" 259] [clojure.test$default_fixture invokeStatic "test.clj" 687] [clojure.test$default_fixture invoke "test.clj" 683] [cider.nrepl.middleware.test$test_vars$fn__134687 invoke "test.clj" 259] [clojure.test$default_fixture invokeStatic "test.clj" 687] [clojure.test$default_fixture invoke "test.clj" 683] [cider.nrepl.middleware.test$test_vars invokeStatic "test.clj" 256] [cider.nrepl.middleware.test$test_vars invoke "test.clj" 250] [cider.nrepl.middleware.test$test_ns invokeStatic "test.clj" 272] [cider.nrepl.middleware.test$test_ns invoke "test.clj" 263] [cider.nrepl.middleware.test$test_var_query invokeStatic "test.clj" 283] [cider.nrepl.middleware.test$test_var_query invoke "test.clj" 276] [cider.nrepl.middleware.test$handle_test_var_query_op$fn__134731$fn__134732 invoke "test.clj" 321] [clojure.lang.AFn applyToHelper "AFn.java" 152] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$with_bindings_STAR_ invokeStatic "core.clj" 1990] [clojure.core$with_bindings_STAR_ doInvoke "core.clj" 1990] [clojure.lang.RestFn invoke "RestFn.java" 425] [cider.nrepl.middleware.test$handle_test_var_query_op$fn__134731 invoke "test.clj" 313] [clojure.lang.AFn run "AFn.java" 22] [nrepl.middleware.session$session_exec$main_loop__9666$fn__9670 invoke "session.clj" 218] [nrepl.middleware.session$session_exec$main_loop__9666 invoke "session.clj" 217] [clojure.lang.AFn run "AFn.java" 22] [java.lang.Thread run "Thread.java" 1589]], :cause "[\"Not a valid base type: :metabase.lib.schema.expression/type.unknown, received: :metabase.lib.schema.expression/type.unknown\"]", :data {:type :malli.core/invalid-output, :data {:output :metabase.lib.schema.common/base-type, :value :metabase.lib.schema.expression/type.unknown, :args [{:lib/type :mbql/query, :lib/metadata (metabase.lib.metadata.cached-provider/cached-metadata-provider (metabase.lib.metadata.jvm/->UncachedApplicationDatabaseMetadataProvider 2368)), :database 2368, :type :pipeline, :stages [{:lib/type :mbql.stage/mbql, :source-table 32407, :aggregation [[:metric #:lib{:uuid "06174580-70bd-474a-b167-95a9ef3899cc"} 318]]}]} -1 [:case #:lib{:uuid "01480a56-78ee-4b7f-9407-23a6493cdd7d"} [[[:< #:lib{:uuid "e078b90f-de46-4dbb-9903-930c3891c375"} [:field #:lib{:uuid "3454c908-194a-447d-af44-cbedee1ba44d"} 133191] 4] [:field #:lib{:uuid "4d3ae622-03b7-40f6-af20-1ae413065640"} 133191]]]]], :schema [:=> [:cat :metabase.lib.schema/query :int :any] :metabase.lib.schema.common/base-type], :fn-name lib.metadata.calculation/type-of}, :humanized ["Not a valid base type: :metabase.lib.schema.expression/type.unknown, received: :metabase.lib.schema.expression/type.unknown"], :link "https://malli.io?schema=%3Ametabase.lib.schema.common%2Fbase-type%0A&value=%3Ametabase.lib.schema.expression%2Ftype.unknown%0A"}} ```
1.0
[MLv2] [Bug] Metadata calculation not working correctly for aggregation that contains a `:metric` clause - Failing test in `metabase.query-processor-test.case-test` ```clj Can we use case in metric with temporary :metabase.models.metric/Metric with attributes {:creator_id 2, :definition {:source-table 32407, :aggregation [:sum [:case [[[:< [:field 133191 nil] 4] [:field 133191 nil]]]]]}, :description "Lookin' for a blueberry", :name "Toucans in the rainforest", :table_id 32407} legacy query = {:database 2368, :type :query, :query {:aggregation [[:metric 316]], :source-table 32407}} pMBQL = {:lib/type :mbql/query, :database 2368, :type :pipeline, :stages [{:lib/type :mbql.stage/mbql, :source-table 32407, :aggregation [[:metric {:lib/uuid "6a455487-6033-4f35-b159-91558080ac1b"} 316]]}]} Failed to calculated metadata for query {:via [{:type clojure.lang.ExceptionInfo, :message "[\"Not a valid base type: :metabase.lib.schema.expression/type.unknown, received: :metabase.lib.schema.expression/type.unknown\"]", :data {:type :malli.core/invalid-output, :data {:output :metabase.lib.schema.common/base-type, :value :metabase.lib.schema.expression/type.unknown, :args [{:lib/type :mbql/query, :lib/metadata (metabase.lib.metadata.cached-provider/cached-metadata-provider (metabase.lib.metadata.jvm/->UncachedApplicationDatabaseMetadataProvider 2368)), :database 2368, :type :pipeline, :stages [{:lib/type :mbql.stage/mbql, :source-table 32407, :aggregation [[:metric #:lib{:uuid "06174580-70bd-474a-b167-95a9ef3899cc"} 318]]}]} -1 [:case #:lib{:uuid "01480a56-78ee-4b7f-9407-23a6493cdd7d"} [[[:< #:lib{:uuid "e078b90f-de46-4dbb-9903-930c3891c375"} [:field #:lib{:uuid "3454c908-194a-447d-af44-cbedee1ba44d"} 133191] 4] [:field #:lib{:uuid "4d3ae622-03b7-40f6-af20-1ae413065640"} 133191]]]]], :schema [:=> [:cat :metabase.lib.schema/query :int :any] :metabase.lib.schema.common/base-type], :fn-name lib.metadata.calculation/type-of}, :humanized ["Not a valid base type: :metabase.lib.schema.expression/type.unknown, received: :metabase.lib.schema.expression/type.unknown"], :link "https://malli.io?schema=%3Ametabase.lib.schema.common%2Fbase-type%0A&value=%3Ametabase.lib.schema.expression%2Ftype.unknown%0A"}, :at [metabase.util.malli$explain_fn_fail_BANG_ invokeStatic "malli.cljc" 58]}], :trace [[metabase.util.malli$explain_fn_fail_BANG_ invokeStatic "malli.cljc" 58] [metabase.util.malli$explain_fn_fail_BANG_ invoke "malli.cljc" 50] [malli.instrument$_strument_BANG_$fn__30546$fn__30547 invoke "instrument.clj" 23] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2529] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.metadata.calculation$eval32823$fn__32825 invoke "calculation.cljc" 155] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32809$type_of__32810 invoke "calculation.cljc" 141] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.metric$eval75577$fn__75578 invoke "metric.cljc" 34] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32809$type_of__32810 invoke "calculation.cljc" 141] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.metric$eval75586$fn__75588 invoke "metric.cljc" 40] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32809$type_of__32810 invoke "calculation.cljc" 141] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.metadata.calculation$eval32843$fn__32844 invoke "calculation.cljc" 173] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32848$metadata__32849 invoke "calculation.cljc" 197] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.aggregation$eval73541$aggregations__73542$fn__73543 invoke "aggregation.cljc" 217] [clojure.core$map_indexed$mapi__8638$fn__8639 invoke "core.clj" 7376] [clojure.lang.LazySeq sval "LazySeq.java" 42] [clojure.lang.LazySeq seq "LazySeq.java" 51] [clojure.lang.RT seq "RT.java" 535] [clojure.core$seq__5467 invokeStatic "core.clj" 139] [clojure.core.protocols$seq_reduce invokeStatic "protocols.clj" 24] [clojure.core.protocols$fn__8236 invokeStatic "protocols.clj" 75] [clojure.core.protocols$fn__8236 invoke "protocols.clj" 75] [clojure.core.protocols$fn__8178$G__8173__8191 invoke "protocols.clj" 13] [clojure.core$reduce invokeStatic "core.clj" 6886] [clojure.core$reduce invoke "core.clj" 6868] [malli.core$_collection_schema$reify$reify__28325$fn__28326 invoke "core.cljc" 1194] [malli.core$_maybe_schema$reify$reify__28467$fn__28468 invoke "core.cljc" 1466] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2528] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 421] [metabase.lib.stage$eval75811$aggregations_columns__75812 invoke "stage.cljc" 107] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.stage$eval75861$breakout_ags_fields_columns__75862$fn__75863 invoke "stage.cljc" 139] [clojure.core$map$fn__5931$fn__5932 invoke "core.clj" 2759] [clojure.lang.PersistentVector reduce "PersistentVector.java" 343] [clojure.core$transduce invokeStatic "core.clj" 6946] [clojure.core$into invokeStatic "core.clj" 6962] [clojure.core$into invoke "core.clj" 6950] [metabase.lib.stage$eval75861$breakout_ags_fields_columns__75862 invoke "stage.cljc" 137] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.stage$eval76011$stage_metadata__76012 invoke "stage.cljc" 315] [clojure.lang.AFn applyToHelper "AFn.java" 156] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn invoke "RestFn.java" 421] [metabase.lib.stage$eval76022$fn__76023 invoke "stage.cljc" 322] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32848$metadata__32849 invoke "calculation.cljc" 197] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.query$eval33071$fn__33072 invoke "query.cljc" 66] [clojure.lang.MultiFn invoke "MultiFn.java" 239] [metabase.lib.metadata.calculation$eval32848$metadata__32849 invoke "calculation.cljc" 197] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.lib.metadata.calculation$eval32848$metadata__32849 invoke "calculation.cljc" 191] [clojure.lang.AFn applyToHelper "AFn.java" 154] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29103 doInvoke "core.cljc" 2526] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [malli.core$_instrument$fn__29091 doInvoke "core.cljc" 2543] [clojure.lang.RestFn invoke "RestFn.java" 408] [metabase.query_processor_test.test_mlv2$test_mlv2_metadata$fn__142634$fn__142635 invoke "test_mlv2.clj" 97] [metabase.query_processor_test.test_mlv2$do_with_pMBQL_query_testing_context invokeStatic "test_mlv2.clj" 24] [metabase.query_processor_test.test_mlv2$do_with_pMBQL_query_testing_context invoke "test_mlv2.clj" 22] [metabase.query_processor_test.test_mlv2$test_mlv2_metadata$fn__142634 invoke "test_mlv2.clj" 91] [metabase.query_processor_test.test_mlv2$do_with_legacy_query_testing_context invokeStatic "test_mlv2.clj" 20] [metabase.query_processor_test.test_mlv2$do_with_legacy_query_testing_context invoke "test_mlv2.clj" 18] [metabase.query_processor_test.test_mlv2$test_mlv2_metadata invokeStatic "test_mlv2.clj" 83] [metabase.query_processor_test.test_mlv2$test_mlv2_metadata invoke "test_mlv2.clj" 80] [metabase.query_processor_test.test_mlv2$post_processing_middleware$fn__142666 invoke "test_mlv2.clj" 127] [metabase.query_processor.middleware.annotate$add_column_info$add_column_info_rff_STAR___88584 invoke "annotate.clj" 711] [metabase.query_processor.middleware.mbql_to_native$mbql__GT_native$fn__98111$fn__98112 invoke "mbql_to_native.clj" 27] [metabase_enterprise.advanced_permissions.query_processor.middleware.permissions$check_download_permissions$fn__100598$fn__100599 invoke "permissions.clj" 126] [metabase.query_processor.context.default$default_reducef invokeStatic "default.clj" 55] [metabase.query_processor.context.default$default_reducef invoke "default.clj" 49] [metabase.query_processor.context$reducef invokeStatic "context.clj" 70] [metabase.query_processor.context$reducef invoke "context.clj" 63] [metabase.query_processor.context.default$default_runf$respond_STAR___72458 invoke "default.clj" 69] [metabase.driver.sql_jdbc.execute$execute_reducible_query invokeStatic "execute.clj" 513] [metabase.driver.sql_jdbc.execute$execute_reducible_query invoke "execute.clj" 489] [metabase.driver.sql_jdbc.execute$execute_reducible_query invokeStatic "execute.clj" 497] [metabase.driver.sql_jdbc.execute$execute_reducible_query invoke "execute.clj" 489] [metabase.driver.sql_jdbc$eval127004$fn__127005 invoke "sql_jdbc.clj" 63] [metabase.driver.h2$eval127896$fn__127897 invoke "h2.clj" 221] [clojure.lang.MultiFn invoke "MultiFn.java" 244] [metabase.query_processor.context$executef invokeStatic "context.clj" 60] [metabase.query_processor.context$executef invoke "context.clj" 49] [metabase.query_processor.context.default$default_runf invokeStatic "default.clj" 68] [metabase.query_processor.context.default$default_runf invoke "default.clj" 66] [metabase.query_processor.context$runf invokeStatic "context.clj" 46] [metabase.query_processor.context$runf invoke "context.clj" 40] [metabase.query_processor.reducible$identity_qp invokeStatic "reducible.clj" 15] [metabase.query_processor.reducible$identity_qp invoke "reducible.clj" 12] [metabase.query_processor.middleware.cache$maybe_return_cached_results$maybe_return_cached_results_STAR___96299 invoke "cache.clj" 224] [metabase.query_processor.middleware.permissions$check_query_permissions$fn__96639 invoke "permissions.clj" 126] [metabase_enterprise.advanced_permissions.query_processor.middleware.permissions$check_download_permissions$fn__100598 invoke "permissions.clj" 125] [metabase_enterprise.sandbox.query_processor.middleware.column_level_perms_check$maybe_apply_column_level_perms_check$fn__100845 invoke "column_level_perms_check.clj" 31] [metabase.query_processor.middleware.mbql_to_native$mbql__GT_native$fn__98111 invoke "mbql_to_native.clj" 24] [metabase.query_processor$fn__101672$combined_post_process__101677$combined_post_process_STAR___101678 invoke "query_processor.clj" 254] [metabase.query_processor$fn__101672$combined_pre_process__101673$combined_pre_process_STAR___101674 invoke "query_processor.clj" 251] [metabase.query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__99381$fn__99386 invoke "resolve_database_and_driver.clj" 36] [metabase.driver$do_with_driver invokeStatic "driver.clj" 90] [metabase.driver$do_with_driver invoke "driver.clj" 86] [metabase.query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__99381 invoke "resolve_database_and_driver.clj" 35] [metabase.query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__97914 invoke "fetch_source_query.clj" 310] [metabase.query_processor.middleware.store$initialize_store$fn__100367$fn__100368 invoke "store.clj" 12] [metabase.query_processor.store$do_with_store invokeStatic "store.clj" 46] [metabase.query_processor.store$do_with_store invoke "store.clj" 40] [metabase.query_processor.middleware.store$initialize_store$fn__100367 invoke "store.clj" 11] [metabase.query_processor_test.test_mlv2$around_middleware$fn__142669 invoke "test_mlv2.clj" 136] [metabase.query_processor.middleware.normalize_query$normalize$fn__98141 invoke "normalize_query.clj" 36] [metabase_enterprise.audit_app.query_processor.middleware.handle_audit_queries$handle_internal_queries$fn__100811 invoke "handle_audit_queries.clj" 131] [metabase.query_processor.reducible$async_qp$qp_STAR___72537$thunk__72539 invoke "reducible.clj" 103] [metabase.query_processor.reducible$async_qp$qp_STAR___72537 invoke "reducible.clj" 109] [metabase.query_processor.reducible$async_qp$qp_STAR___72537 invoke "reducible.clj" 94] [metabase.query_processor.reducible$async_qp$qp_STAR___72537 invoke "reducible.clj" 91] [clojure.lang.AFn applyToHelper "AFn.java" 154] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [metabase.query_processor$base_qp$fn__101684 doInvoke "query_processor.clj" 272] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$apply invoke "core.clj" 662] [metabase.query_processor.reducible$sync_qp$qp_STAR___72549 doInvoke "reducible.clj" 129] [clojure.lang.RestFn applyTo "RestFn.java" 137] [clojure.core$apply invokeStatic "core.clj" 669] [clojure.core$apply invoke "core.clj" 662] [metabase.query_processor$process_query invokeStatic "query_processor.clj" 289] [metabase.query_processor$process_query doInvoke "query_processor.clj" 283] [clojure.lang.RestFn invoke "RestFn.java" 410] [metabase.test.data$run_mbql_query_STAR_ invokeStatic "data.clj" 190] [metabase.test.data$run_mbql_query_STAR_ invoke "data.clj" 187] [metabase.query_processor_test.case_test$test_case invokeStatic "case_test.clj" 9] [metabase.query_processor_test.case_test$test_case invoke "case_test.clj" 7] [metabase.query_processor_test.case_test$fn__143175$fn__143192$fn__143193 invoke "case_test.clj" 51] [toucan2.tools.with_temp$do_with_temp_STAR__primary_method_default invokeStatic "with_temp.clj" 62] [toucan2.tools.with_temp$do_with_temp_STAR__primary_method_default invoke "with_temp.clj" 44] [clojure.lang.AFn applyToHelper "AFn.java" 165] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.lang.AFunction$1 doInvoke "AFunction.java" 31] [clojure.lang.RestFn invoke "RestFn.java" 457] [clojure.core$partial$fn__5908 invoke "core.clj" 2643] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.RestFn applyTo "RestFn.java" 132] [clojure.lang.AFunction$1 doInvoke "AFunction.java" 31] [clojure.lang.RestFn invoke "RestFn.java" 436] [methodical.impl.combo.threaded$eval2207$fn__2208$fn__2209$fn__2212 invoke "threaded.clj" 71] [methodical.impl.combo.threaded$reducer_fn$fn__2175$fn__2179 invoke "threaded.clj" 23] [clojure.lang.ArrayChunk reduce "ArrayChunk.java" 58] [clojure.core.protocols$fn__8244 invokeStatic "protocols.clj" 136] [clojure.core.protocols$fn__8244 invoke "protocols.clj" 124] [clojure.core.protocols$fn__8204$G__8199__8213 invoke "protocols.clj" 19] [clojure.core.protocols$seq_reduce invokeStatic "protocols.clj" 31] [clojure.core.protocols$fn__8236 invokeStatic "protocols.clj" 75] [clojure.core.protocols$fn__8236 invoke "protocols.clj" 75] [clojure.core.protocols$fn__8178$G__8173__8191 invoke "protocols.clj" 13] [clojure.core$reduce invokeStatic "core.clj" 6886] [clojure.core$reduce invoke "core.clj" 6868] [methodical.impl.combo.threaded$reducer_fn$fn__2175 invoke "threaded.clj" 21] [clojure.core$comp$fn__5876 invoke "core.clj" 2588] [methodical.impl.combo.threaded$combine_with_threader$fn__2185 invoke "threaded.clj" 44] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.RestFn applyTo "RestFn.java" 132] [clojure.lang.AFunction$1 doInvoke "AFunction.java" 31] [clojure.lang.RestFn invoke "RestFn.java" 436] [metabase.test.redefs$do_with_temp_STAR__around_method_default invokeStatic "redefs.clj" 23] [metabase.test.redefs$do_with_temp_STAR__around_method_default invoke "redefs.clj" 10] [clojure.lang.AFn applyToHelper "AFn.java" 165] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.lang.AFunction$1 doInvoke "AFunction.java" 31] [clojure.lang.RestFn invoke "RestFn.java" 457] [clojure.core$partial$fn__5908 invoke "core.clj" 2643] [clojure.lang.AFn applyToHelper "AFn.java" 160] [clojure.lang.RestFn applyTo "RestFn.java" 132] [clojure.lang.AFunction$1 doInvoke "AFunction.java" 31] [clojure.lang.RestFn invoke "RestFn.java" 436] [methodical.impl.standard$invoke_multifn invokeStatic "standard.clj" 58] [methodical.impl.standard$invoke_multifn invoke "standard.clj" 47] [methodical.impl.standard.StandardMultiFn invoke "standard.clj" 195] [toucan2.tools.with_temp$do_with_temp invokeStatic "with_temp.clj" 68] [toucan2.tools.with_temp$do_with_temp invoke "with_temp.clj" 66] [metabase.query_processor_test.case_test$fn__143175$fn__143192 invoke "case_test.clj" 46] [metabase.test.data.datasets$do_with_driver_when_testing$fn__110589$fn__110590 invoke "datasets.clj" 42] [metabase.driver$do_with_driver invokeStatic "driver.clj" 90] [metabase.driver$do_with_driver invoke "driver.clj" 86] [metabase.test.data.datasets$do_with_driver_when_testing$fn__110589 invoke "datasets.clj" 41] [metabase.test.data.datasets$do_when_testing_driver invokeStatic "datasets.clj" 30] [metabase.test.data.datasets$do_when_testing_driver invoke "datasets.clj" 22] [metabase.test.data.datasets$do_with_driver_when_testing invokeStatic "datasets.clj" 40] [metabase.test.data.datasets$do_with_driver_when_testing invoke "datasets.clj" 39] [metabase.query_processor_test.case_test$fn__143175 invokeStatic "case_test.clj" 44] [metabase.query_processor_test.case_test$fn__143175 invoke "case_test.clj" 43] [cider.nrepl.middleware.test$test_var$fn__134683 invoke "test.clj" 244] [cider.nrepl.middleware.test$test_var invokeStatic "test.clj" 244] [cider.nrepl.middleware.test$test_var invoke "test.clj" 236] [cider.nrepl.middleware.test$test_vars$fn__134687$fn__134692 invoke "test.clj" 259] [clojure.test$default_fixture invokeStatic "test.clj" 687] [clojure.test$default_fixture invoke "test.clj" 683] [cider.nrepl.middleware.test$test_vars$fn__134687 invoke "test.clj" 259] [clojure.test$default_fixture invokeStatic "test.clj" 687] [clojure.test$default_fixture invoke "test.clj" 683] [cider.nrepl.middleware.test$test_vars invokeStatic "test.clj" 256] [cider.nrepl.middleware.test$test_vars invoke "test.clj" 250] [cider.nrepl.middleware.test$test_ns invokeStatic "test.clj" 272] [cider.nrepl.middleware.test$test_ns invoke "test.clj" 263] [cider.nrepl.middleware.test$test_var_query invokeStatic "test.clj" 283] [cider.nrepl.middleware.test$test_var_query invoke "test.clj" 276] [cider.nrepl.middleware.test$handle_test_var_query_op$fn__134731$fn__134732 invoke "test.clj" 321] [clojure.lang.AFn applyToHelper "AFn.java" 152] [clojure.lang.AFn applyTo "AFn.java" 144] [clojure.core$apply invokeStatic "core.clj" 667] [clojure.core$with_bindings_STAR_ invokeStatic "core.clj" 1990] [clojure.core$with_bindings_STAR_ doInvoke "core.clj" 1990] [clojure.lang.RestFn invoke "RestFn.java" 425] [cider.nrepl.middleware.test$handle_test_var_query_op$fn__134731 invoke "test.clj" 313] [clojure.lang.AFn run "AFn.java" 22] [nrepl.middleware.session$session_exec$main_loop__9666$fn__9670 invoke "session.clj" 218] [nrepl.middleware.session$session_exec$main_loop__9666 invoke "session.clj" 217] [clojure.lang.AFn run "AFn.java" 22] [java.lang.Thread run "Thread.java" 1589]], :cause "[\"Not a valid base type: :metabase.lib.schema.expression/type.unknown, received: :metabase.lib.schema.expression/type.unknown\"]", :data {:type :malli.core/invalid-output, :data {:output :metabase.lib.schema.common/base-type, :value :metabase.lib.schema.expression/type.unknown, :args [{:lib/type :mbql/query, :lib/metadata (metabase.lib.metadata.cached-provider/cached-metadata-provider (metabase.lib.metadata.jvm/->UncachedApplicationDatabaseMetadataProvider 2368)), :database 2368, :type :pipeline, :stages [{:lib/type :mbql.stage/mbql, :source-table 32407, :aggregation [[:metric #:lib{:uuid "06174580-70bd-474a-b167-95a9ef3899cc"} 318]]}]} -1 [:case #:lib{:uuid "01480a56-78ee-4b7f-9407-23a6493cdd7d"} [[[:< #:lib{:uuid "e078b90f-de46-4dbb-9903-930c3891c375"} [:field #:lib{:uuid "3454c908-194a-447d-af44-cbedee1ba44d"} 133191] 4] [:field #:lib{:uuid "4d3ae622-03b7-40f6-af20-1ae413065640"} 133191]]]]], :schema [:=> [:cat :metabase.lib.schema/query :int :any] :metabase.lib.schema.common/base-type], :fn-name lib.metadata.calculation/type-of}, :humanized ["Not a valid base type: :metabase.lib.schema.expression/type.unknown, received: :metabase.lib.schema.expression/type.unknown"], :link "https://malli.io?schema=%3Ametabase.lib.schema.common%2Fbase-type%0A&value=%3Ametabase.lib.schema.expression%2Ftype.unknown%0A"}} ```
process
metadata calculation not working correctly for aggregation that contains a metric clause failing test in metabase query processor test case test clj can we use case in metric with temporary metabase models metric metric with attributes creator id definition source table aggregation description lookin for a blueberry name toucans in the rainforest table id legacy query database type query query aggregation source table pmbql lib type mbql query database type pipeline stages lib type mbql stage mbql source table aggregation failed to calculated metadata for query via type clojure lang exceptioninfo message data type malli core invalid output data output metabase lib schema common base type value metabase lib schema expression type unknown args lib type mbql query lib metadata metabase lib metadata cached provider cached metadata provider metabase lib metadata jvm uncachedapplicationdatabasemetadataprovider database type pipeline stages lib type mbql stage mbql source table aggregation case lib uuid lib uuid schema metabase lib schema common base type fn name lib metadata calculation type of humanized link at trace metabase query processor middleware annotate add column info add column info rff star invoke annotate clj metabase query processor middleware mbql to native mbql gt native fn fn invoke mbql to native clj metabase enterprise advanced permissions query processor middleware permissions check download permissions fn fn invoke permissions clj metabase query processor middleware cache maybe return cached results maybe return cached results star invoke cache clj metabase enterprise advanced permissions query processor middleware permissions check download permissions fn invoke permissions clj metabase enterprise sandbox query processor middleware column level perms check maybe apply column level perms check fn invoke column level perms check clj metabase query processor fn combined post process combined post process star invoke query processor clj metabase query processor fn combined pre process combined pre process star invoke query processor clj metabase query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj metabase query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj metabase query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj metabase enterprise audit app query processor middleware handle audit queries handle internal queries fn invoke handle audit queries clj cause data type malli core invalid output data output metabase lib schema common base type value metabase lib schema expression type unknown args lib type mbql query lib metadata metabase lib metadata cached provider cached metadata provider metabase lib metadata jvm uncachedapplicationdatabasemetadataprovider database type pipeline stages lib type mbql stage mbql source table aggregation case lib uuid lib uuid schema metabase lib schema common base type fn name lib metadata calculation type of humanized link
1
651,455
21,479,265,387
IssuesEvent
2022-04-26 16:07:41
OpenDAX/OpenDAX
https://api.github.com/repos/OpenDAX/OpenDAX
closed
Lua module char array
Low Priority Bug
The Lua module returns a table when the tag is a CHAR array. It might make more sense to return a string made up of those characters. This seems like it would be the expected behavior.
1.0
Lua module char array - The Lua module returns a table when the tag is a CHAR array. It might make more sense to return a string made up of those characters. This seems like it would be the expected behavior.
non_process
lua module char array the lua module returns a table when the tag is a char array it might make more sense to return a string made up of those characters this seems like it would be the expected behavior
0
22,759
32,079,783,053
IssuesEvent
2023-09-25 13:17:17
h4sh5/npm-auto-scanner
https://api.github.com/repos/h4sh5/npm-auto-scanner
opened
monaca 4.2.9 has 2 guarddog issues
npm-install-script npm-silent-process-execution
```{"npm-install-script":[{"code":" \"preinstall\": \"node bin/preinstall\",","location":"package/package.json:9","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const childProcess = spawn('npm.cmd', ['run', 'monaca:preview'], {stdio: 'ignore', detached: true});","location":"package/src/serve.js:27","message":"This package is silently executing another executable"}]}```
1.0
monaca 4.2.9 has 2 guarddog issues - ```{"npm-install-script":[{"code":" \"preinstall\": \"node bin/preinstall\",","location":"package/package.json:9","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const childProcess = spawn('npm.cmd', ['run', 'monaca:preview'], {stdio: 'ignore', detached: true});","location":"package/src/serve.js:27","message":"This package is silently executing another executable"}]}```
process
monaca has guarddog issues npm install script npm silent process execution stdio ignore detached true location package src serve js message this package is silently executing another executable
1
661,568
22,060,937,448
IssuesEvent
2022-05-30 17:43:34
DXgovernance/dxvote
https://api.github.com/repos/DXgovernance/dxvote
closed
Simplify decode text string replacement
Enhancement High Priority Refactor
Currently we use a parameter index to replace values in the decodeText, this works but is not very intuitive to work out what is being used where as devs have to count the param index to work it out. This caused a regression bug with swapr campaign creation where the starting at 0 value was not considered and the decoded string was incorrect causing confusion and doubt for users. A better solution may be changing PARAM_X to the name of the parameter, a value we already have and is unique for the most part. If anyone has any better solutions these are also welcome, not sure if we could pass parameters in like we would with i18n string templates.
1.0
Simplify decode text string replacement - Currently we use a parameter index to replace values in the decodeText, this works but is not very intuitive to work out what is being used where as devs have to count the param index to work it out. This caused a regression bug with swapr campaign creation where the starting at 0 value was not considered and the decoded string was incorrect causing confusion and doubt for users. A better solution may be changing PARAM_X to the name of the parameter, a value we already have and is unique for the most part. If anyone has any better solutions these are also welcome, not sure if we could pass parameters in like we would with i18n string templates.
non_process
simplify decode text string replacement currently we use a parameter index to replace values in the decodetext this works but is not very intuitive to work out what is being used where as devs have to count the param index to work it out this caused a regression bug with swapr campaign creation where the starting at value was not considered and the decoded string was incorrect causing confusion and doubt for users a better solution may be changing param x to the name of the parameter a value we already have and is unique for the most part if anyone has any better solutions these are also welcome not sure if we could pass parameters in like we would with string templates
0
217
2,644,624,793
IssuesEvent
2015-03-12 17:54:33
ContaoDMS/dms
https://api.github.com/repos/ContaoDMS/dms
closed
Remove Contao 2 compatibility
Exercise ⚙ - Processed
Completly remove Contao 2 compatibility (database.sql, html Folder, .htaccess, etc.)
1.0
Remove Contao 2 compatibility - Completly remove Contao 2 compatibility (database.sql, html Folder, .htaccess, etc.)
process
remove contao compatibility completly remove contao compatibility database sql html folder htaccess etc
1
19,377
25,506,686,496
IssuesEvent
2022-11-28 09:59:17
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] Enrollment flow/updated consent flow > Review consent screen > Study name is not getting displayed for the Custom consent
Bug P1 Process: Fixed Process: Tested dev
1)Study name is not getting displayed for the Custom consent 2)Pdf is generated without study name
2.0
[iOS] Enrollment flow/updated consent flow > Review consent screen > Study name is not getting displayed for the Custom consent - 1)Study name is not getting displayed for the Custom consent 2)Pdf is generated without study name
process
enrollment flow updated consent flow review consent screen study name is not getting displayed for the custom consent study name is not getting displayed for the custom consent pdf is generated without study name
1
16,415
21,191,863,017
IssuesEvent
2022-04-08 18:23:57
wayfair-incubator/gator
https://api.github.com/repos/wayfair-incubator/gator
opened
Port Response Functionality from Internal Gator
pre-processing
# Description Gator will rely heavily on the `Response` objects defined as part of dynamic processing in Internal Gator. Even "static" changesets, without asynchronous resources, will rely on the structure of the Response to instruct `PostProcessing` what to do - which PRs to create, issues to create, whether or not the process failed, and why. Copy this functionality over from internally hosted Gator, including tests.
1.0
Port Response Functionality from Internal Gator - # Description Gator will rely heavily on the `Response` objects defined as part of dynamic processing in Internal Gator. Even "static" changesets, without asynchronous resources, will rely on the structure of the Response to instruct `PostProcessing` what to do - which PRs to create, issues to create, whether or not the process failed, and why. Copy this functionality over from internally hosted Gator, including tests.
process
port response functionality from internal gator description gator will rely heavily on the response objects defined as part of dynamic processing in internal gator even static changesets without asynchronous resources will rely on the structure of the response to instruct postprocessing what to do which prs to create issues to create whether or not the process failed and why copy this functionality over from internally hosted gator including tests
1
579,354
17,189,670,495
IssuesEvent
2021-07-16 09:07:47
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.ebay.de - design is broken
browser-firefox engine-gecko priority-important
<!-- @browser: Firefox 89.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:89.0) Gecko/20100101 Firefox/89.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/80299 --> **URL**: https://www.ebay.de/ **Browser / Version**: Firefox 89.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Design is broken **Description**: Items are misaligned **Steps to Reproduce**: On Chrome the site is working, only on Firefox there is a problem, whole page style is broken. On chrome is working properly <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.ebay.de - design is broken - <!-- @browser: Firefox 89.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:89.0) Gecko/20100101 Firefox/89.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/80299 --> **URL**: https://www.ebay.de/ **Browser / Version**: Firefox 89.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Design is broken **Description**: Items are misaligned **Steps to Reproduce**: On Chrome the site is working, only on Firefox there is a problem, whole page style is broken. On chrome is working properly <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
design is broken url browser version firefox operating system windows tested another browser yes chrome problem type design is broken description items are misaligned steps to reproduce on chrome the site is working only on firefox there is a problem whole page style is broken on chrome is working properly browser configuration none from with ❤️
0
22,266
30,819,745,957
IssuesEvent
2023-08-01 15:32:42
symfony/symfony-docs
https://api.github.com/repos/symfony/symfony-docs
closed
[Process] Support using `Process::findExecutable()` independently of `o…
Process
| Q | A | ------------ | --- | Feature PR | symfony/symfony#47422 | PR author(s) | @BlackbitDevs | Merged in | 6.4 We created this issue to not forget to document this new feature. We would really appreciate if you can help us with this task. If you are not sure how to do it, please ask us and we will help you. To fix this issue, please create a PR against the 6.4 branch in the [symfony-docs repository](https://github.com/symfony/symfony-docs). Thank you! :smiley:
1.0
[Process] Support using `Process::findExecutable()` independently of `o… - | Q | A | ------------ | --- | Feature PR | symfony/symfony#47422 | PR author(s) | @BlackbitDevs | Merged in | 6.4 We created this issue to not forget to document this new feature. We would really appreciate if you can help us with this task. If you are not sure how to do it, please ask us and we will help you. To fix this issue, please create a PR against the 6.4 branch in the [symfony-docs repository](https://github.com/symfony/symfony-docs). Thank you! :smiley:
process
support using process findexecutable independently of o… q a feature pr symfony symfony pr author s blackbitdevs merged in we created this issue to not forget to document this new feature we would really appreciate if you can help us with this task if you are not sure how to do it please ask us and we will help you to fix this issue please create a pr against the branch in the thank you smiley
1
137,344
18,752,690,262
IssuesEvent
2021-11-05 05:49:59
madhans23/linux-4.15
https://api.github.com/repos/madhans23/linux-4.15
opened
CVE-2019-12984 (Medium) detected in linux-stablev4.17.12
security vulnerability
## CVE-2019-12984 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stablev4.17.12</b></p></summary> <p> <p>Linux kernel stable tree</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git>https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.15/commit/d96ee498864d1a0b6222cfb17d64ca8196014940">d96ee498864d1a0b6222cfb17d64ca8196014940</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/netlink.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/netlink.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A NULL pointer dereference vulnerability in the function nfc_genl_deactivate_target() in net/nfc/netlink.c in the Linux kernel before 5.1.13 can be triggered by a malicious user-mode program that omits certain NFC attributes, leading to denial of service. <p>Publish Date: 2019-06-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12984>CVE-2019-12984</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12984">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12984</a></p> <p>Release Date: 2019-06-26</p> <p>Fix Resolution: v5.1-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-12984 (Medium) detected in linux-stablev4.17.12 - ## CVE-2019-12984 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stablev4.17.12</b></p></summary> <p> <p>Linux kernel stable tree</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git>https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.15/commit/d96ee498864d1a0b6222cfb17d64ca8196014940">d96ee498864d1a0b6222cfb17d64ca8196014940</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/netlink.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/nfc/netlink.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A NULL pointer dereference vulnerability in the function nfc_genl_deactivate_target() in net/nfc/netlink.c in the Linux kernel before 5.1.13 can be triggered by a malicious user-mode program that omits certain NFC attributes, leading to denial of service. <p>Publish Date: 2019-06-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12984>CVE-2019-12984</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12984">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12984</a></p> <p>Release Date: 2019-06-26</p> <p>Fix Resolution: v5.1-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux cve medium severity vulnerability vulnerable library linux linux kernel stable tree library home page a href found in head commit a href found in base branch master vulnerable source files net nfc netlink c net nfc netlink c vulnerability details a null pointer dereference vulnerability in the function nfc genl deactivate target in net nfc netlink c in the linux kernel before can be triggered by a malicious user mode program that omits certain nfc attributes leading to denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
2,666
5,445,832,243
IssuesEvent
2017-03-07 08:57:16
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
Process fails inside apache mod_php
Bug Process Status: Needs Review Status: Waiting feedback Unconfirmed
## Environment symfony/symfony v2.8.4 PHP `5.5.9-1ubuntu4.14` / `API220121212,NTS` Apache modules: `core mod_so mod_watchdog http_core mod_log_config mod_logio mod_version mod_unixd mod_access_compat mod_alias mod_auth_basic mod_authn_core mod_authn_file mod_authz_core mod_authz_groupfile mod_authz_host mod_authz_user mod_autoindex mod_bw mod_cgi mod_deflate mod_dir mod_env mod_expires mod_filter mod_headers mod_mime prefork mod_negotiation mod_php5 mod_rewrite mod_setenvif mod_socache_shmcb mod_ssl mod_status` PHP Extensions: `bcmath, bz2, calendar, Core, ctype, curl, date, dba, dom, ereg, exif, fileinfo, filter, ftp, gd, gettext, hash, iconv, imagick, intl, json, libxml, mbstring, mcrypt, memcachememcached, mhash, mysql, mysqli, newrelic, openssl, pcntl, pcre, PDO, pdo_mysql, Phar, posix, readline, Reflection, session, shmop, SimpleXML, soap, sockets, SPL, standardsysvmsg, sysvsem, sysvshm, tokenizer, wddx, xml, xmlreader, xmlwriter, Zend OPcache, zip, zlib` ## Problem ``` php <?php include __DIR__.'/../vendor/autoload.php'; $proc = new \Symfony\Component\Process\Process('echo TEST'); var_dump( $proc->start(), $proc->wait(), $proc->getExitCode(), $proc->getOutput(), $proc->getErrorOutput() ); /* outputs: NULL NULL NULL string(0) "" string(0) "" */ $proc2 = proc_open('echo TEST2', [1=>['pipe', 'w']], $pipes); var_dump(is_resource($proc2)); if (is_resource($proc2)) { var_dump(stream_get_contents($pipes[1])); fclose($pipes[1]); var_dump(proc_close($proc2)); } /* outputs bool(true) string(6) "TEST2 " NULL */ ``` As you can see, proc_open _kinda_ works at least the output works, but the exit code is null, which is already pretty bad. But then with the Process component I don't even get the output to work. Note that if I run the script from CLI, it works fine, I get 0 exit codes and output for both examples. Running via Apache mod_php fails.
1.0
Process fails inside apache mod_php - ## Environment symfony/symfony v2.8.4 PHP `5.5.9-1ubuntu4.14` / `API220121212,NTS` Apache modules: `core mod_so mod_watchdog http_core mod_log_config mod_logio mod_version mod_unixd mod_access_compat mod_alias mod_auth_basic mod_authn_core mod_authn_file mod_authz_core mod_authz_groupfile mod_authz_host mod_authz_user mod_autoindex mod_bw mod_cgi mod_deflate mod_dir mod_env mod_expires mod_filter mod_headers mod_mime prefork mod_negotiation mod_php5 mod_rewrite mod_setenvif mod_socache_shmcb mod_ssl mod_status` PHP Extensions: `bcmath, bz2, calendar, Core, ctype, curl, date, dba, dom, ereg, exif, fileinfo, filter, ftp, gd, gettext, hash, iconv, imagick, intl, json, libxml, mbstring, mcrypt, memcachememcached, mhash, mysql, mysqli, newrelic, openssl, pcntl, pcre, PDO, pdo_mysql, Phar, posix, readline, Reflection, session, shmop, SimpleXML, soap, sockets, SPL, standardsysvmsg, sysvsem, sysvshm, tokenizer, wddx, xml, xmlreader, xmlwriter, Zend OPcache, zip, zlib` ## Problem ``` php <?php include __DIR__.'/../vendor/autoload.php'; $proc = new \Symfony\Component\Process\Process('echo TEST'); var_dump( $proc->start(), $proc->wait(), $proc->getExitCode(), $proc->getOutput(), $proc->getErrorOutput() ); /* outputs: NULL NULL NULL string(0) "" string(0) "" */ $proc2 = proc_open('echo TEST2', [1=>['pipe', 'w']], $pipes); var_dump(is_resource($proc2)); if (is_resource($proc2)) { var_dump(stream_get_contents($pipes[1])); fclose($pipes[1]); var_dump(proc_close($proc2)); } /* outputs bool(true) string(6) "TEST2 " NULL */ ``` As you can see, proc_open _kinda_ works at least the output works, but the exit code is null, which is already pretty bad. But then with the Process component I don't even get the output to work. Note that if I run the script from CLI, it works fine, I get 0 exit codes and output for both examples. Running via Apache mod_php fails.
process
process fails inside apache mod php environment symfony symfony php nts apache modules core mod so mod watchdog http core mod log config mod logio mod version mod unixd mod access compat mod alias mod auth basic mod authn core mod authn file mod authz core mod authz groupfile mod authz host mod authz user mod autoindex mod bw mod cgi mod deflate mod dir mod env mod expires mod filter mod headers mod mime prefork mod negotiation mod mod rewrite mod setenvif mod socache shmcb mod ssl mod status php extensions bcmath calendar core ctype curl date dba dom ereg exif fileinfo filter ftp gd gettext hash iconv imagick intl json libxml mbstring mcrypt memcachememcached mhash mysql mysqli newrelic openssl pcntl pcre pdo pdo mysql phar posix readline reflection session shmop simplexml soap sockets spl standardsysvmsg sysvsem sysvshm tokenizer wddx xml xmlreader xmlwriter zend opcache zip zlib problem php php include dir vendor autoload php proc new symfony component process process echo test var dump proc start proc wait proc getexitcode proc getoutput proc geterroroutput outputs null null null string string proc open echo pipes var dump is resource if is resource var dump stream get contents pipes fclose pipes var dump proc close outputs bool true string null as you can see proc open kinda works at least the output works but the exit code is null which is already pretty bad but then with the process component i don t even get the output to work note that if i run the script from cli it works fine i get exit codes and output for both examples running via apache mod php fails
1
338
2,793,239,736
IssuesEvent
2015-05-11 09:36:17
ecodistrict/IDSSDashboard
https://api.github.com/repos/ecodistrict/IDSSDashboard
closed
Select best alternative tab needed?
form feedback 09102014 process step: select best alternatives question
Do we even need this tab? We could merge this with assess alternatives
1.0
Select best alternative tab needed? - Do we even need this tab? We could merge this with assess alternatives
process
select best alternative tab needed do we even need this tab we could merge this with assess alternatives
1
17,753
23,666,244,329
IssuesEvent
2022-08-26 21:19:23
googleapis/google-api-python-client
https://api.github.com/repos/googleapis/google-api-python-client
closed
Dependency Dashboard
type: process
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Edited/Blocked These updates have been manually edited so Renovate will no longer make changes. To discard all commits and start over, click on a checkbox. - [ ] <!-- rebase-branch=renovate/distlib-0.x -->[chore(deps): update dependency distlib to v0.3.6](../pull/1887) - [ ] <!-- rebase-branch=renovate/protobuf-4.x -->[chore(deps): update dependency protobuf to v4](../pull/1886) ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/click-8.x -->[chore(deps): update dependency click to v8.1.3](../pull/1884) - [ ] <!-- recreate-branch=renovate/setuptools-65.x -->[chore(deps): update dependency setuptools to v65.3.0](../pull/1885) ## Detected dependencies <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/main.yml</summary> - `actions/checkout v3` - `actions/setup-python v4` - `actions/github-script v6.2.0` </details> </blockquote> </details> <details><summary>pip_requirements</summary> <blockquote> <details><summary>.kokoro/requirements.txt</summary> - `argcomplete ==2.0.0` - `attrs ==22.1.0` - `bleach ==5.0.1` - `cachetools ==5.2.0` - `certifi ==2022.6.15` - `cffi ==1.15.1` - `charset-normalizer ==2.1.1` - `click ==8.0.4` - `colorlog ==6.6.0` - `commonmark ==0.9.1` - `cryptography ==37.0.4` - `distlib ==0.3.5` - `docutils ==0.19` - `filelock ==3.8.0` - `gcp-docuploader ==0.6.3` - `gcp-releasetool ==1.8.6` - `google-api-core ==2.8.2` - `google-auth ==2.11.0` - `google-cloud-core ==2.3.2` - `google-cloud-storage ==2.5.0` - `google-crc32c ==1.3.0` - `google-resumable-media ==2.3.3` - `googleapis-common-protos ==1.56.4` - `idna ==3.3` - `importlib-metadata ==4.12.0` - `jeepney ==0.8.0` - `jinja2 ==3.1.2` - `keyring ==23.8.2` - `markupsafe ==2.1.1` - `nox ==2022.8.7` - `packaging ==21.3` - `pkginfo ==1.8.3` - `platformdirs ==2.5.2` - `protobuf ==3.20.1` - `py ==1.11.0` - `pyasn1 ==0.4.8` - `pyasn1-modules ==0.2.8` - `pycparser ==2.21` - `pygments ==2.13.0` - `pyjwt ==2.4.0` - `pyparsing ==3.0.9` - `pyperclip ==1.8.2` - `python-dateutil ==2.8.2` - `readme-renderer ==37.0` - `requests ==2.28.1` - `requests-toolbelt ==0.9.1` - `rfc3986 ==2.0.0` - `rich ==12.5.1` - `rsa ==4.9` - `secretstorage ==3.3.3` - `six ==1.16.0` - `twine ==4.0.1` - `typing-extensions ==4.3.0` - `urllib3 ==1.26.12` - `virtualenv ==20.16.3` - `webencodings ==0.5.1` - `wheel ==0.37.1` - `zipp ==3.8.1` - `setuptools ==65.2.0` </details> <details><summary>samples/compute/requirements.txt</summary> - `google-api-python-client ==2.58.0` - `google-auth ==2.11.0` - `google-auth-httplib2 ==0.1.0` </details> <details><summary>scripts/requirements.txt</summary> - `pandas ==1.4.3` </details> </blockquote> </details> <details><summary>pip_setup</summary> <blockquote> <details><summary>setup.py</summary> - `httplib2 >=0.15.0,<1dev` - `google-auth >=1.19.0,<3.0.0dev` - `google-auth-httplib2 >=0.1.0` - `google-api-core >= 1.31.5, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.0` - `uritemplate >=3.0.1,<5` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Edited/Blocked These updates have been manually edited so Renovate will no longer make changes. To discard all commits and start over, click on a checkbox. - [ ] <!-- rebase-branch=renovate/distlib-0.x -->[chore(deps): update dependency distlib to v0.3.6](../pull/1887) - [ ] <!-- rebase-branch=renovate/protobuf-4.x -->[chore(deps): update dependency protobuf to v4](../pull/1886) ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/click-8.x -->[chore(deps): update dependency click to v8.1.3](../pull/1884) - [ ] <!-- recreate-branch=renovate/setuptools-65.x -->[chore(deps): update dependency setuptools to v65.3.0](../pull/1885) ## Detected dependencies <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/main.yml</summary> - `actions/checkout v3` - `actions/setup-python v4` - `actions/github-script v6.2.0` </details> </blockquote> </details> <details><summary>pip_requirements</summary> <blockquote> <details><summary>.kokoro/requirements.txt</summary> - `argcomplete ==2.0.0` - `attrs ==22.1.0` - `bleach ==5.0.1` - `cachetools ==5.2.0` - `certifi ==2022.6.15` - `cffi ==1.15.1` - `charset-normalizer ==2.1.1` - `click ==8.0.4` - `colorlog ==6.6.0` - `commonmark ==0.9.1` - `cryptography ==37.0.4` - `distlib ==0.3.5` - `docutils ==0.19` - `filelock ==3.8.0` - `gcp-docuploader ==0.6.3` - `gcp-releasetool ==1.8.6` - `google-api-core ==2.8.2` - `google-auth ==2.11.0` - `google-cloud-core ==2.3.2` - `google-cloud-storage ==2.5.0` - `google-crc32c ==1.3.0` - `google-resumable-media ==2.3.3` - `googleapis-common-protos ==1.56.4` - `idna ==3.3` - `importlib-metadata ==4.12.0` - `jeepney ==0.8.0` - `jinja2 ==3.1.2` - `keyring ==23.8.2` - `markupsafe ==2.1.1` - `nox ==2022.8.7` - `packaging ==21.3` - `pkginfo ==1.8.3` - `platformdirs ==2.5.2` - `protobuf ==3.20.1` - `py ==1.11.0` - `pyasn1 ==0.4.8` - `pyasn1-modules ==0.2.8` - `pycparser ==2.21` - `pygments ==2.13.0` - `pyjwt ==2.4.0` - `pyparsing ==3.0.9` - `pyperclip ==1.8.2` - `python-dateutil ==2.8.2` - `readme-renderer ==37.0` - `requests ==2.28.1` - `requests-toolbelt ==0.9.1` - `rfc3986 ==2.0.0` - `rich ==12.5.1` - `rsa ==4.9` - `secretstorage ==3.3.3` - `six ==1.16.0` - `twine ==4.0.1` - `typing-extensions ==4.3.0` - `urllib3 ==1.26.12` - `virtualenv ==20.16.3` - `webencodings ==0.5.1` - `wheel ==0.37.1` - `zipp ==3.8.1` - `setuptools ==65.2.0` </details> <details><summary>samples/compute/requirements.txt</summary> - `google-api-python-client ==2.58.0` - `google-auth ==2.11.0` - `google-auth-httplib2 ==0.1.0` </details> <details><summary>scripts/requirements.txt</summary> - `pandas ==1.4.3` </details> </blockquote> </details> <details><summary>pip_setup</summary> <blockquote> <details><summary>setup.py</summary> - `httplib2 >=0.15.0,<1dev` - `google-auth >=1.19.0,<3.0.0dev` - `google-auth-httplib2 >=0.1.0` - `google-api-core >= 1.31.5, <3.0.0dev,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.0` - `uritemplate >=3.0.1,<5` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more edited blocked these updates have been manually edited so renovate will no longer make changes to discard all commits and start over click on a checkbox pull pull ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull pull detected dependencies github actions github workflows main yml actions checkout actions setup python actions github script pip requirements kokoro requirements txt argcomplete attrs bleach cachetools certifi cffi charset normalizer click colorlog commonmark cryptography distlib docutils filelock gcp docuploader gcp releasetool google api core google auth google cloud core google cloud storage google google resumable media googleapis common protos idna importlib metadata jeepney keyring markupsafe nox packaging pkginfo platformdirs protobuf py modules pycparser pygments pyjwt pyparsing pyperclip python dateutil readme renderer requests requests toolbelt rich rsa secretstorage six twine typing extensions virtualenv webencodings wheel zipp setuptools samples compute requirements txt google api python client google auth google auth scripts requirements txt pandas pip setup setup py google auth google auth google api core uritemplate check this box to trigger a request for renovate to run again on this repository
1
33,041
12,165,853,719
IssuesEvent
2020-04-27 08:17:08
Baneeishaque/ask-med-pharma_website
https://api.github.com/repos/Baneeishaque/ask-med-pharma_website
opened
WS-2020-0042 (High) detected in acorn-6.3.0.tgz, acorn-5.7.3.tgz
security vulnerability
## WS-2020-0042 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>acorn-6.3.0.tgz</b>, <b>acorn-5.7.3.tgz</b></p></summary> <p> <details><summary><b>acorn-6.3.0.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-6.3.0.tgz">https://registry.npmjs.org/acorn/-/acorn-6.3.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/ask-med-pharma_website/wp-content/themes/twentytwenty/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/ask-med-pharma_website/wp-content/themes/twentytwenty/node_modules/acorn-globals/node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - scripts-5.0.0.tgz (Root Library) - jest-24.9.0.tgz - jest-cli-24.9.0.tgz - jest-config-24.9.0.tgz - jest-environment-jsdom-24.9.0.tgz - jsdom-11.12.0.tgz - acorn-globals-4.3.4.tgz - :x: **acorn-6.3.0.tgz** (Vulnerable Library) </details> <details><summary><b>acorn-5.7.3.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/ask-med-pharma_website/wp-content/themes/twentytwenty/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/ask-med-pharma_website/wp-content/themes/twentytwenty/node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - scripts-5.0.0.tgz (Root Library) - webpack-4.8.3.tgz - :x: **acorn-5.7.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/ask-med-pharma_website/commit/c7f5a051704dd823a801e2402d6a6ddf574962a2">c7f5a051704dd823a801e2402d6a6ddf574962a2</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser. <p>Publish Date: 2020-03-08 <p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p> <p>Release Date: 2020-03-08</p> <p>Fix Resolution: 7.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2020-0042 (High) detected in acorn-6.3.0.tgz, acorn-5.7.3.tgz - ## WS-2020-0042 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>acorn-6.3.0.tgz</b>, <b>acorn-5.7.3.tgz</b></p></summary> <p> <details><summary><b>acorn-6.3.0.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-6.3.0.tgz">https://registry.npmjs.org/acorn/-/acorn-6.3.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/ask-med-pharma_website/wp-content/themes/twentytwenty/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/ask-med-pharma_website/wp-content/themes/twentytwenty/node_modules/acorn-globals/node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - scripts-5.0.0.tgz (Root Library) - jest-24.9.0.tgz - jest-cli-24.9.0.tgz - jest-config-24.9.0.tgz - jest-environment-jsdom-24.9.0.tgz - jsdom-11.12.0.tgz - acorn-globals-4.3.4.tgz - :x: **acorn-6.3.0.tgz** (Vulnerable Library) </details> <details><summary><b>acorn-5.7.3.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/ask-med-pharma_website/wp-content/themes/twentytwenty/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/ask-med-pharma_website/wp-content/themes/twentytwenty/node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - scripts-5.0.0.tgz (Root Library) - webpack-4.8.3.tgz - :x: **acorn-5.7.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/ask-med-pharma_website/commit/c7f5a051704dd823a801e2402d6a6ddf574962a2">c7f5a051704dd823a801e2402d6a6ddf574962a2</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser. <p>Publish Date: 2020-03-08 <p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p> <p>Release Date: 2020-03-08</p> <p>Fix Resolution: 7.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws high detected in acorn tgz acorn tgz ws high severity vulnerability vulnerable libraries acorn tgz acorn tgz acorn tgz ecmascript parser library home page a href path to dependency file tmp ws scm ask med pharma website wp content themes twentytwenty package json path to vulnerable library tmp ws scm ask med pharma website wp content themes twentytwenty node modules acorn globals node modules acorn package json dependency hierarchy scripts tgz root library jest tgz jest cli tgz jest config tgz jest environment jsdom tgz jsdom tgz acorn globals tgz x acorn tgz vulnerable library acorn tgz ecmascript parser library home page a href path to dependency file tmp ws scm ask med pharma website wp content themes twentytwenty package json path to vulnerable library tmp ws scm ask med pharma website wp content themes twentytwenty node modules acorn package json dependency hierarchy scripts tgz root library webpack tgz x acorn tgz vulnerable library found in head commit a href vulnerability details acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
390,228
26,854,169,717
IssuesEvent
2023-02-03 13:25:00
roots-id/rootswallet
https://api.github.com/repos/roots-id/rootswallet
closed
Review existing iOS file/directory structure and generate issues
documentation enhancement
Potentially also run RootsWallet in iOS and see what breaks
1.0
Review existing iOS file/directory structure and generate issues - Potentially also run RootsWallet in iOS and see what breaks
non_process
review existing ios file directory structure and generate issues potentially also run rootswallet in ios and see what breaks
0
161,790
20,155,358,842
IssuesEvent
2022-02-09 15:59:44
jgeraigery/experian-java
https://api.github.com/repos/jgeraigery/experian-java
closed
CVE-2020-11112 (High) detected in jackson-databind-2.9.2.jar - autoclosed
security vulnerability
## CVE-2020-11112 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p> <p>Path to vulnerable library: ository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9c836fb688a6591b590fe6154680b73ceed58629">9c836fb688a6591b590fe6154680b73ceed58629</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.proxy.provider.remoting.RmiProvider (aka apache/commons-proxy). <p>Publish Date: 2020-03-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11112>CVE-2020-11112</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112</a></p> <p>Release Date: 2020-03-31</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11112","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.proxy.provider.remoting.RmiProvider (aka apache/commons-proxy).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11112","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-11112 (High) detected in jackson-databind-2.9.2.jar - autoclosed - ## CVE-2020-11112 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p> <p>Path to vulnerable library: ository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9c836fb688a6591b590fe6154680b73ceed58629">9c836fb688a6591b590fe6154680b73ceed58629</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.proxy.provider.remoting.RmiProvider (aka apache/commons-proxy). <p>Publish Date: 2020-03-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11112>CVE-2020-11112</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11112</a></p> <p>Release Date: 2020-03-31</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11112","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.proxy.provider.remoting.RmiProvider (aka apache/commons-proxy).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11112","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file experian java mavenworkspace bis services lib bis services base pom xml path to vulnerable library ository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons proxy provider remoting rmiprovider aka apache commons proxy publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons proxy provider remoting rmiprovider aka apache commons proxy vulnerabilityurl
0
6,967
10,119,384,217
IssuesEvent
2019-07-31 11:20:29
linnovate/root
https://api.github.com/repos/linnovate/root
closed
tasks inheritance from discussion
Process bug Tasks
create a task create a disucssion add users as partners with different permissions **example meeting:** ![image](https://user-images.githubusercontent.com/38312178/62207242-51ca6700-b39c-11e9-8311-0ef882ff747d.png) go to the created task assign the discussion to it the partners' permissions doesnt change : ![image](https://user-images.githubusercontent.com/38312178/62207901-01540900-b39e-11e9-8612-d02a80ce6b1e.png)
1.0
tasks inheritance from discussion - create a task create a disucssion add users as partners with different permissions **example meeting:** ![image](https://user-images.githubusercontent.com/38312178/62207242-51ca6700-b39c-11e9-8311-0ef882ff747d.png) go to the created task assign the discussion to it the partners' permissions doesnt change : ![image](https://user-images.githubusercontent.com/38312178/62207901-01540900-b39e-11e9-8612-d02a80ce6b1e.png)
process
tasks inheritance from discussion create a task create a disucssion add users as partners with different permissions example meeting go to the created task assign the discussion to it the partners permissions doesnt change
1
14,847
18,242,123,449
IssuesEvent
2021-10-01 14:04:10
pycaret/pycaret
https://api.github.com/repos/pycaret/pycaret
closed
ValueError: dtype for the target variable should be int32 or int64 only
bug preprocessing
**Describe the bug** <!-- --> In regression it is usual to use independent variables in float. In Pycaret version 2.3.4 I get the following error in regression when activating "Remove Multicollinearity=True", in "False" it does not give the problem. ERROR: pycaret regression ValueError: dtype for the target variable should be int32 or int64 only I understand that it should support float and integer values. FILE: /usr/local/lib/python3.7/dist-packages/pycaret/internal/preprocess.py rows 2515, 2516 class Fix_multicollinearity(BaseEstimator, TransformerMixin): def fit(self, data, y=None): ... if data[self.target_variable].dtype not in ["int32", "int64"]: raise ValueError('dtype for the target variable should be int32 or int64 only') ... **To Reproduce** <!-- --> from pycaret.datasets import get_data dataset = get_data('diamond') dataset['Price'] = dataset['Price'].astype('float32') from pycaret.regression import * exp_reg101 = setup(data = dataset, target = 'Price', session_id=123, remove_multicollinearity = True, multicollinearity_threshold = 0.95) --------------------------------------------------------------------------- ValueError Traceback (most recent call last) <ipython-input-9-f41871135e22> in <module>() 1 from pycaret.regression import * 2 exp_reg101 = setup(data = dataset, target = 'Price', session_id=123, ----> 3 remove_multicollinearity = True, multicollinearity_threshold = 0.95) 7 frames /usr/local/lib/python3.7/dist-packages/pycaret/internal/preprocess.py in fit(self, data, y) 2514 2515 if data[self.target_variable].dtype not in ["int32", "int64"]: -> 2516 raise ValueError('dtype for the target variable should be int32 or int64 only') 2517 2518 # global data1 ValueError: dtype for the target variable should be int32 or int64 only ```python from pycaret.datasets import get_data dataset = get_data('diamond') dataset['Price'] = dataset['Price'].astype('float32') from pycaret.regression import * exp_reg101 = setup(data = dataset, target = 'Price', session_id=123, remove_multicollinearity = True, multicollinearity_threshold = 0.95) ``` **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> **Additional context** <!-- Add any other context about the problem here. --> I understand that it should support float and integer values. **Versions** <2.3.4> <!-- Please run the following code snippet and paste the output here: import pycaret pycaret.__version__ --> </details> <!-- Thanks for contributing! -->
1.0
ValueError: dtype for the target variable should be int32 or int64 only - **Describe the bug** <!-- --> In regression it is usual to use independent variables in float. In Pycaret version 2.3.4 I get the following error in regression when activating "Remove Multicollinearity=True", in "False" it does not give the problem. ERROR: pycaret regression ValueError: dtype for the target variable should be int32 or int64 only I understand that it should support float and integer values. FILE: /usr/local/lib/python3.7/dist-packages/pycaret/internal/preprocess.py rows 2515, 2516 class Fix_multicollinearity(BaseEstimator, TransformerMixin): def fit(self, data, y=None): ... if data[self.target_variable].dtype not in ["int32", "int64"]: raise ValueError('dtype for the target variable should be int32 or int64 only') ... **To Reproduce** <!-- --> from pycaret.datasets import get_data dataset = get_data('diamond') dataset['Price'] = dataset['Price'].astype('float32') from pycaret.regression import * exp_reg101 = setup(data = dataset, target = 'Price', session_id=123, remove_multicollinearity = True, multicollinearity_threshold = 0.95) --------------------------------------------------------------------------- ValueError Traceback (most recent call last) <ipython-input-9-f41871135e22> in <module>() 1 from pycaret.regression import * 2 exp_reg101 = setup(data = dataset, target = 'Price', session_id=123, ----> 3 remove_multicollinearity = True, multicollinearity_threshold = 0.95) 7 frames /usr/local/lib/python3.7/dist-packages/pycaret/internal/preprocess.py in fit(self, data, y) 2514 2515 if data[self.target_variable].dtype not in ["int32", "int64"]: -> 2516 raise ValueError('dtype for the target variable should be int32 or int64 only') 2517 2518 # global data1 ValueError: dtype for the target variable should be int32 or int64 only ```python from pycaret.datasets import get_data dataset = get_data('diamond') dataset['Price'] = dataset['Price'].astype('float32') from pycaret.regression import * exp_reg101 = setup(data = dataset, target = 'Price', session_id=123, remove_multicollinearity = True, multicollinearity_threshold = 0.95) ``` **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> **Additional context** <!-- Add any other context about the problem here. --> I understand that it should support float and integer values. **Versions** <2.3.4> <!-- Please run the following code snippet and paste the output here: import pycaret pycaret.__version__ --> </details> <!-- Thanks for contributing! -->
process
valueerror dtype for the target variable should be or only describe the bug in regression it is usual to use independent variables in float in pycaret version i get the following error in regression when activating remove multicollinearity true in false it does not give the problem error pycaret regression valueerror dtype for the target variable should be or only i understand that it should support float and integer values file usr local lib dist packages pycaret internal preprocess py rows class fix multicollinearity baseestimator transformermixin def fit self data y none if data dtype not in raise valueerror dtype for the target variable should be or only to reproduce from pycaret datasets import get data dataset get data diamond dataset dataset astype from pycaret regression import exp setup data dataset target price session id remove multicollinearity true multicollinearity threshold valueerror traceback most recent call last in from pycaret regression import exp setup data dataset target price session id remove multicollinearity true multicollinearity threshold frames usr local lib dist packages pycaret internal preprocess py in fit self data y if data dtype not in raise valueerror dtype for the target variable should be or only global valueerror dtype for the target variable should be or only python from pycaret datasets import get data dataset get data diamond dataset dataset astype from pycaret regression import exp setup data dataset target price session id remove multicollinearity true multicollinearity threshold expected behavior a clear and concise description of what you expected to happen additional context add any other context about the problem here i understand that it should support float and integer values versions please run the following code snippet and paste the output here import pycaret pycaret version
1
46,662
11,866,052,983
IssuesEvent
2020-03-26 02:26:26
spack/spack
https://api.github.com/repos/spack/spack
opened
autoconf and other packages on ppc64le
build-error
### Spack version <!-- Add the output to the command below --> ```console [kai@longhorn ~]$ spack --version ``` ### Steps to reproduce the issue ```console [kai@longhorn ~]$ spack spec autoconf Input spec -------------------------------- autoconf Concretized -------------------------------- autoconf@2.69%gcc@7.3.0 arch=linux-rhel7-power9le ^m4@1.4.18%gcc@7.3.0 patches=3877ab548f88597ab2327a2230ee048d2d07ace1062efe81fc92e91b7f39cd00,fc9b61654a3ba1a8d6cd78ce087e7c96366c290bc8d2c299f09828d793b853c8 +sigsegv arch=linux-rhel7-power9le ^libsigsegv@2.12%gcc@7.3.0 arch=linux-rhel7-power9le ^perl@5.30.1%gcc@7.3.0+cpanm+shared+threads arch=linux-rhel7-power9le ^gdbm@1.18.1%gcc@7.3.0 arch=linux-rhel7-power9le ^readline@8.0%gcc@7.3.0 arch=linux-rhel7-power9le ^ncurses@6.2%gcc@7.3.0~symlinks+termlib arch=linux-rhel7-power9le ^pkgconf@1.6.3%gcc@7.3.0 arch=linux-rhel7-power9le [kai@longhorn ~]$ spack install autoconf [...] See build log for details: /tmp/kai/spack-stage/spack-stage-autoconf-2.69-ftyunbfd663jlfj24legpgewbdsjygse/spack-build-out.txt Traceback (most recent call last): File "/home/01537/kai/build/spack/lib/spack/spack/build_environment.py", line 801, in child_process return_value = function() File "/home/01537/kai/build/spack/lib/spack/spack/installer.py", line 1113, in build_process phase(pkg.spec, pkg.prefix) File "/home/01537/kai/build/spack/lib/spack/spack/package.py", line 112, in phase_wrapper callback(instance) File "/home/01537/kai/build/spack/lib/spack/spack/build_systems/autotools.py", line 160, in _do_patch_config_guess raise RuntimeError('Failed to find suitable config.guess') RuntimeError: Failed to find suitable config.guess ``` ### Platform and user environment ```console [kai@longhorn ~]$ uname -a Linux login1.longhorn.tacc.utexas.edu 4.14.0-115.10.1.el7a.ppc64le #1 SMP Wed Jun 26 09:32:17 UTC 2019 ppc64le ppc64le ppc64le GNU/Linux [kai@longhorn ~]$ cat /etc/redhat-release Red Hat Enterprise Linux Server release 7.6 (Maipo) ``` This machine (longhorn) is similar to Summit, except that it has very little software installed. I've tracked the issue down already: When trying to install `autoconf`, spack checks whether the included `config.guess` works. It does not, because it does not recognize what should be `powerpc64le-unknown-linux-gnu`. So spack is looking for a newer `config.guess` to replace the included one. On Summit, it'll find one in `/usr/share/automake-x.yy`, and things work. On this machine, automake is not installed (at least not in the standard location). If the package depended on `automake`, it'd look at the spack-installed automake to find a newer `config.guess`. That doesn't apply here. I don't have access to install anything in `/usr/share`, so the question is, how can I provide an updated `config.guess`? I've hacked around it for now by modifying spack's source so that it searches in my home directory, but clearly that's not a sustainable solution. [This problem isn't really limited to just the autoconf package, either, but any package that ships with an outdated `config.guess`, which doesn't depend on `automake`: `libsodium` was another manifestation of the same issue.] One way to make it possible to work around this problem would be for spack to search for `config.guess` in a user-specified location, though I still don't like it, since the user would still have to go find an appropriate `config.guess`, and point spack to it. I'd much rather have something that works out of the box, but I don't have any good idea on how to get it done.
1.0
autoconf and other packages on ppc64le - ### Spack version <!-- Add the output to the command below --> ```console [kai@longhorn ~]$ spack --version ``` ### Steps to reproduce the issue ```console [kai@longhorn ~]$ spack spec autoconf Input spec -------------------------------- autoconf Concretized -------------------------------- autoconf@2.69%gcc@7.3.0 arch=linux-rhel7-power9le ^m4@1.4.18%gcc@7.3.0 patches=3877ab548f88597ab2327a2230ee048d2d07ace1062efe81fc92e91b7f39cd00,fc9b61654a3ba1a8d6cd78ce087e7c96366c290bc8d2c299f09828d793b853c8 +sigsegv arch=linux-rhel7-power9le ^libsigsegv@2.12%gcc@7.3.0 arch=linux-rhel7-power9le ^perl@5.30.1%gcc@7.3.0+cpanm+shared+threads arch=linux-rhel7-power9le ^gdbm@1.18.1%gcc@7.3.0 arch=linux-rhel7-power9le ^readline@8.0%gcc@7.3.0 arch=linux-rhel7-power9le ^ncurses@6.2%gcc@7.3.0~symlinks+termlib arch=linux-rhel7-power9le ^pkgconf@1.6.3%gcc@7.3.0 arch=linux-rhel7-power9le [kai@longhorn ~]$ spack install autoconf [...] See build log for details: /tmp/kai/spack-stage/spack-stage-autoconf-2.69-ftyunbfd663jlfj24legpgewbdsjygse/spack-build-out.txt Traceback (most recent call last): File "/home/01537/kai/build/spack/lib/spack/spack/build_environment.py", line 801, in child_process return_value = function() File "/home/01537/kai/build/spack/lib/spack/spack/installer.py", line 1113, in build_process phase(pkg.spec, pkg.prefix) File "/home/01537/kai/build/spack/lib/spack/spack/package.py", line 112, in phase_wrapper callback(instance) File "/home/01537/kai/build/spack/lib/spack/spack/build_systems/autotools.py", line 160, in _do_patch_config_guess raise RuntimeError('Failed to find suitable config.guess') RuntimeError: Failed to find suitable config.guess ``` ### Platform and user environment ```console [kai@longhorn ~]$ uname -a Linux login1.longhorn.tacc.utexas.edu 4.14.0-115.10.1.el7a.ppc64le #1 SMP Wed Jun 26 09:32:17 UTC 2019 ppc64le ppc64le ppc64le GNU/Linux [kai@longhorn ~]$ cat /etc/redhat-release Red Hat Enterprise Linux Server release 7.6 (Maipo) ``` This machine (longhorn) is similar to Summit, except that it has very little software installed. I've tracked the issue down already: When trying to install `autoconf`, spack checks whether the included `config.guess` works. It does not, because it does not recognize what should be `powerpc64le-unknown-linux-gnu`. So spack is looking for a newer `config.guess` to replace the included one. On Summit, it'll find one in `/usr/share/automake-x.yy`, and things work. On this machine, automake is not installed (at least not in the standard location). If the package depended on `automake`, it'd look at the spack-installed automake to find a newer `config.guess`. That doesn't apply here. I don't have access to install anything in `/usr/share`, so the question is, how can I provide an updated `config.guess`? I've hacked around it for now by modifying spack's source so that it searches in my home directory, but clearly that's not a sustainable solution. [This problem isn't really limited to just the autoconf package, either, but any package that ships with an outdated `config.guess`, which doesn't depend on `automake`: `libsodium` was another manifestation of the same issue.] One way to make it possible to work around this problem would be for spack to search for `config.guess` in a user-specified location, though I still don't like it, since the user would still have to go find an appropriate `config.guess`, and point spack to it. I'd much rather have something that works out of the box, but I don't have any good idea on how to get it done.
non_process
autoconf and other packages on spack version console spack version steps to reproduce the issue console spack spec autoconf input spec autoconf concretized autoconf gcc arch linux gcc patches sigsegv arch linux libsigsegv gcc arch linux perl gcc cpanm shared threads arch linux gdbm gcc arch linux readline gcc arch linux ncurses gcc symlinks termlib arch linux pkgconf gcc arch linux spack install autoconf see build log for details tmp kai spack stage spack stage autoconf spack build out txt traceback most recent call last file home kai build spack lib spack spack build environment py line in child process return value function file home kai build spack lib spack spack installer py line in build process phase pkg spec pkg prefix file home kai build spack lib spack spack package py line in phase wrapper callback instance file home kai build spack lib spack spack build systems autotools py line in do patch config guess raise runtimeerror failed to find suitable config guess runtimeerror failed to find suitable config guess platform and user environment console uname a linux longhorn tacc utexas edu smp wed jun utc gnu linux cat etc redhat release red hat enterprise linux server release maipo this machine longhorn is similar to summit except that it has very little software installed i ve tracked the issue down already when trying to install autoconf spack checks whether the included config guess works it does not because it does not recognize what should be unknown linux gnu so spack is looking for a newer config guess to replace the included one on summit it ll find one in usr share automake x yy and things work on this machine automake is not installed at least not in the standard location if the package depended on automake it d look at the spack installed automake to find a newer config guess that doesn t apply here i don t have access to install anything in usr share so the question is how can i provide an updated config guess i ve hacked around it for now by modifying spack s source so that it searches in my home directory but clearly that s not a sustainable solution one way to make it possible to work around this problem would be for spack to search for config guess in a user specified location though i still don t like it since the user would still have to go find an appropriate config guess and point spack to it i d much rather have something that works out of the box but i don t have any good idea on how to get it done
0
10,702
13,497,560,421
IssuesEvent
2020-09-12 08:16:10
metabase/metabase
https://api.github.com/repos/metabase/metabase
reopened
Download full results of a Query to CSV or XLSX - Excel does not recognize the Timestamp format
.Backend Difficulty:Easy Priority:P1 Querying/Processor Reporting/Export Type:Bug
Download full results of a Query to CSV or XLSX - Excel does not recognize the Timestamp format **To Reproduce** Steps to reproduce the behavior: 1. Go to any query you may have that has a timestamp field (at least in Oracle database as data source). 2. Click on 'Download full results' in XLSX or CSV. 3. Open with Excel. 4. See column containing timestamp, it is shown in the cell as: 2019-03-18T17:24:31.000+01:00 (type 'General', i.e. not recognized). It cannot be converted either or get recognized if cell is changed to date, time, ... **Expected behavior** When exporting results containing dates, times or datetimes (timestamps fields), the formatting should be recognized by Excel on Open without any need to convert or tweak the data. **Screenshots** ![image](https://user-images.githubusercontent.com/10819908/64324141-10922c00-cfc6-11e9-947f-35197fb400d6.png) **Information about your Metabase Installation:** - Your browser and the version: Chrome version 76.0.3809.132 (Official Build) (32-bit) - Your operating system: Windows 10.0.17134 Build 17134 - Your databases: Oracle, MySQL - Metabase version: v.0.33 - Metabase hosting environment: Jar-file on CentOS - Metabase internal database: MySQL **Severity** This annoying and prevents that some users can make use of Metabase to complete some of their tasks. ⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
1.0
Download full results of a Query to CSV or XLSX - Excel does not recognize the Timestamp format - Download full results of a Query to CSV or XLSX - Excel does not recognize the Timestamp format **To Reproduce** Steps to reproduce the behavior: 1. Go to any query you may have that has a timestamp field (at least in Oracle database as data source). 2. Click on 'Download full results' in XLSX or CSV. 3. Open with Excel. 4. See column containing timestamp, it is shown in the cell as: 2019-03-18T17:24:31.000+01:00 (type 'General', i.e. not recognized). It cannot be converted either or get recognized if cell is changed to date, time, ... **Expected behavior** When exporting results containing dates, times or datetimes (timestamps fields), the formatting should be recognized by Excel on Open without any need to convert or tweak the data. **Screenshots** ![image](https://user-images.githubusercontent.com/10819908/64324141-10922c00-cfc6-11e9-947f-35197fb400d6.png) **Information about your Metabase Installation:** - Your browser and the version: Chrome version 76.0.3809.132 (Official Build) (32-bit) - Your operating system: Windows 10.0.17134 Build 17134 - Your databases: Oracle, MySQL - Metabase version: v.0.33 - Metabase hosting environment: Jar-file on CentOS - Metabase internal database: MySQL **Severity** This annoying and prevents that some users can make use of Metabase to complete some of their tasks. ⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
process
download full results of a query to csv or xlsx excel does not recognize the timestamp format download full results of a query to csv or xlsx excel does not recognize the timestamp format to reproduce steps to reproduce the behavior go to any query you may have that has a timestamp field at least in oracle database as data source click on download full results in xlsx or csv open with excel see column containing timestamp it is shown in the cell as type general i e not recognized it cannot be converted either or get recognized if cell is changed to date time expected behavior when exporting results containing dates times or datetimes timestamps fields the formatting should be recognized by excel on open without any need to convert or tweak the data screenshots information about your metabase installation your browser and the version chrome version official build bit your operating system windows build your databases oracle mysql metabase version v metabase hosting environment jar file on centos metabase internal database mysql severity this annoying and prevents that some users can make use of metabase to complete some of their tasks ⬇️ please click the 👍 reaction instead of leaving a or 👍 comment
1
7,047
10,208,508,980
IssuesEvent
2019-08-14 10:17:31
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
New Devops Domain
Pri2 assigned-to-author automation/svc process-automation/subsvc product-question triaged
Does not support the new devops domain https://aka.ms/new-domain --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Source Control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration#feedback) * Content Source: [articles/automation/source-control-integration.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
1.0
New Devops Domain - Does not support the new devops domain https://aka.ms/new-domain --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Source Control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration#feedback) * Content Source: [articles/automation/source-control-integration.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
process
new devops domain does not support the new devops domain document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login bobbytreed microsoft alias robreed
1
299,088
9,205,047,204
IssuesEvent
2019-03-08 09:28:22
richelbilderbeek/djog_unos_2018
https://api.github.com/repos/richelbilderbeek/djog_unos_2018
closed
Lions and crocodiles eat cows only when close
medium priority
**Is your feature request related to a problem? Please describe.** Currently, the lions and crocodiles eat cows from approx 5 meters. In reality, this is zero meters. **Describe the solution you'd like** Let the lions and crocodiles eat cows only when close **Describe alternatives you've considered** None. **Additional context** None,
1.0
Lions and crocodiles eat cows only when close - **Is your feature request related to a problem? Please describe.** Currently, the lions and crocodiles eat cows from approx 5 meters. In reality, this is zero meters. **Describe the solution you'd like** Let the lions and crocodiles eat cows only when close **Describe alternatives you've considered** None. **Additional context** None,
non_process
lions and crocodiles eat cows only when close is your feature request related to a problem please describe currently the lions and crocodiles eat cows from approx meters in reality this is zero meters describe the solution you d like let the lions and crocodiles eat cows only when close describe alternatives you ve considered none additional context none
0
6,314
9,320,787,232
IssuesEvent
2019-03-27 00:50:55
tc39/proposal-promise-allSettled
https://api.github.com/repos/tc39/proposal-promise-allSettled
closed
Advance to stage 3
process
Criteria taken from [the TC39 process document](https://tc39.github.io/process-document/) minus those from previous stages: > - [x] Complete spec text https://github.com/tc39/proposal-promise-allSettled#specification > - [x] Designated reviewers have signed off on the current spec text - [x] @codehag - [x] @domenic - [x] @gibson042 > - [x] All ECMAScript editors have signed off on the current spec text - [x] @ljharb - [x] @zenparsing
1.0
Advance to stage 3 - Criteria taken from [the TC39 process document](https://tc39.github.io/process-document/) minus those from previous stages: > - [x] Complete spec text https://github.com/tc39/proposal-promise-allSettled#specification > - [x] Designated reviewers have signed off on the current spec text - [x] @codehag - [x] @domenic - [x] @gibson042 > - [x] All ECMAScript editors have signed off on the current spec text - [x] @ljharb - [x] @zenparsing
process
advance to stage criteria taken from minus those from previous stages complete spec text designated reviewers have signed off on the current spec text codehag domenic all ecmascript editors have signed off on the current spec text ljharb zenparsing
1
47,765
12,110,519,856
IssuesEvent
2020-04-21 10:33:34
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
opened
composer.workflows.bq_copy_across_locations_test: test_dag failed
buildcop: issue priority: p1 type: bug
This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: fa38579cbf0946fd8ce0610bf86cb19cbe9ab782 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/d94a37ce-d68d-48d2-a1e4-3e74086e9c77), [Sponge](http://sponge2/d94a37ce-d68d-48d2-a1e4-3e74086e9c77) status: failed
1.0
composer.workflows.bq_copy_across_locations_test: test_dag failed - This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: fa38579cbf0946fd8ce0610bf86cb19cbe9ab782 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/d94a37ce-d68d-48d2-a1e4-3e74086e9c77), [Sponge](http://sponge2/d94a37ce-d68d-48d2-a1e4-3e74086e9c77) status: failed
non_process
composer workflows bq copy across locations test test dag failed this test failed to configure my behavior see if i m commenting on this issue too often add the buildcop quiet label and i will stop commenting commit buildurl status failed
0
19,925
26,391,855,419
IssuesEvent
2023-01-12 16:14:00
python/cpython
https://api.github.com/repos/python/cpython
closed
"OSError not raised by Listener" in test_asyncio test_multiprocessing_* with Linux kernel 6.0.16-6.1.3 (kernel bug)
type-bug tests expert-asyncio expert-multiprocessing
# Bug report A few tests in CPython's test suite have suddenly started failing. I've first experienced the problem rebuilding 3.8.16 on 2023-01-11, then confirmed it on top of 3.12.0a3, 3.12.0a4 and main today. The same 3.8.16 and 3.12.0a3 versions used to pass the respective tests on 2012-12-07. I'm currently trying to investigate what have changed since but unfortunately it's hard for such a large span of time. Any ideas what to look for would be appreciated. The failing tests are: ``` 4 tests failed: test_asyncio test_multiprocessing_fork test_multiprocessing_forkserver test_multiprocessing_spawn ``` FWICS all the failures resemble this one: ```pytb ====================================================================== FAIL: test_multiple_bind (test.test_multiprocessing_spawn.WithProcessesTestListener.test_multiple_bind) ---------------------------------------------------------------------- Traceback (most recent call last): File "/home/mgorny/git/cpython/Lib/test/_test_multiprocessing.py", line 3377, in test_multiple_bind self.assertRaises(OSError, self.connection.Listener, AssertionError: OSError not raised by Listener ``` I've run these tests using: ``` ./python -u -W default -bb -E -E -m test -v test_asyncio test_multiprocessing_fork test_multiprocessing_forkserver test_multiprocessing_spawn ``` Full log (415k): [test.txt](https://github.com/python/cpython/files/10393849/test.txt) # Your environment - CPython versions tested on: 3.8.16, 3.12.0a3, 3.12.0a4, 762745a124cbc297cf2fe6f3ec9ca1840bb2e873 - Operating system and architecture: Gentoo Linux amd64, glibc 2.36-r6 p6, kernel 6.0.18-gentoo-dist
1.0
"OSError not raised by Listener" in test_asyncio test_multiprocessing_* with Linux kernel 6.0.16-6.1.3 (kernel bug) - # Bug report A few tests in CPython's test suite have suddenly started failing. I've first experienced the problem rebuilding 3.8.16 on 2023-01-11, then confirmed it on top of 3.12.0a3, 3.12.0a4 and main today. The same 3.8.16 and 3.12.0a3 versions used to pass the respective tests on 2012-12-07. I'm currently trying to investigate what have changed since but unfortunately it's hard for such a large span of time. Any ideas what to look for would be appreciated. The failing tests are: ``` 4 tests failed: test_asyncio test_multiprocessing_fork test_multiprocessing_forkserver test_multiprocessing_spawn ``` FWICS all the failures resemble this one: ```pytb ====================================================================== FAIL: test_multiple_bind (test.test_multiprocessing_spawn.WithProcessesTestListener.test_multiple_bind) ---------------------------------------------------------------------- Traceback (most recent call last): File "/home/mgorny/git/cpython/Lib/test/_test_multiprocessing.py", line 3377, in test_multiple_bind self.assertRaises(OSError, self.connection.Listener, AssertionError: OSError not raised by Listener ``` I've run these tests using: ``` ./python -u -W default -bb -E -E -m test -v test_asyncio test_multiprocessing_fork test_multiprocessing_forkserver test_multiprocessing_spawn ``` Full log (415k): [test.txt](https://github.com/python/cpython/files/10393849/test.txt) # Your environment - CPython versions tested on: 3.8.16, 3.12.0a3, 3.12.0a4, 762745a124cbc297cf2fe6f3ec9ca1840bb2e873 - Operating system and architecture: Gentoo Linux amd64, glibc 2.36-r6 p6, kernel 6.0.18-gentoo-dist
process
oserror not raised by listener in test asyncio test multiprocessing with linux kernel kernel bug bug report a few tests in cpython s test suite have suddenly started failing i ve first experienced the problem rebuilding on then confirmed it on top of and main today the same and versions used to pass the respective tests on i m currently trying to investigate what have changed since but unfortunately it s hard for such a large span of time any ideas what to look for would be appreciated the failing tests are tests failed test asyncio test multiprocessing fork test multiprocessing forkserver test multiprocessing spawn fwics all the failures resemble this one pytb fail test multiple bind test test multiprocessing spawn withprocessestestlistener test multiple bind traceback most recent call last file home mgorny git cpython lib test test multiprocessing py line in test multiple bind self assertraises oserror self connection listener assertionerror oserror not raised by listener i ve run these tests using python u w default bb e e m test v test asyncio test multiprocessing fork test multiprocessing forkserver test multiprocessing spawn full log your environment cpython versions tested on operating system and architecture gentoo linux glibc kernel gentoo dist
1
270,321
28,960,497,082
IssuesEvent
2023-05-10 01:47:02
artsking/linux-4.19.72
https://api.github.com/repos/artsking/linux-4.19.72
reopened
WS-2021-0213 (High) detected in linux-yoctov5.4.51
Mend: dependency security vulnerability
## WS-2021-0213 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.19.72/commit/8519fe4185f1a7567a708f01b476f195b0f1046c">8519fe4185f1a7567a708f01b476f195b0f1046c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/ti/wl1251/cmd.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/ti/wl1251/cmd.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Linux/kernel in versions v3.5-rc1--v5.13 is vulnerable to possible buffer overflow in wl1251_cmd_scan <p>Publish Date: 2021-05-31 <p>URL: <a href=https://github.com/torvalds/linux/commit/57ad99ae3c6738ba87bad259bb57c641ca68ebf6>WS-2021-0213</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1001488">https://osv.dev/vulnerability/UVI-2021-1001488</a></p> <p>Release Date: 2021-05-31</p> <p>Fix Resolution: v5.14-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0213 (High) detected in linux-yoctov5.4.51 - ## WS-2021-0213 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.19.72/commit/8519fe4185f1a7567a708f01b476f195b0f1046c">8519fe4185f1a7567a708f01b476f195b0f1046c</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/ti/wl1251/cmd.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/ti/wl1251/cmd.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Linux/kernel in versions v3.5-rc1--v5.13 is vulnerable to possible buffer overflow in wl1251_cmd_scan <p>Publish Date: 2021-05-31 <p>URL: <a href=https://github.com/torvalds/linux/commit/57ad99ae3c6738ba87bad259bb57c641ca68ebf6>WS-2021-0213</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1001488">https://osv.dev/vulnerability/UVI-2021-1001488</a></p> <p>Release Date: 2021-05-31</p> <p>Fix Resolution: v5.14-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws high detected in linux ws high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers net wireless ti cmd c drivers net wireless ti cmd c vulnerability details linux kernel in versions is vulnerable to possible buffer overflow in cmd scan publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
17,733
23,641,756,644
IssuesEvent
2022-08-25 17:45:37
carbon-design-system/ibm-cloud-cognitive
https://api.github.com/repos/carbon-design-system/ibm-cloud-cognitive
closed
Revisiting motion libraries for component animations
type: process improvement needs: investigation
## What will this achieve? After working with motion on components in Carbon for IBM Products, we noticed that trying to animate elements that leave the DOM can be challenging. @matthewgallo has added utilities to help with this but they can quickly become buggy (relying on state and adding handlers in places to undo "onAnimationEnd" handler.) We should consider Framer motion because it is easier to use and has a ton of features for reduce motion and hooks that the team can benefit from.
1.0
Revisiting motion libraries for component animations - ## What will this achieve? After working with motion on components in Carbon for IBM Products, we noticed that trying to animate elements that leave the DOM can be challenging. @matthewgallo has added utilities to help with this but they can quickly become buggy (relying on state and adding handlers in places to undo "onAnimationEnd" handler.) We should consider Framer motion because it is easier to use and has a ton of features for reduce motion and hooks that the team can benefit from.
process
revisiting motion libraries for component animations what will this achieve after working with motion on components in carbon for ibm products we noticed that trying to animate elements that leave the dom can be challenging matthewgallo has added utilities to help with this but they can quickly become buggy relying on state and adding handlers in places to undo onanimationend handler we should consider framer motion because it is easier to use and has a ton of features for reduce motion and hooks that the team can benefit from
1
20,138
26,686,597,572
IssuesEvent
2023-01-26 22:34:20
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Not is clear the price and parallelism for the deployment group
doc-enhancement devops/prod Pri2 devops-cicd-process/tech
Hello team, I am looking for the following information in the documentation, but it is not clear, you can add the following points to be clearer: - Is there a charge to add a deployment group - Maximum amount of parallelism to deploy in the deployment groups - In a release pipeline I have several stages and each stage has a deployment group, is there a parallelism limit to execute the pipeline and also there is any charge? Thank you so much! --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9dd8a19f-54c2-d052-8560-9f179b24a0b3 * Version Independent ID: cf217d69-791f-3375-bc43-f57483427c46 * Content: [(Classic) Deployment group jobs in release pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/deployment-group-phases?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/deployment-group-phases.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/deployment-group-phases.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Not is clear the price and parallelism for the deployment group - Hello team, I am looking for the following information in the documentation, but it is not clear, you can add the following points to be clearer: - Is there a charge to add a deployment group - Maximum amount of parallelism to deploy in the deployment groups - In a release pipeline I have several stages and each stage has a deployment group, is there a parallelism limit to execute the pipeline and also there is any charge? Thank you so much! --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9dd8a19f-54c2-d052-8560-9f179b24a0b3 * Version Independent ID: cf217d69-791f-3375-bc43-f57483427c46 * Content: [(Classic) Deployment group jobs in release pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/deployment-group-phases?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/deployment-group-phases.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/deployment-group-phases.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
not is clear the price and parallelism for the deployment group hello team i am looking for the following information in the documentation but it is not clear you can add the following points to be clearer is there a charge to add a deployment group maximum amount of parallelism to deploy in the deployment groups in a release pipeline i have several stages and each stage has a deployment group is there a parallelism limit to execute the pipeline and also there is any charge thank you so much document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
10,255
13,108,770,758
IssuesEvent
2020-08-04 17:26:20
googleapis/code-suggester
https://api.github.com/repos/googleapis/code-suggester
closed
Support non-"master" default primary branch
enhancement type: process
A branch can have multiple origin branches, and also have the non-"master" branch as the default branch. Framework-core should allow user configuration of which branches are the origin branch. - [x] CLI supports - [x] framework-core supports
1.0
Support non-"master" default primary branch - A branch can have multiple origin branches, and also have the non-"master" branch as the default branch. Framework-core should allow user configuration of which branches are the origin branch. - [x] CLI supports - [x] framework-core supports
process
support non master default primary branch a branch can have multiple origin branches and also have the non master branch as the default branch framework core should allow user configuration of which branches are the origin branch cli supports framework core supports
1
6,692
9,811,181,789
IssuesEvent
2019-06-12 22:39:49
ArctosDB/new-collections
https://api.github.com/repos/ArctosDB/new-collections
closed
NFWFL Draft MOU
Committed to Arctos MOU draft in process Prospective Collection Form Complete
Draft MOU Work with new collection to complete Draft MOU, answer any questions about migration, Arctos operating procedures, and costs; (download sample template include collection contact info).
1.0
NFWFL Draft MOU - Draft MOU Work with new collection to complete Draft MOU, answer any questions about migration, Arctos operating procedures, and costs; (download sample template include collection contact info).
process
nfwfl draft mou draft mou work with new collection to complete draft mou answer any questions about migration arctos operating procedures and costs download sample template include collection contact info
1
9,784
12,801,167,878
IssuesEvent
2020-07-02 18:34:52
solid/process
https://api.github.com/repos/solid/process
closed
Unclear which editorial assignment would review proposals from some panels
process proposal
There are several panels without editorial assignments including: * accessibility * artificial intelligence * caching * explaining the vision * external interoperability and outreach * internationalisation * privacy and individuals rights protection * self hosting * specification entry document This means it's unclear which editors would review proposals coming out of these panels. I propose staring editorial assignments that match these panels or defining which existing editorial assignments would review proposals coming out fo these panels.
1.0
Unclear which editorial assignment would review proposals from some panels - There are several panels without editorial assignments including: * accessibility * artificial intelligence * caching * explaining the vision * external interoperability and outreach * internationalisation * privacy and individuals rights protection * self hosting * specification entry document This means it's unclear which editors would review proposals coming out of these panels. I propose staring editorial assignments that match these panels or defining which existing editorial assignments would review proposals coming out fo these panels.
process
unclear which editorial assignment would review proposals from some panels there are several panels without editorial assignments including accessibility artificial intelligence caching explaining the vision external interoperability and outreach internationalisation privacy and individuals rights protection self hosting specification entry document this means it s unclear which editors would review proposals coming out of these panels i propose staring editorial assignments that match these panels or defining which existing editorial assignments would review proposals coming out fo these panels
1
21,795
30,303,018,030
IssuesEvent
2023-07-10 07:35:16
pycaret/pycaret
https://api.github.com/repos/pycaret/pycaret
closed
[BUG]: Numeric imputation "drop"
bug classification preprocessing
### pycaret version checks - [X] I have checked that this issue has not already been reported [here](https://github.com/pycaret/pycaret/issues). - [X] I have confirmed this bug exists on the [latest version](https://github.com/pycaret/pycaret/releases) of pycaret. - [ ] I have confirmed this bug exists on the master branch of pycaret (pip install -U git+https://github.com/pycaret/pycaret.git@master). ### Issue Description Hello, I have an error with numeric_imputation='drop'. If I use for example numeric_imputation='mean' it works, but with numeric_imputation='drop' the training fails. ### Reproducible Example ```python from pycaret.datasets import get_data data = get_data('titanic') from pycaret.classification import ClassificationExperiment exp = ClassificationExperiment() exp.setup(data, target = 'Survived', session_id = 123, ignore_features=['PassengerId', 'Name', 'Ticket'], numeric_imputation='drop', memory=None) model = exp.create_model('lr') ``` ### Expected Behavior I expect the training to work. Instead the metrics table is empty ![image](https://github.com/pycaret/pycaret/assets/28063509/4c13ea45-6d73-441c-96dd-193b8da07cba) and in the log I have the error below. ### Actual Results ```python-traceback WARNING:/opt/conda/lib/python3.10/site-packages/sklearn/model_selection/_validation.py:794: UserWarning: Scoring failed. The score on this train-test partition for these parameters will be set to 0. Details: Traceback (most recent call last): File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_scorer.py", line 115, in __call__ score = scorer._score(cached_call, estimator, *args, **kwargs) File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_scorer.py", line 282, in _score return self._sign * self._score_func(y_true, y_pred, **self._kwargs) File "/opt/conda/lib/python3.10/site-packages/pycaret/internal/metrics.py", line 35, in __call__ return self.score_func(y_true, y_pred, **kwargs) File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_classification.py", line 1954, in precision_score p, _, _, _ = precision_recall_fscore_support( File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_classification.py", line 1573, in precision_recall_fscore_support labels = _check_set_wise_labels(y_true, y_pred, average, labels, pos_label) File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_classification.py", line 1374, in _check_set_wise_labels y_type, y_true, y_pred = _check_targets(y_true, y_pred) File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_classification.py", line 86, in _check_targets check_consistent_length(y_true, y_pred) File "/opt/conda/lib/python3.10/site-packages/sklearn/utils/validation.py", line 397, in check_consistent_length raise ValueError( ValueError: Found input variables with inconsistent numbers of samples: [62, 45] ``` ### Installed Versions <details> System: python: 3.10.11 | packaged by conda-forge | (main, May 10 2023, 18:58:44) [GCC 11.3.0] executable: /opt/conda/bin/python machine: Linux-5.15.0-76-generic-x86_64-with-glibc2.35 PyCaret required dependencies: pip: 23.1.2 setuptools: 67.7.2 pycaret: 3.0.4 IPython: 8.13.2 ipywidgets: 8.0.6 tqdm: 4.65.0 numpy: 1.23.5 pandas: 1.5.3 jinja2: 3.1.2 scipy: 1.10.1 joblib: 1.2.0 sklearn: 1.2.2 pyod: 1.0.9 imblearn: 0.10.1 category_encoders: 2.6.1 lightgbm: 3.3.5 numba: 0.56.4 requests: 2.29.0 matplotlib: 3.7.1 scikitplot: 0.3.7 yellowbrick: 1.5 plotly: 5.13.1 plotly-resampler: Not installed kaleido: 0.2.1 schemdraw: 0.15 statsmodels: 0.14.0 sktime: 0.17.0 tbats: 1.1.3 pmdarima: 2.0.3 psutil: 5.9.5 markupsafe: 2.1.2 pickle5: Not installed cloudpickle: 2.2.1 deprecation: 2.1.0 xxhash: 3.2.0 wurlitzer: 3.0.3 PyCaret optional dependencies: shap: 0.41.0 interpret: 0.4.2 umap: Not installed pandas_profiling: 4.2.0 explainerdashboard: Not installed autoviz: Not installed fairlearn: Not installed deepchecks: Not installed xgboost: Not installed catboost: Not installed kmodes: Not installed mlxtend: Not installed statsforecast: Not installed tune_sklearn: Not installed ray: Not installed hyperopt: Not installed optuna: Not installed skopt: Not installed mlflow: Not installed gradio: Not installed fastapi: Not installed uvicorn: Not installed m2cgen: Not installed evidently: Not installed fugue: Not installed streamlit: Not installed prophet: 1.1.3</details>
1.0
[BUG]: Numeric imputation "drop" - ### pycaret version checks - [X] I have checked that this issue has not already been reported [here](https://github.com/pycaret/pycaret/issues). - [X] I have confirmed this bug exists on the [latest version](https://github.com/pycaret/pycaret/releases) of pycaret. - [ ] I have confirmed this bug exists on the master branch of pycaret (pip install -U git+https://github.com/pycaret/pycaret.git@master). ### Issue Description Hello, I have an error with numeric_imputation='drop'. If I use for example numeric_imputation='mean' it works, but with numeric_imputation='drop' the training fails. ### Reproducible Example ```python from pycaret.datasets import get_data data = get_data('titanic') from pycaret.classification import ClassificationExperiment exp = ClassificationExperiment() exp.setup(data, target = 'Survived', session_id = 123, ignore_features=['PassengerId', 'Name', 'Ticket'], numeric_imputation='drop', memory=None) model = exp.create_model('lr') ``` ### Expected Behavior I expect the training to work. Instead the metrics table is empty ![image](https://github.com/pycaret/pycaret/assets/28063509/4c13ea45-6d73-441c-96dd-193b8da07cba) and in the log I have the error below. ### Actual Results ```python-traceback WARNING:/opt/conda/lib/python3.10/site-packages/sklearn/model_selection/_validation.py:794: UserWarning: Scoring failed. The score on this train-test partition for these parameters will be set to 0. Details: Traceback (most recent call last): File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_scorer.py", line 115, in __call__ score = scorer._score(cached_call, estimator, *args, **kwargs) File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_scorer.py", line 282, in _score return self._sign * self._score_func(y_true, y_pred, **self._kwargs) File "/opt/conda/lib/python3.10/site-packages/pycaret/internal/metrics.py", line 35, in __call__ return self.score_func(y_true, y_pred, **kwargs) File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_classification.py", line 1954, in precision_score p, _, _, _ = precision_recall_fscore_support( File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_classification.py", line 1573, in precision_recall_fscore_support labels = _check_set_wise_labels(y_true, y_pred, average, labels, pos_label) File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_classification.py", line 1374, in _check_set_wise_labels y_type, y_true, y_pred = _check_targets(y_true, y_pred) File "/opt/conda/lib/python3.10/site-packages/sklearn/metrics/_classification.py", line 86, in _check_targets check_consistent_length(y_true, y_pred) File "/opt/conda/lib/python3.10/site-packages/sklearn/utils/validation.py", line 397, in check_consistent_length raise ValueError( ValueError: Found input variables with inconsistent numbers of samples: [62, 45] ``` ### Installed Versions <details> System: python: 3.10.11 | packaged by conda-forge | (main, May 10 2023, 18:58:44) [GCC 11.3.0] executable: /opt/conda/bin/python machine: Linux-5.15.0-76-generic-x86_64-with-glibc2.35 PyCaret required dependencies: pip: 23.1.2 setuptools: 67.7.2 pycaret: 3.0.4 IPython: 8.13.2 ipywidgets: 8.0.6 tqdm: 4.65.0 numpy: 1.23.5 pandas: 1.5.3 jinja2: 3.1.2 scipy: 1.10.1 joblib: 1.2.0 sklearn: 1.2.2 pyod: 1.0.9 imblearn: 0.10.1 category_encoders: 2.6.1 lightgbm: 3.3.5 numba: 0.56.4 requests: 2.29.0 matplotlib: 3.7.1 scikitplot: 0.3.7 yellowbrick: 1.5 plotly: 5.13.1 plotly-resampler: Not installed kaleido: 0.2.1 schemdraw: 0.15 statsmodels: 0.14.0 sktime: 0.17.0 tbats: 1.1.3 pmdarima: 2.0.3 psutil: 5.9.5 markupsafe: 2.1.2 pickle5: Not installed cloudpickle: 2.2.1 deprecation: 2.1.0 xxhash: 3.2.0 wurlitzer: 3.0.3 PyCaret optional dependencies: shap: 0.41.0 interpret: 0.4.2 umap: Not installed pandas_profiling: 4.2.0 explainerdashboard: Not installed autoviz: Not installed fairlearn: Not installed deepchecks: Not installed xgboost: Not installed catboost: Not installed kmodes: Not installed mlxtend: Not installed statsforecast: Not installed tune_sklearn: Not installed ray: Not installed hyperopt: Not installed optuna: Not installed skopt: Not installed mlflow: Not installed gradio: Not installed fastapi: Not installed uvicorn: Not installed m2cgen: Not installed evidently: Not installed fugue: Not installed streamlit: Not installed prophet: 1.1.3</details>
process
numeric imputation drop pycaret version checks i have checked that this issue has not already been reported i have confirmed this bug exists on the of pycaret i have confirmed this bug exists on the master branch of pycaret pip install u git issue description hello i have an error with numeric imputation drop if i use for example numeric imputation mean it works but with numeric imputation drop the training fails reproducible example python from pycaret datasets import get data data get data titanic from pycaret classification import classificationexperiment exp classificationexperiment exp setup data target survived session id ignore features numeric imputation drop memory none model exp create model lr expected behavior i expect the training to work instead the metrics table is empty and in the log i have the error below actual results python traceback warning opt conda lib site packages sklearn model selection validation py userwarning scoring failed the score on this train test partition for these parameters will be set to details traceback most recent call last file opt conda lib site packages sklearn metrics scorer py line in call score scorer score cached call estimator args kwargs file opt conda lib site packages sklearn metrics scorer py line in score return self sign self score func y true y pred self kwargs file opt conda lib site packages pycaret internal metrics py line in call return self score func y true y pred kwargs file opt conda lib site packages sklearn metrics classification py line in precision score p precision recall fscore support file opt conda lib site packages sklearn metrics classification py line in precision recall fscore support labels check set wise labels y true y pred average labels pos label file opt conda lib site packages sklearn metrics classification py line in check set wise labels y type y true y pred check targets y true y pred file opt conda lib site packages sklearn metrics classification py line in check targets check consistent length y true y pred file opt conda lib site packages sklearn utils validation py line in check consistent length raise valueerror valueerror found input variables with inconsistent numbers of samples installed versions system python packaged by conda forge main may executable opt conda bin python machine linux generic with pycaret required dependencies pip setuptools pycaret ipython ipywidgets tqdm numpy pandas scipy joblib sklearn pyod imblearn category encoders lightgbm numba requests matplotlib scikitplot yellowbrick plotly plotly resampler not installed kaleido schemdraw statsmodels sktime tbats pmdarima psutil markupsafe not installed cloudpickle deprecation xxhash wurlitzer pycaret optional dependencies shap interpret umap not installed pandas profiling explainerdashboard not installed autoviz not installed fairlearn not installed deepchecks not installed xgboost not installed catboost not installed kmodes not installed mlxtend not installed statsforecast not installed tune sklearn not installed ray not installed hyperopt not installed optuna not installed skopt not installed mlflow not installed gradio not installed fastapi not installed uvicorn not installed not installed evidently not installed fugue not installed streamlit not installed prophet
1
17,556
23,370,316,128
IssuesEvent
2022-08-10 19:10:46
GoogleCloudPlatform/cloud-ops-sandbox
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
closed
Add build-cop bot to Sandbox repo
type: process priority: p3 lang: yaml
Adding [build-cop bot](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop) will allow automated creation/closure of bugs associated with failed CI runs.
1.0
Add build-cop bot to Sandbox repo - Adding [build-cop bot](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop) will allow automated creation/closure of bugs associated with failed CI runs.
process
add build cop bot to sandbox repo adding will allow automated creation closure of bugs associated with failed ci runs
1
13,560
16,103,417,816
IssuesEvent
2021-04-27 12:21:16
osstotalsoft/nbb
https://api.github.com/repos/osstotalsoft/nbb
closed
Process manager - state should not be contrained to beeing a struct
enhancement good first issue process manager
Currently, the state in process manager is a struct. It should be a class. Using a struct is a way to insure the state does not have behaviour. However, the "framework" should not be THAT opinionated and should allow the programmer to have the responsibility of its own code.
1.0
Process manager - state should not be contrained to beeing a struct - Currently, the state in process manager is a struct. It should be a class. Using a struct is a way to insure the state does not have behaviour. However, the "framework" should not be THAT opinionated and should allow the programmer to have the responsibility of its own code.
process
process manager state should not be contrained to beeing a struct currently the state in process manager is a struct it should be a class using a struct is a way to insure the state does not have behaviour however the framework should not be that opinionated and should allow the programmer to have the responsibility of its own code
1
380,954
11,271,494,378
IssuesEvent
2020-01-14 13:10:44
googleapis/elixir-google-api
https://api.github.com/repos/googleapis/elixir-google-api
opened
Synthesis failed for OAuth2
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate OAuth2. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-oauth2' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main last_synth_commit_hash = get_last_metadata_commit(args.metadata_path) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit text=True, File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run with Popen(*popenargs, **kwargs) as process: TypeError: __init__() got an unexpected keyword argument 'text' ``` Google internal developers can see the full log [here](https://sponge/164839b8-e54e-409d-bb48-012d4d868e86).
1.0
Synthesis failed for OAuth2 - Hello! Autosynth couldn't regenerate OAuth2. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-oauth2' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main last_synth_commit_hash = get_last_metadata_commit(args.metadata_path) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit text=True, File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run with Popen(*popenargs, **kwargs) as process: TypeError: __init__() got an unexpected keyword argument 'text' ``` Google internal developers can see the full log [here](https://sponge/164839b8-e54e-409d-bb48-012d4d868e86).
non_process
synthesis failed for hello autosynth couldn t regenerate broken heart here s the output from running synth py cloning into working repo switched to branch autosynth traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main last synth commit hash get last metadata commit args metadata path file tmpfs src git autosynth autosynth synth py line in get last metadata commit text true file home kbuilder pyenv versions lib subprocess py line in run with popen popenargs kwargs as process typeerror init got an unexpected keyword argument text google internal developers can see the full log
0
17,029
22,406,802,748
IssuesEvent
2022-06-18 04:41:58
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[feature][processing] Add "Offset" parameter to "Drape" algorithms (Request in QGIS)
Processing Alg 3.26
### Request for documentation From pull request QGIS/qgis#48772 Author: @agiudiceandrea QGIS version: 3.26 **[feature][processing] Add "Offset" parameter to "Drape" algorithms** ### PR Description: ## Description Adds the "Offset" parameter (`QgsProcessingParameterNumber::Double`, default to 0) to the "Drape (set Z value from raster)" and "Set M value from raster" algorithms. The "Offset" value is algebraically added to the raster value after applying the "Scale factor". Fixes #48651. <!-- BEFORE HITTING SUBMIT -- Please BUILD AND TEST your changes thoroughly. This is YOUR responsibility! Do NOT rely on the QGIS code maintainers to do this for you!! IMPORTANT NOTES FOR FIRST TIME CONTRIBUTORS =========================================== Congratulations, you are about to make a pull request to QGIS! To make this as easy and pleasurable for everyone, please take the time to read these lines before opening the pull request. Include a few sentences describing the overall goals for this pull request (PR). If applicable also add screenshots or - even better - screencasts. Include both: *what* you changed and *why* you changed it. If this is a pull request that adds new functionality which needs documentation, give an especially detailed explanation. In this case, start with a short abstract and then write some text that can be copied 1:1 to the documentation in the best case. Also mention if you think this PR needs to be backported. And list relevant or fixed issues. ------------------------ Reviewing is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by checking the following list. Feel free to ask in a comment if you have troubles with any of them. - Commit messages are descriptive and explain the rationale for changes. - Commits which fix bugs include `Fixes #11111` at the bottom of the commit message. If this is your first pull request and you forgot to do this, write the same statement into this text field with the pull request description. - New unit tests have been added for relevant changes - You have run the `scripts/prepare_commit.sh` script (https://github.com/qgis/QGIS/blob/master/.github/CONTRIBUTING.md#contributing-to-qgis) before each commit. If you didn't do this, you can also run `./scripts/astyle_all.sh` from your source folder. - You have read the QGIS Coding Standards (https://docs.qgis.org/testing/en/docs/developers_guide/codingstandards.html) and this PR complies with them --> ### Commits tagged with [need-docs] or [FEATURE]
1.0
[feature][processing] Add "Offset" parameter to "Drape" algorithms (Request in QGIS) - ### Request for documentation From pull request QGIS/qgis#48772 Author: @agiudiceandrea QGIS version: 3.26 **[feature][processing] Add "Offset" parameter to "Drape" algorithms** ### PR Description: ## Description Adds the "Offset" parameter (`QgsProcessingParameterNumber::Double`, default to 0) to the "Drape (set Z value from raster)" and "Set M value from raster" algorithms. The "Offset" value is algebraically added to the raster value after applying the "Scale factor". Fixes #48651. <!-- BEFORE HITTING SUBMIT -- Please BUILD AND TEST your changes thoroughly. This is YOUR responsibility! Do NOT rely on the QGIS code maintainers to do this for you!! IMPORTANT NOTES FOR FIRST TIME CONTRIBUTORS =========================================== Congratulations, you are about to make a pull request to QGIS! To make this as easy and pleasurable for everyone, please take the time to read these lines before opening the pull request. Include a few sentences describing the overall goals for this pull request (PR). If applicable also add screenshots or - even better - screencasts. Include both: *what* you changed and *why* you changed it. If this is a pull request that adds new functionality which needs documentation, give an especially detailed explanation. In this case, start with a short abstract and then write some text that can be copied 1:1 to the documentation in the best case. Also mention if you think this PR needs to be backported. And list relevant or fixed issues. ------------------------ Reviewing is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by checking the following list. Feel free to ask in a comment if you have troubles with any of them. - Commit messages are descriptive and explain the rationale for changes. - Commits which fix bugs include `Fixes #11111` at the bottom of the commit message. If this is your first pull request and you forgot to do this, write the same statement into this text field with the pull request description. - New unit tests have been added for relevant changes - You have run the `scripts/prepare_commit.sh` script (https://github.com/qgis/QGIS/blob/master/.github/CONTRIBUTING.md#contributing-to-qgis) before each commit. If you didn't do this, you can also run `./scripts/astyle_all.sh` from your source folder. - You have read the QGIS Coding Standards (https://docs.qgis.org/testing/en/docs/developers_guide/codingstandards.html) and this PR complies with them --> ### Commits tagged with [need-docs] or [FEATURE]
process
add offset parameter to drape algorithms request in qgis request for documentation from pull request qgis qgis author agiudiceandrea qgis version add offset parameter to drape algorithms pr description description adds the offset parameter qgsprocessingparameternumber double default to to the drape set z value from raster and set m value from raster algorithms the offset value is algebraically added to the raster value after applying the scale factor fixes before hitting submit please build and test your changes thoroughly this is your responsibility do not rely on the qgis code maintainers to do this for you important notes for first time contributors congratulations you are about to make a pull request to qgis to make this as easy and pleasurable for everyone please take the time to read these lines before opening the pull request include a few sentences describing the overall goals for this pull request pr if applicable also add screenshots or even better screencasts include both what you changed and why you changed it if this is a pull request that adds new functionality which needs documentation give an especially detailed explanation in this case start with a short abstract and then write some text that can be copied to the documentation in the best case also mention if you think this pr needs to be backported and list relevant or fixed issues reviewing is a process done by project maintainers mostly on a volunteer basis we try to keep the overhead as small as possible and appreciate if you help us to do so by checking the following list feel free to ask in a comment if you have troubles with any of them commit messages are descriptive and explain the rationale for changes commits which fix bugs include fixes at the bottom of the commit message if this is your first pull request and you forgot to do this write the same statement into this text field with the pull request description new unit tests have been added for relevant changes you have run the scripts prepare commit sh script before each commit if you didn t do this you can also run scripts astyle all sh from your source folder you have read the qgis coding standards and this pr complies with them commits tagged with or
1
160,994
6,106,736,475
IssuesEvent
2017-06-21 05:42:28
pmem/issues
https://api.github.com/repos/pmem/issues
closed
tests: cannot set UNITTEST_QUIET before running RUNTESTS
Exposure: Low OS: Linux Priority: 4 low State: To be verified Type: Bug
Cannot set UNITTEST_QUIET before running RUNTESTS. Reason: RUNTESTS sets UNITTEST_QUIET to 1 in line 252 and so make pcheck silently skips some tests.
1.0
tests: cannot set UNITTEST_QUIET before running RUNTESTS - Cannot set UNITTEST_QUIET before running RUNTESTS. Reason: RUNTESTS sets UNITTEST_QUIET to 1 in line 252 and so make pcheck silently skips some tests.
non_process
tests cannot set unittest quiet before running runtests cannot set unittest quiet before running runtests reason runtests sets unittest quiet to in line and so make pcheck silently skips some tests
0
151,662
23,854,242,741
IssuesEvent
2022-09-06 21:08:56
hong-yura/WorkForUs-Project
https://api.github.com/repos/hong-yura/WorkForUs-Project
opened
[FE] : 일정관리 화면
🎨DESIGN
## 🔔 요구 사항 일정관리 페이지의 화면을 만든다. ## ✅ 수행 작업 - [ ] Toast UI Calendar 적용 - [ ] 버튼에 캘린더 조작 기능 추가
1.0
[FE] : 일정관리 화면 - ## 🔔 요구 사항 일정관리 페이지의 화면을 만든다. ## ✅ 수행 작업 - [ ] Toast UI Calendar 적용 - [ ] 버튼에 캘린더 조작 기능 추가
non_process
일정관리 화면 🔔 요구 사항 일정관리 페이지의 화면을 만든다 ✅ 수행 작업 toast ui calendar 적용 버튼에 캘린더 조작 기능 추가
0
4,314
4,972,700,875
IssuesEvent
2016-12-05 22:20:33
dotnet/wcf
https://api.github.com/repos/dotnet/wcf
closed
Use TestProperties and to parse command line arguments to configure ports
Infrastructure
Pull request https://github.com/dotnet/wcf/pull/1011 recommended using TestProperties and to parse command line arguments to configure ports for self hosted service. We need to evaluate if it's needed to configure ports as there is a debate to use the same port IIS uses.
1.0
Use TestProperties and to parse command line arguments to configure ports - Pull request https://github.com/dotnet/wcf/pull/1011 recommended using TestProperties and to parse command line arguments to configure ports for self hosted service. We need to evaluate if it's needed to configure ports as there is a debate to use the same port IIS uses.
non_process
use testproperties and to parse command line arguments to configure ports pull request recommended using testproperties and to parse command line arguments to configure ports for self hosted service we need to evaluate if it s needed to configure ports as there is a debate to use the same port iis uses
0
57,764
3,083,767,938
IssuesEvent
2015-08-24 11:10:34
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
opened
[Autoupdate] Выполняем лишнюю закачку файлов цифровой подписи
bug imported Priority-Medium
_From [Pavel.Pimenov@gmail.com](https://code.google.com/u/Pavel.Pimenov@gmail.com/) on July 15, 2013 23:19:36_ Лог на стороне сервера при старте приложения. [16/Jul/2013:01:02:07 +0400] "GET /etc/flylinkdc-config-r5xx.xml HTTP/1.1" 200 2029 "-" "FlylinkDC++ r502 -beta96 build 14508" [16/Jul/2013:01:02:43 +0400] "GET /update/5xx/beta/Update5_beta.xml HTTP/1.1" 200 153431 "-" " r502 -beta96 build 14508" [16/Jul/2013:01:02:43 +0400] "GET /update/5xx/beta/Update5_beta.sign HTTP/1.1" 200 128 "-" " r502 -beta96 build 14508" [16/Jul/2013:01:02:43 +0400] "GET /update/alluser/UpdateAU.xml HTTP/1.1" 200 1280 "-" " r502 -beta96 build 14508" [16/Jul/2013:01:02:43 +0400] "GET /update/alluser/UpdateAU.sign HTTP/1.1" 200 128 "-" " r502 -beta96 build 14508" 1. Файлы *.sign качаются сразу после *.xml даже в том случае если система в обновленном состоянии. скачивать sign файл только если пользователь захотел обновиться или есть кандидаты для обновления. 2. TODO в файле Update5_beta.xml на каждом файле очень избыточно храниться полный URL в тэге Url= Подумать и сократить постоянную часть переместив ее в корень xml постоянная части имеет вид: http://www.fly-server.ru/update/5xx/beta/ Если ее убрать размер xml сокращается с 153431 до 126084 но при этом придется изменить имя конфига... может и не стоит это делать - файл сейчас с сервера отдается в сжатом виде (прозрачно жмется для всех клиентов от Vista и выше) Может оптимальнее EmoPacks хранить в zip архиве а не россыпью? _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1080_
1.0
[Autoupdate] Выполняем лишнюю закачку файлов цифровой подписи - _From [Pavel.Pimenov@gmail.com](https://code.google.com/u/Pavel.Pimenov@gmail.com/) on July 15, 2013 23:19:36_ Лог на стороне сервера при старте приложения. [16/Jul/2013:01:02:07 +0400] "GET /etc/flylinkdc-config-r5xx.xml HTTP/1.1" 200 2029 "-" "FlylinkDC++ r502 -beta96 build 14508" [16/Jul/2013:01:02:43 +0400] "GET /update/5xx/beta/Update5_beta.xml HTTP/1.1" 200 153431 "-" " r502 -beta96 build 14508" [16/Jul/2013:01:02:43 +0400] "GET /update/5xx/beta/Update5_beta.sign HTTP/1.1" 200 128 "-" " r502 -beta96 build 14508" [16/Jul/2013:01:02:43 +0400] "GET /update/alluser/UpdateAU.xml HTTP/1.1" 200 1280 "-" " r502 -beta96 build 14508" [16/Jul/2013:01:02:43 +0400] "GET /update/alluser/UpdateAU.sign HTTP/1.1" 200 128 "-" " r502 -beta96 build 14508" 1. Файлы *.sign качаются сразу после *.xml даже в том случае если система в обновленном состоянии. скачивать sign файл только если пользователь захотел обновиться или есть кандидаты для обновления. 2. TODO в файле Update5_beta.xml на каждом файле очень избыточно храниться полный URL в тэге Url= Подумать и сократить постоянную часть переместив ее в корень xml постоянная части имеет вид: http://www.fly-server.ru/update/5xx/beta/ Если ее убрать размер xml сокращается с 153431 до 126084 но при этом придется изменить имя конфига... может и не стоит это делать - файл сейчас с сервера отдается в сжатом виде (прозрачно жмется для всех клиентов от Vista и выше) Может оптимальнее EmoPacks хранить в zip архиве а не россыпью? _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=1080_
non_process
выполняем лишнюю закачку файлов цифровой подписи from on july лог на стороне сервера при старте приложения get etc flylinkdc config xml http flylinkdc build get update beta beta xml http build get update beta beta sign http build get update alluser updateau xml http build get update alluser updateau sign http build файлы sign качаются сразу после xml даже в том случае если система в обновленном состоянии скачивать sign файл только если пользователь захотел обновиться или есть кандидаты для обновления todo в файле beta xml на каждом файле очень избыточно храниться полный url в тэге url подумать и сократить постоянную часть переместив ее в корень xml постоянная части имеет вид если ее убрать размер xml сокращается с до но при этом придется изменить имя конфига может и не стоит это делать файл сейчас с сервера отдается в сжатом виде прозрачно жмется для всех клиентов от vista и выше может оптимальнее emopacks хранить в zip архиве а не россыпью original issue
0
323,637
9,866,890,036
IssuesEvent
2019-06-21 08:48:52
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
outlook.live.com - see bug description
browser-firefox engine-gecko priority-critical
<!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: addon-reporter-firefox --> **URL**: https://outlook.live.com/mail/ **Browser / Version**: Firefox 68.0 **Operating System**: Windows 7 **Tested Another Browser**: No **Problem type**: Something else **Description**: site not opening **Steps to Reproduce**: Outlook is turning and turning, but doesn't open to mailbox [![Screenshot Description](https://webcompat.com/uploads/2019/6/74c26d9f-0fae-4458-8455-dde6c29d9023-thumb.jpg)](https://webcompat.com/uploads/2019/6/74c26d9f-0fae-4458-8455-dde6c29d9023.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
outlook.live.com - see bug description - <!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: addon-reporter-firefox --> **URL**: https://outlook.live.com/mail/ **Browser / Version**: Firefox 68.0 **Operating System**: Windows 7 **Tested Another Browser**: No **Problem type**: Something else **Description**: site not opening **Steps to Reproduce**: Outlook is turning and turning, but doesn't open to mailbox [![Screenshot Description](https://webcompat.com/uploads/2019/6/74c26d9f-0fae-4458-8455-dde6c29d9023-thumb.jpg)](https://webcompat.com/uploads/2019/6/74c26d9f-0fae-4458-8455-dde6c29d9023.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
outlook live com see bug description url browser version firefox operating system windows tested another browser no problem type something else description site not opening steps to reproduce outlook is turning and turning but doesn t open to mailbox browser configuration none from with ❤️
0
158,896
20,035,502,160
IssuesEvent
2022-02-02 11:25:44
kapseliboi/coronavirus-dashboard
https://api.github.com/repos/kapseliboi/coronavirus-dashboard
opened
CVE-2021-23368 (Medium) detected in postcss-7.0.21.tgz, postcss-7.0.32.tgz
security vulnerability
## CVE-2021-23368 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>postcss-7.0.21.tgz</b>, <b>postcss-7.0.32.tgz</b></p></summary> <p> <details><summary><b>postcss-7.0.21.tgz</b></p></summary> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.21.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.21.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/resolve-url-loader/node_modules/postcss/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.4.4.tgz (Root Library) - resolve-url-loader-3.1.2.tgz - :x: **postcss-7.0.21.tgz** (Vulnerable Library) </details> <details><summary><b>postcss-7.0.32.tgz</b></p></summary> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/postcss/package.json</p> <p> Dependency Hierarchy: - stylelint-processor-styled-components-1.10.0.tgz (Root Library) - :x: **postcss-7.0.32.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/coronavirus-dashboard/commit/4c0d18fe8dc22b989e9017e943fc5b7d5ef1d8d7">4c0d18fe8dc22b989e9017e943fc5b7d5ef1d8d7</a></p> <p>Found in base branch: <b>v3-development</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package postcss from 7.0.0 and before 8.2.10 are vulnerable to Regular Expression Denial of Service (ReDoS) during source map parsing. <p>Publish Date: 2021-04-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23368>CVE-2021-23368</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368</a></p> <p>Release Date: 2021-04-12</p> <p>Fix Resolution (postcss): 7.0.36</p> <p>Direct dependency fix Resolution (react-scripts): 4.0.0-next.116</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23368 (Medium) detected in postcss-7.0.21.tgz, postcss-7.0.32.tgz - ## CVE-2021-23368 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>postcss-7.0.21.tgz</b>, <b>postcss-7.0.32.tgz</b></p></summary> <p> <details><summary><b>postcss-7.0.21.tgz</b></p></summary> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.21.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.21.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/resolve-url-loader/node_modules/postcss/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.4.4.tgz (Root Library) - resolve-url-loader-3.1.2.tgz - :x: **postcss-7.0.21.tgz** (Vulnerable Library) </details> <details><summary><b>postcss-7.0.32.tgz</b></p></summary> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/postcss/package.json</p> <p> Dependency Hierarchy: - stylelint-processor-styled-components-1.10.0.tgz (Root Library) - :x: **postcss-7.0.32.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/coronavirus-dashboard/commit/4c0d18fe8dc22b989e9017e943fc5b7d5ef1d8d7">4c0d18fe8dc22b989e9017e943fc5b7d5ef1d8d7</a></p> <p>Found in base branch: <b>v3-development</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package postcss from 7.0.0 and before 8.2.10 are vulnerable to Regular Expression Denial of Service (ReDoS) during source map parsing. <p>Publish Date: 2021-04-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23368>CVE-2021-23368</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368</a></p> <p>Release Date: 2021-04-12</p> <p>Fix Resolution (postcss): 7.0.36</p> <p>Direct dependency fix Resolution (react-scripts): 4.0.0-next.116</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in postcss tgz postcss tgz cve medium severity vulnerability vulnerable libraries postcss tgz postcss tgz postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file package json path to vulnerable library node modules resolve url loader node modules postcss package json dependency hierarchy react scripts tgz root library resolve url loader tgz x postcss tgz vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file package json path to vulnerable library node modules postcss package json dependency hierarchy stylelint processor styled components tgz root library x postcss tgz vulnerable library found in head commit a href found in base branch development vulnerability details the package postcss from and before are vulnerable to regular expression denial of service redos during source map parsing publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss direct dependency fix resolution react scripts next step up your open source security game with whitesource
0
322,113
23,891,647,802
IssuesEvent
2022-09-08 11:58:40
COPRS/rs-issues
https://api.github.com/repos/COPRS/rs-issues
opened
[BUG][PRO] ingestion-trigger: if two regex are set for a same directory, files are compared randomly to either
bug documentation ivv pro CCB
<!-- Note: Please search to see if an issue already exists for the bug you encountered. Note: A closed bug can be reopened and affected to a new version of the software. --> **Environment:** <!-- - Delivery tag: release/0.1.0 - Platform: IVV Orange Cloud - Configuration: --> - Delivery tag: 1.4.0-rc1 - Platform: IVV Orange Cloud **Test:** <!-- - Name: TST_INFRA_DEP_orange - Traçability (requirements): NA --> - Name: TST_PRO_WF_S1L0 **Current Behavior:** <!-- A concise description of what you're experiencing. --> In the ingestion-trigger configuration, if two regex are configured for the same directory, files are compared randomly to either regex. Thus, some files may be ignored because they were not compared to the correct regex. **Expected Behavior:** <!-- A concise description of what you expected to happen. --> Each file should be compared to each regex in the ingestion-trigger configuration. **Steps To Reproduce:** <!-- Example: steps to reproduce the behavior: 1. In this environment... 2. With this config... 3. Run '...' 4. See error... --> 1 - In this environment, deploy an ingestion rs-core, with the following configuration for the ingestion-trigger: ```properties app.ingestion-trigger.ingestion-trigger.polling.inbox1.directory=https:/src1/NOMINAL/ app.ingestion-trigger.ingestion-trigger.polling.inbox1.matchRegex={regex1} ... app.ingestion-trigger.ingestion-trigger.polling.inbox2.directory=https:/src1/NOMINAL/ app.ingestion-trigger.ingestion-trigger.polling.inbox2.matchRegex={regex2} ... ``` 2 - In the mock source, put files in the directory that matches either regex 3 - Check the trigger logs. Some files will be ignored. **Workaround** <!-- A concise description of the first analysis. --> While it is possible to avoid this issue by specifying different folder for the source directory, a documentation warning would be welcomed. **Bug Generic Definition of Ready (DoR)** - [X] The affect version in which the bug has been found is mentioned - [X] The context and environment of the bug is detailed - [X] The description of the bug is clear and unambiguous - [X] The procedure (steps) to reproduce the bug is clearly detailed - [ ] The failed tests is linked to the bug : failed result % expected result - [ ] The tested User Story / features is linked to the bug - [ ] Logs are attached if available - [ ] A data set attached if available - [X] Category label is link to the bug <!-- infra, mon, pro, perfo, hmi, secu --> **Bug Generic Definition of Done (DoD)** - [ ] the modification implemented (the solution to fix the bug) is described in the bug. - [ ] Unit tests & Continuous integration performed - Test results available - Structural Test coverage reported by SONAR - [ ] Code committed in GIT with right tag or Analysis/Trade Off documentation up-to-date in reference-system-documentation repository - [ ] Code is compliant with coding rules (SONAR Report as evidence) - [ ] Acceptance criteria of the related User story are checked and Passed
1.0
[BUG][PRO] ingestion-trigger: if two regex are set for a same directory, files are compared randomly to either - <!-- Note: Please search to see if an issue already exists for the bug you encountered. Note: A closed bug can be reopened and affected to a new version of the software. --> **Environment:** <!-- - Delivery tag: release/0.1.0 - Platform: IVV Orange Cloud - Configuration: --> - Delivery tag: 1.4.0-rc1 - Platform: IVV Orange Cloud **Test:** <!-- - Name: TST_INFRA_DEP_orange - Traçability (requirements): NA --> - Name: TST_PRO_WF_S1L0 **Current Behavior:** <!-- A concise description of what you're experiencing. --> In the ingestion-trigger configuration, if two regex are configured for the same directory, files are compared randomly to either regex. Thus, some files may be ignored because they were not compared to the correct regex. **Expected Behavior:** <!-- A concise description of what you expected to happen. --> Each file should be compared to each regex in the ingestion-trigger configuration. **Steps To Reproduce:** <!-- Example: steps to reproduce the behavior: 1. In this environment... 2. With this config... 3. Run '...' 4. See error... --> 1 - In this environment, deploy an ingestion rs-core, with the following configuration for the ingestion-trigger: ```properties app.ingestion-trigger.ingestion-trigger.polling.inbox1.directory=https:/src1/NOMINAL/ app.ingestion-trigger.ingestion-trigger.polling.inbox1.matchRegex={regex1} ... app.ingestion-trigger.ingestion-trigger.polling.inbox2.directory=https:/src1/NOMINAL/ app.ingestion-trigger.ingestion-trigger.polling.inbox2.matchRegex={regex2} ... ``` 2 - In the mock source, put files in the directory that matches either regex 3 - Check the trigger logs. Some files will be ignored. **Workaround** <!-- A concise description of the first analysis. --> While it is possible to avoid this issue by specifying different folder for the source directory, a documentation warning would be welcomed. **Bug Generic Definition of Ready (DoR)** - [X] The affect version in which the bug has been found is mentioned - [X] The context and environment of the bug is detailed - [X] The description of the bug is clear and unambiguous - [X] The procedure (steps) to reproduce the bug is clearly detailed - [ ] The failed tests is linked to the bug : failed result % expected result - [ ] The tested User Story / features is linked to the bug - [ ] Logs are attached if available - [ ] A data set attached if available - [X] Category label is link to the bug <!-- infra, mon, pro, perfo, hmi, secu --> **Bug Generic Definition of Done (DoD)** - [ ] the modification implemented (the solution to fix the bug) is described in the bug. - [ ] Unit tests & Continuous integration performed - Test results available - Structural Test coverage reported by SONAR - [ ] Code committed in GIT with right tag or Analysis/Trade Off documentation up-to-date in reference-system-documentation repository - [ ] Code is compliant with coding rules (SONAR Report as evidence) - [ ] Acceptance criteria of the related User story are checked and Passed
non_process
ingestion trigger if two regex are set for a same directory files are compared randomly to either note please search to see if an issue already exists for the bug you encountered note a closed bug can be reopened and affected to a new version of the software environment delivery tag release platform ivv orange cloud configuration delivery tag platform ivv orange cloud test name tst infra dep orange traçability requirements na name tst pro wf current behavior in the ingestion trigger configuration if two regex are configured for the same directory files are compared randomly to either regex thus some files may be ignored because they were not compared to the correct regex expected behavior each file should be compared to each regex in the ingestion trigger configuration steps to reproduce example steps to reproduce the behavior in this environment with this config run see error in this environment deploy an ingestion rs core with the following configuration for the ingestion trigger properties app ingestion trigger ingestion trigger polling directory https nominal app ingestion trigger ingestion trigger polling matchregex app ingestion trigger ingestion trigger polling directory https nominal app ingestion trigger ingestion trigger polling matchregex in the mock source put files in the directory that matches either regex check the trigger logs some files will be ignored workaround while it is possible to avoid this issue by specifying different folder for the source directory a documentation warning would be welcomed bug generic definition of ready dor the affect version in which the bug has been found is mentioned the context and environment of the bug is detailed the description of the bug is clear and unambiguous the procedure steps to reproduce the bug is clearly detailed the failed tests is linked to the bug failed result expected result the tested user story features is linked to the bug logs are attached if available a data set attached if available category label is link to the bug bug generic definition of done dod the modification implemented the solution to fix the bug is described in the bug unit tests continuous integration performed test results available structural test coverage reported by sonar code committed in git with right tag or analysis trade off documentation up to date in reference system documentation repository code is compliant with coding rules sonar report as evidence acceptance criteria of the related user story are checked and passed
0
772,042
27,103,449,426
IssuesEvent
2023-02-15 10:20:28
FastcampusMini/mini-project
https://api.github.com/repos/FastcampusMini/mini-project
opened
페이징 기능 구현
For: API For: Backend Priority: High Status: In Progress Type: Feature
## Title 페이징 기능 구현 ## Description 신청한 상품내역 GET /user/addList 구매 내역 조회 GET /product/order/check 검색서비스 GET /search 메인페이지(금융상품) GET /products 추천상품 GET /products/recommended 위의 api들에서 페이징 구현 필요 하지만 다른분들이 맡은 부분이랑 추후 수정될수도 있고 conflict 가능성도 있어 일단은 검색서비스에만 페이징 적용후 기능 완성한 후 나머지는 추후 feature/paging 브랜치로 구현할 예정 ## Tasks 페이징이 적용된 api 목록 - [ ] 검색서비스 GET /search - [ ] 구매 내역 조회 GET /product/order/check - [ ] 신청한 상품내역 GET /user/addList - [ ] 메인페이지(금융상품) GET /products - [ ] 추천상품 GET /products/recommended
1.0
페이징 기능 구현 - ## Title 페이징 기능 구현 ## Description 신청한 상품내역 GET /user/addList 구매 내역 조회 GET /product/order/check 검색서비스 GET /search 메인페이지(금융상품) GET /products 추천상품 GET /products/recommended 위의 api들에서 페이징 구현 필요 하지만 다른분들이 맡은 부분이랑 추후 수정될수도 있고 conflict 가능성도 있어 일단은 검색서비스에만 페이징 적용후 기능 완성한 후 나머지는 추후 feature/paging 브랜치로 구현할 예정 ## Tasks 페이징이 적용된 api 목록 - [ ] 검색서비스 GET /search - [ ] 구매 내역 조회 GET /product/order/check - [ ] 신청한 상품내역 GET /user/addList - [ ] 메인페이지(금융상품) GET /products - [ ] 추천상품 GET /products/recommended
non_process
페이징 기능 구현 title 페이징 기능 구현 description 신청한 상품내역 get user addlist 구매 내역 조회 get product order check 검색서비스 get search 메인페이지 금융상품 get products 추천상품 get products recommended 위의 api들에서 페이징 구현 필요 하지만 다른분들이 맡은 부분이랑 추후 수정될수도 있고 conflict 가능성도 있어 일단은 검색서비스에만 페이징 적용후 기능 완성한 후 나머지는 추후 feature paging 브랜치로 구현할 예정 tasks 페이징이 적용된 api 목록 검색서비스 get search 구매 내역 조회 get product order check 신청한 상품내역 get user addlist 메인페이지 금융상품 get products 추천상품 get products recommended
0
67,018
3,265,297,472
IssuesEvent
2015-10-22 15:40:09
cs2103aug2015-w15-3j/main
https://api.github.com/repos/cs2103aug2015-w15-3j/main
closed
Advanced features of autocomplete
priority.medium type.enhancement
## Features - [ ] autocomplete displayed ID to details of task for `EDIT` command - [x] update view while autocompleting task name - [x] pressing up will retrieve previous used command
1.0
Advanced features of autocomplete - ## Features - [ ] autocomplete displayed ID to details of task for `EDIT` command - [x] update view while autocompleting task name - [x] pressing up will retrieve previous used command
non_process
advanced features of autocomplete features autocomplete displayed id to details of task for edit command update view while autocompleting task name pressing up will retrieve previous used command
0
127,910
17,378,571,106
IssuesEvent
2021-07-31 07:38:19
Joystream/atlas
https://api.github.com/repos/Joystream/atlas
opened
Merging sections into one content section component
design
There are components with simmilar build which are channel section, video section and carousel - after discussion with klaudiusz we decided to merge it into one component: content section.
1.0
Merging sections into one content section component - There are components with simmilar build which are channel section, video section and carousel - after discussion with klaudiusz we decided to merge it into one component: content section.
non_process
merging sections into one content section component there are components with simmilar build which are channel section video section and carousel after discussion with klaudiusz we decided to merge it into one component content section
0
224,535
7,471,635,660
IssuesEvent
2018-04-03 09:53:26
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Thread Requirements on RFC6282
area: Networking enhancement priority: high
**_Reported by Gajinder Vij:_** The Thread Protocol imposes the following requirements on RFC6282: RFC6282.2: Specific Updates to RFC 4944 RFC6282.3: IPv6 Header Compression RFC6282.3.1: LOWPAN_IPHC Encoding Format RFC6282.3.1.1: Base Format RFC6282.3.1.2: Context Identifier Extension RFC6282.3.2: IPV6 Header Encoding RFC6282.3.2.1: Traffic Class and Flow Label Compression RFC6282.3.2.2: Deriving IID's from the encapsulating Header RFC6282.3.2.3: Stateless Multicast Address Compression RFC6282.3.2.4: Stateful Multicast Address Compression RFC6282.4: IPv6 Next Header Compression RFC6282.4.1: LOWPAN_NHC Format RFC6282.4.2: IPv6 Extension Header Compression � Modified: When there is a single trailing Pad1 or PadN option of 7 octets or less and the containing header is a multiple of 8 octets, the trailing Pad1 or PadN option SHOULD be elided by the compressor. RFC6282.4.3: UDP Header Compression RFC6282.4.3.1: Compressing UDP Ports RFC6282.4.3.2: Compressing UDP Checksum RFC6282.4.3.3: UDP LOWPAN_NHC Format RFC6282.5: IANA Considerations (Imported from Jira ZEP-839)
1.0
Thread Requirements on RFC6282 - **_Reported by Gajinder Vij:_** The Thread Protocol imposes the following requirements on RFC6282: RFC6282.2: Specific Updates to RFC 4944 RFC6282.3: IPv6 Header Compression RFC6282.3.1: LOWPAN_IPHC Encoding Format RFC6282.3.1.1: Base Format RFC6282.3.1.2: Context Identifier Extension RFC6282.3.2: IPV6 Header Encoding RFC6282.3.2.1: Traffic Class and Flow Label Compression RFC6282.3.2.2: Deriving IID's from the encapsulating Header RFC6282.3.2.3: Stateless Multicast Address Compression RFC6282.3.2.4: Stateful Multicast Address Compression RFC6282.4: IPv6 Next Header Compression RFC6282.4.1: LOWPAN_NHC Format RFC6282.4.2: IPv6 Extension Header Compression � Modified: When there is a single trailing Pad1 or PadN option of 7 octets or less and the containing header is a multiple of 8 octets, the trailing Pad1 or PadN option SHOULD be elided by the compressor. RFC6282.4.3: UDP Header Compression RFC6282.4.3.1: Compressing UDP Ports RFC6282.4.3.2: Compressing UDP Checksum RFC6282.4.3.3: UDP LOWPAN_NHC Format RFC6282.5: IANA Considerations (Imported from Jira ZEP-839)
non_process
thread requirements on reported by gajinder vij the thread protocol imposes the following requirements on specific updates to rfc header compression lowpan iphc encoding format base format context identifier extension header encoding traffic class and flow label compression deriving iid s from the encapsulating header stateless multicast address compression stateful multicast address compression next header compression lowpan nhc format extension header compression � modified when there is a single trailing or padn option of octets or less and the containing header is a multiple of octets the trailing or padn option should be elided by the compressor udp header compression compressing udp ports compressing udp checksum udp lowpan nhc format iana considerations imported from jira zep
0
11,738
14,581,585,741
IssuesEvent
2020-12-18 10:57:19
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
multiple identical compound unique fields with different types not generating multiple types
bug/2-confirmed process/candidate team/client tech/typescript topic: types
<!-- Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client. Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports --> ## Bug description I've got 2 models with the following structure, both use the same unique field mapping on type and token, but use different enum types. There is only 1 type generated for these compounded fields (or the other is overridden), but it's causing typescript to complain when querying via the compounded unique fields since the types are incorrect. ```prisma model Token { id String @id @default(cuid()) type TokenType token String created_at DateTime @default(now()) updated_at DateTime @updatedAt @@unique([type, token]) } model UserCredential { id String @id @default(cuid()) type CredentialType token String @unique created_at DateTime @default(now()) updated_at DateTime @updatedAt user User @relation(fields: [user_id], references: [id]) user_id String @@unique([type, token]) } ``` Generated Type ```typescript export type TypeTokenCompoundUniqueInput = { type: CredentialType token: string } ``` ## How to reproduce Create two models with the same compounded unique constraints that both share the same name then run prisma generate ## Expected behaviour I'd expect two types to be exports, one for each of the compounded unique constraints that follow the correct types ## Environment & setup <!-- In which environment does the problem occur --> - OS: Windows - Database: PostgreSQL - Node.js version: v12.16.2 - Prisma version: 6.12.1
1.0
multiple identical compound unique fields with different types not generating multiple types - <!-- Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client. Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports --> ## Bug description I've got 2 models with the following structure, both use the same unique field mapping on type and token, but use different enum types. There is only 1 type generated for these compounded fields (or the other is overridden), but it's causing typescript to complain when querying via the compounded unique fields since the types are incorrect. ```prisma model Token { id String @id @default(cuid()) type TokenType token String created_at DateTime @default(now()) updated_at DateTime @updatedAt @@unique([type, token]) } model UserCredential { id String @id @default(cuid()) type CredentialType token String @unique created_at DateTime @default(now()) updated_at DateTime @updatedAt user User @relation(fields: [user_id], references: [id]) user_id String @@unique([type, token]) } ``` Generated Type ```typescript export type TypeTokenCompoundUniqueInput = { type: CredentialType token: string } ``` ## How to reproduce Create two models with the same compounded unique constraints that both share the same name then run prisma generate ## Expected behaviour I'd expect two types to be exports, one for each of the compounded unique constraints that follow the correct types ## Environment & setup <!-- In which environment does the problem occur --> - OS: Windows - Database: PostgreSQL - Node.js version: v12.16.2 - Prisma version: 6.12.1
process
multiple identical compound unique fields with different types not generating multiple types thanks for helping us improve prisma 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by setting the debug environment variable and enabling additional logging output in prisma client learn more about writing proper bug reports here bug description i ve got models with the following structure both use the same unique field mapping on type and token but use different enum types there is only type generated for these compounded fields or the other is overridden but it s causing typescript to complain when querying via the compounded unique fields since the types are incorrect prisma model token id string id default cuid type tokentype token string created at datetime default now updated at datetime updatedat unique model usercredential id string id default cuid type credentialtype token string unique created at datetime default now updated at datetime updatedat user user relation fields references user id string unique generated type typescript export type typetokencompounduniqueinput type credentialtype token string how to reproduce create two models with the same compounded unique constraints that both share the same name then run prisma generate expected behaviour i d expect two types to be exports one for each of the compounded unique constraints that follow the correct types environment setup os windows database postgresql node js version prisma version
1
445,484
12,831,608,845
IssuesEvent
2020-07-07 05:50:46
minio/minio-js
https://api.github.com/repos/minio/minio-js
closed
Error when using minio-js in ant design , TypeError: _fs.default.stat is not a function
community priority: medium
## An error occurs when I use minio-js in the ant design project, the error message is TypeError: _fs.default.stat is not a function ![{4EDE6528-708C-49AA-B941-8752A6F83B62}_20200702144145](https://user-images.githubusercontent.com/38524062/86325152-303fc280-bc72-11ea-82f6-5a0d779ebb07.jpg)
1.0
Error when using minio-js in ant design , TypeError: _fs.default.stat is not a function - ## An error occurs when I use minio-js in the ant design project, the error message is TypeError: _fs.default.stat is not a function ![{4EDE6528-708C-49AA-B941-8752A6F83B62}_20200702144145](https://user-images.githubusercontent.com/38524062/86325152-303fc280-bc72-11ea-82f6-5a0d779ebb07.jpg)
non_process
error when using minio js in ant design typeerror fs default stat is not a function an error occurs when i use minio js in the ant design project the error message is typeerror fs default stat is not a function
0
15,942
20,161,363,418
IssuesEvent
2022-02-09 21:57:51
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
opened
Further small Electrotech tweaks
Process Heating
I didn't wnat to keep taging on to the exisiting issue... NON-EAF It looks like that Electrical power input doesn't do anything except cause errors. Confirm and then either hide field or remove from validation. Electrical Heat Delivered is calcuated and can be negative and that is bad. Make it so if negative it shows up as zero and show an error "More heat than necessary is being delivered via burners. Check fuel inputs or estimate other losses" ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/870f433a-4be7-468b-9466-daecfd0d5b05)
1.0
Further small Electrotech tweaks - I didn't wnat to keep taging on to the exisiting issue... NON-EAF It looks like that Electrical power input doesn't do anything except cause errors. Confirm and then either hide field or remove from validation. Electrical Heat Delivered is calcuated and can be negative and that is bad. Make it so if negative it shows up as zero and show an error "More heat than necessary is being delivered via burners. Check fuel inputs or estimate other losses" ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/870f433a-4be7-468b-9466-daecfd0d5b05)
process
further small electrotech tweaks i didn t wnat to keep taging on to the exisiting issue non eaf it looks like that electrical power input doesn t do anything except cause errors confirm and then either hide field or remove from validation electrical heat delivered is calcuated and can be negative and that is bad make it so if negative it shows up as zero and show an error more heat than necessary is being delivered via burners check fuel inputs or estimate other losses
1
15,371
19,552,766,008
IssuesEvent
2022-01-03 01:49:57
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
Support json manifest
TYPE: enhancement SYSTEM: resource processing AREA: server health-monitor FREQUENCY: level 1 STATE: Stale
https://developer.mozilla.org/en-US/docs/Web/Manifest https://www.trulia.com contains json manifest file, but with non-standard content type: ```content-type: text/cache-manifest; charset=UTF-8```
1.0
Support json manifest - https://developer.mozilla.org/en-US/docs/Web/Manifest https://www.trulia.com contains json manifest file, but with non-standard content type: ```content-type: text/cache-manifest; charset=UTF-8```
process
support json manifest contains json manifest file but with non standard content type content type text cache manifest charset utf
1
16,955
22,307,990,839
IssuesEvent
2022-06-13 14:34:21
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Acceptance tests failing on docker timeout
bug P1 process
### Description Acceptance tests started failing on ``` node:events:505 throw er; // Unhandled 'error' event ^ Error: connect ECONNREFUSED 127.0.0.1:5600 at TCPConnectWrap.afterConnect [as oncomplete] (node:net:1187:16) Emitted 'error' event on Socket instance at: at emitErrorNT (node:internal/streams/destroy:157:8) at emitErrorCloseNT (node:internal/streams/destroy:122:3) at processTicksAndRejections (node:internal/process/task_queues:83:21) { errno: -111, code: 'ECONNREFUSED', syscall: 'connect', address: '127.0.0.1', port: 5600 } ``` This results in a ``` Error: Timeout of 240000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/home/runner/work/hedera-json-rpc-relay/hedera-json-rpc-relay/packages/server/tests/acceptance.spec.ts) ``` ### Steps to reproduce 1. Run acceptance tests in github action ### Additional context _No response_ ### Hedera network other ### Version v0.2.0-SNAPSHOT ### Operating system _No response_
1.0
Acceptance tests failing on docker timeout - ### Description Acceptance tests started failing on ``` node:events:505 throw er; // Unhandled 'error' event ^ Error: connect ECONNREFUSED 127.0.0.1:5600 at TCPConnectWrap.afterConnect [as oncomplete] (node:net:1187:16) Emitted 'error' event on Socket instance at: at emitErrorNT (node:internal/streams/destroy:157:8) at emitErrorCloseNT (node:internal/streams/destroy:122:3) at processTicksAndRejections (node:internal/process/task_queues:83:21) { errno: -111, code: 'ECONNREFUSED', syscall: 'connect', address: '127.0.0.1', port: 5600 } ``` This results in a ``` Error: Timeout of 240000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/home/runner/work/hedera-json-rpc-relay/hedera-json-rpc-relay/packages/server/tests/acceptance.spec.ts) ``` ### Steps to reproduce 1. Run acceptance tests in github action ### Additional context _No response_ ### Hedera network other ### Version v0.2.0-SNAPSHOT ### Operating system _No response_
process
acceptance tests failing on docker timeout description acceptance tests started failing on node events throw er unhandled error event error connect econnrefused at tcpconnectwrap afterconnect node net emitted error event on socket instance at at emiterrornt node internal streams destroy at emiterrorclosent node internal streams destroy at processticksandrejections node internal process task queues errno code econnrefused syscall connect address port this results in a error timeout of exceeded for async tests and hooks ensure done is called if returning a promise ensure it resolves home runner work hedera json rpc relay hedera json rpc relay packages server tests acceptance spec ts steps to reproduce run acceptance tests in github action additional context no response hedera network other version snapshot operating system no response
1
12,538
14,972,535,265
IssuesEvent
2021-01-27 23:02:58
BootBlock/FileSieve
https://api.github.com/repos/BootBlock/FileSieve
closed
Limit how deep sub-folder processing can go
processing
Within a Source Item, add a numerical property that limits how far "down / deep" sub-folder processing can go. This'll require changes to the Directory.**GetFiles**() method(s). ---- *Thanks to Phil P. for suggesting!*
1.0
Limit how deep sub-folder processing can go - Within a Source Item, add a numerical property that limits how far "down / deep" sub-folder processing can go. This'll require changes to the Directory.**GetFiles**() method(s). ---- *Thanks to Phil P. for suggesting!*
process
limit how deep sub folder processing can go within a source item add a numerical property that limits how far down deep sub folder processing can go this ll require changes to the directory getfiles method s thanks to phil p for suggesting
1
13,050
15,387,992,049
IssuesEvent
2021-03-03 10:12:56
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Consider not abbreviating “Integer”
kind/discussion process/candidate team/migrations topic: types
I just started trying Prisma and got to this code snippet in the Quickstart. ``` model Post { post_id Int @id @default(autoincrement()) content String? created_at DateTime @default(now()) title String author User? } ``` Why is “Integer” abbreviated to “Int”? I looked up the data types and it appears “Integer” is the only one that is abbreviated, which is unexpected/inconsistent. If it’s too difficult to change it now to be consistent, could it be aliased so both “Int” and “Integer” work?
1.0
Consider not abbreviating “Integer” - I just started trying Prisma and got to this code snippet in the Quickstart. ``` model Post { post_id Int @id @default(autoincrement()) content String? created_at DateTime @default(now()) title String author User? } ``` Why is “Integer” abbreviated to “Int”? I looked up the data types and it appears “Integer” is the only one that is abbreviated, which is unexpected/inconsistent. If it’s too difficult to change it now to be consistent, could it be aliased so both “Int” and “Integer” work?
process
consider not abbreviating “integer” i just started trying prisma and got to this code snippet in the quickstart model post post id int id default autoincrement content string created at datetime default now title string author user why is “integer” abbreviated to “int” i looked up the data types and it appears “integer” is the only one that is abbreviated which is unexpected inconsistent if it’s too difficult to change it now to be consistent could it be aliased so both “int” and “integer” work
1
4,228
7,181,225,812
IssuesEvent
2018-02-01 03:35:05
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
block enhanced blooms are wrong at xxx90 in each case less than 700000
apps-blockScrape status-inprocess type-bug
Because of the way I was running the bloom creator, I lost blooms from blocks xxxx90 through xxxx99. It ran to the end at xxx100, then re-ran at xxxx99 which blew away 90-98. To fix: re-write blocks between 90-99 for blocks less than 700,000
1.0
block enhanced blooms are wrong at xxx90 in each case less than 700000 - Because of the way I was running the bloom creator, I lost blooms from blocks xxxx90 through xxxx99. It ran to the end at xxx100, then re-ran at xxxx99 which blew away 90-98. To fix: re-write blocks between 90-99 for blocks less than 700,000
process
block enhanced blooms are wrong at in each case less than because of the way i was running the bloom creator i lost blooms from blocks through it ran to the end at then re ran at which blew away to fix re write blocks between for blocks less than
1
5,632
8,483,516,359
IssuesEvent
2018-10-25 21:51:47
Rokid/ShadowNode
https://api.github.com/repos/Rokid/ShadowNode
closed
child_process: crashed while sending data
bug child_process
```js var isChild = process.env.IS_CHILD === 'true' function ref() { setInterval(() => { }, 10000) } function runWithChild() { process.on('message', chunk => { console.log('on parent message', chunk) }) } function runWithParent() { var path = require('path') var child_process = require('child_process') var selfPath = path.resolve(__dirname, __filename) var child = child_process.fork(selfPath, ['~~parent~~'], { env: { IS_CHILD: 'true', } }) child.on('message', chunk => { console.log('on child message', chunk) }) child.on('disconnect', () => { console.log('child disconnected') }) setInterval(() => { child.send({ cloud: false, intent: 'play_random', slots: { keyword: { pinyin: '', type: 'keyword', value: '歌' }, want: { pinyin: '', type: 'want', value: '要' }, me: { pinyin: '', type: 'me', value: '我' }, play5: { pinyin: '', type: 'play5', value: '听' } }, pattern: '^($trigger)?($please)?($you|$me)?($want)?($at)?($room)?($inside)?($play5)($other)?[($adj)($de)]?($keyword)($ah)?$', asr: '我要听歌', appId: 'xxxxxxxxxxxxxxxxxxx', appName: 'xxx音乐' }) }, 30) console.log('forked child') } isChild ? runWithChild() : runWithParent() ref() ``` after running the above code for a while, the child exit with below error: ```shell TypeError: Invalid UTF-8 string child disconnected Error: ERR_IPC_CHANNEL_CLOSED at anonymous (/data/test.js:26:16) ``` this is caused by tcp multi packet, We should handle the case of a message in multiple packets in ipc
1.0
child_process: crashed while sending data - ```js var isChild = process.env.IS_CHILD === 'true' function ref() { setInterval(() => { }, 10000) } function runWithChild() { process.on('message', chunk => { console.log('on parent message', chunk) }) } function runWithParent() { var path = require('path') var child_process = require('child_process') var selfPath = path.resolve(__dirname, __filename) var child = child_process.fork(selfPath, ['~~parent~~'], { env: { IS_CHILD: 'true', } }) child.on('message', chunk => { console.log('on child message', chunk) }) child.on('disconnect', () => { console.log('child disconnected') }) setInterval(() => { child.send({ cloud: false, intent: 'play_random', slots: { keyword: { pinyin: '', type: 'keyword', value: '歌' }, want: { pinyin: '', type: 'want', value: '要' }, me: { pinyin: '', type: 'me', value: '我' }, play5: { pinyin: '', type: 'play5', value: '听' } }, pattern: '^($trigger)?($please)?($you|$me)?($want)?($at)?($room)?($inside)?($play5)($other)?[($adj)($de)]?($keyword)($ah)?$', asr: '我要听歌', appId: 'xxxxxxxxxxxxxxxxxxx', appName: 'xxx音乐' }) }, 30) console.log('forked child') } isChild ? runWithChild() : runWithParent() ref() ``` after running the above code for a while, the child exit with below error: ```shell TypeError: Invalid UTF-8 string child disconnected Error: ERR_IPC_CHANNEL_CLOSED at anonymous (/data/test.js:26:16) ``` this is caused by tcp multi packet, We should handle the case of a message in multiple packets in ipc
process
child process crashed while sending data js var ischild process env is child true function ref setinterval function runwithchild process on message chunk console log on parent message chunk function runwithparent var path require path var child process require child process var selfpath path resolve dirname filename var child child process fork selfpath env is child true child on message chunk console log on child message chunk child on disconnect console log child disconnected setinterval child send cloud false intent play random slots keyword pinyin type keyword value 歌 want pinyin type want value 要 me pinyin type me value 我 pinyin type value 听 pattern trigger please you me want at room inside other keyword ah asr 我要听歌 appid xxxxxxxxxxxxxxxxxxx appname xxx音乐 console log forked child ischild runwithchild runwithparent ref after running the above code for a while the child exit with below error shell typeerror invalid utf string child disconnected error err ipc channel closed at anonymous data test js this is caused by tcp multi packet we should handle the case of a message in multiple packets in ipc
1
28,710
11,671,975,372
IssuesEvent
2020-03-04 05:07:45
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
Share authentication cookie between two sites produces exception caused by ISecurityStampValidator is missing
Needs: Attention :wave: area-security
I have one API and one frontend app where I share the data protection keys using the file system. This is how my apps is configured. **Frontend**: ```c# public void ConfigureServices(IServiceCollection services) { services .AddDataProtection() .SetApplicationName("SharedCookieApp") .PersistKeysToFileSystem(new DirectoryInfo("/dataprotection")); services .AddAuthentication(options => { options.DefaultScheme = IdentityConstants.ApplicationScheme; options.DefaultSignInScheme = IdentityConstants.ExternalScheme; }) .AddOAuth("GitHub", "GitHub", o => { ... }) .AddIdentityCookies(options => { options.ApplicationCookie.Configure(o => { o.Cookie.Name = ".AspNet.SharedCookie"; o.Cookie.Domain = ".example.app"; o.LoginPath = new PathString("/Account/Login/"); o.AccessDeniedPath = new PathString("/Account/Forbidden/"); }); }); services.AddIdentityCore<User>(options => { options.Password.RequireDigit = true; options.Password.RequiredLength = 8; options.Password.RequireNonAlphanumeric = false; options.Password.RequireUppercase = true; options.Password.RequireLowercase = true; }) .AddRoles<Raven.Identity.IdentityRole>() .AddRoleStore<RoleStore<Raven.Identity.IdentityRole>>() .AddUserStore<UserStore<User, Raven.Identity.IdentityRole>>() .AddSignInManager(); } public void Configure(IApplicationBuilder app, IWebHostEnvironment env) { if (env.IsDevelopment()) { app.UseDeveloperExceptionPage(); } app.UseForwardedHeaders(); app.UseRouting(); app.UseCookiePolicy(); app.UseAuthentication(); app.UseAuthorization(); ... } ``` And from the API I need to use the authorize attribute to make sure that the user is authenticated when making a request. **API**: ```c# public void ConfigureServices(IServiceCollection services) { services.AddDataProtection() .SetApplicationName("SharedCookieApp") .PersistKeysToFileSystem(new DirectoryInfo("/dataprotection")); services.AddAuthentication(options => { options.DefaultScheme = IdentityConstants.ApplicationScheme; }) .AddIdentityCookies(options => { options.ApplicationCookie.Configure(a => { a.Cookie.Name = ".AspNet.SharedCookie"; a.Cookie.Domain = ".example.app"; }); }); services.AddControllers(); } public void Configure(IApplicationBuilder app, IWebHostEnvironment env) { if (env.IsDevelopment()) { app.UseDeveloperExceptionPage(); } app.UseHttpsRedirection(); app.UseCookiePolicy(); app.UseAuthentication(); app.UseRouting(); app.UseAuthorization(); app.UseEndpoints(endpoints => { endpoints.MapControllers(); }); } ``` The authentication works correct but when I try to access the api I get the following exception. ``` fail: Microsoft.AspNetCore.Server.Kestrel[13] Connection id "0HLTP6L6AV4FV", Request id "0HLTP6L6AV4FV:00000001": An unhandled exception was thrown by the application. System.InvalidOperationException: No service for type 'Microsoft.AspNetCore.Identity.ISecurityStampValidator' has been registered. at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService(IServiceProvider provider, Type serviceType) at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService[T](IServiceProvider provider) at Microsoft.AspNetCore.Identity.SecurityStampValidator.ValidateAsync[TValidator](CookieValidatePrincipalContext context) at Microsoft.AspNetCore.Identity.SecurityStampValidator.ValidatePrincipalAsync(CookieValidatePrincipalContext context) at Microsoft.AspNetCore.Authentication.Cookies.CookieAuthenticationEvents.ValidatePrincipal(CookieValidatePrincipalContext context) at Microsoft.AspNetCore.Authentication.Cookies.CookieAuthenticationHandler.HandleAuthenticateAsync() at Microsoft.AspNetCore.Authentication.AuthenticationHandler`1.AuthenticateAsync() at Microsoft.AspNetCore.Authentication.AuthenticationService.AuthenticateAsync(HttpContext context, String scheme) at Microsoft.AspNetCore.Authentication.AuthenticationMiddleware.Invoke(HttpContext context) at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.Http.HttpProtocol.ProcessRequests[TContext](IHttpApplication`1 application) ``` I can't figure out what the problem is so I appreciate all the help I can get.
True
Share authentication cookie between two sites produces exception caused by ISecurityStampValidator is missing - I have one API and one frontend app where I share the data protection keys using the file system. This is how my apps is configured. **Frontend**: ```c# public void ConfigureServices(IServiceCollection services) { services .AddDataProtection() .SetApplicationName("SharedCookieApp") .PersistKeysToFileSystem(new DirectoryInfo("/dataprotection")); services .AddAuthentication(options => { options.DefaultScheme = IdentityConstants.ApplicationScheme; options.DefaultSignInScheme = IdentityConstants.ExternalScheme; }) .AddOAuth("GitHub", "GitHub", o => { ... }) .AddIdentityCookies(options => { options.ApplicationCookie.Configure(o => { o.Cookie.Name = ".AspNet.SharedCookie"; o.Cookie.Domain = ".example.app"; o.LoginPath = new PathString("/Account/Login/"); o.AccessDeniedPath = new PathString("/Account/Forbidden/"); }); }); services.AddIdentityCore<User>(options => { options.Password.RequireDigit = true; options.Password.RequiredLength = 8; options.Password.RequireNonAlphanumeric = false; options.Password.RequireUppercase = true; options.Password.RequireLowercase = true; }) .AddRoles<Raven.Identity.IdentityRole>() .AddRoleStore<RoleStore<Raven.Identity.IdentityRole>>() .AddUserStore<UserStore<User, Raven.Identity.IdentityRole>>() .AddSignInManager(); } public void Configure(IApplicationBuilder app, IWebHostEnvironment env) { if (env.IsDevelopment()) { app.UseDeveloperExceptionPage(); } app.UseForwardedHeaders(); app.UseRouting(); app.UseCookiePolicy(); app.UseAuthentication(); app.UseAuthorization(); ... } ``` And from the API I need to use the authorize attribute to make sure that the user is authenticated when making a request. **API**: ```c# public void ConfigureServices(IServiceCollection services) { services.AddDataProtection() .SetApplicationName("SharedCookieApp") .PersistKeysToFileSystem(new DirectoryInfo("/dataprotection")); services.AddAuthentication(options => { options.DefaultScheme = IdentityConstants.ApplicationScheme; }) .AddIdentityCookies(options => { options.ApplicationCookie.Configure(a => { a.Cookie.Name = ".AspNet.SharedCookie"; a.Cookie.Domain = ".example.app"; }); }); services.AddControllers(); } public void Configure(IApplicationBuilder app, IWebHostEnvironment env) { if (env.IsDevelopment()) { app.UseDeveloperExceptionPage(); } app.UseHttpsRedirection(); app.UseCookiePolicy(); app.UseAuthentication(); app.UseRouting(); app.UseAuthorization(); app.UseEndpoints(endpoints => { endpoints.MapControllers(); }); } ``` The authentication works correct but when I try to access the api I get the following exception. ``` fail: Microsoft.AspNetCore.Server.Kestrel[13] Connection id "0HLTP6L6AV4FV", Request id "0HLTP6L6AV4FV:00000001": An unhandled exception was thrown by the application. System.InvalidOperationException: No service for type 'Microsoft.AspNetCore.Identity.ISecurityStampValidator' has been registered. at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService(IServiceProvider provider, Type serviceType) at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetRequiredService[T](IServiceProvider provider) at Microsoft.AspNetCore.Identity.SecurityStampValidator.ValidateAsync[TValidator](CookieValidatePrincipalContext context) at Microsoft.AspNetCore.Identity.SecurityStampValidator.ValidatePrincipalAsync(CookieValidatePrincipalContext context) at Microsoft.AspNetCore.Authentication.Cookies.CookieAuthenticationEvents.ValidatePrincipal(CookieValidatePrincipalContext context) at Microsoft.AspNetCore.Authentication.Cookies.CookieAuthenticationHandler.HandleAuthenticateAsync() at Microsoft.AspNetCore.Authentication.AuthenticationHandler`1.AuthenticateAsync() at Microsoft.AspNetCore.Authentication.AuthenticationService.AuthenticateAsync(HttpContext context, String scheme) at Microsoft.AspNetCore.Authentication.AuthenticationMiddleware.Invoke(HttpContext context) at Microsoft.AspNetCore.Server.Kestrel.Core.Internal.Http.HttpProtocol.ProcessRequests[TContext](IHttpApplication`1 application) ``` I can't figure out what the problem is so I appreciate all the help I can get.
non_process
share authentication cookie between two sites produces exception caused by isecuritystampvalidator is missing i have one api and one frontend app where i share the data protection keys using the file system this is how my apps is configured frontend c public void configureservices iservicecollection services services adddataprotection setapplicationname sharedcookieapp persistkeystofilesystem new directoryinfo dataprotection services addauthentication options options defaultscheme identityconstants applicationscheme options defaultsigninscheme identityconstants externalscheme addoauth github github o addidentitycookies options options applicationcookie configure o o cookie name aspnet sharedcookie o cookie domain example app o loginpath new pathstring account login o accessdeniedpath new pathstring account forbidden services addidentitycore options options password requiredigit true options password requiredlength options password requirenonalphanumeric false options password requireuppercase true options password requirelowercase true addroles addrolestore adduserstore addsigninmanager public void configure iapplicationbuilder app iwebhostenvironment env if env isdevelopment app usedeveloperexceptionpage app useforwardedheaders app userouting app usecookiepolicy app useauthentication app useauthorization and from the api i need to use the authorize attribute to make sure that the user is authenticated when making a request api c public void configureservices iservicecollection services services adddataprotection setapplicationname sharedcookieapp persistkeystofilesystem new directoryinfo dataprotection services addauthentication options options defaultscheme identityconstants applicationscheme addidentitycookies options options applicationcookie configure a a cookie name aspnet sharedcookie a cookie domain example app services addcontrollers public void configure iapplicationbuilder app iwebhostenvironment env if env isdevelopment app usedeveloperexceptionpage app usehttpsredirection app usecookiepolicy app useauthentication app userouting app useauthorization app useendpoints endpoints endpoints mapcontrollers the authentication works correct but when i try to access the api i get the following exception fail microsoft aspnetcore server kestrel connection id request id an unhandled exception was thrown by the application system invalidoperationexception no service for type microsoft aspnetcore identity isecuritystampvalidator has been registered at microsoft extensions dependencyinjection serviceproviderserviceextensions getrequiredservice iserviceprovider provider type servicetype at microsoft extensions dependencyinjection serviceproviderserviceextensions getrequiredservice iserviceprovider provider at microsoft aspnetcore identity securitystampvalidator validateasync cookievalidateprincipalcontext context at microsoft aspnetcore identity securitystampvalidator validateprincipalasync cookievalidateprincipalcontext context at microsoft aspnetcore authentication cookies cookieauthenticationevents validateprincipal cookievalidateprincipalcontext context at microsoft aspnetcore authentication cookies cookieauthenticationhandler handleauthenticateasync at microsoft aspnetcore authentication authenticationhandler authenticateasync at microsoft aspnetcore authentication authenticationservice authenticateasync httpcontext context string scheme at microsoft aspnetcore authentication authenticationmiddleware invoke httpcontext context at microsoft aspnetcore server kestrel core internal http httpprotocol processrequests ihttpapplication application i can t figure out what the problem is so i appreciate all the help i can get
0
8,841
11,948,369,637
IssuesEvent
2020-04-03 11:43:43
digitalmethodsinitiative/4cat
https://api.github.com/repos/digitalmethodsinitiative/4cat
closed
Make workers do less when starting
(mostly) back-end processors
Scrapers and other workers have a significant delay in starting up (as a worker) because they run the same bootstrapping code as the 'full' daemon, which builds a map of all workers, initializes everything, etc. Either they should run less startup code or the general startup code should be reduced.
1.0
Make workers do less when starting - Scrapers and other workers have a significant delay in starting up (as a worker) because they run the same bootstrapping code as the 'full' daemon, which builds a map of all workers, initializes everything, etc. Either they should run less startup code or the general startup code should be reduced.
process
make workers do less when starting scrapers and other workers have a significant delay in starting up as a worker because they run the same bootstrapping code as the full daemon which builds a map of all workers initializes everything etc either they should run less startup code or the general startup code should be reduced
1
256,811
22,102,384,596
IssuesEvent
2022-06-01 14:36:16
vellvm/vellvm
https://api.github.com/repos/vellvm/vellvm
opened
QuickChick Performance
bug testing quickchick performance
There seems to be a couple of performance issues with the QC generator. Here's some things I've noticed: - [ ] We can't seem to generate large code blocks --- I think `failGen` is getting called somewhere: `Sample (run_GenLLVM (gen_code_length 10)).` - [ ] After some fixes to the generator it now seems to get stuck when generating larger programs. This could be because of a loop in one of the generators? I don't think it can just be calling `failGen` constantly, because I think it will just stop testing after a while, but that might be another possibility. - [ ] Some things don't seem to be generated... Like the `insert` / `extract` instructions. Could be related to the performance issues here.
1.0
QuickChick Performance - There seems to be a couple of performance issues with the QC generator. Here's some things I've noticed: - [ ] We can't seem to generate large code blocks --- I think `failGen` is getting called somewhere: `Sample (run_GenLLVM (gen_code_length 10)).` - [ ] After some fixes to the generator it now seems to get stuck when generating larger programs. This could be because of a loop in one of the generators? I don't think it can just be calling `failGen` constantly, because I think it will just stop testing after a while, but that might be another possibility. - [ ] Some things don't seem to be generated... Like the `insert` / `extract` instructions. Could be related to the performance issues here.
non_process
quickchick performance there seems to be a couple of performance issues with the qc generator here s some things i ve noticed we can t seem to generate large code blocks i think failgen is getting called somewhere sample run genllvm gen code length after some fixes to the generator it now seems to get stuck when generating larger programs this could be because of a loop in one of the generators i don t think it can just be calling failgen constantly because i think it will just stop testing after a while but that might be another possibility some things don t seem to be generated like the insert extract instructions could be related to the performance issues here
0
2,587
5,345,489,806
IssuesEvent
2017-02-17 17:04:46
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
closed
Track number of SET NAMES and CHANGE_USER
CONNECTION POOL QUERY PROCESSOR
This is related to #818 . As `SET NAMES` and `CHANGE_USER` are internally handled, it is useful to track their number
1.0
Track number of SET NAMES and CHANGE_USER - This is related to #818 . As `SET NAMES` and `CHANGE_USER` are internally handled, it is useful to track their number
process
track number of set names and change user this is related to as set names and change user are internally handled it is useful to track their number
1
19,338
13,888,936,425
IssuesEvent
2020-10-19 07:09:10
pingcap/tiup
https://api.github.com/repos/pingcap/tiup
closed
Add an option to skip the forced cluster start operation during scaling out
category/stability category/usability status/TODO type/enhancement type/feature-request
## Feature Request **Is your feature request related to a problem? Please describe:** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> We observed a case (#730) that the "start cluster" might fail on some certain environment if the cluster has large amount instances, and if it fails to start the cluster, the whole scale out process is interrupted. **Describe the feature you'd like:** <!-- A clear and concise description of what you want to happen. --> If the cluster can be confirmed to be already started, the "start cluster" operation could be safely skipped, but we don't have that option for user now. **Describe alternatives you've considered:** <!-- A clear and concise description of any alternative solutions or features you've considered. --> It's also possible to use a `-f/--force` flag to skip all operations that not critically necessary during an operation, like various checks or preparations. This could be a more generic solution. **Teachability, Documentation, Adoption, Migration Strategy:** <!-- If you can, explain some scenarios how users might use this, situations it would be helpful in. Any API designs, mockups, or diagrams are also helpful. -->
True
Add an option to skip the forced cluster start operation during scaling out - ## Feature Request **Is your feature request related to a problem? Please describe:** <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> We observed a case (#730) that the "start cluster" might fail on some certain environment if the cluster has large amount instances, and if it fails to start the cluster, the whole scale out process is interrupted. **Describe the feature you'd like:** <!-- A clear and concise description of what you want to happen. --> If the cluster can be confirmed to be already started, the "start cluster" operation could be safely skipped, but we don't have that option for user now. **Describe alternatives you've considered:** <!-- A clear and concise description of any alternative solutions or features you've considered. --> It's also possible to use a `-f/--force` flag to skip all operations that not critically necessary during an operation, like various checks or preparations. This could be a more generic solution. **Teachability, Documentation, Adoption, Migration Strategy:** <!-- If you can, explain some scenarios how users might use this, situations it would be helpful in. Any API designs, mockups, or diagrams are also helpful. -->
non_process
add an option to skip the forced cluster start operation during scaling out feature request is your feature request related to a problem please describe we observed a case that the start cluster might fail on some certain environment if the cluster has large amount instances and if it fails to start the cluster the whole scale out process is interrupted describe the feature you d like if the cluster can be confirmed to be already started the start cluster operation could be safely skipped but we don t have that option for user now describe alternatives you ve considered it s also possible to use a f force flag to skip all operations that not critically necessary during an operation like various checks or preparations this could be a more generic solution teachability documentation adoption migration strategy
0
11,842
14,657,406,525
IssuesEvent
2020-12-28 15:31:42
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
Highlighting of the data in blue colour should be removed where ever row pop-out is present
Bug P2 Participant manager Process: Dev Process: Tested QA Process: Tested dev UI
1. Highlighting of the data in blue colour should be removed where ever row pop-out is present 2. Row pop-out should be as per invision screen [Note : It should be handled in all the pages where ever row pop-out is present] ![row pop-up](https://user-images.githubusercontent.com/71445210/102527056-96a28b00-40c2-11eb-8fa5-9188ba427fa0.png)
3.0
Highlighting of the data in blue colour should be removed where ever row pop-out is present - 1. Highlighting of the data in blue colour should be removed where ever row pop-out is present 2. Row pop-out should be as per invision screen [Note : It should be handled in all the pages where ever row pop-out is present] ![row pop-up](https://user-images.githubusercontent.com/71445210/102527056-96a28b00-40c2-11eb-8fa5-9188ba427fa0.png)
process
highlighting of the data in blue colour should be removed where ever row pop out is present highlighting of the data in blue colour should be removed where ever row pop out is present row pop out should be as per invision screen
1
16,728
21,891,031,186
IssuesEvent
2022-05-20 01:39:28
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
bazel 3.7.2 can't build TensorFlow Lite 2.7.0 for iOS
type: support / not a bug (process) untriaged team-OSS
### Description of the bug: I am trying to build TensorFlow Lite 2.7.0 for iOS on macOS 12.3.1 m1 (arm64 CPU). But it failed to build it successfully. ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. git clone https://github.com/tensorflow/tensorflow tensorflow-2.7.0_ios cd tensorflow-2.7.0_ios git checkout r2.7 ./configure default for all except the following option Do you wish to build TensorFlow with iOS support? [y/N]: y iOS support will be enabled for TensorFlow. bazel build --config=ios_fat -c opt //tensorflow/lite/ios:TensorFlowLiteC_framework ### Which operating system are you running Bazel on? macOS 12.3.1 m1 (arm64 CPU) ### What is the output of `bazel info release`? release 3.7.2 ### Have you found anything relevant by searching the web? No ### Any other information, logs, or outputs that you want to share? ERROR: /Users/andyueng/samba/workspace_TensorFlow2/tensorflow-2.7.0_ios/tensorflow/lite/ios/BUILD:104:21: Bundling Preprocessed_TensorFlowLiteC_framework failed (Exit 127): bundletool failed: error executing command (cd /private/var/tmp/_bazel_andyueng/6c55a359b06f2067a8d1d18d335cb317/execroot/org_tensorflow && \ exec env - \ bazel-out/host/bin/external/build_bazel_rules_apple/tools/bundletool/bundletool bazel-out/darwin-opt/bin/tensorflow/lite/ios/Preprocessed_TensorFlowLiteC_framework-intermediates/bundletool_control.json) Execution platform: @local_execution_config_platform//:platform env: python: No such file or directory Target //tensorflow/lite/ios:TensorFlowLiteC_framework failed to build INFO: Elapsed time: 197.909s, Critical Path: 37.87s INFO: 1960 processes: 150 internal, 1810 local. FAILED: Build did NOT complete successfully
1.0
bazel 3.7.2 can't build TensorFlow Lite 2.7.0 for iOS - ### Description of the bug: I am trying to build TensorFlow Lite 2.7.0 for iOS on macOS 12.3.1 m1 (arm64 CPU). But it failed to build it successfully. ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. git clone https://github.com/tensorflow/tensorflow tensorflow-2.7.0_ios cd tensorflow-2.7.0_ios git checkout r2.7 ./configure default for all except the following option Do you wish to build TensorFlow with iOS support? [y/N]: y iOS support will be enabled for TensorFlow. bazel build --config=ios_fat -c opt //tensorflow/lite/ios:TensorFlowLiteC_framework ### Which operating system are you running Bazel on? macOS 12.3.1 m1 (arm64 CPU) ### What is the output of `bazel info release`? release 3.7.2 ### Have you found anything relevant by searching the web? No ### Any other information, logs, or outputs that you want to share? ERROR: /Users/andyueng/samba/workspace_TensorFlow2/tensorflow-2.7.0_ios/tensorflow/lite/ios/BUILD:104:21: Bundling Preprocessed_TensorFlowLiteC_framework failed (Exit 127): bundletool failed: error executing command (cd /private/var/tmp/_bazel_andyueng/6c55a359b06f2067a8d1d18d335cb317/execroot/org_tensorflow && \ exec env - \ bazel-out/host/bin/external/build_bazel_rules_apple/tools/bundletool/bundletool bazel-out/darwin-opt/bin/tensorflow/lite/ios/Preprocessed_TensorFlowLiteC_framework-intermediates/bundletool_control.json) Execution platform: @local_execution_config_platform//:platform env: python: No such file or directory Target //tensorflow/lite/ios:TensorFlowLiteC_framework failed to build INFO: Elapsed time: 197.909s, Critical Path: 37.87s INFO: 1960 processes: 150 internal, 1810 local. FAILED: Build did NOT complete successfully
process
bazel can t build tensorflow lite for ios description of the bug i am trying to build tensorflow lite for ios on macos cpu but it failed to build it successfully what s the simplest easiest way to reproduce this bug please provide a minimal example if possible git clone tensorflow ios cd tensorflow ios git checkout configure default for all except the following option do you wish to build tensorflow with ios support y ios support will be enabled for tensorflow bazel build config ios fat c opt tensorflow lite ios tensorflowlitec framework which operating system are you running bazel on macos cpu what is the output of bazel info release release have you found anything relevant by searching the web no any other information logs or outputs that you want to share error users andyueng samba workspace tensorflow ios tensorflow lite ios build bundling preprocessed tensorflowlitec framework failed exit bundletool failed error executing command cd private var tmp bazel andyueng execroot org tensorflow exec env bazel out host bin external build bazel rules apple tools bundletool bundletool bazel out darwin opt bin tensorflow lite ios preprocessed tensorflowlitec framework intermediates bundletool control json execution platform local execution config platform platform env python no such file or directory target tensorflow lite ios tensorflowlitec framework failed to build info elapsed time critical path info processes internal local failed build did not complete successfully
1
252,999
27,286,050,925
IssuesEvent
2023-02-23 13:34:42
lukebrogan-mend/c-goof
https://api.github.com/repos/lukebrogan-mend/c-goof
opened
velocity-tools-2.0.jar: 14 vulnerabilities (highest severity is: 9.8)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>velocity-tools-2.0.jar</b></p></summary> <p>VelocityTools is an integrated collection of Velocity subprojects with the common goal of creating tools and infrastructure to speed and ease development of both web and non-web applications using the Velocity template engine.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/velocity/velocity-tools/2.0/velocity-tools-2.0.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (velocity-tools version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-13116](https://www.mend.io/vulnerability-database/CVE-2019-13116) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | commons-collections-3.2.jar | Transitive | N/A* | &#10060; | | [CVE-2020-10683](https://www.mend.io/vulnerability-database/CVE-2020-10683) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | dom4j-1.1.jar | Transitive | N/A* | &#10060; | | [CVE-2017-15708](https://www.mend.io/vulnerability-database/CVE-2017-15708) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | commons-collections-3.2.jar | Transitive | N/A* | &#10060; | | [CVE-2015-7501](https://www.mend.io/vulnerability-database/CVE-2015-7501) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | commons-collections-3.2.jar | Transitive | N/A* | &#10060; | | [CVE-2016-1182](https://www.mend.io/vulnerability-database/CVE-2016-1182) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.2 | struts-core-1.3.8.jar | Transitive | N/A* | &#10060; | | [CVE-2016-1181](https://www.mend.io/vulnerability-database/CVE-2016-1181) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | struts-core-1.3.8.jar | Transitive | N/A* | &#10060; | | [CVE-2015-0899](https://www.mend.io/vulnerability-database/CVE-2015-0899) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | struts-core-1.3.8.jar | Transitive | N/A* | &#10060; | | [CVE-2018-1000632](https://www.mend.io/vulnerability-database/CVE-2018-1000632) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | dom4j-1.1.jar | Transitive | N/A* | &#10060; | | [CVE-2015-4852](https://www.mend.io/vulnerability-database/CVE-2015-4852) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-collections-3.2.jar | Transitive | N/A* | &#10060; | | [CVE-2019-10086](https://www.mend.io/vulnerability-database/CVE-2019-10086) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-beanutils-1.7.0.jar | Transitive | N/A* | &#10060; | | [CVE-2014-0114](https://www.mend.io/vulnerability-database/CVE-2014-0114) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | detected in multiple dependencies | Transitive | N/A* | &#10060; | | [CVE-2015-6420](https://www.mend.io/vulnerability-database/CVE-2015-6420) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-collections-3.2.jar | Transitive | N/A* | &#10060; | | [CVE-2020-13959](https://www.mend.io/vulnerability-database/CVE-2020-13959) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | velocity-tools-2.0.jar | Direct | org.apache.velocity.tools:velocity-tools-view:3.1 | &#9989; | | [CVE-2008-2025](https://www.mend.io/vulnerability-database/CVE-2008-2025) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.7 | struts-taglib-1.3.8.jar | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2019-13116</summary> ### Vulnerable Library - <b>commons-collections-3.2.jar</b></p> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2/commons-collections-3.2.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-collections-3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The MuleSoft Mule Community Edition runtime engine before 3.8 allows remote attackers to execute arbitrary code because of Java Deserialization, related to Apache Commons Collections <p>Publish Date: 2019-10-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-13116>CVE-2019-13116</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13116</a></p> <p>Release Date: 2019-10-16</p> <p>Fix Resolution: commons-collections:commons-collections:3.2.2</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2020-10683</summary> ### Vulnerable Library - <b>dom4j-1.1.jar</b></p> <p></p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/dom4j/dom4j/1.1/dom4j-1.1.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **dom4j-1.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> dom4j before 2.0.3 and 2.1.x before 2.1.3 allows external DTDs and External Entities by default, which might enable XXE attacks. However, there is popular external documentation from OWASP showing how to enable the safe, non-default behavior in any application that uses dom4j. <p>Publish Date: 2020-05-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-10683>CVE-2020-10683</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-05-01</p> <p>Fix Resolution: org.dom4j:dom4j:2.1.3,org.dom4j:dom4j:2.0.3</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2017-15708</summary> ### Vulnerable Library - <b>commons-collections-3.2.jar</b></p> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2/commons-collections-3.2.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-collections-3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Synapse, by default no authentication is required for Java Remote Method Invocation (RMI). So Apache Synapse 3.0.1 or all previous releases (3.0.0, 2.1.0, 2.0.0, 1.2, 1.1.2, 1.1.1) allows remote code execution attacks that can be performed by injecting specially crafted serialized objects. And the presence of Apache Commons Collections 3.2.1 (commons-collections-3.2.1.jar) or previous versions in Synapse distribution makes this exploitable. To mitigate the issue, we need to limit RMI access to trusted users only. Further upgrading to 3.0.1 version will eliminate the risk of having said Commons Collection version. In Synapse 3.0.1, Commons Collection has been updated to 3.2.2 version. <p>Publish Date: 2017-12-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-15708>CVE-2017-15708</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-15708">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-15708</a></p> <p>Release Date: 2017-12-10</p> <p>Fix Resolution: org.apache.synapse:Apache-Synapse:3.0.1;commons-collections:commons-collections:3.2.2</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2015-7501</summary> ### Vulnerable Library - <b>commons-collections-3.2.jar</b></p> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2/commons-collections-3.2.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-collections-3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Red Hat JBoss A-MQ 6.x; BPM Suite (BPMS) 6.x; BRMS 6.x and 5.x; Data Grid (JDG) 6.x; Data Virtualization (JDV) 6.x and 5.x; Enterprise Application Platform 6.x, 5.x, and 4.3.x; Fuse 6.x; Fuse Service Works (FSW) 6.x; Operations Network (JBoss ON) 3.x; Portal 6.x; SOA Platform (SOA-P) 5.x; Web Server (JWS) 3.x; Red Hat OpenShift/xPAAS 3.x; and Red Hat Subscription Asset Manager 1.3 allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library. <p>Publish Date: 2017-11-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-7501>CVE-2015-7501</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1279330">https://bugzilla.redhat.com/show_bug.cgi?id=1279330</a></p> <p>Release Date: 2017-11-09</p> <p>Fix Resolution: commons-collections:commons-collections:3.2.2;org.apache.commons:commons-collections4:4.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2016-1182</summary> ### Vulnerable Library - <b>struts-core-1.3.8.jar</b></p> <p>Apache Struts</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/struts-core/1.3.8/struts-core-1.3.8.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **struts-core-1.3.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> ActionServlet.java in Apache Struts 1 1.x through 1.3.10 does not properly restrict the Validator configuration, which allows remote attackers to conduct cross-site scripting (XSS) attacks or cause a denial of service via crafted input, a related issue to CVE-2015-0899. <p>Publish Date: 2016-07-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-1182>CVE-2016-1182</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.2</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-1182">https://nvd.nist.gov/vuln/detail/CVE-2016-1182</a></p> <p>Release Date: 2016-07-04</p> <p>Fix Resolution: struts:struts - 1.1-b3</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2016-1181</summary> ### Vulnerable Library - <b>struts-core-1.3.8.jar</b></p> <p>Apache Struts</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/struts-core/1.3.8/struts-core-1.3.8.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **struts-core-1.3.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> ActionServlet.java in Apache Struts 1 1.x through 1.3.10 mishandles multithreaded access to an ActionForm instance, which allows remote attackers to execute arbitrary code or cause a denial of service (unexpected memory access) via a multipart request, a related issue to CVE-2015-0899. <p>Publish Date: 2016-07-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-1181>CVE-2016-1181</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-1181">https://nvd.nist.gov/vuln/detail/CVE-2016-1181</a></p> <p>Release Date: 2016-07-04</p> <p>Fix Resolution: struts:struts - 1.1-b3</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-0899</summary> ### Vulnerable Library - <b>struts-core-1.3.8.jar</b></p> <p>Apache Struts</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/struts-core/1.3.8/struts-core-1.3.8.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **struts-core-1.3.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The MultiPageValidator implementation in Apache Struts 1 1.1 through 1.3.10 allows remote attackers to bypass intended access restrictions via a modified page parameter. <p>Publish Date: 2016-07-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-0899>CVE-2015-0899</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-0899">https://nvd.nist.gov/vuln/detail/CVE-2015-0899</a></p> <p>Release Date: 2016-07-04</p> <p>Fix Resolution: struts:struts - 1.1-beta-2,1.1-b2-20021124</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-1000632</summary> ### Vulnerable Library - <b>dom4j-1.1.jar</b></p> <p></p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/dom4j/dom4j/1.1/dom4j-1.1.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **dom4j-1.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> dom4j version prior to version 2.1.1 contains a CWE-91: XML Injection vulnerability in Class: Element. Methods: addElement, addAttribute that can result in an attacker tampering with XML documents through XML injection. This attack appear to be exploitable via an attacker specifying attributes or elements in the XML document. This vulnerability appears to have been fixed in 2.1.1 or later. <p>Publish Date: 2018-08-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-1000632>CVE-2018-1000632</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000632/">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000632/</a></p> <p>Release Date: 2018-08-20</p> <p>Fix Resolution: org.dom4j:dom4j:2.0.3</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-4852</summary> ### Vulnerable Library - <b>commons-collections-3.2.jar</b></p> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2/commons-collections-3.2.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-collections-3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product. <p>Publish Date: 2015-11-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-4852>CVE-2015-4852</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.openwall.com/lists/oss-security/2015/11/17/19">https://www.openwall.com/lists/oss-security/2015/11/17/19</a></p> <p>Release Date: 2015-11-18</p> <p>Fix Resolution: commons-collections:commons-collections:3.2.2</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10086</summary> ### Vulnerable Library - <b>commons-beanutils-1.7.0.jar</b></p> <p></p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-beanutils-1.7.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean. <p>Publish Date: 2019-08-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-10086>CVE-2019-10086</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-08-20</p> <p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2014-0114</summary> ### Vulnerable Libraries - <b>struts-core-1.3.8.jar</b>, <b>commons-beanutils-1.7.0.jar</b></p> <p> ### <b>struts-core-1.3.8.jar</b></p> <p>Apache Struts</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/struts-core/1.3.8/struts-core-1.3.8.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **struts-core-1.3.8.jar** (Vulnerable Library) ### <b>commons-beanutils-1.7.0.jar</b></p> <p></p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-beanutils-1.7.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1. <p>Publish Date: 2014-04-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-0114>CVE-2014-0114</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p> <p>Release Date: 2014-04-30</p> <p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-6420</summary> ### Vulnerable Library - <b>commons-collections-3.2.jar</b></p> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2/commons-collections-3.2.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-collections-3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Serialized-object interfaces in certain Cisco Collaboration and Social Media; Endpoint Clients and Client Software; Network Application, Service, and Acceleration; Network and Content Security Devices; Network Management and Provisioning; Routing and Switching - Enterprise and Service Provider; Unified Computing; Voice and Unified Communications Devices; Video, Streaming, TelePresence, and Transcoding Devices; Wireless; and Cisco Hosted Services products allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library. <p>Publish Date: 2015-12-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-6420>CVE-2015-6420</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2015-12-15</p> <p>Fix Resolution: commons-collections:commons-collections3.2.2,org.apache.commons:commons-collections4:4.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-13959</summary> ### Vulnerable Library - <b>velocity-tools-2.0.jar</b></p> <p>VelocityTools is an integrated collection of Velocity subprojects with the common goal of creating tools and infrastructure to speed and ease development of both web and non-web applications using the Velocity template engine.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/velocity/velocity-tools/2.0/velocity-tools-2.0.jar</p> <p> Dependency Hierarchy: - :x: **velocity-tools-2.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The default error page for VelocityView in Apache Velocity Tools prior to 3.1 reflects back the vm file that was entered as part of the URL. An attacker can set an XSS payload file as this vm file in the URL which results in this payload being executed. XSS vulnerabilities allow attackers to execute arbitrary JavaScript in the context of the attacked website and the attacked user. This can be abused to steal session cookies, perform requests in the name of the victim or for phishing attacks. <p>Publish Date: 2021-03-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-13959>CVE-2020-13959</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-fh63-4r66-jc7v">https://github.com/advisories/GHSA-fh63-4r66-jc7v</a></p> <p>Release Date: 2021-03-10</p> <p>Fix Resolution: org.apache.velocity.tools:velocity-tools-view:3.1</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2008-2025</summary> ### Vulnerable Library - <b>struts-taglib-1.3.8.jar</b></p> <p>Apache Struts</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/struts-taglib/1.3.8/struts-taglib-1.3.8.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **struts-taglib-1.3.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Cross-site scripting (XSS) vulnerability in Apache Struts before 1.2.9-162.31.1 on SUSE Linux Enterprise (SLE) 11, before 1.2.9-108.2 on SUSE openSUSE 10.3, before 1.2.9-198.2 on SUSE openSUSE 11.0, and before 1.2.9-162.163.2 on SUSE openSUSE 11.1 allows remote attackers to inject arbitrary web script or HTML via unspecified vectors related to "insufficient quoting of parameters." <p>Publish Date: 2009-04-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2008-2025>CVE-2008-2025</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>3.7</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2008-2025">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2008-2025</a></p> <p>Release Date: 2009-04-09</p> <p>Fix Resolution: org.apache.struts:struts-taglib:1.4.0</p> </p> <p></p> </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
True
velocity-tools-2.0.jar: 14 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>velocity-tools-2.0.jar</b></p></summary> <p>VelocityTools is an integrated collection of Velocity subprojects with the common goal of creating tools and infrastructure to speed and ease development of both web and non-web applications using the Velocity template engine.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/velocity/velocity-tools/2.0/velocity-tools-2.0.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (velocity-tools version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-13116](https://www.mend.io/vulnerability-database/CVE-2019-13116) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | commons-collections-3.2.jar | Transitive | N/A* | &#10060; | | [CVE-2020-10683](https://www.mend.io/vulnerability-database/CVE-2020-10683) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | dom4j-1.1.jar | Transitive | N/A* | &#10060; | | [CVE-2017-15708](https://www.mend.io/vulnerability-database/CVE-2017-15708) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | commons-collections-3.2.jar | Transitive | N/A* | &#10060; | | [CVE-2015-7501](https://www.mend.io/vulnerability-database/CVE-2015-7501) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> Critical | 9.8 | commons-collections-3.2.jar | Transitive | N/A* | &#10060; | | [CVE-2016-1182](https://www.mend.io/vulnerability-database/CVE-2016-1182) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.2 | struts-core-1.3.8.jar | Transitive | N/A* | &#10060; | | [CVE-2016-1181](https://www.mend.io/vulnerability-database/CVE-2016-1181) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | struts-core-1.3.8.jar | Transitive | N/A* | &#10060; | | [CVE-2015-0899](https://www.mend.io/vulnerability-database/CVE-2015-0899) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | struts-core-1.3.8.jar | Transitive | N/A* | &#10060; | | [CVE-2018-1000632](https://www.mend.io/vulnerability-database/CVE-2018-1000632) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | dom4j-1.1.jar | Transitive | N/A* | &#10060; | | [CVE-2015-4852](https://www.mend.io/vulnerability-database/CVE-2015-4852) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-collections-3.2.jar | Transitive | N/A* | &#10060; | | [CVE-2019-10086](https://www.mend.io/vulnerability-database/CVE-2019-10086) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-beanutils-1.7.0.jar | Transitive | N/A* | &#10060; | | [CVE-2014-0114](https://www.mend.io/vulnerability-database/CVE-2014-0114) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | detected in multiple dependencies | Transitive | N/A* | &#10060; | | [CVE-2015-6420](https://www.mend.io/vulnerability-database/CVE-2015-6420) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-collections-3.2.jar | Transitive | N/A* | &#10060; | | [CVE-2020-13959](https://www.mend.io/vulnerability-database/CVE-2020-13959) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | velocity-tools-2.0.jar | Direct | org.apache.velocity.tools:velocity-tools-view:3.1 | &#9989; | | [CVE-2008-2025](https://www.mend.io/vulnerability-database/CVE-2008-2025) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.7 | struts-taglib-1.3.8.jar | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2019-13116</summary> ### Vulnerable Library - <b>commons-collections-3.2.jar</b></p> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2/commons-collections-3.2.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-collections-3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The MuleSoft Mule Community Edition runtime engine before 3.8 allows remote attackers to execute arbitrary code because of Java Deserialization, related to Apache Commons Collections <p>Publish Date: 2019-10-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-13116>CVE-2019-13116</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13116">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13116</a></p> <p>Release Date: 2019-10-16</p> <p>Fix Resolution: commons-collections:commons-collections:3.2.2</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2020-10683</summary> ### Vulnerable Library - <b>dom4j-1.1.jar</b></p> <p></p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/dom4j/dom4j/1.1/dom4j-1.1.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **dom4j-1.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> dom4j before 2.0.3 and 2.1.x before 2.1.3 allows external DTDs and External Entities by default, which might enable XXE attacks. However, there is popular external documentation from OWASP showing how to enable the safe, non-default behavior in any application that uses dom4j. <p>Publish Date: 2020-05-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-10683>CVE-2020-10683</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-05-01</p> <p>Fix Resolution: org.dom4j:dom4j:2.1.3,org.dom4j:dom4j:2.0.3</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2017-15708</summary> ### Vulnerable Library - <b>commons-collections-3.2.jar</b></p> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2/commons-collections-3.2.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-collections-3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Synapse, by default no authentication is required for Java Remote Method Invocation (RMI). So Apache Synapse 3.0.1 or all previous releases (3.0.0, 2.1.0, 2.0.0, 1.2, 1.1.2, 1.1.1) allows remote code execution attacks that can be performed by injecting specially crafted serialized objects. And the presence of Apache Commons Collections 3.2.1 (commons-collections-3.2.1.jar) or previous versions in Synapse distribution makes this exploitable. To mitigate the issue, we need to limit RMI access to trusted users only. Further upgrading to 3.0.1 version will eliminate the risk of having said Commons Collection version. In Synapse 3.0.1, Commons Collection has been updated to 3.2.2 version. <p>Publish Date: 2017-12-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-15708>CVE-2017-15708</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-15708">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-15708</a></p> <p>Release Date: 2017-12-10</p> <p>Fix Resolution: org.apache.synapse:Apache-Synapse:3.0.1;commons-collections:commons-collections:3.2.2</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png' width=19 height=20> CVE-2015-7501</summary> ### Vulnerable Library - <b>commons-collections-3.2.jar</b></p> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2/commons-collections-3.2.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-collections-3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Red Hat JBoss A-MQ 6.x; BPM Suite (BPMS) 6.x; BRMS 6.x and 5.x; Data Grid (JDG) 6.x; Data Virtualization (JDV) 6.x and 5.x; Enterprise Application Platform 6.x, 5.x, and 4.3.x; Fuse 6.x; Fuse Service Works (FSW) 6.x; Operations Network (JBoss ON) 3.x; Portal 6.x; SOA Platform (SOA-P) 5.x; Web Server (JWS) 3.x; Red Hat OpenShift/xPAAS 3.x; and Red Hat Subscription Asset Manager 1.3 allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library. <p>Publish Date: 2017-11-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-7501>CVE-2015-7501</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1279330">https://bugzilla.redhat.com/show_bug.cgi?id=1279330</a></p> <p>Release Date: 2017-11-09</p> <p>Fix Resolution: commons-collections:commons-collections:3.2.2;org.apache.commons:commons-collections4:4.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2016-1182</summary> ### Vulnerable Library - <b>struts-core-1.3.8.jar</b></p> <p>Apache Struts</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/struts-core/1.3.8/struts-core-1.3.8.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **struts-core-1.3.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> ActionServlet.java in Apache Struts 1 1.x through 1.3.10 does not properly restrict the Validator configuration, which allows remote attackers to conduct cross-site scripting (XSS) attacks or cause a denial of service via crafted input, a related issue to CVE-2015-0899. <p>Publish Date: 2016-07-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-1182>CVE-2016-1182</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.2</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-1182">https://nvd.nist.gov/vuln/detail/CVE-2016-1182</a></p> <p>Release Date: 2016-07-04</p> <p>Fix Resolution: struts:struts - 1.1-b3</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2016-1181</summary> ### Vulnerable Library - <b>struts-core-1.3.8.jar</b></p> <p>Apache Struts</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/struts-core/1.3.8/struts-core-1.3.8.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **struts-core-1.3.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> ActionServlet.java in Apache Struts 1 1.x through 1.3.10 mishandles multithreaded access to an ActionForm instance, which allows remote attackers to execute arbitrary code or cause a denial of service (unexpected memory access) via a multipart request, a related issue to CVE-2015-0899. <p>Publish Date: 2016-07-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-1181>CVE-2016-1181</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>8.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-1181">https://nvd.nist.gov/vuln/detail/CVE-2016-1181</a></p> <p>Release Date: 2016-07-04</p> <p>Fix Resolution: struts:struts - 1.1-b3</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-0899</summary> ### Vulnerable Library - <b>struts-core-1.3.8.jar</b></p> <p>Apache Struts</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/struts-core/1.3.8/struts-core-1.3.8.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **struts-core-1.3.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The MultiPageValidator implementation in Apache Struts 1 1.1 through 1.3.10 allows remote attackers to bypass intended access restrictions via a modified page parameter. <p>Publish Date: 2016-07-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-0899>CVE-2015-0899</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-0899">https://nvd.nist.gov/vuln/detail/CVE-2015-0899</a></p> <p>Release Date: 2016-07-04</p> <p>Fix Resolution: struts:struts - 1.1-beta-2,1.1-b2-20021124</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-1000632</summary> ### Vulnerable Library - <b>dom4j-1.1.jar</b></p> <p></p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/dom4j/dom4j/1.1/dom4j-1.1.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **dom4j-1.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> dom4j version prior to version 2.1.1 contains a CWE-91: XML Injection vulnerability in Class: Element. Methods: addElement, addAttribute that can result in an attacker tampering with XML documents through XML injection. This attack appear to be exploitable via an attacker specifying attributes or elements in the XML document. This vulnerability appears to have been fixed in 2.1.1 or later. <p>Publish Date: 2018-08-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-1000632>CVE-2018-1000632</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000632/">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000632/</a></p> <p>Release Date: 2018-08-20</p> <p>Fix Resolution: org.dom4j:dom4j:2.0.3</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-4852</summary> ### Vulnerable Library - <b>commons-collections-3.2.jar</b></p> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2/commons-collections-3.2.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-collections-3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product. <p>Publish Date: 2015-11-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-4852>CVE-2015-4852</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.openwall.com/lists/oss-security/2015/11/17/19">https://www.openwall.com/lists/oss-security/2015/11/17/19</a></p> <p>Release Date: 2015-11-18</p> <p>Fix Resolution: commons-collections:commons-collections:3.2.2</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10086</summary> ### Vulnerable Library - <b>commons-beanutils-1.7.0.jar</b></p> <p></p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-beanutils-1.7.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean. <p>Publish Date: 2019-08-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-10086>CVE-2019-10086</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-08-20</p> <p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2014-0114</summary> ### Vulnerable Libraries - <b>struts-core-1.3.8.jar</b>, <b>commons-beanutils-1.7.0.jar</b></p> <p> ### <b>struts-core-1.3.8.jar</b></p> <p>Apache Struts</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/struts-core/1.3.8/struts-core-1.3.8.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **struts-core-1.3.8.jar** (Vulnerable Library) ### <b>commons-beanutils-1.7.0.jar</b></p> <p></p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-beanutils-1.7.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1. <p>Publish Date: 2014-04-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-0114>CVE-2014-0114</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p> <p>Release Date: 2014-04-30</p> <p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4;org.apache.struts:struts2-core:2.0.5</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2015-6420</summary> ### Vulnerable Library - <b>commons-collections-3.2.jar</b></p> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-collections/commons-collections/3.2/commons-collections-3.2.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **commons-collections-3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Serialized-object interfaces in certain Cisco Collaboration and Social Media; Endpoint Clients and Client Software; Network Application, Service, and Acceleration; Network and Content Security Devices; Network Management and Provisioning; Routing and Switching - Enterprise and Service Provider; Unified Computing; Voice and Unified Communications Devices; Video, Streaming, TelePresence, and Transcoding Devices; Wireless; and Cisco Hosted Services products allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library. <p>Publish Date: 2015-12-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-6420>CVE-2015-6420</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2015-12-15</p> <p>Fix Resolution: commons-collections:commons-collections3.2.2,org.apache.commons:commons-collections4:4.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-13959</summary> ### Vulnerable Library - <b>velocity-tools-2.0.jar</b></p> <p>VelocityTools is an integrated collection of Velocity subprojects with the common goal of creating tools and infrastructure to speed and ease development of both web and non-web applications using the Velocity template engine.</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/velocity/velocity-tools/2.0/velocity-tools-2.0.jar</p> <p> Dependency Hierarchy: - :x: **velocity-tools-2.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> The default error page for VelocityView in Apache Velocity Tools prior to 3.1 reflects back the vm file that was entered as part of the URL. An attacker can set an XSS payload file as this vm file in the URL which results in this payload being executed. XSS vulnerabilities allow attackers to execute arbitrary JavaScript in the context of the attacked website and the attacked user. This can be abused to steal session cookies, perform requests in the name of the victim or for phishing attacks. <p>Publish Date: 2021-03-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-13959>CVE-2020-13959</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-fh63-4r66-jc7v">https://github.com/advisories/GHSA-fh63-4r66-jc7v</a></p> <p>Release Date: 2021-03-10</p> <p>Fix Resolution: org.apache.velocity.tools:velocity-tools-view:3.1</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2008-2025</summary> ### Vulnerable Library - <b>struts-taglib-1.3.8.jar</b></p> <p>Apache Struts</p> <p>Path to dependency file: /vendor/aws-sdk-cpp-1.3.50/code-generation/generator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/struts-taglib/1.3.8/struts-taglib-1.3.8.jar</p> <p> Dependency Hierarchy: - velocity-tools-2.0.jar (Root Library) - :x: **struts-taglib-1.3.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebrogan-mend/c-goof/commit/802457f036ab4eb3ba012bdede243dbbc52f72ec">802457f036ab4eb3ba012bdede243dbbc52f72ec</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Cross-site scripting (XSS) vulnerability in Apache Struts before 1.2.9-162.31.1 on SUSE Linux Enterprise (SLE) 11, before 1.2.9-108.2 on SUSE openSUSE 10.3, before 1.2.9-198.2 on SUSE openSUSE 11.0, and before 1.2.9-162.163.2 on SUSE openSUSE 11.1 allows remote attackers to inject arbitrary web script or HTML via unspecified vectors related to "insufficient quoting of parameters." <p>Publish Date: 2009-04-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2008-2025>CVE-2008-2025</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>3.7</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2008-2025">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2008-2025</a></p> <p>Release Date: 2009-04-09</p> <p>Fix Resolution: org.apache.struts:struts-taglib:1.4.0</p> </p> <p></p> </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
non_process
velocity tools jar vulnerabilities highest severity is vulnerable library velocity tools jar velocitytools is an integrated collection of velocity subprojects with the common goal of creating tools and infrastructure to speed and ease development of both web and non web applications using the velocity template engine path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository org apache velocity velocity tools velocity tools jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in velocity tools version remediation available critical commons collections jar transitive n a critical jar transitive n a critical commons collections jar transitive n a critical commons collections jar transitive n a high struts core jar transitive n a high struts core jar transitive n a high struts core jar transitive n a high jar transitive n a high commons collections jar transitive n a high commons beanutils jar transitive n a high detected in multiple dependencies transitive n a high commons collections jar transitive n a medium velocity tools jar direct org apache velocity tools velocity tools view low struts taglib jar transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library commons collections jar types that extend and augment the java collections framework path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository commons collections commons collections commons collections jar dependency hierarchy velocity tools jar root library x commons collections jar vulnerable library found in head commit a href found in base branch main vulnerability details the mulesoft mule community edition runtime engine before allows remote attackers to execute arbitrary code because of java deserialization related to apache commons collections publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution commons collections commons collections cve vulnerable library jar path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository jar dependency hierarchy velocity tools jar root library x jar vulnerable library found in head commit a href found in base branch main vulnerability details before and x before allows external dtds and external entities by default which might enable xxe attacks however there is popular external documentation from owasp showing how to enable the safe non default behavior in any application that uses publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution org org cve vulnerable library commons collections jar types that extend and augment the java collections framework path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository commons collections commons collections commons collections jar dependency hierarchy velocity tools jar root library x commons collections jar vulnerable library found in head commit a href found in base branch main vulnerability details in apache synapse by default no authentication is required for java remote method invocation rmi so apache synapse or all previous releases allows remote code execution attacks that can be performed by injecting specially crafted serialized objects and the presence of apache commons collections commons collections jar or previous versions in synapse distribution makes this exploitable to mitigate the issue we need to limit rmi access to trusted users only further upgrading to version will eliminate the risk of having said commons collection version in synapse commons collection has been updated to version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache synapse apache synapse commons collections commons collections cve vulnerable library commons collections jar types that extend and augment the java collections framework path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository commons collections commons collections commons collections jar dependency hierarchy velocity tools jar root library x commons collections jar vulnerable library found in head commit a href found in base branch main vulnerability details red hat jboss a mq x bpm suite bpms x brms x and x data grid jdg x data virtualization jdv x and x enterprise application platform x x and x fuse x fuse service works fsw x operations network jboss on x portal x soa platform soa p x web server jws x red hat openshift xpaas x and red hat subscription asset manager allow remote attackers to execute arbitrary commands via a crafted serialized java object related to the apache commons collections acc library publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution commons collections commons collections org apache commons commons cve vulnerable library struts core jar apache struts path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository org apache struts struts core struts core jar dependency hierarchy velocity tools jar root library x struts core jar vulnerable library found in head commit a href found in base branch main vulnerability details actionservlet java in apache struts x through does not properly restrict the validator configuration which allows remote attackers to conduct cross site scripting xss attacks or cause a denial of service via crafted input a related issue to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution struts struts cve vulnerable library struts core jar apache struts path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository org apache struts struts core struts core jar dependency hierarchy velocity tools jar root library x struts core jar vulnerable library found in head commit a href found in base branch main vulnerability details actionservlet java in apache struts x through mishandles multithreaded access to an actionform instance which allows remote attackers to execute arbitrary code or cause a denial of service unexpected memory access via a multipart request a related issue to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution struts struts cve vulnerable library struts core jar apache struts path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository org apache struts struts core struts core jar dependency hierarchy velocity tools jar root library x struts core jar vulnerable library found in head commit a href found in base branch main vulnerability details the multipagevalidator implementation in apache struts through allows remote attackers to bypass intended access restrictions via a modified page parameter publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution struts struts beta cve vulnerable library jar path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository jar dependency hierarchy velocity tools jar root library x jar vulnerable library found in head commit a href found in base branch main vulnerability details version prior to version contains a cwe xml injection vulnerability in class element methods addelement addattribute that can result in an attacker tampering with xml documents through xml injection this attack appear to be exploitable via an attacker specifying attributes or elements in the xml document this vulnerability appears to have been fixed in or later publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org cve vulnerable library commons collections jar types that extend and augment the java collections framework path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository commons collections commons collections commons collections jar dependency hierarchy velocity tools jar root library x commons collections jar vulnerable library found in head commit a href found in base branch main vulnerability details the wls security component in oracle weblogic server and allows remote attackers to execute arbitrary commands via a crafted serialized java object in protocol traffic to tcp port related to oracle common modules com bea core apache commons collections jar note the scope of this cve is limited to the weblogic server product publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution commons collections commons collections cve vulnerable library commons beanutils jar path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy velocity tools jar root library x commons beanutils jar vulnerable library found in head commit a href found in base branch main vulnerability details in apache commons beanutils a special beanintrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all java objects we however were not using this by default characteristic of the propertyutilsbean publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution commons beanutils commons beanutils cve vulnerable libraries struts core jar commons beanutils jar struts core jar apache struts path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository org apache struts struts core struts core jar dependency hierarchy velocity tools jar root library x struts core jar vulnerable library commons beanutils jar path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy velocity tools jar root library x commons beanutils jar vulnerable library found in head commit a href found in base branch main vulnerability details apache commons beanutils as distributed in lib commons beanutils jar in apache struts x through and in other products requiring commons beanutils through does not suppress the class property which allows remote attackers to manipulate the classloader and execute arbitrary code via the class parameter as demonstrated by the passing of this parameter to the getclass method of the actionform object in struts publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution commons beanutils commons beanutils org apache struts core cve vulnerable library commons collections jar types that extend and augment the java collections framework path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository commons collections commons collections commons collections jar dependency hierarchy velocity tools jar root library x commons collections jar vulnerable library found in head commit a href found in base branch main vulnerability details serialized object interfaces in certain cisco collaboration and social media endpoint clients and client software network application service and acceleration network and content security devices network management and provisioning routing and switching enterprise and service provider unified computing voice and unified communications devices video streaming telepresence and transcoding devices wireless and cisco hosted services products allow remote attackers to execute arbitrary commands via a crafted serialized java object related to the apache commons collections acc library publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution commons collections commons org apache commons commons cve vulnerable library velocity tools jar velocitytools is an integrated collection of velocity subprojects with the common goal of creating tools and infrastructure to speed and ease development of both web and non web applications using the velocity template engine path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository org apache velocity velocity tools velocity tools jar dependency hierarchy x velocity tools jar vulnerable library found in head commit a href found in base branch main vulnerability details the default error page for velocityview in apache velocity tools prior to reflects back the vm file that was entered as part of the url an attacker can set an xss payload file as this vm file in the url which results in this payload being executed xss vulnerabilities allow attackers to execute arbitrary javascript in the context of the attacked website and the attacked user this can be abused to steal session cookies perform requests in the name of the victim or for phishing attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache velocity tools velocity tools view rescue worker helmet automatic remediation is available for this issue cve vulnerable library struts taglib jar apache struts path to dependency file vendor aws sdk cpp code generation generator pom xml path to vulnerable library home wss scanner repository org apache struts struts taglib struts taglib jar dependency hierarchy velocity tools jar root library x struts taglib jar vulnerable library found in head commit a href found in base branch main vulnerability details cross site scripting xss vulnerability in apache struts before on suse linux enterprise sle before on suse opensuse before on suse opensuse and before on suse opensuse allows remote attackers to inject arbitrary web script or html via unspecified vectors related to insufficient quoting of parameters publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache struts struts taglib rescue worker helmet automatic remediation is available for this issue
0
14,202
17,101,046,859
IssuesEvent
2021-07-09 11:17:08
prisma/prisma-engines
https://api.github.com/repos/prisma/prisma-engines
closed
Chore: change grammar of messages from `can not` to `cannot`
engines/other process/candidate team/migrations topic: error
> Don’t use can not when you mean cannot. The only time you’re likely to see can not written as separate words is when the word “can” happens to precede some other phrase that happens to start with “not”: It's only valid in these cases for example: > We can not only break even, but also turn a profit. > > The company’s new product can not only reduce emissions, but also trap some of the existing greenhouse gasses. It seems that all/most cases should be `cannot` https://sourcegraph.com/search?q=context:global+repo:github.com/prisma/prisma-engines+%22can+not%22&patternType=regexp
1.0
Chore: change grammar of messages from `can not` to `cannot` - > Don’t use can not when you mean cannot. The only time you’re likely to see can not written as separate words is when the word “can” happens to precede some other phrase that happens to start with “not”: It's only valid in these cases for example: > We can not only break even, but also turn a profit. > > The company’s new product can not only reduce emissions, but also trap some of the existing greenhouse gasses. It seems that all/most cases should be `cannot` https://sourcegraph.com/search?q=context:global+repo:github.com/prisma/prisma-engines+%22can+not%22&patternType=regexp
process
chore change grammar of messages from can not to cannot don’t use can not when you mean cannot the only time you’re likely to see can not written as separate words is when the word “can” happens to precede some other phrase that happens to start with “not” it s only valid in these cases for example we can not only break even but also turn a profit the company’s new product can not only reduce emissions but also trap some of the existing greenhouse gasses it seems that all most cases should be cannot
1
140,386
5,401,025,853
IssuesEvent
2017-02-27 23:42:20
BCDevExchange/devex
https://api.github.com/repos/BCDevExchange/devex
closed
Set up pages for enhanced Sign Up process
enhancement priority ui-ux
@ccoldwell, could you please set up as much as you can according to the following process. If you can get all of the elements inserted and linked, I can mess with the UI: ![sign up process](https://cloud.githubusercontent.com/assets/16946297/23189371/778c4f2a-f847-11e6-8f59-fa498884199d.jpg) Here's a mockup of a modified Profile Detail (Edit Profile) page: ![profile detail](https://cloud.githubusercontent.com/assets/16946297/23189397/95ebe318-f847-11e6-8eec-9fce91a10981.jpg) For **confirmations**, not sure what you'd recommend (I'm thinking a [modal](http://getbootstrap.com/javascript/#modals-examples) would be nice, but not sure if these are bad for mobile?).....I just think we need something more blatant than the current alerts that only appear for a couple seconds in the bottom right corner. ![confirmation](https://cloud.githubusercontent.com/assets/16946297/23189608/94f10ece-f848-11e6-9fd9-30251742ec45.jpg)
1.0
Set up pages for enhanced Sign Up process - @ccoldwell, could you please set up as much as you can according to the following process. If you can get all of the elements inserted and linked, I can mess with the UI: ![sign up process](https://cloud.githubusercontent.com/assets/16946297/23189371/778c4f2a-f847-11e6-8f59-fa498884199d.jpg) Here's a mockup of a modified Profile Detail (Edit Profile) page: ![profile detail](https://cloud.githubusercontent.com/assets/16946297/23189397/95ebe318-f847-11e6-8eec-9fce91a10981.jpg) For **confirmations**, not sure what you'd recommend (I'm thinking a [modal](http://getbootstrap.com/javascript/#modals-examples) would be nice, but not sure if these are bad for mobile?).....I just think we need something more blatant than the current alerts that only appear for a couple seconds in the bottom right corner. ![confirmation](https://cloud.githubusercontent.com/assets/16946297/23189608/94f10ece-f848-11e6-9fd9-30251742ec45.jpg)
non_process
set up pages for enhanced sign up process ccoldwell could you please set up as much as you can according to the following process if you can get all of the elements inserted and linked i can mess with the ui here s a mockup of a modified profile detail edit profile page for confirmations not sure what you d recommend i m thinking a would be nice but not sure if these are bad for mobile i just think we need something more blatant than the current alerts that only appear for a couple seconds in the bottom right corner
0
9,079
12,149,841,895
IssuesEvent
2020-04-24 16:51:58
CERT-Polska/drakvuf-sandbox
https://api.github.com/repos/CERT-Polska/drakvuf-sandbox
closed
Parse procmon log to the form of process tree
critical path drakrun/postprocessing enhancement
Parse `procmon` output into a form of a process tree that can be easily read by human (and visualized). E.g. output structure: ``` [ { "pid":2408, "process":"C:\\Users\\Administrator\\AppData\\Local\\Temp\\cmd.exe", "children":[ { "pid":2968, "process":"C:\\Users\\Administrator\\AppData\\Local\\Temp\\emo2.exe", "children":[ { "pid":3048, "process":"C:\\Users\\Administrator\\AppData\\Local\\Temp\\emo3.exe", "children":[ ] } ] }, { "pid":3028, "process":"C:\\Users\\Administrator\\AppData\\Local\\Temp\\emo4.exe", "children":[ ] } ] } ] ```
1.0
Parse procmon log to the form of process tree - Parse `procmon` output into a form of a process tree that can be easily read by human (and visualized). E.g. output structure: ``` [ { "pid":2408, "process":"C:\\Users\\Administrator\\AppData\\Local\\Temp\\cmd.exe", "children":[ { "pid":2968, "process":"C:\\Users\\Administrator\\AppData\\Local\\Temp\\emo2.exe", "children":[ { "pid":3048, "process":"C:\\Users\\Administrator\\AppData\\Local\\Temp\\emo3.exe", "children":[ ] } ] }, { "pid":3028, "process":"C:\\Users\\Administrator\\AppData\\Local\\Temp\\emo4.exe", "children":[ ] } ] } ] ```
process
parse procmon log to the form of process tree parse procmon output into a form of a process tree that can be easily read by human and visualized e g output structure pid process c users administrator appdata local temp cmd exe children pid process c users administrator appdata local temp exe children pid process c users administrator appdata local temp exe children pid process c users administrator appdata local temp exe children
1
8,776
11,899,692,648
IssuesEvent
2020-03-30 09:25:24
prisma/prisma2
https://api.github.com/repos/prisma/prisma2
closed
Add `cli` as executable for @prisma/cli?
kind/improvement process/candidate topic: cli
## Problem When using npx without having `@prisma/cli` installed before we need to do `npx -p @prisma/cli@alpha prisma2 --version` ## Solution If we put another name for the executable in the package.json `cli` npx should work like: `npx @prisma/cli@alpha --version` which is a lot easier (no parameters!) ## Additional context We may want to test how it's doing with a global install.
1.0
Add `cli` as executable for @prisma/cli? - ## Problem When using npx without having `@prisma/cli` installed before we need to do `npx -p @prisma/cli@alpha prisma2 --version` ## Solution If we put another name for the executable in the package.json `cli` npx should work like: `npx @prisma/cli@alpha --version` which is a lot easier (no parameters!) ## Additional context We may want to test how it's doing with a global install.
process
add cli as executable for prisma cli problem when using npx without having prisma cli installed before we need to do npx p prisma cli alpha version solution if we put another name for the executable in the package json cli npx should work like npx prisma cli alpha version which is a lot easier no parameters additional context we may want to test how it s doing with a global install
1
12,410
14,918,142,839
IssuesEvent
2021-01-22 21:07:35
c4gnv/meta
https://api.github.com/repos/c4gnv/meta
closed
Schedule meetings to kick off 2018
process
More structured meetings around specific projects (Josh, Bjorn) - Manage own schedule per project, in person or virtual, led by project champion, working meetings Once per month community-oriented meeting (Andrew, Samantha) - Facebook or Meet-ups, try for a few months then try different social media channels - Avenue for volunteers to plug into projects then begin participating in project-specific meetings
1.0
Schedule meetings to kick off 2018 - More structured meetings around specific projects (Josh, Bjorn) - Manage own schedule per project, in person or virtual, led by project champion, working meetings Once per month community-oriented meeting (Andrew, Samantha) - Facebook or Meet-ups, try for a few months then try different social media channels - Avenue for volunteers to plug into projects then begin participating in project-specific meetings
process
schedule meetings to kick off more structured meetings around specific projects josh bjorn manage own schedule per project in person or virtual led by project champion working meetings once per month community oriented meeting andrew samantha facebook or meet ups try for a few months then try different social media channels avenue for volunteers to plug into projects then begin participating in project specific meetings
1
5,758
8,598,730,546
IssuesEvent
2018-11-15 22:46:57
gfrebello/qs-trip-planning-procedure
https://api.github.com/repos/gfrebello/qs-trip-planning-procedure
closed
Code front end for flight reservations
Priority:Very High Process:Implement Requirement
For the user to be able to reserve flights, they need to be able to interact with the planning page, which is mostly front end functionality. This task is related to the front end of the flight reservations. https://stackoverflow.com/questions/11743392/check-if-array-is-empty-or-exists https://stackoverflow.com/questions/42576198/get-object-data-and-target-element-from-onclick-event-in-react-js https://stackoverflow.com/questions/9612758/add-a-css-border-on-hover-without-moving-the-element https://stackoverflow.com/questions/49386324/componentdidupdate-vs-componentwillreceiveprops-use-case-in-react https://stackoverflow.com/questions/114543/how-to-horizontally-center-a-div https://stackoverflow.com/questions/6783902/offset-div-from-center https://github.com/reactstrap/reactstrap/issues/925 https://reactstrap.github.io/components/listgroup/ http://www.redbitdev.com/getting-started-with-react-flexbox-grid/
1.0
Code front end for flight reservations - For the user to be able to reserve flights, they need to be able to interact with the planning page, which is mostly front end functionality. This task is related to the front end of the flight reservations. https://stackoverflow.com/questions/11743392/check-if-array-is-empty-or-exists https://stackoverflow.com/questions/42576198/get-object-data-and-target-element-from-onclick-event-in-react-js https://stackoverflow.com/questions/9612758/add-a-css-border-on-hover-without-moving-the-element https://stackoverflow.com/questions/49386324/componentdidupdate-vs-componentwillreceiveprops-use-case-in-react https://stackoverflow.com/questions/114543/how-to-horizontally-center-a-div https://stackoverflow.com/questions/6783902/offset-div-from-center https://github.com/reactstrap/reactstrap/issues/925 https://reactstrap.github.io/components/listgroup/ http://www.redbitdev.com/getting-started-with-react-flexbox-grid/
process
code front end for flight reservations for the user to be able to reserve flights they need to be able to interact with the planning page which is mostly front end functionality this task is related to the front end of the flight reservations
1
257,764
19,531,059,947
IssuesEvent
2021-12-30 16:53:50
GoogleCloudPlatform/dfcx-scrapi
https://api.github.com/repos/GoogleCloudPlatform/dfcx-scrapi
closed
Pre-configured Issue and Pull Request Templates for this project
documentation
## Background I noticed that this repository does not have an issue template or PR template yet, so I was thinking of following these guides to create the respective templates. This can make writing bug reports, issues, PRs quicker and less ambiguous. ### Proposed solution Issue Template: - [Guide for Issue Templates](https://docs.github.com/en/communities/using-templates-to-encourage-useful-issues-and-pull-requests/configuring-issue-templates-for-your-repository) - Two types of issue templates: - Bug report [proposed template](https://github.com/stevemao/github-issue-templates/blob/master/bugs-only/ISSUE_TEMPLATE.md) - Feature request [proposed template](https://github.com/wagtail/wagtail/issues/5560) - Alternatively, Github's default templates can also be used. Pull Request Template: - [Guide for PR templates](https://docs.github.com/en/communities/using-templates-to-encourage-useful-issues-and-pull-requests/creating-a-pull-request-template-for-your-repository) - [Proposed template](https://github.com/embeddedartistry/templates/blob/master/oss_docs/PULL_REQUEST_TEMPLATE.md) ### Questions - Should these templates be public or hidden?
1.0
Pre-configured Issue and Pull Request Templates for this project - ## Background I noticed that this repository does not have an issue template or PR template yet, so I was thinking of following these guides to create the respective templates. This can make writing bug reports, issues, PRs quicker and less ambiguous. ### Proposed solution Issue Template: - [Guide for Issue Templates](https://docs.github.com/en/communities/using-templates-to-encourage-useful-issues-and-pull-requests/configuring-issue-templates-for-your-repository) - Two types of issue templates: - Bug report [proposed template](https://github.com/stevemao/github-issue-templates/blob/master/bugs-only/ISSUE_TEMPLATE.md) - Feature request [proposed template](https://github.com/wagtail/wagtail/issues/5560) - Alternatively, Github's default templates can also be used. Pull Request Template: - [Guide for PR templates](https://docs.github.com/en/communities/using-templates-to-encourage-useful-issues-and-pull-requests/creating-a-pull-request-template-for-your-repository) - [Proposed template](https://github.com/embeddedartistry/templates/blob/master/oss_docs/PULL_REQUEST_TEMPLATE.md) ### Questions - Should these templates be public or hidden?
non_process
pre configured issue and pull request templates for this project background i noticed that this repository does not have an issue template or pr template yet so i was thinking of following these guides to create the respective templates this can make writing bug reports issues prs quicker and less ambiguous proposed solution issue template two types of issue templates bug report feature request alternatively github s default templates can also be used pull request template questions should these templates be public or hidden
0
78,896
9,808,586,954
IssuesEvent
2019-06-12 15:55:21
fecgov/fec-cms
https://api.github.com/repos/fecgov/fec-cms
closed
Design interaction for terminated committees
Work: Content Work: Front-end Work: UX/Design
**What we’re after**: Showing terminated and administratively terminated committees should not be our default state. We need to provide users with committee filing frequency information and allow them to decide if want to see these committees. **Current state**: We include terminated and administratively terminated committees. We do not alert users that they might be viewing terminated committees. We do not provide users a way to exclude terminated committees. https://www.fec.gov/data/committees/ **Completion criteria**:
1.0
Design interaction for terminated committees - **What we’re after**: Showing terminated and administratively terminated committees should not be our default state. We need to provide users with committee filing frequency information and allow them to decide if want to see these committees. **Current state**: We include terminated and administratively terminated committees. We do not alert users that they might be viewing terminated committees. We do not provide users a way to exclude terminated committees. https://www.fec.gov/data/committees/ **Completion criteria**:
non_process
design interaction for terminated committees what we’re after showing terminated and administratively terminated committees should not be our default state we need to provide users with committee filing frequency information and allow them to decide if want to see these committees current state we include terminated and administratively terminated committees we do not alert users that they might be viewing terminated committees we do not provide users a way to exclude terminated committees completion criteria
0
760
3,244,601,230
IssuesEvent
2015-10-16 03:54:31
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Cache accesses to process.env?
process
I've found in multiple projects that accessing `process.env` within a hot section of the code leads to major slowdown. This really hurts in React server-side rendering ([issue](https://github.com/facebook/react/issues/812)) and has caused them to rearrange how they access the env. It would sense to cache already-accessed properties rather than reaching out to the actual environment. And of course, update the cache on assignment. I would be happy to contribute a patch if anyone could point me in the right direction to get started, as I'm new to Node core dev.
1.0
Cache accesses to process.env? - I've found in multiple projects that accessing `process.env` within a hot section of the code leads to major slowdown. This really hurts in React server-side rendering ([issue](https://github.com/facebook/react/issues/812)) and has caused them to rearrange how they access the env. It would sense to cache already-accessed properties rather than reaching out to the actual environment. And of course, update the cache on assignment. I would be happy to contribute a patch if anyone could point me in the right direction to get started, as I'm new to Node core dev.
process
cache accesses to process env i ve found in multiple projects that accessing process env within a hot section of the code leads to major slowdown this really hurts in react server side rendering and has caused them to rearrange how they access the env it would sense to cache already accessed properties rather than reaching out to the actual environment and of course update the cache on assignment i would be happy to contribute a patch if anyone could point me in the right direction to get started as i m new to node core dev
1
659,122
21,916,684,996
IssuesEvent
2022-05-21 23:40:54
AxerBot/axer-bot
https://api.github.com/repos/AxerBot/axer-bot
closed
convert bot responses to embeds when possible
enhancement medium priority
currently, some bot replies use an old approach that doesn't involve embeds, would be nicer to make it consistent
1.0
convert bot responses to embeds when possible - currently, some bot replies use an old approach that doesn't involve embeds, would be nicer to make it consistent
non_process
convert bot responses to embeds when possible currently some bot replies use an old approach that doesn t involve embeds would be nicer to make it consistent
0
228,500
18,238,688,603
IssuesEvent
2021-10-01 10:09:44
akka/akka
https://api.github.com/repos/akka/akka
closed
Give BehaviorTestKit-spawned actors access to a real configuration
1 - triaged help wanted t:testing
It's reasonably common for actors to leverage configuration via `context.system.settings.config` (which avoids every actor/class doing its own calls to `ConfigFactory.load`; this trick was taught to me by an ELE on a consulting engagement some years ago :) ). When run by the `BehaviorTestKit`, however, `context.system.settings.config` is a config that's solely the `reference.conf` files from the classpath. For testing a behavior in full isolation, this is fine: a pattern of def behaviorFor(fixture: Fixture, state: State): Behavior[Command] = ??? def apply(...): Behavior[Command] = Behaviors.setup { ctx => // assemble values from config etc., build into a fixture/state instance behaviorFor(...) } lets you build fixtures and states in your test (maybe even using property-based testing...) and build the behavior under test with `behaviorFor`. This gets complicated when the behavior under test spawns another actor defined with this pattern: the `spawn` will be via `apply`, which will pull in a config which is unlikely to be suitable (and if you want things to fail-fast on bad config, the spawn itself will probably fail) In asynchronous testing, the `ActorTestKit`'s `ActorSystem` loads `application-test.conf`, so it makes sense (to me) to have `BehaviorTestKit`'s `ActorSystemStub` do likewise.
1.0
Give BehaviorTestKit-spawned actors access to a real configuration - It's reasonably common for actors to leverage configuration via `context.system.settings.config` (which avoids every actor/class doing its own calls to `ConfigFactory.load`; this trick was taught to me by an ELE on a consulting engagement some years ago :) ). When run by the `BehaviorTestKit`, however, `context.system.settings.config` is a config that's solely the `reference.conf` files from the classpath. For testing a behavior in full isolation, this is fine: a pattern of def behaviorFor(fixture: Fixture, state: State): Behavior[Command] = ??? def apply(...): Behavior[Command] = Behaviors.setup { ctx => // assemble values from config etc., build into a fixture/state instance behaviorFor(...) } lets you build fixtures and states in your test (maybe even using property-based testing...) and build the behavior under test with `behaviorFor`. This gets complicated when the behavior under test spawns another actor defined with this pattern: the `spawn` will be via `apply`, which will pull in a config which is unlikely to be suitable (and if you want things to fail-fast on bad config, the spawn itself will probably fail) In asynchronous testing, the `ActorTestKit`'s `ActorSystem` loads `application-test.conf`, so it makes sense (to me) to have `BehaviorTestKit`'s `ActorSystemStub` do likewise.
non_process
give behaviortestkit spawned actors access to a real configuration it s reasonably common for actors to leverage configuration via context system settings config which avoids every actor class doing its own calls to configfactory load this trick was taught to me by an ele on a consulting engagement some years ago when run by the behaviortestkit however context system settings config is a config that s solely the reference conf files from the classpath for testing a behavior in full isolation this is fine a pattern of def behaviorfor fixture fixture state state behavior def apply behavior behaviors setup ctx assemble values from config etc build into a fixture state instance behaviorfor lets you build fixtures and states in your test maybe even using property based testing and build the behavior under test with behaviorfor this gets complicated when the behavior under test spawns another actor defined with this pattern the spawn will be via apply which will pull in a config which is unlikely to be suitable and if you want things to fail fast on bad config the spawn itself will probably fail in asynchronous testing the actortestkit s actorsystem loads application test conf so it makes sense to me to have behaviortestkit s actorsystemstub do likewise
0
142,592
19,099,363,236
IssuesEvent
2021-11-29 20:27:11
harrinry/stream-chat-react
https://api.github.com/repos/harrinry/stream-chat-react
opened
CVE-2018-19797 (Medium) detected in node-sass-4.14.1.tgz
security vulnerability
## CVE-2018-19797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: stream-chat-react/package.json</p> <p>Path to vulnerable library: stream-chat-react/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - rollup-plugin-scss-2.6.1.tgz (Root Library) - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/harrinry/stream-chat-react/commit/a80df0ece7581dc566b7c05d5f5ae6ecc66b730f">a80df0ece7581dc566b7c05d5f5ae6ecc66b730f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file. <p>Publish Date: 2018-12-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19797>CVE-2018-19797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.6.0">https://github.com/sass/libsass/releases/tag/3.6.0</a></p> <p>Release Date: 2018-12-03</p> <p>Fix Resolution: libsass - 3.6.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-sass","packageVersion":"4.14.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"rollup-plugin-scss:2.6.1;node-sass:4.14.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"libsass - 3.6.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-19797","vulnerabilityDetails":"In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19797","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-19797 (Medium) detected in node-sass-4.14.1.tgz - ## CVE-2018-19797 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: stream-chat-react/package.json</p> <p>Path to vulnerable library: stream-chat-react/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - rollup-plugin-scss-2.6.1.tgz (Root Library) - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/harrinry/stream-chat-react/commit/a80df0ece7581dc566b7c05d5f5ae6ecc66b730f">a80df0ece7581dc566b7c05d5f5ae6ecc66b730f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file. <p>Publish Date: 2018-12-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19797>CVE-2018-19797</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.6.0">https://github.com/sass/libsass/releases/tag/3.6.0</a></p> <p>Release Date: 2018-12-03</p> <p>Fix Resolution: libsass - 3.6.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-sass","packageVersion":"4.14.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"rollup-plugin-scss:2.6.1;node-sass:4.14.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"libsass - 3.6.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-19797","vulnerabilityDetails":"In LibSass 3.5.5, a NULL Pointer Dereference in the function Sass::Selector_List::populate_extends in SharedPtr.hpp (used by ast.cpp and ast_selectors.cpp) may cause a Denial of Service (application crash) via a crafted sass input file.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19797","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in node sass tgz cve medium severity vulnerability vulnerable library node sass tgz wrapper around libsass library home page a href path to dependency file stream chat react package json path to vulnerable library stream chat react node modules node sass package json dependency hierarchy rollup plugin scss tgz root library x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details in libsass a null pointer dereference in the function sass selector list populate extends in sharedptr hpp used by ast cpp and ast selectors cpp may cause a denial of service application crash via a crafted sass input file publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree rollup plugin scss node sass isminimumfixversionavailable true minimumfixversion libsass isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in libsass a null pointer dereference in the function sass selector list populate extends in sharedptr hpp used by ast cpp and ast selectors cpp may cause a denial of service application crash via a crafted sass input file vulnerabilityurl
0
50,355
21,076,589,045
IssuesEvent
2022-04-02 08:21:59
emergenzeHack/ukrainehelp.emergenzehack.info_segnalazioni
https://api.github.com/repos/emergenzeHack/ukrainehelp.emergenzehack.info_segnalazioni
opened
https://www.raiplay.it/benvenuti-bambini Cartoni animati in lingua italiana e ucraina (contenuti gr
Services translation Children
<pre><yamldata> servicetypes: materialGoods: false hospitality: false transport: false healthcare: false Legal: false translation: true job: false psychologicalSupport: false Children: true disability: false women: false education: false offerFromWho: Raiplay title: https://www.raiplay.it/benvenuti-bambini Cartoni animati in lingua italiana e ucraina (contenuti gratuiti). recipients: '' description: '' url: https://www.raiplay.it/benvenuti-bambini address: mode: autocomplete address: place_id: 283767136 licence: Data © OpenStreetMap contributors, ODbL 1.0. https://osm.org/copyright osm_type: relation osm_id: 41485 boundingbox: - '41.6556417' - '42.1410285' - '12.2344669' - '12.8557603' lat: '41.8933203' lon: '12.4829321' display_name: Roma, Roma Capitale, Lazio, Italia class: boundary type: administrative importance: 0.7896107180689524 icon: https://nominatim.openstreetmap.org/ui/mapicons//poi_boundary_administrative.p.20.png address: city: Roma county: Roma Capitale state: Lazio country: Italia country_code: it iConfirmToHaveReadAndAcceptedInformativeToThreatPersonalData: true label: services submit: true </yamldata></pre>
1.0
https://www.raiplay.it/benvenuti-bambini Cartoni animati in lingua italiana e ucraina (contenuti gr - <pre><yamldata> servicetypes: materialGoods: false hospitality: false transport: false healthcare: false Legal: false translation: true job: false psychologicalSupport: false Children: true disability: false women: false education: false offerFromWho: Raiplay title: https://www.raiplay.it/benvenuti-bambini Cartoni animati in lingua italiana e ucraina (contenuti gratuiti). recipients: '' description: '' url: https://www.raiplay.it/benvenuti-bambini address: mode: autocomplete address: place_id: 283767136 licence: Data © OpenStreetMap contributors, ODbL 1.0. https://osm.org/copyright osm_type: relation osm_id: 41485 boundingbox: - '41.6556417' - '42.1410285' - '12.2344669' - '12.8557603' lat: '41.8933203' lon: '12.4829321' display_name: Roma, Roma Capitale, Lazio, Italia class: boundary type: administrative importance: 0.7896107180689524 icon: https://nominatim.openstreetmap.org/ui/mapicons//poi_boundary_administrative.p.20.png address: city: Roma county: Roma Capitale state: Lazio country: Italia country_code: it iConfirmToHaveReadAndAcceptedInformativeToThreatPersonalData: true label: services submit: true </yamldata></pre>
non_process
cartoni animati in lingua italiana e ucraina contenuti gr servicetypes materialgoods false hospitality false transport false healthcare false legal false translation true job false psychologicalsupport false children true disability false women false education false offerfromwho raiplay title cartoni animati in lingua italiana e ucraina contenuti gratuiti recipients description url address mode autocomplete address place id licence data © openstreetmap contributors odbl osm type relation osm id boundingbox lat lon display name roma roma capitale lazio italia class boundary type administrative importance icon address city roma county roma capitale state lazio country italia country code it iconfirmtohavereadandacceptedinformativetothreatpersonaldata true label services submit true
0
1,313
3,865,477,292
IssuesEvent
2016-04-08 17:34:08
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Process.Modules tests are hanging in CI on OS X
2 - In Progress System.Diagnostics.Process X-Plat
Locally then run fine, and they made it through a couple of CI runs. But something about shelling out to vmmap is causing them to hang in CI.
1.0
Process.Modules tests are hanging in CI on OS X - Locally then run fine, and they made it through a couple of CI runs. But something about shelling out to vmmap is causing them to hang in CI.
process
process modules tests are hanging in ci on os x locally then run fine and they made it through a couple of ci runs but something about shelling out to vmmap is causing them to hang in ci
1
24,435
11,035,129,981
IssuesEvent
2019-12-07 11:23:47
Ignitus/Ignitus-client
https://api.github.com/repos/Ignitus/Ignitus-client
opened
CVE-2019-10747 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz
security vulnerability
## CVE-2019-10747 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary> <p> <details><summary><b>set-value-2.0.0.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/Ignitus-client/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/Ignitus-client/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - jest-cli-24.7.1.tgz (Root Library) - core-24.7.1.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>set-value-0.4.3.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/Ignitus-client/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/Ignitus-client/node_modules/union-value/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - jest-cli-24.7.1.tgz (Root Library) - core-24.7.1.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - union-value-1.0.0.tgz - :x: **set-value-0.4.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Ignitus/Ignitus-client/commit/4a136622e36d4bca4d34d3a5d332b6d73cdda58d">4a136622e36d4bca4d34d3a5d332b6d73cdda58d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads. <p>Publish Date: 2019-08-23 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10747>CVE-2019-10747</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f">https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f</a></p> <p>Release Date: 2019-07-24</p> <p>Fix Resolution: 2.0.1,3.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-10747 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz - ## CVE-2019-10747 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary> <p> <details><summary><b>set-value-2.0.0.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/Ignitus-client/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/Ignitus-client/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - jest-cli-24.7.1.tgz (Root Library) - core-24.7.1.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>set-value-0.4.3.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/Ignitus-client/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/Ignitus-client/node_modules/union-value/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - jest-cli-24.7.1.tgz (Root Library) - core-24.7.1.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - union-value-1.0.0.tgz - :x: **set-value-0.4.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Ignitus/Ignitus-client/commit/4a136622e36d4bca4d34d3a5d332b6d73cdda58d">4a136622e36d4bca4d34d3a5d332b6d73cdda58d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads. <p>Publish Date: 2019-08-23 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10747>CVE-2019-10747</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f">https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f</a></p> <p>Release Date: 2019-07-24</p> <p>Fix Resolution: 2.0.1,3.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in set value tgz set value tgz cve high severity vulnerability vulnerable libraries set value tgz set value tgz set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file tmp ws scm ignitus client package json path to vulnerable library tmp ws scm ignitus client node modules set value package json dependency hierarchy jest cli tgz root library core tgz micromatch tgz snapdragon tgz base tgz cache base tgz x set value tgz vulnerable library set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file tmp ws scm ignitus client package json path to vulnerable library tmp ws scm ignitus client node modules union value node modules set value package json dependency hierarchy jest cli tgz root library core tgz micromatch tgz snapdragon tgz base tgz cache base tgz union value tgz x set value tgz vulnerable library found in head commit a href vulnerability details set value is vulnerable to prototype pollution in versions lower than the function mixin deep could be tricked into adding or modifying properties of object prototype using any of the constructor prototype and proto payloads publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
5,163
7,934,001,209
IssuesEvent
2018-07-08 14:05:21
frc4571/The-Beelzebub
https://api.github.com/repos/frc4571/The-Beelzebub
closed
Convert coordinates obtained from GRIP into a motion profile
drive-subsystem vision-processing
This is to use the GRIP output to move the drive train along a particular motion profile consisting of a series of distance and angle values
1.0
Convert coordinates obtained from GRIP into a motion profile - This is to use the GRIP output to move the drive train along a particular motion profile consisting of a series of distance and angle values
process
convert coordinates obtained from grip into a motion profile this is to use the grip output to move the drive train along a particular motion profile consisting of a series of distance and angle values
1
320,906
27,492,837,846
IssuesEvent
2023-03-04 20:49:32
4ian/GDevelop
https://api.github.com/repos/4ian/GDevelop
closed
Crash while using an editor ->I tried to undo something while writing the name of a variable, but I believe it gives an error whenever I type something and try to use Undo/Redo.
💥crash 👋 Needs confirmation/testing
=> Please write here a short description of when the error occured and how to reproduce it. You also may have to create an account on GitHub before posting. When you're ready, click on "Submit new issue". Don't change the rest of the message. Thanks! -> **I tried to undo something while writing the name of a variable, but I believe it gives an error whenever I type something and try to use Undo/Redo.** ## Error stack (don't write anything here) ``` RuntimeError: function signature mismatch at wasm-function[3807]:0x17dbbd at wasm-function[1873]:0x4b2d3 at wasm-function[2286]:0x6924f at wasm-function[1340]:0x3482f at InstructionSentenceFormatter.GetAsFormattedText.InstructionSentenceFormatter.GetAsFormattedText [as getAsFormattedText] (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/libGD.js?cache-buster=5.0.0-beta115-e4fc065dc13f86cee1d996de6201ed470a04ea66:9:1333508) at file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/1.aea6426a.chunk.js:1:710754 at file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/1.aea6426a.chunk.js:1:711574 at file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/1.aea6426a.chunk.js:1:322331 at Bi (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:220738) at _a (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:227578) at Sa (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:234326) at Uo (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:258320) at Fo (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:258704) at Eu (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:265740) at wu (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:265120) at yu (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:264141) at qo (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:263010) at Object.enqueueForceUpdate (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:212721) at n.w.forceUpdate (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:162773) at n.value (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/1.aea6426a.chunk.js:1:704946) at file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/1.aea6426a.chunk.js:1:804348 at i (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:306395) ``` ## Component stack (don't write anything here) ``` in Unknown in DropTarget(Component) in DragSource(DropTarget(Component)) in Unknown in Unknown in div in Unknown in DropTarget(Component) in Unknown in f in div in Unknown in n in div in n in span in div in div in div in div in div in t in DragSource(t) in div in div in t in DropTarget(t) in div in div in t in t in Scrolling(t) in div in t in div in t in re in Unknown in L in div in n in div in n in u in a in n in n in t in Unknown in Unknown in div in n in div in Unknown in n in m in m in m in m in m in s in n in t in n in Unknown in v in t in m in n in n in n in DragDropContext(n) in n in s ``` ## Other details * IDE version: 5.0.0-beta115-e4fc065dc13f86cee1d996de6201ed470a04ea66 * Arch: x64, * Platform Name: win32, * System Version: 10.0.19043, * User Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) GDevelop5/5.0.0-beta115 Chrome/80.0.3987.165 Electron/8.2.5 Safari/537.36,
1.0
Crash while using an editor ->I tried to undo something while writing the name of a variable, but I believe it gives an error whenever I type something and try to use Undo/Redo. - => Please write here a short description of when the error occured and how to reproduce it. You also may have to create an account on GitHub before posting. When you're ready, click on "Submit new issue". Don't change the rest of the message. Thanks! -> **I tried to undo something while writing the name of a variable, but I believe it gives an error whenever I type something and try to use Undo/Redo.** ## Error stack (don't write anything here) ``` RuntimeError: function signature mismatch at wasm-function[3807]:0x17dbbd at wasm-function[1873]:0x4b2d3 at wasm-function[2286]:0x6924f at wasm-function[1340]:0x3482f at InstructionSentenceFormatter.GetAsFormattedText.InstructionSentenceFormatter.GetAsFormattedText [as getAsFormattedText] (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/libGD.js?cache-buster=5.0.0-beta115-e4fc065dc13f86cee1d996de6201ed470a04ea66:9:1333508) at file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/1.aea6426a.chunk.js:1:710754 at file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/1.aea6426a.chunk.js:1:711574 at file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/1.aea6426a.chunk.js:1:322331 at Bi (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:220738) at _a (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:227578) at Sa (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:234326) at Uo (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:258320) at Fo (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:258704) at Eu (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:265740) at wu (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:265120) at yu (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:264141) at qo (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:263010) at Object.enqueueForceUpdate (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:212721) at n.w.forceUpdate (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:162773) at n.value (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/1.aea6426a.chunk.js:1:704946) at file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/1.aea6426a.chunk.js:1:804348 at i (file:///C:/Users/Overl/AppData/Local/Programs/gdevelop/resources/app.asar/www/static/js/60.be227616.chunk.js:2:306395) ``` ## Component stack (don't write anything here) ``` in Unknown in DropTarget(Component) in DragSource(DropTarget(Component)) in Unknown in Unknown in div in Unknown in DropTarget(Component) in Unknown in f in div in Unknown in n in div in n in span in div in div in div in div in div in t in DragSource(t) in div in div in t in DropTarget(t) in div in div in t in t in Scrolling(t) in div in t in div in t in re in Unknown in L in div in n in div in n in u in a in n in n in t in Unknown in Unknown in div in n in div in Unknown in n in m in m in m in m in m in s in n in t in n in Unknown in v in t in m in n in n in n in DragDropContext(n) in n in s ``` ## Other details * IDE version: 5.0.0-beta115-e4fc065dc13f86cee1d996de6201ed470a04ea66 * Arch: x64, * Platform Name: win32, * System Version: 10.0.19043, * User Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) GDevelop5/5.0.0-beta115 Chrome/80.0.3987.165 Electron/8.2.5 Safari/537.36,
non_process
crash while using an editor i tried to undo something while writing the name of a variable but i believe it gives an error whenever i type something and try to use undo redo please write here a short description of when the error occured and how to reproduce it you also may have to create an account on github before posting when you re ready click on submit new issue don t change the rest of the message thanks i tried to undo something while writing the name of a variable but i believe it gives an error whenever i type something and try to use undo redo error stack don t write anything here runtimeerror function signature mismatch at wasm function at wasm function at wasm function at wasm function at instructionsentenceformatter getasformattedtext instructionsentenceformatter getasformattedtext file c users overl appdata local programs gdevelop resources app asar www libgd js cache buster at file c users overl appdata local programs gdevelop resources app asar www static js chunk js at file c users overl appdata local programs gdevelop resources app asar www static js chunk js at file c users overl appdata local programs gdevelop resources app asar www static js chunk js at bi file c users overl appdata local programs gdevelop resources app asar www static js chunk js at a file c users overl appdata local programs gdevelop resources app asar www static js chunk js at sa file c users overl appdata local programs gdevelop resources app asar www static js chunk js at uo file c users overl appdata local programs gdevelop resources app asar www static js chunk js at fo file c users overl appdata local programs gdevelop resources app asar www static js chunk js at eu file c users overl appdata local programs gdevelop resources app asar www static js chunk js at wu file c users overl appdata local programs gdevelop resources app asar www static js chunk js at yu file c users overl appdata local programs gdevelop resources app asar www static js chunk js at qo file c users overl appdata local programs gdevelop resources app asar www static js chunk js at object enqueueforceupdate file c users overl appdata local programs gdevelop resources app asar www static js chunk js at n w forceupdate file c users overl appdata local programs gdevelop resources app asar www static js chunk js at n value file c users overl appdata local programs gdevelop resources app asar www static js chunk js at file c users overl appdata local programs gdevelop resources app asar www static js chunk js at i file c users overl appdata local programs gdevelop resources app asar www static js chunk js component stack don t write anything here in unknown in droptarget component in dragsource droptarget component in unknown in unknown in div in unknown in droptarget component in unknown in f in div in unknown in n in div in n in span in div in div in div in div in div in t in dragsource t in div in div in t in droptarget t in div in div in t in t in scrolling t in div in t in div in t in re in unknown in l in div in n in div in n in u in a in n in n in t in unknown in unknown in div in n in div in unknown in n in m in m in m in m in m in s in n in t in n in unknown in v in t in m in n in n in n in dragdropcontext n in n in s other details ide version arch platform name system version user agent mozilla windows nt applewebkit khtml like gecko chrome electron safari
0
322,903
9,829,794,945
IssuesEvent
2019-06-16 01:14:22
sqlalchemy/sqlalchemy
https://api.github.com/repos/sqlalchemy/sqlalchemy
closed
all_orm_descriptors is including __mapper__
bug high priority orm
the check is looking for `InspectionAttr` and not verifying that it's a descriptor ``` from sqlalchemy import Column from sqlalchemy import Integer from sqlalchemy import String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy import inspect Base = declarative_base() class HasAsDict: def _asdict(self): return { key: getattr(self, key) for key in inspect(self).mapper.all_orm_descriptors.keys() } class A(HasAsDict, Base): __tablename__ = "a" id = Column(Integer, primary_key=True) x = Column(String) y = Column(String) z = Column(String) a1 = A() a1.y = "y" a1.x = "x" a1.z = "z" print(a1._asdict()) ``` `{'__mapper__': <Mapper at 0x7fd7a6abb9e8; A>, 'id': None, 'y': 'y', 'z': 'z', 'x': 'x'} `
1.0
all_orm_descriptors is including __mapper__ - the check is looking for `InspectionAttr` and not verifying that it's a descriptor ``` from sqlalchemy import Column from sqlalchemy import Integer from sqlalchemy import String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy import inspect Base = declarative_base() class HasAsDict: def _asdict(self): return { key: getattr(self, key) for key in inspect(self).mapper.all_orm_descriptors.keys() } class A(HasAsDict, Base): __tablename__ = "a" id = Column(Integer, primary_key=True) x = Column(String) y = Column(String) z = Column(String) a1 = A() a1.y = "y" a1.x = "x" a1.z = "z" print(a1._asdict()) ``` `{'__mapper__': <Mapper at 0x7fd7a6abb9e8; A>, 'id': None, 'y': 'y', 'z': 'z', 'x': 'x'} `
non_process
all orm descriptors is including mapper the check is looking for inspectionattr and not verifying that it s a descriptor from sqlalchemy import column from sqlalchemy import integer from sqlalchemy import string from sqlalchemy ext declarative import declarative base from sqlalchemy import inspect base declarative base class hasasdict def asdict self return key getattr self key for key in inspect self mapper all orm descriptors keys class a hasasdict base tablename a id column integer primary key true x column string y column string z column string a y y x x z z print asdict mapper id none y y z z x x
0
853
3,316,098,316
IssuesEvent
2015-11-06 15:30:29
OmniLayer/omnicore
https://api.github.com/repos/OmniLayer/omnicore
closed
List all requirements, and tag issues with milestone for 0.0.10
process
Using one issue to track the progress and requirements for the next release seems to have limited value, so I'd like to try to use the milestone feature of GitHub, which is probably a much cleaner solution. I would like to gather *all* requirements necessary for the release, and I created a [milestone](https://github.com/OmniLayer/omnicore/milestones/0.0.10) for this purpose. Here are four simple guidelines: 1. once there is *no* open issue tagged with [0.0.10](https://github.com/OmniLayer/omnicore/milestones/0.0.10) tag left, it means [0.0.10](https://github.com/OmniLayer/omnicore/milestones/0.0.10) is *ready* and we can publish a release candidate 2. if there is an unresolved bug, or feature that *must* be resolved or included in the release, and there is no related issue for it, a [new issue](https://github.com/OmniLayer/omnicore/issues/new) should be created 3. if there is an open issue, which is *necessary* for the release, but *without* the [0.0.10](https://github.com/OmniLayer/omnicore/milestones/0.0.10) milestone tag, the tag should be added 4. if there is an open issue, which is tagged with [0.0.10](https://github.com/OmniLayer/omnicore/milestones/0.0.10), but *not* considered as necessary for the release, the milestone tag should be removed This means as consequence that as long as there are open issues with milestone tag, there won't be a release (or the process failed). This should help us to prioritize tasks and to get a better understanding of what still needs to be done. Personally I'm going to be rather conservative, and I won't start to tag issues, which may, or may not be required for the release, and I would really welcome, if someone from the "planning committee" takes a moment to tackle it. I'm especially thinking about items such as the "send all" transaction type, or whether documentation is considered as a requirement, which I can't evaluate from my point of view, because I have no insight of the business side. Maintainers can add and remove milestone tags similar to issue labels: ![set_milestone](https://cloud.githubusercontent.com/assets/5836089/9428478/5b58e09a-49af-11e5-8e79-b0da4158db2a.png) Other contributors are welcome, and can leave a note in this thread, so a maintainer may tag the related issue. The process itself may also be discussed here of course.
1.0
List all requirements, and tag issues with milestone for 0.0.10 - Using one issue to track the progress and requirements for the next release seems to have limited value, so I'd like to try to use the milestone feature of GitHub, which is probably a much cleaner solution. I would like to gather *all* requirements necessary for the release, and I created a [milestone](https://github.com/OmniLayer/omnicore/milestones/0.0.10) for this purpose. Here are four simple guidelines: 1. once there is *no* open issue tagged with [0.0.10](https://github.com/OmniLayer/omnicore/milestones/0.0.10) tag left, it means [0.0.10](https://github.com/OmniLayer/omnicore/milestones/0.0.10) is *ready* and we can publish a release candidate 2. if there is an unresolved bug, or feature that *must* be resolved or included in the release, and there is no related issue for it, a [new issue](https://github.com/OmniLayer/omnicore/issues/new) should be created 3. if there is an open issue, which is *necessary* for the release, but *without* the [0.0.10](https://github.com/OmniLayer/omnicore/milestones/0.0.10) milestone tag, the tag should be added 4. if there is an open issue, which is tagged with [0.0.10](https://github.com/OmniLayer/omnicore/milestones/0.0.10), but *not* considered as necessary for the release, the milestone tag should be removed This means as consequence that as long as there are open issues with milestone tag, there won't be a release (or the process failed). This should help us to prioritize tasks and to get a better understanding of what still needs to be done. Personally I'm going to be rather conservative, and I won't start to tag issues, which may, or may not be required for the release, and I would really welcome, if someone from the "planning committee" takes a moment to tackle it. I'm especially thinking about items such as the "send all" transaction type, or whether documentation is considered as a requirement, which I can't evaluate from my point of view, because I have no insight of the business side. Maintainers can add and remove milestone tags similar to issue labels: ![set_milestone](https://cloud.githubusercontent.com/assets/5836089/9428478/5b58e09a-49af-11e5-8e79-b0da4158db2a.png) Other contributors are welcome, and can leave a note in this thread, so a maintainer may tag the related issue. The process itself may also be discussed here of course.
process
list all requirements and tag issues with milestone for using one issue to track the progress and requirements for the next release seems to have limited value so i d like to try to use the milestone feature of github which is probably a much cleaner solution i would like to gather all requirements necessary for the release and i created a for this purpose here are four simple guidelines once there is no open issue tagged with tag left it means is ready and we can publish a release candidate if there is an unresolved bug or feature that must be resolved or included in the release and there is no related issue for it a should be created if there is an open issue which is necessary for the release but without the milestone tag the tag should be added if there is an open issue which is tagged with but not considered as necessary for the release the milestone tag should be removed this means as consequence that as long as there are open issues with milestone tag there won t be a release or the process failed this should help us to prioritize tasks and to get a better understanding of what still needs to be done personally i m going to be rather conservative and i won t start to tag issues which may or may not be required for the release and i would really welcome if someone from the planning committee takes a moment to tackle it i m especially thinking about items such as the send all transaction type or whether documentation is considered as a requirement which i can t evaluate from my point of view because i have no insight of the business side maintainers can add and remove milestone tags similar to issue labels other contributors are welcome and can leave a note in this thread so a maintainer may tag the related issue the process itself may also be discussed here of course
1
38,131
10,142,322,650
IssuesEvent
2019-08-03 23:01:13
zurb/foundation-sites
https://api.github.com/repos/zurb/foundation-sites
closed
build: clone with less commits
PR open 🚚 build
<!--- --------------------------------------------------------------------- --> <!--- Please fill the following template --> <!--- Your issue may be ignored otherwise --> <!--- --------------------------------------------------------------------- --> <!--- Only submit bug or feature requests here. For help or questions to --> <!--- the community, see the forum: https://foundation.zurb.com/forum --> ## Expected Behavior <!--- If describing a bug, tell us what should happen. --> <!--- If suggesting a change/feature, tell us why and how it should work. --> Travis CI should not clone the last 50 commits but much less to improve the performance. ## Current Behavior <!--- If describing a bug, tell us what happens instead. --> <!--- If suggesting a change/feature, explain the difference from --> <!--- current behavior. --> Currently the default depth (50) is used for `git clone`. ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change. --> Set the depth to `3`. ## Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most --> <!--- useful in the real world --> Faster builds. ## Checklist (all required): <!--- Go over all the following points, and put an `x` in the boxes. --> <!--- If you're unsure about any of these, don't hesitate to ask. --> - [x] I have read and follow the [CONTRIBUTING](CONTRIBUTING.md) document. - [ ] This is a bug report or a feature request. - [x] There are no other issues similar to this one. - [x] The issue title is descriptive. - [x] The template is fully and correctly filled. <!--- --------------------------------------------------------------------- --> <!--- For more information, see the CONTRIBUTING.md document --> <!--- Thank you for your issue and happy coding ;) --> <!--- --------------------------------------------------------------------- -->
1.0
build: clone with less commits - <!--- --------------------------------------------------------------------- --> <!--- Please fill the following template --> <!--- Your issue may be ignored otherwise --> <!--- --------------------------------------------------------------------- --> <!--- Only submit bug or feature requests here. For help or questions to --> <!--- the community, see the forum: https://foundation.zurb.com/forum --> ## Expected Behavior <!--- If describing a bug, tell us what should happen. --> <!--- If suggesting a change/feature, tell us why and how it should work. --> Travis CI should not clone the last 50 commits but much less to improve the performance. ## Current Behavior <!--- If describing a bug, tell us what happens instead. --> <!--- If suggesting a change/feature, explain the difference from --> <!--- current behavior. --> Currently the default depth (50) is used for `git clone`. ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change. --> Set the depth to `3`. ## Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most --> <!--- useful in the real world --> Faster builds. ## Checklist (all required): <!--- Go over all the following points, and put an `x` in the boxes. --> <!--- If you're unsure about any of these, don't hesitate to ask. --> - [x] I have read and follow the [CONTRIBUTING](CONTRIBUTING.md) document. - [ ] This is a bug report or a feature request. - [x] There are no other issues similar to this one. - [x] The issue title is descriptive. - [x] The template is fully and correctly filled. <!--- --------------------------------------------------------------------- --> <!--- For more information, see the CONTRIBUTING.md document --> <!--- Thank you for your issue and happy coding ;) --> <!--- --------------------------------------------------------------------- -->
non_process
build clone with less commits expected behavior travis ci should not clone the last commits but much less to improve the performance current behavior currently the default depth is used for git clone possible solution set the depth to context faster builds checklist all required i have read and follow the contributing md document this is a bug report or a feature request there are no other issues similar to this one the issue title is descriptive the template is fully and correctly filled
0
113,985
11,834,616,201
IssuesEvent
2020-03-23 09:13:46
uStudioCompany/ustudio-ui
https://api.github.com/repos/uStudioCompany/ustudio-ui
closed
[BUG] (Tooltip) incorrect classNames list
bug documentation
**Describe the bug** Tooltip's classNames property type isn`t equal to it is specified in the PropTypes. **Screenshots** ![image](https://user-images.githubusercontent.com/23137619/77154331-25d33180-6aa4-11ea-93d8-90b40737e04b.png) ![image](https://user-images.githubusercontent.com/23137619/77154355-31265d00-6aa4-11ea-9e5f-7f64f0ce08e7.png)
1.0
[BUG] (Tooltip) incorrect classNames list - **Describe the bug** Tooltip's classNames property type isn`t equal to it is specified in the PropTypes. **Screenshots** ![image](https://user-images.githubusercontent.com/23137619/77154331-25d33180-6aa4-11ea-93d8-90b40737e04b.png) ![image](https://user-images.githubusercontent.com/23137619/77154355-31265d00-6aa4-11ea-9e5f-7f64f0ce08e7.png)
non_process
tooltip incorrect classnames list describe the bug tooltip s classnames property type isn t equal to it is specified in the proptypes screenshots
0
9,239
12,268,965,089
IssuesEvent
2020-05-07 13:22:53
bisq-network/bisq
https://api.github.com/repos/bisq-network/bisq
closed
Bisq does not add "dust" outputs to miner fee
a:bug in:trade-process in:wallet is:critical
### Background According to Bitcoin consensus rules, no output can be less than 546 sats. If a transaction with such an output is broadcast, it will be rejected. ### Description Bisq attempts to broadcast Bitcoin transactions containing "dust" outputs, instead of simply adding the dust to the miner fee like most wallets do. #### Version v1.2.7 ### Steps to reproduce This is the difficult part, as users randomly do this if their transaction coincidentally results in an output of less than 546 sats. However, since Bisq assumes transactions it creates are valid (without checking the dust rule), this results in triggering other random Bisq issues. Sometimes it results in a failed trade, sometimes it results in a corrupted wallet. ### Expected behaviour The dust output should not be created, and instead added to fees. ### Actual behaviour The dust output is created, and the Bitcoin transaction is rejected, causing further issues in Bisq depending on the situation.
1.0
Bisq does not add "dust" outputs to miner fee - ### Background According to Bitcoin consensus rules, no output can be less than 546 sats. If a transaction with such an output is broadcast, it will be rejected. ### Description Bisq attempts to broadcast Bitcoin transactions containing "dust" outputs, instead of simply adding the dust to the miner fee like most wallets do. #### Version v1.2.7 ### Steps to reproduce This is the difficult part, as users randomly do this if their transaction coincidentally results in an output of less than 546 sats. However, since Bisq assumes transactions it creates are valid (without checking the dust rule), this results in triggering other random Bisq issues. Sometimes it results in a failed trade, sometimes it results in a corrupted wallet. ### Expected behaviour The dust output should not be created, and instead added to fees. ### Actual behaviour The dust output is created, and the Bitcoin transaction is rejected, causing further issues in Bisq depending on the situation.
process
bisq does not add dust outputs to miner fee background according to bitcoin consensus rules no output can be less than sats if a transaction with such an output is broadcast it will be rejected description bisq attempts to broadcast bitcoin transactions containing dust outputs instead of simply adding the dust to the miner fee like most wallets do version steps to reproduce this is the difficult part as users randomly do this if their transaction coincidentally results in an output of less than sats however since bisq assumes transactions it creates are valid without checking the dust rule this results in triggering other random bisq issues sometimes it results in a failed trade sometimes it results in a corrupted wallet expected behaviour the dust output should not be created and instead added to fees actual behaviour the dust output is created and the bitcoin transaction is rejected causing further issues in bisq depending on the situation
1
18,066
24,079,660,254
IssuesEvent
2022-09-19 04:33:24
cmadland/assessment
https://api.github.com/repos/cmadland/assessment
opened
delucaEstablishingFoundationValid2013
conceptions of assessment assessment as testing assessment as purpose assessment as process assessment as format teacher education assessment confidence
# Reference DeLuca, C., Chavez, T., & Cao, C. (2013). Establishing a foundation for valid teacher judgement on student learning: The role of pre-service assessment education. Assessment in Education: Principles, Policy & Practice, 20(1), 107–126. https://doi.org/10/gj5v98 # Summary DeLuca et al. surveyed 48 pre-service teachers before and after an undergraduate course in educational measurement. They used a framework of four primary conceptions of assessment as testing, format, purpose, and process, which are consecutively ordered by complexity. They found that participants conceptions of assessment shifted from more simple views of assessment as testing to more complex and integrated views of assessment as purpose and process, and, to a lesser extent, assessment as format. Interestingly, they found that assessment as format was prominent as a secondary conception. # Abstract - The purpose of this research was to explore changes in teacher candidates’ confidence in assessment as a foundation for promoting valid teacher judgements on student learning. Based on a sample of 48 teacher candidates enrolled in a semester-long measurement course at one teacher education programme in Florida, this survey-based study provided evidence that teacher candidates can develop confidence in assessment whilst deepening their conceptions of assessment. The majority of teacher candidates expanded their conception of assessment from one based solely on testing to one that recognised multiple forms of assessment that serve varying purposes. Simultaneously, teacher candidates expressed greater confidence in practical assessment approaches and factors related to assessment praxis. Based on study findings, we argue that pre-service teacher education has a critical role to play in promoting assessment literacy in beginning teachers and providing a foundation for teachers’ continued learning about assessment throughout their careers. The paper concludes with implications for future research in the area of assessment education. - The purpose of this study was to examine pre-service teacher candidates’ confidence levels in relation to their changing conceptions of assessment during the term of a senior-level assessment course. The specific research questions guiding this study were the following: - 1. How do teacher candidates’ conceptions of assessment differ at the beginning and end of a pre-service course on student assessment? 2. How does teacher candidates’ confidence toward assessment practices change in relation to their shifting conceptions of assessment? # Related - # Quotes - # Developing teacher confidence in assessment - Pre-service assessment education has been identified as a fundamental structure for preparing teachers to assess student learning (Chen 2005). Research conducted on the impact of explicit assessment courses has generally found that these courses have the potential to complement field-based teaching experience (i.e. practical) and develop a foundation for assessment literacy in teacher candidates (Graham 2005). - # Conceptions of assessment - Brown, Lake, and Matters (2011) assert that teachers maintain multiple conceptions of classroom assessment based on assessment format, purpose and use. Derived from formal assessment education as well as from the practicalities and policies that govern teachers’ daily work, teachers’ #[[conceptions of assessment]] shape their in-class assessment practices (Brown 2009; Cizek et al. 1995). We present four dominant conceptions of assessment as an analytic framework for our study on the linkage between teacher candidates’ confidence and their changing understandings about assessment. Specifically, the four conceptions of assessment that we describe are: (a) #[[assessment as testing]], (b) #[[assessment as format]], (c) #[[assessment as purpose]] and (d) #[[assessment as process]] . We assert that while none of these conceptions represent simple measurement processes, each conception increases in complexity by connecting together assessment concepts and extending toward other constructs in teaching and learning. - ## Asssessment as testing - Our first conception of assessment assimilates assessment with traditional, paperand-pencil testing. Within the USA, tests are often used as summative measures of student learning and can be either teacher-constructed or external and standardised. - Moreover, due to efficiency and a lack of knowledge about other forms of assessment, teachers continue to construct and use tests as the dominant means to measure student learning in their classrooms. - Although classical and modern test theories are not simple, when teachers are not aware of the psychometric underpinnings of testing or educated on issues related to test reliability, fairness and validity, they may perceive this form of assessment as easy and straightforward and use it with confidence (DeLuca and Klinger 2010). - ## Assessment as format - our second conception expands on this association to include alternative forms of assessment and to delineate a conception based on assessment format. - A typical classification of assessments is that of performance, product and objectively scored assessments (Airasian and Russell 2008). Performance and product assessments can be further delineated as contrived or authentic assessments, depending upon their applicability to ‘real-world’ experiences (Bartz, Anderson-Robinson, and Hillman 1994). - Conceptions of assessment by format typically differentiate assessment practices based on the cognitive level of the standard/curriculum expectations being assessed. - Higher order thinking skills are generally associated with performance and product assessments that enable students to construct responses and assert reasoning, while lower order thinking skills are associated with objectively scored assessments (McMillan 2007). - ## Assessment as purpose - Third, one of the most common conceptions of assessment is based on purpose and use, with the distinction between formative and summative assessment purposes widely used across assessment literature (Pellegrino, Chudowsky, and Glaser 2001). - Formative assessment refers to data collected throughout the instructional process and used to provide feedback to inform teaching and learning processes. - Summative assessment refers to data collected at the end of a unit or term in order to justify judgements on student achievement and report on student learning (i.e. contribute to student grades). - Another distinction on assessment purpose and use is based on the seminal work of Black and Wiliam (1998), which described purposes of assessment of learning and assessment for learning. - Subsumed within AFL, Earl (2003) articulated a third purpose of assessment, assessment as learning, which focuses on using assessments to support the development of students’ metacognition and selfregulation abilities. - ## Assessment as process - Finally, closely linked to assessment as, for, and of learning is assessment conceived as a process of interpretation and integration. Pellegrino, Chudowsky, and Glaser (2001, 36) state: ‘assessment is a process of reasoning from evidence’. This conception of assessment includes an awareness of the limitations of various assessment methods (including considerations of validity, reliability and bias) in relation to the specific context of teaching and learning. Hence, teachers’ ability to make meaning from assessment data becomes increasingly important to this conception as assessment is viewed as an essential component of teaching and integrally connected with curriculum and learning (Airasian and Russell 2008; Pellegrino, Chudowsky, and Glaser 2001). - # Data analyses - Results were analysed in relation to: (a) participants’ changing conceptions of assessment, (b) participants’ changing confidence in assessment and (c) the interaction between participants’ changing conceptions and confidence in assessment. - data were coded based on participants’ primary and, when available, secondary conceptions of assessment as testing, purpose, format and process. Dual coding of data by primary and secondary conceptions enabled a co-occurrence frequency analysis that provided a more complex analysis of participants’ understandings of assessment and which was helpful in ‘understanding how thematic domains, concepts, or ideas were distributed within a data set’ (Namey et al. 2008, 145). - Quantitatively, #[[descriptive statistics]] were calculated related to the gender, age, programme of study and year of study distribution across the sample. Then, changes in participants’ confidence levels were analysed through an #[[exploratory factor analysis]] using a promax (oblique) rotation. Although close-ended items had a theoretical foundation (i.e. Popham 2008), a #[[principal component factor analysis]] was conducted for each separate section of the pre-questionnaire to identify confidence factors across the Likert scale items based on responses from the 48 participants. #[[Factor loadings]] were determined by the greatest value for each item, with loadings of at least 0.40. No items maintained lower loading values. #[[Cronbach alpha]] values were also calculated for #reliability indices of each factor. We recognise that given the sample size of 48 participants, the generalisability and internal stability of factors are somewhat problematic. Accordingly, we caution readers to interpret factors with the caveat that the results from this study are preliminary and require further validation based on a large sample of participants. - In order to determine changes over time and the interaction between participants’ assessment conceptions and their confidence, we conducted a #[[repeated measures analysis]] across both questionnaire administrations (Stevens 2001). In preparation for the repeated measures analysis, assessment conception codes were converted to numerical values (indicating the ordinal progression from the most simplistic conception to the most complex) and placed on a proportional scale to confidence factors. A repeated measures analysis (two within group) was conducted for confidence and conception factors across the two data time points. The repeated measures analysis also allowed for explorations of any interaction effects between the teacher candidates’ conceptions and the confidence factors. A follow-up #[[dependent t-test]] addressed the research questions on how the candidates’ conceptions of assessment and confidence toward assessment practices differed between the beginning and end of the measurement course (Stevens 2001). #Significance was based on a #[[Bonferroni adjusted p-value]] of 0.007. As a caveat to this data analysis strategy, we recognise that the sample size limits the reliability of the measure, the internal stability of factor loadings and the resulting effect sizes, which mitigates the generalisability of results from factor analyses and significance analyses (i.e. t-test and ANOVA). We urge that results not be interpreted as generalised findings but rather as a basis for further validation research on these suggested factors and observed trends. - # Results - ## Conceptions of assessment - Data supported that participants’ conceptions of assessment developed from an initial dominant view of assessment as testing to a more complex view of assessment as format, purpose and process. - In analysing primary and secondary conceptions of assessment, data pointed to participants’ not only changing their conceptions from assessment as testing to assessment as purpose and as process, but also understanding assessment in more complex and interconnected ways. - As evidenced through this study, explicit assessment education can enhance teacher candidates’ confidence in assessment praxis and approaches, while deepening their conceptual and theoretical understandings of educational assessment. As a result of this learning, teacher candidates may foster a realistic and optimistic orientation for engaging in more valid forms of assessment within their future schools and classrooms. - - # Citations - - mentioning cite from section 'review of related literature' of publication: - Yidana, M. B., P., A. P. (2018). Economics Teachers’ Conceptions of Classroom Assessment. Int J Innov Educ Res, 10(6), 153-174. https://doi.org/10.31686/ijier.vol6.iss10.1176 --- “…Similar studies (Brown & Hirschfeld, 2008;Levy-Vered & Alhija, 2015) have found that attending an assessment course or having more training in assessment did not improve teachers' conceptions of assessment. However, there are studies that suggest that teachers' conception of assessment developed after receiving periodic training in assessment (DeLuca, Chavez & Cao, 2013;Smith, Hill, Cowie & Gilmore, 2014). Interestingly, the findings of the study also showed that after teachers had attended the course on assessment, their irrelevance conception of assessment improved significantly.…”
1.0
delucaEstablishingFoundationValid2013 - # Reference DeLuca, C., Chavez, T., & Cao, C. (2013). Establishing a foundation for valid teacher judgement on student learning: The role of pre-service assessment education. Assessment in Education: Principles, Policy & Practice, 20(1), 107–126. https://doi.org/10/gj5v98 # Summary DeLuca et al. surveyed 48 pre-service teachers before and after an undergraduate course in educational measurement. They used a framework of four primary conceptions of assessment as testing, format, purpose, and process, which are consecutively ordered by complexity. They found that participants conceptions of assessment shifted from more simple views of assessment as testing to more complex and integrated views of assessment as purpose and process, and, to a lesser extent, assessment as format. Interestingly, they found that assessment as format was prominent as a secondary conception. # Abstract - The purpose of this research was to explore changes in teacher candidates’ confidence in assessment as a foundation for promoting valid teacher judgements on student learning. Based on a sample of 48 teacher candidates enrolled in a semester-long measurement course at one teacher education programme in Florida, this survey-based study provided evidence that teacher candidates can develop confidence in assessment whilst deepening their conceptions of assessment. The majority of teacher candidates expanded their conception of assessment from one based solely on testing to one that recognised multiple forms of assessment that serve varying purposes. Simultaneously, teacher candidates expressed greater confidence in practical assessment approaches and factors related to assessment praxis. Based on study findings, we argue that pre-service teacher education has a critical role to play in promoting assessment literacy in beginning teachers and providing a foundation for teachers’ continued learning about assessment throughout their careers. The paper concludes with implications for future research in the area of assessment education. - The purpose of this study was to examine pre-service teacher candidates’ confidence levels in relation to their changing conceptions of assessment during the term of a senior-level assessment course. The specific research questions guiding this study were the following: - 1. How do teacher candidates’ conceptions of assessment differ at the beginning and end of a pre-service course on student assessment? 2. How does teacher candidates’ confidence toward assessment practices change in relation to their shifting conceptions of assessment? # Related - # Quotes - # Developing teacher confidence in assessment - Pre-service assessment education has been identified as a fundamental structure for preparing teachers to assess student learning (Chen 2005). Research conducted on the impact of explicit assessment courses has generally found that these courses have the potential to complement field-based teaching experience (i.e. practical) and develop a foundation for assessment literacy in teacher candidates (Graham 2005). - # Conceptions of assessment - Brown, Lake, and Matters (2011) assert that teachers maintain multiple conceptions of classroom assessment based on assessment format, purpose and use. Derived from formal assessment education as well as from the practicalities and policies that govern teachers’ daily work, teachers’ #[[conceptions of assessment]] shape their in-class assessment practices (Brown 2009; Cizek et al. 1995). We present four dominant conceptions of assessment as an analytic framework for our study on the linkage between teacher candidates’ confidence and their changing understandings about assessment. Specifically, the four conceptions of assessment that we describe are: (a) #[[assessment as testing]], (b) #[[assessment as format]], (c) #[[assessment as purpose]] and (d) #[[assessment as process]] . We assert that while none of these conceptions represent simple measurement processes, each conception increases in complexity by connecting together assessment concepts and extending toward other constructs in teaching and learning. - ## Asssessment as testing - Our first conception of assessment assimilates assessment with traditional, paperand-pencil testing. Within the USA, tests are often used as summative measures of student learning and can be either teacher-constructed or external and standardised. - Moreover, due to efficiency and a lack of knowledge about other forms of assessment, teachers continue to construct and use tests as the dominant means to measure student learning in their classrooms. - Although classical and modern test theories are not simple, when teachers are not aware of the psychometric underpinnings of testing or educated on issues related to test reliability, fairness and validity, they may perceive this form of assessment as easy and straightforward and use it with confidence (DeLuca and Klinger 2010). - ## Assessment as format - our second conception expands on this association to include alternative forms of assessment and to delineate a conception based on assessment format. - A typical classification of assessments is that of performance, product and objectively scored assessments (Airasian and Russell 2008). Performance and product assessments can be further delineated as contrived or authentic assessments, depending upon their applicability to ‘real-world’ experiences (Bartz, Anderson-Robinson, and Hillman 1994). - Conceptions of assessment by format typically differentiate assessment practices based on the cognitive level of the standard/curriculum expectations being assessed. - Higher order thinking skills are generally associated with performance and product assessments that enable students to construct responses and assert reasoning, while lower order thinking skills are associated with objectively scored assessments (McMillan 2007). - ## Assessment as purpose - Third, one of the most common conceptions of assessment is based on purpose and use, with the distinction between formative and summative assessment purposes widely used across assessment literature (Pellegrino, Chudowsky, and Glaser 2001). - Formative assessment refers to data collected throughout the instructional process and used to provide feedback to inform teaching and learning processes. - Summative assessment refers to data collected at the end of a unit or term in order to justify judgements on student achievement and report on student learning (i.e. contribute to student grades). - Another distinction on assessment purpose and use is based on the seminal work of Black and Wiliam (1998), which described purposes of assessment of learning and assessment for learning. - Subsumed within AFL, Earl (2003) articulated a third purpose of assessment, assessment as learning, which focuses on using assessments to support the development of students’ metacognition and selfregulation abilities. - ## Assessment as process - Finally, closely linked to assessment as, for, and of learning is assessment conceived as a process of interpretation and integration. Pellegrino, Chudowsky, and Glaser (2001, 36) state: ‘assessment is a process of reasoning from evidence’. This conception of assessment includes an awareness of the limitations of various assessment methods (including considerations of validity, reliability and bias) in relation to the specific context of teaching and learning. Hence, teachers’ ability to make meaning from assessment data becomes increasingly important to this conception as assessment is viewed as an essential component of teaching and integrally connected with curriculum and learning (Airasian and Russell 2008; Pellegrino, Chudowsky, and Glaser 2001). - # Data analyses - Results were analysed in relation to: (a) participants’ changing conceptions of assessment, (b) participants’ changing confidence in assessment and (c) the interaction between participants’ changing conceptions and confidence in assessment. - data were coded based on participants’ primary and, when available, secondary conceptions of assessment as testing, purpose, format and process. Dual coding of data by primary and secondary conceptions enabled a co-occurrence frequency analysis that provided a more complex analysis of participants’ understandings of assessment and which was helpful in ‘understanding how thematic domains, concepts, or ideas were distributed within a data set’ (Namey et al. 2008, 145). - Quantitatively, #[[descriptive statistics]] were calculated related to the gender, age, programme of study and year of study distribution across the sample. Then, changes in participants’ confidence levels were analysed through an #[[exploratory factor analysis]] using a promax (oblique) rotation. Although close-ended items had a theoretical foundation (i.e. Popham 2008), a #[[principal component factor analysis]] was conducted for each separate section of the pre-questionnaire to identify confidence factors across the Likert scale items based on responses from the 48 participants. #[[Factor loadings]] were determined by the greatest value for each item, with loadings of at least 0.40. No items maintained lower loading values. #[[Cronbach alpha]] values were also calculated for #reliability indices of each factor. We recognise that given the sample size of 48 participants, the generalisability and internal stability of factors are somewhat problematic. Accordingly, we caution readers to interpret factors with the caveat that the results from this study are preliminary and require further validation based on a large sample of participants. - In order to determine changes over time and the interaction between participants’ assessment conceptions and their confidence, we conducted a #[[repeated measures analysis]] across both questionnaire administrations (Stevens 2001). In preparation for the repeated measures analysis, assessment conception codes were converted to numerical values (indicating the ordinal progression from the most simplistic conception to the most complex) and placed on a proportional scale to confidence factors. A repeated measures analysis (two within group) was conducted for confidence and conception factors across the two data time points. The repeated measures analysis also allowed for explorations of any interaction effects between the teacher candidates’ conceptions and the confidence factors. A follow-up #[[dependent t-test]] addressed the research questions on how the candidates’ conceptions of assessment and confidence toward assessment practices differed between the beginning and end of the measurement course (Stevens 2001). #Significance was based on a #[[Bonferroni adjusted p-value]] of 0.007. As a caveat to this data analysis strategy, we recognise that the sample size limits the reliability of the measure, the internal stability of factor loadings and the resulting effect sizes, which mitigates the generalisability of results from factor analyses and significance analyses (i.e. t-test and ANOVA). We urge that results not be interpreted as generalised findings but rather as a basis for further validation research on these suggested factors and observed trends. - # Results - ## Conceptions of assessment - Data supported that participants’ conceptions of assessment developed from an initial dominant view of assessment as testing to a more complex view of assessment as format, purpose and process. - In analysing primary and secondary conceptions of assessment, data pointed to participants’ not only changing their conceptions from assessment as testing to assessment as purpose and as process, but also understanding assessment in more complex and interconnected ways. - As evidenced through this study, explicit assessment education can enhance teacher candidates’ confidence in assessment praxis and approaches, while deepening their conceptual and theoretical understandings of educational assessment. As a result of this learning, teacher candidates may foster a realistic and optimistic orientation for engaging in more valid forms of assessment within their future schools and classrooms. - - # Citations - - mentioning cite from section 'review of related literature' of publication: - Yidana, M. B., P., A. P. (2018). Economics Teachers’ Conceptions of Classroom Assessment. Int J Innov Educ Res, 10(6), 153-174. https://doi.org/10.31686/ijier.vol6.iss10.1176 --- “…Similar studies (Brown & Hirschfeld, 2008;Levy-Vered & Alhija, 2015) have found that attending an assessment course or having more training in assessment did not improve teachers' conceptions of assessment. However, there are studies that suggest that teachers' conception of assessment developed after receiving periodic training in assessment (DeLuca, Chavez & Cao, 2013;Smith, Hill, Cowie & Gilmore, 2014). Interestingly, the findings of the study also showed that after teachers had attended the course on assessment, their irrelevance conception of assessment improved significantly.…”
process
reference deluca c chavez t cao c establishing a foundation for valid teacher judgement on student learning the role of pre service assessment education assessment in education principles policy practice – summary deluca et al surveyed pre service teachers before and after an undergraduate course in educational measurement they used a framework of four primary conceptions of assessment as testing format purpose and process which are consecutively ordered by complexity they found that participants conceptions of assessment shifted from more simple views of assessment as testing to more complex and integrated views of assessment as purpose and process and to a lesser extent assessment as format interestingly they found that assessment as format was prominent as a secondary conception abstract the purpose of this research was to explore changes in teacher candidates’ confidence in assessment as a foundation for promoting valid teacher judgements on student learning based on a sample of teacher candidates enrolled in a semester long measurement course at one teacher education programme in florida this survey based study provided evidence that teacher candidates can develop confidence in assessment whilst deepening their conceptions of assessment the majority of teacher candidates expanded their conception of assessment from one based solely on testing to one that recognised multiple forms of assessment that serve varying purposes simultaneously teacher candidates expressed greater confidence in practical assessment approaches and factors related to assessment praxis based on study findings we argue that pre service teacher education has a critical role to play in promoting assessment literacy in beginning teachers and providing a foundation for teachers’ continued learning about assessment throughout their careers the paper concludes with implications for future research in the area of assessment education the purpose of this study was to examine pre service teacher candidates’ confidence levels in relation to their changing conceptions of assessment during the term of a senior level assessment course the specific research questions guiding this study were the following how do teacher candidates’ conceptions of assessment differ at the beginning and end of a pre service course on student assessment how does teacher candidates’ confidence toward assessment practices change in relation to their shifting conceptions of assessment related quotes developing teacher confidence in assessment pre service assessment education has been identified as a fundamental structure for preparing teachers to assess student learning chen research conducted on the impact of explicit assessment courses has generally found that these courses have the potential to complement field based teaching experience i e practical and develop a foundation for assessment literacy in teacher candidates graham conceptions of assessment brown lake and matters assert that teachers maintain multiple conceptions of classroom assessment based on assessment format purpose and use derived from formal assessment education as well as from the practicalities and policies that govern teachers’ daily work teachers’ shape their in class assessment practices brown cizek et al we present four dominant conceptions of assessment as an analytic framework for our study on the linkage between teacher candidates’ confidence and their changing understandings about assessment specifically the four conceptions of assessment that we describe are a b c and d we assert that while none of these conceptions represent simple measurement processes each conception increases in complexity by connecting together assessment concepts and extending toward other constructs in teaching and learning asssessment as testing our first conception of assessment assimilates assessment with traditional paperand pencil testing within the usa tests are often used as summative measures of student learning and can be either teacher constructed or external and standardised moreover due to efficiency and a lack of knowledge about other forms of assessment teachers continue to construct and use tests as the dominant means to measure student learning in their classrooms although classical and modern test theories are not simple when teachers are not aware of the psychometric underpinnings of testing or educated on issues related to test reliability fairness and validity they may perceive this form of assessment as easy and straightforward and use it with confidence deluca and klinger assessment as format our second conception expands on this association to include alternative forms of assessment and to delineate a conception based on assessment format a typical classification of assessments is that of performance product and objectively scored assessments airasian and russell performance and product assessments can be further delineated as contrived or authentic assessments depending upon their applicability to ‘real world’ experiences bartz anderson robinson and hillman conceptions of assessment by format typically differentiate assessment practices based on the cognitive level of the standard curriculum expectations being assessed higher order thinking skills are generally associated with performance and product assessments that enable students to construct responses and assert reasoning while lower order thinking skills are associated with objectively scored assessments mcmillan assessment as purpose third one of the most common conceptions of assessment is based on purpose and use with the distinction between formative and summative assessment purposes widely used across assessment literature pellegrino chudowsky and glaser formative assessment refers to data collected throughout the instructional process and used to provide feedback to inform teaching and learning processes summative assessment refers to data collected at the end of a unit or term in order to justify judgements on student achievement and report on student learning i e contribute to student grades another distinction on assessment purpose and use is based on the seminal work of black and wiliam which described purposes of assessment of learning and assessment for learning subsumed within afl earl articulated a third purpose of assessment assessment as learning which focuses on using assessments to support the development of students’ metacognition and selfregulation abilities assessment as process finally closely linked to assessment as for and of learning is assessment conceived as a process of interpretation and integration pellegrino chudowsky and glaser state ‘assessment is a process of reasoning from evidence’ this conception of assessment includes an awareness of the limitations of various assessment methods including considerations of validity reliability and bias in relation to the specific context of teaching and learning hence teachers’ ability to make meaning from assessment data becomes increasingly important to this conception as assessment is viewed as an essential component of teaching and integrally connected with curriculum and learning airasian and russell pellegrino chudowsky and glaser data analyses results were analysed in relation to a participants’ changing conceptions of assessment b participants’ changing confidence in assessment and c the interaction between participants’ changing conceptions and confidence in assessment data were coded based on participants’ primary and when available secondary conceptions of assessment as testing purpose format and process dual coding of data by primary and secondary conceptions enabled a co occurrence frequency analysis that provided a more complex analysis of participants’ understandings of assessment and which was helpful in ‘understanding how thematic domains concepts or ideas were distributed within a data set’ namey et al quantitatively were calculated related to the gender age programme of study and year of study distribution across the sample then changes in participants’ confidence levels were analysed through an using a promax oblique rotation although close ended items had a theoretical foundation i e popham a was conducted for each separate section of the pre questionnaire to identify confidence factors across the likert scale items based on responses from the participants were determined by the greatest value for each item with loadings of at least no items maintained lower loading values values were also calculated for reliability indices of each factor we recognise that given the sample size of participants the generalisability and internal stability of factors are somewhat problematic accordingly we caution readers to interpret factors with the caveat that the results from this study are preliminary and require further validation based on a large sample of participants in order to determine changes over time and the interaction between participants’ assessment conceptions and their confidence we conducted a across both questionnaire administrations stevens in preparation for the repeated measures analysis assessment conception codes were converted to numerical values indicating the ordinal progression from the most simplistic conception to the most complex and placed on a proportional scale to confidence factors a repeated measures analysis two within group was conducted for confidence and conception factors across the two data time points the repeated measures analysis also allowed for explorations of any interaction effects between the teacher candidates’ conceptions and the confidence factors a follow up addressed the research questions on how the candidates’ conceptions of assessment and confidence toward assessment practices differed between the beginning and end of the measurement course stevens significance was based on a of as a caveat to this data analysis strategy we recognise that the sample size limits the reliability of the measure the internal stability of factor loadings and the resulting effect sizes which mitigates the generalisability of results from factor analyses and significance analyses i e t test and anova we urge that results not be interpreted as generalised findings but rather as a basis for further validation research on these suggested factors and observed trends results conceptions of assessment data supported that participants’ conceptions of assessment developed from an initial dominant view of assessment as testing to a more complex view of assessment as format purpose and process in analysing primary and secondary conceptions of assessment data pointed to participants’ not only changing their conceptions from assessment as testing to assessment as purpose and as process but also understanding assessment in more complex and interconnected ways as evidenced through this study explicit assessment education can enhance teacher candidates’ confidence in assessment praxis and approaches while deepening their conceptual and theoretical understandings of educational assessment as a result of this learning teacher candidates may foster a realistic and optimistic orientation for engaging in more valid forms of assessment within their future schools and classrooms citations mentioning cite from section review of related literature of publication yidana m b p a p economics teachers’ conceptions of classroom assessment int j innov educ res “…similar studies brown hirschfeld levy vered alhija have found that attending an assessment course or having more training in assessment did not improve teachers conceptions of assessment however there are studies that suggest that teachers conception of assessment developed after receiving periodic training in assessment deluca chavez cao smith hill cowie gilmore interestingly the findings of the study also showed that after teachers had attended the course on assessment their irrelevance conception of assessment improved significantly …”
1
20,736
27,436,170,105
IssuesEvent
2023-03-02 07:37:16
TUM-Dev/NavigaTUM
https://api.github.com/repos/TUM-Dev/NavigaTUM
opened
[General] Raumnummer nicht existent
webform delete-after-processing general
0501.02.119 Sehr geehrte Damen und Herren, bei der Eingabe der Telefonnebenstelle in TUM-Online kann die Raumnummer nicht bearbeitet oder korrigiert werden. Fehlercode: Raumnummer existiert nicht. Die Raumnummer 0501.02.112 müsste zu den Nebenstellen: 22110 + 22112 + 22114 + 22790 geändert werden. Die richtige Raumnummer ist 0501.02.119. Diese wurde zu den Nebenstellen 25507 + 25508 eingetragen, ist dennoch zu o.g. Nummern nicht existent. Bitte kontaktieren Sie uns unter 089 289 25580 oder unter fernmeldeservice.za4@tum.de telefongebuehrenabrechnung.za4@tum.de Mit herzlichen Grüßen Mariam Djafari
1.0
[General] Raumnummer nicht existent - 0501.02.119 Sehr geehrte Damen und Herren, bei der Eingabe der Telefonnebenstelle in TUM-Online kann die Raumnummer nicht bearbeitet oder korrigiert werden. Fehlercode: Raumnummer existiert nicht. Die Raumnummer 0501.02.112 müsste zu den Nebenstellen: 22110 + 22112 + 22114 + 22790 geändert werden. Die richtige Raumnummer ist 0501.02.119. Diese wurde zu den Nebenstellen 25507 + 25508 eingetragen, ist dennoch zu o.g. Nummern nicht existent. Bitte kontaktieren Sie uns unter 089 289 25580 oder unter fernmeldeservice.za4@tum.de telefongebuehrenabrechnung.za4@tum.de Mit herzlichen Grüßen Mariam Djafari
process
raumnummer nicht existent sehr geehrte damen und herren bei der eingabe der telefonnebenstelle in tum online kann die raumnummer nicht bearbeitet oder korrigiert werden fehlercode raumnummer existiert nicht die raumnummer müsste zu den nebenstellen geändert werden die richtige raumnummer ist diese wurde zu den nebenstellen eingetragen ist dennoch zu o g nummern nicht existent bitte kontaktieren sie uns unter oder unter fernmeldeservice tum de telefongebuehrenabrechnung tum de mit herzlichen grüßen mariam djafari
1
18,832
10,295,675,833
IssuesEvent
2019-08-27 21:53:20
alansanie/wingtips
https://api.github.com/repos/alansanie/wingtips
opened
WS-2009-0001 (Low) detected in commons-codec-1.9.jar, commons-codec-1.11.jar
security vulnerability
## WS-2009-0001 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>commons-codec-1.9.jar</b>, <b>commons-codec-1.11.jar</b></p></summary> <p> <details><summary><b>commons-codec-1.9.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Library home page: <a href="http://commons.apache.org/proper/commons-codec/">http://commons.apache.org/proper/commons-codec/</a></p> <p>Path to dependency file: /wingtips/wingtips-zipkin-spring-boot/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.9/9ce04e34240f674bc72680f8b843b1457383161a/commons-codec-1.9.jar,/root/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.9/9ce04e34240f674bc72680f8b843b1457383161a/commons-codec-1.9.jar</p> <p> Dependency Hierarchy: - httpclient-4.4.1.jar (Root Library) - :x: **commons-codec-1.9.jar** (Vulnerable Library) </details> <details><summary><b>commons-codec-1.11.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Library home page: <a href="http://commons.apache.org/proper/commons-codec/">http://commons.apache.org/proper/commons-codec/</a></p> <p>Path to dependency file: /wingtips/wingtips-lightstep/build.gradle</p> <p>Path to vulnerable library: radle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.11/3acb4705652e16236558f0f4f2192cc33c3bd189/commons-codec-1.11.jar,radle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.11/3acb4705652e16236558f0f4f2192cc33c3bd189/commons-codec-1.11.jar</p> <p> Dependency Hierarchy: - :x: **commons-codec-1.11.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/alansanie/wingtips/commit/183e8ad413608960359e84065f37e537d5142cdd">183e8ad413608960359e84065f37e537d5142cdd</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Not all "business" method implementations of public API in Apache Commons Codec 1.x are thread safe, which might disclose the wrong data or allow an attacker to change non-private fields. Updated 2018-10-07 - an additional review by WhiteSource research team could not indicate on a clear security vulnerability <p>Publish Date: 2007-10-07 <p>URL: <a href=https://issues.apache.org/jira/browse/CODEC-55>WS-2009-0001</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>0.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p>
True
WS-2009-0001 (Low) detected in commons-codec-1.9.jar, commons-codec-1.11.jar - ## WS-2009-0001 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>commons-codec-1.9.jar</b>, <b>commons-codec-1.11.jar</b></p></summary> <p> <details><summary><b>commons-codec-1.9.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Library home page: <a href="http://commons.apache.org/proper/commons-codec/">http://commons.apache.org/proper/commons-codec/</a></p> <p>Path to dependency file: /wingtips/wingtips-zipkin-spring-boot/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.9/9ce04e34240f674bc72680f8b843b1457383161a/commons-codec-1.9.jar,/root/.gradle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.9/9ce04e34240f674bc72680f8b843b1457383161a/commons-codec-1.9.jar</p> <p> Dependency Hierarchy: - httpclient-4.4.1.jar (Root Library) - :x: **commons-codec-1.9.jar** (Vulnerable Library) </details> <details><summary><b>commons-codec-1.11.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Library home page: <a href="http://commons.apache.org/proper/commons-codec/">http://commons.apache.org/proper/commons-codec/</a></p> <p>Path to dependency file: /wingtips/wingtips-lightstep/build.gradle</p> <p>Path to vulnerable library: radle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.11/3acb4705652e16236558f0f4f2192cc33c3bd189/commons-codec-1.11.jar,radle/caches/modules-2/files-2.1/commons-codec/commons-codec/1.11/3acb4705652e16236558f0f4f2192cc33c3bd189/commons-codec-1.11.jar</p> <p> Dependency Hierarchy: - :x: **commons-codec-1.11.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/alansanie/wingtips/commit/183e8ad413608960359e84065f37e537d5142cdd">183e8ad413608960359e84065f37e537d5142cdd</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Not all "business" method implementations of public API in Apache Commons Codec 1.x are thread safe, which might disclose the wrong data or allow an attacker to change non-private fields. Updated 2018-10-07 - an additional review by WhiteSource research team could not indicate on a clear security vulnerability <p>Publish Date: 2007-10-07 <p>URL: <a href=https://issues.apache.org/jira/browse/CODEC-55>WS-2009-0001</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>0.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p>
non_process
ws low detected in commons codec jar commons codec jar ws low severity vulnerability vulnerable libraries commons codec jar commons codec jar commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities library home page a href path to dependency file wingtips wingtips zipkin spring boot build gradle path to vulnerable library root gradle caches modules files commons codec commons codec commons codec jar root gradle caches modules files commons codec commons codec commons codec jar dependency hierarchy httpclient jar root library x commons codec jar vulnerable library commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities library home page a href path to dependency file wingtips wingtips lightstep build gradle path to vulnerable library radle caches modules files commons codec commons codec commons codec jar radle caches modules files commons codec commons codec commons codec jar dependency hierarchy x commons codec jar vulnerable library found in head commit a href vulnerability details not all business method implementations of public api in apache commons codec x are thread safe which might disclose the wrong data or allow an attacker to change non private fields updated an additional review by whitesource research team could not indicate on a clear security vulnerability publish date url a href cvss score details base score metrics not available
0
8,797
11,908,244,374
IssuesEvent
2020-03-31 00:23:18
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
OTB provider setup
Bug Feedback Processing
Author Name: **Paolo Cavallini** (@pcav) Original Redmine Issue: [21372](https://issues.qgis.org/issues/21372) Affected QGIS version: 3.7(master) Redmine category:processing/otb Assignee: Rashad Kanavath --- In Debian, the paths to add to the config are `/usr/lib/x86_64-linux-gnu/otb/applications/` and `/usr`. This is not documented, and it is difficult to find for the user. Better add it as a default (similarly for other OSs). More generally: why is this necessary? Other backends do not need a config, and this is way easier for the user.
1.0
OTB provider setup - Author Name: **Paolo Cavallini** (@pcav) Original Redmine Issue: [21372](https://issues.qgis.org/issues/21372) Affected QGIS version: 3.7(master) Redmine category:processing/otb Assignee: Rashad Kanavath --- In Debian, the paths to add to the config are `/usr/lib/x86_64-linux-gnu/otb/applications/` and `/usr`. This is not documented, and it is difficult to find for the user. Better add it as a default (similarly for other OSs). More generally: why is this necessary? Other backends do not need a config, and this is way easier for the user.
process
otb provider setup author name paolo cavallini pcav original redmine issue affected qgis version master redmine category processing otb assignee rashad kanavath in debian the paths to add to the config are usr lib linux gnu otb applications and usr this is not documented and it is difficult to find for the user better add it as a default similarly for other oss more generally why is this necessary other backends do not need a config and this is way easier for the user
1
72,973
7,319,562,328
IssuesEvent
2018-03-02 01:31:36
couchbase/couchbase-lite-core
https://api.github.com/repos/couchbase/couchbase-lite-core
closed
Race conditions pushing obsoleted revisions
P1: high :scream_cat: functional-test-failure :-1: in progress known-issue ready
CBL - 2.0- DB 22 -> ios build 574, Android 299 client -> ios, Android sg -> 2.0.0- 773 ### Steps to reproduce : Create docs in CBL DB1, CBL DB2, CBL DB3. Configure replication from CBL db1 to CBL db2 with push_pull and continuous. Configure replication from CBL db1 to CBL db3 with push_pull and continuous. Start the replication for 2 and 3. update docs on CBL DB1, CBL DB2, CBL DB3. WAit for replication to complete Have multithreading and update docs in all 3 cbl dbs. Wait for replication to complete for step 3 and step 4 replications. Configure replication from CBL db3 to sync-gateway and start replication. Wait for replication to complete. Verify all docs from CBL db3 replicated to sync-gateway ### Result : 1 doc is missing in sync-gateway , I see that conflict resolution failed in CBL logs as a 'Sync warning' Note : out of 3 times, it passed 2 times and failed 1 time. ###Test code : https://github.com/couchbaselabs/mobile-testkit/blob/feature/cbl20-ver1/testsuites/CBLTester/CBL_Functional_tests/test_no_conflicts_cbl.py#L696 ### Logs : 2018-02-01 20:24:56.673208-0800 CBLTestServer-iOS[55966:1608762] CouchbaseLite DB Info: DataFile: abort transaction 2018-02-01 20:24:56.673387-0800 CBLTestServer-iOS[55966:1666581] CouchbaseLite Sync Info: {DBWorker#4268} Looking up 11 revisions in the db (seq '2710'..'2720') 2018-02-01 20:24:56.673706-0800 CBLTestServer-iOS[55966:1666313] CouchbaseLite BLIP Verbose: {BLIPIO#89} Received frame: REQ #1420 ---C, length 548 2018-02-01 20:24:56.673908-0800 CBLTestServer-iOS[55966:1608762] CouchbaseLite Sync WARNING: Conflict resolution of 'no-conflicts2_4' failed: (null) Is this expected? This is regression bug , but the failure is not consitent, it is transient failure
1.0
Race conditions pushing obsoleted revisions - CBL - 2.0- DB 22 -> ios build 574, Android 299 client -> ios, Android sg -> 2.0.0- 773 ### Steps to reproduce : Create docs in CBL DB1, CBL DB2, CBL DB3. Configure replication from CBL db1 to CBL db2 with push_pull and continuous. Configure replication from CBL db1 to CBL db3 with push_pull and continuous. Start the replication for 2 and 3. update docs on CBL DB1, CBL DB2, CBL DB3. WAit for replication to complete Have multithreading and update docs in all 3 cbl dbs. Wait for replication to complete for step 3 and step 4 replications. Configure replication from CBL db3 to sync-gateway and start replication. Wait for replication to complete. Verify all docs from CBL db3 replicated to sync-gateway ### Result : 1 doc is missing in sync-gateway , I see that conflict resolution failed in CBL logs as a 'Sync warning' Note : out of 3 times, it passed 2 times and failed 1 time. ###Test code : https://github.com/couchbaselabs/mobile-testkit/blob/feature/cbl20-ver1/testsuites/CBLTester/CBL_Functional_tests/test_no_conflicts_cbl.py#L696 ### Logs : 2018-02-01 20:24:56.673208-0800 CBLTestServer-iOS[55966:1608762] CouchbaseLite DB Info: DataFile: abort transaction 2018-02-01 20:24:56.673387-0800 CBLTestServer-iOS[55966:1666581] CouchbaseLite Sync Info: {DBWorker#4268} Looking up 11 revisions in the db (seq '2710'..'2720') 2018-02-01 20:24:56.673706-0800 CBLTestServer-iOS[55966:1666313] CouchbaseLite BLIP Verbose: {BLIPIO#89} Received frame: REQ #1420 ---C, length 548 2018-02-01 20:24:56.673908-0800 CBLTestServer-iOS[55966:1608762] CouchbaseLite Sync WARNING: Conflict resolution of 'no-conflicts2_4' failed: (null) Is this expected? This is regression bug , but the failure is not consitent, it is transient failure
non_process
race conditions pushing obsoleted revisions cbl db ios build android client ios android sg steps to reproduce create docs in cbl cbl cbl configure replication from cbl to cbl with push pull and continuous configure replication from cbl to cbl with push pull and continuous start the replication for and update docs on cbl cbl cbl wait for replication to complete have multithreading and update docs in all cbl dbs wait for replication to complete for step and step replications configure replication from cbl to sync gateway and start replication wait for replication to complete verify all docs from cbl replicated to sync gateway result doc is missing in sync gateway i see that conflict resolution failed in cbl logs as a sync warning note out of times it passed times and failed time test code logs cbltestserver ios couchbaselite db info datafile abort transaction cbltestserver ios couchbaselite sync info dbworker looking up revisions in the db seq cbltestserver ios couchbaselite blip verbose blipio received frame req c length cbltestserver ios couchbaselite sync warning conflict resolution of no failed null is this expected this is regression bug but the failure is not consitent it is transient failure
0
192,655
15,355,135,947
IssuesEvent
2021-03-01 10:41:45
zeek/spicy
https://api.github.com/repos/zeek/spicy
closed
Add language support to explicitly trigger a parse error
Documentation
Wasn't sure if it's intended for `&requires` to be evaluated before field hooks: ``` module Test; public type MyUnit = unit { s: bytes &size=4 &requires=(self.flag); on s { self.flag = True; } var flag: bool; }; ``` ``` $ printf 'test' | spicy-dump test.spicy [fatal error] terminating with uncaught exception of type spicy::rt::ParseError: parse error: &requires failed ($$ == test) (test.spicy:4:22) ``` Initially thought that field hooks might get evaluated first in real example here and found otherwise: https://github.com/zeek/spicy/compare/topic/jsiwek/pe#diff-36ceb1dc5c1f8f65e1fb3553b2d552ef12f58b3651f1df59aafa9e6c3fafc63eR64-R67 * Generally seems to be lack of documentation on evaluation order and adding that could help. * For what I was trying to do, would be convenient if the order was different, but realize it's kind of arbitrary/implicit, so might generally be better to have an explicit way to induce parsing errors from within arbitrary hook-code itself. E.g. either with a `throw` or special `return` value. Kind of like mentioned in previous error-handling discussions https://github.com/zeek/spicy/issues/89#issuecomment-648637582 and https://github.com/zeek/spicy/discussions/616#discussioncomment-185009
1.0
Add language support to explicitly trigger a parse error - Wasn't sure if it's intended for `&requires` to be evaluated before field hooks: ``` module Test; public type MyUnit = unit { s: bytes &size=4 &requires=(self.flag); on s { self.flag = True; } var flag: bool; }; ``` ``` $ printf 'test' | spicy-dump test.spicy [fatal error] terminating with uncaught exception of type spicy::rt::ParseError: parse error: &requires failed ($$ == test) (test.spicy:4:22) ``` Initially thought that field hooks might get evaluated first in real example here and found otherwise: https://github.com/zeek/spicy/compare/topic/jsiwek/pe#diff-36ceb1dc5c1f8f65e1fb3553b2d552ef12f58b3651f1df59aafa9e6c3fafc63eR64-R67 * Generally seems to be lack of documentation on evaluation order and adding that could help. * For what I was trying to do, would be convenient if the order was different, but realize it's kind of arbitrary/implicit, so might generally be better to have an explicit way to induce parsing errors from within arbitrary hook-code itself. E.g. either with a `throw` or special `return` value. Kind of like mentioned in previous error-handling discussions https://github.com/zeek/spicy/issues/89#issuecomment-648637582 and https://github.com/zeek/spicy/discussions/616#discussioncomment-185009
non_process
add language support to explicitly trigger a parse error wasn t sure if it s intended for requires to be evaluated before field hooks module test public type myunit unit s bytes size requires self flag on s self flag true var flag bool printf test spicy dump test spicy terminating with uncaught exception of type spicy rt parseerror parse error requires failed test test spicy initially thought that field hooks might get evaluated first in real example here and found otherwise generally seems to be lack of documentation on evaluation order and adding that could help for what i was trying to do would be convenient if the order was different but realize it s kind of arbitrary implicit so might generally be better to have an explicit way to induce parsing errors from within arbitrary hook code itself e g either with a throw or special return value kind of like mentioned in previous error handling discussions and
0
533,338
15,589,014,586
IssuesEvent
2021-03-18 07:24:20
AY2021S2-CS2103T-T12-2/tp
https://api.github.com/repos/AY2021S2-CS2103T-T12-2/tp
closed
Food intake list implementation inaccurate
priority.High severity.Medium type.Bug
Extra date property for unmatched food intake items Below is a sample json from save, contains a single save file for multiple entries of saved dates ```json { "date" : null, "foodIntakes" : [ { "name" : "test", "fats" : 10.0, "carbos" : 10.0, "proteins" : 10.0, "date" : "0001-01-01" }, { "name" : "test", "fats" : 10.0, "carbos" : 10.0, "proteins" : 10.0, "date" : "2050-06-20" } ] } ```
1.0
Food intake list implementation inaccurate - Extra date property for unmatched food intake items Below is a sample json from save, contains a single save file for multiple entries of saved dates ```json { "date" : null, "foodIntakes" : [ { "name" : "test", "fats" : 10.0, "carbos" : 10.0, "proteins" : 10.0, "date" : "0001-01-01" }, { "name" : "test", "fats" : 10.0, "carbos" : 10.0, "proteins" : 10.0, "date" : "2050-06-20" } ] } ```
non_process
food intake list implementation inaccurate extra date property for unmatched food intake items below is a sample json from save contains a single save file for multiple entries of saved dates json date null foodintakes name test fats carbos proteins date name test fats carbos proteins date
0
372,724
26,017,266,589
IssuesEvent
2022-12-21 09:33:52
sdv-dev/DeepEcho
https://api.github.com/repos/sdv-dev/DeepEcho
closed
DeepEcho Package Maintenance Updates
documentation maintenance internal
_This is a placeholder issue for overall package maintenance updates for the next minor DeepEcho release: `0.4.0`_
1.0
DeepEcho Package Maintenance Updates - _This is a placeholder issue for overall package maintenance updates for the next minor DeepEcho release: `0.4.0`_
non_process
deepecho package maintenance updates this is a placeholder issue for overall package maintenance updates for the next minor deepecho release
0
20,082
26,577,651,763
IssuesEvent
2023-01-22 02:00:08
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Fri, 20 Jan 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Regularizing disparity estimation via multi task learning with structured light reconstruction - **Authors:** Alistair Weld, Joao Cartucho, Chi Xu, Joseph Davids, Stamatia Giannarou - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2301.08140 - **Pdf link:** https://arxiv.org/pdf/2301.08140 - **Abstract** 3D reconstruction is a useful tool for surgical planning and guidance. However, the lack of available medical data stunts research and development in this field, as supervised deep learning methods for accurate disparity estimation rely heavily on large datasets containing ground truth information. Alternative approaches to supervision have been explored, such as self-supervision, which can reduce or remove entirely the need for ground truth. However, no proposed alternatives have demonstrated performance capabilities close to what would be expected from a supervised setup. This work aims to alleviate this issue. In this paper, we investigate the learning of structured light projections to enhance the development of direct disparity estimation networks. We show for the first time that it is possible to accurately learn the projection of structured light on a scene, implicitly learning disparity. Secondly, we \textcolor{black}{explore the use of a multi task learning (MTL) framework for the joint training of structured light and disparity. We present results which show that MTL with structured light improves disparity training; without increasing the number of model parameters. Our MTL setup outperformed the single task learning (STL) network in every validation test. Notably, in the medical generalisation test, the STL error was 1.4 times worse than that of the best MTL performance. The benefit of using MTL is emphasised when the training data is limited.} A dataset containing stereoscopic images, disparity maps and structured light projections on medical phantoms and ex vivo tissue was created for evaluation together with virtual scenes. This dataset will be made publicly available in the future. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW There is no result ## Keyword: raw image There is no result
2.0
New submissions for Fri, 20 Jan 23 - ## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Regularizing disparity estimation via multi task learning with structured light reconstruction - **Authors:** Alistair Weld, Joao Cartucho, Chi Xu, Joseph Davids, Stamatia Giannarou - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2301.08140 - **Pdf link:** https://arxiv.org/pdf/2301.08140 - **Abstract** 3D reconstruction is a useful tool for surgical planning and guidance. However, the lack of available medical data stunts research and development in this field, as supervised deep learning methods for accurate disparity estimation rely heavily on large datasets containing ground truth information. Alternative approaches to supervision have been explored, such as self-supervision, which can reduce or remove entirely the need for ground truth. However, no proposed alternatives have demonstrated performance capabilities close to what would be expected from a supervised setup. This work aims to alleviate this issue. In this paper, we investigate the learning of structured light projections to enhance the development of direct disparity estimation networks. We show for the first time that it is possible to accurately learn the projection of structured light on a scene, implicitly learning disparity. Secondly, we \textcolor{black}{explore the use of a multi task learning (MTL) framework for the joint training of structured light and disparity. We present results which show that MTL with structured light improves disparity training; without increasing the number of model parameters. Our MTL setup outperformed the single task learning (STL) network in every validation test. Notably, in the medical generalisation test, the STL error was 1.4 times worse than that of the best MTL performance. The benefit of using MTL is emphasised when the training data is limited.} A dataset containing stereoscopic images, disparity maps and structured light projections on medical phantoms and ex vivo tissue was created for evaluation together with virtual scenes. This dataset will be made publicly available in the future. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW There is no result ## Keyword: raw image There is no result
process
new submissions for fri jan keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp regularizing disparity estimation via multi task learning with structured light reconstruction authors alistair weld joao cartucho chi xu joseph davids stamatia giannarou subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract reconstruction is a useful tool for surgical planning and guidance however the lack of available medical data stunts research and development in this field as supervised deep learning methods for accurate disparity estimation rely heavily on large datasets containing ground truth information alternative approaches to supervision have been explored such as self supervision which can reduce or remove entirely the need for ground truth however no proposed alternatives have demonstrated performance capabilities close to what would be expected from a supervised setup this work aims to alleviate this issue in this paper we investigate the learning of structured light projections to enhance the development of direct disparity estimation networks we show for the first time that it is possible to accurately learn the projection of structured light on a scene implicitly learning disparity secondly we textcolor black explore the use of a multi task learning mtl framework for the joint training of structured light and disparity we present results which show that mtl with structured light improves disparity training without increasing the number of model parameters our mtl setup outperformed the single task learning stl network in every validation test notably in the medical generalisation test the stl error was times worse than that of the best mtl performance the benefit of using mtl is emphasised when the training data is limited a dataset containing stereoscopic images disparity maps and structured light projections on medical phantoms and ex vivo tissue was created for evaluation together with virtual scenes this dataset will be made publicly available in the future keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw there is no result keyword raw image there is no result
1