Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
8,351
| 11,501,708,762
|
IssuesEvent
|
2020-02-12 17:40:32
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
GO:0140418. effector-mediated modulation of host process by symbiont
|
multi-species process
|
missing parent child relationship
GO:0140418 effector-mediated modulation of host process by symbiont
GO:0140415 effector-mediated modulation of host defenses by symbiont
@pgaudet I am sure you did this. I saw it? Was there some reversion?
|
1.0
|
GO:0140418. effector-mediated modulation of host process by symbiont - missing parent child relationship
GO:0140418 effector-mediated modulation of host process by symbiont
GO:0140415 effector-mediated modulation of host defenses by symbiont
@pgaudet I am sure you did this. I saw it? Was there some reversion?
|
process
|
go effector mediated modulation of host process by symbiont missing parent child relationship go effector mediated modulation of host process by symbiont go effector mediated modulation of host defenses by symbiont pgaudet i am sure you did this i saw it was there some reversion
| 1
|
20,536
| 30,453,353,932
|
IssuesEvent
|
2023-07-16 15:21:11
|
ldtteam/minecolonies
|
https://api.github.com/repos/ldtteam/minecolonies
|
closed
|
[BUG] Rack outline does not appear through walls when searching inventory of a building
|
Bug Compatibility: Forge Bug: Confirmed 1.19
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Are you using the latest MineColonies Verison?
- [X] I am running the latest **alpha** version of MineColonies for my Minecraft version.
### Did you check on the Wiki? or ask on Discord?
- [X] I checked the MineColonies Wiki and made sure my issue is not covered there. Or I was sent from discord to open an issue here.
### Minecraft Version
1.19
### MineColonies Version
1.19.1-1.0.921-ALPHA
### Structurize Version
1.19.1-1.1.423-ALPHA
### Related Mods and their Versions
Forge: 43.1.1
BlockUI: 1.19-0.0.60-ALPHA
Domum: 1.19-1.0.59-ALPHA
Multi-Piston: 1.19-1.2.16-ALPHA
### Current Behavior
When hitting the ? next to the item in the inventory of the main building block (the chest) the outline shows up around the correct rack but does not show through walls. Neither red no green outlines do this
### Expected Behavior
According to Moobien on the Discord, it should show through walls by default
### Reproduction Steps
Start a colony
Place a building
Place an item in a rack(s)
Search for the item
Search for the rack(s)
### Logs
https://gist.github.com/DarthLeahcim/a44eaaf6810f0ad20a161c49aabf11b9
### Anything else?
First noticed on MineColonies v1.19.1-1.0.927-ALPHA with Forge v42.0.9 but could've been an issue before this.
The log may be a bit messy from me accidentally skipping a step or two when quickly creating a town in Creative
* Add a thumbs-up to the bug report if you are also affected. This helps the bug report become more visible to the team and doesn't clutter the comments.
* Add a comment if you have any insights or background information that isn't already part of the conversation.
|
True
|
[BUG] Rack outline does not appear through walls when searching inventory of a building - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Are you using the latest MineColonies Verison?
- [X] I am running the latest **alpha** version of MineColonies for my Minecraft version.
### Did you check on the Wiki? or ask on Discord?
- [X] I checked the MineColonies Wiki and made sure my issue is not covered there. Or I was sent from discord to open an issue here.
### Minecraft Version
1.19
### MineColonies Version
1.19.1-1.0.921-ALPHA
### Structurize Version
1.19.1-1.1.423-ALPHA
### Related Mods and their Versions
Forge: 43.1.1
BlockUI: 1.19-0.0.60-ALPHA
Domum: 1.19-1.0.59-ALPHA
Multi-Piston: 1.19-1.2.16-ALPHA
### Current Behavior
When hitting the ? next to the item in the inventory of the main building block (the chest) the outline shows up around the correct rack but does not show through walls. Neither red no green outlines do this
### Expected Behavior
According to Moobien on the Discord, it should show through walls by default
### Reproduction Steps
Start a colony
Place a building
Place an item in a rack(s)
Search for the item
Search for the rack(s)
### Logs
https://gist.github.com/DarthLeahcim/a44eaaf6810f0ad20a161c49aabf11b9
### Anything else?
First noticed on MineColonies v1.19.1-1.0.927-ALPHA with Forge v42.0.9 but could've been an issue before this.
The log may be a bit messy from me accidentally skipping a step or two when quickly creating a town in Creative
* Add a thumbs-up to the bug report if you are also affected. This helps the bug report become more visible to the team and doesn't clutter the comments.
* Add a comment if you have any insights or background information that isn't already part of the conversation.
|
non_process
|
rack outline does not appear through walls when searching inventory of a building is there an existing issue for this i have searched the existing issues are you using the latest minecolonies verison i am running the latest alpha version of minecolonies for my minecraft version did you check on the wiki or ask on discord i checked the minecolonies wiki and made sure my issue is not covered there or i was sent from discord to open an issue here minecraft version minecolonies version alpha structurize version alpha related mods and their versions forge blockui alpha domum alpha multi piston alpha current behavior when hitting the next to the item in the inventory of the main building block the chest the outline shows up around the correct rack but does not show through walls neither red no green outlines do this expected behavior according to moobien on the discord it should show through walls by default reproduction steps start a colony place a building place an item in a rack s search for the item search for the rack s logs anything else first noticed on minecolonies alpha with forge but could ve been an issue before this the log may be a bit messy from me accidentally skipping a step or two when quickly creating a town in creative add a thumbs up to the bug report if you are also affected this helps the bug report become more visible to the team and doesn t clutter the comments add a comment if you have any insights or background information that isn t already part of the conversation
| 0
|
18,057
| 24,067,640,574
|
IssuesEvent
|
2022-09-17 18:22:38
|
COS301-SE-2022/Pure-LoRa-Tracking
|
https://api.github.com/repos/COS301-SE-2022/Pure-LoRa-Tracking
|
closed
|
(processing): CRON check for data ready
|
(system) Database (system) Server (role) data engineer (system) AI (bus) processing
|
CRON service to check whether data is ready to be processed by AI
|
1.0
|
(processing): CRON check for data ready - CRON service to check whether data is ready to be processed by AI
|
process
|
processing cron check for data ready cron service to check whether data is ready to be processed by ai
| 1
|
14,344
| 17,371,154,813
|
IssuesEvent
|
2021-07-30 14:10:30
|
GoogleCloudPlatform/getting-started-dotnet
|
https://api.github.com/repos/GoogleCloudPlatform/getting-started-dotnet
|
opened
|
Review all renovate PRs after fixing the CI known errors.
|
priority: p3 type: process
|
I first need to fix the errors described in #142 before being able to know for certain that renovate's dependency updates won't be a problem.
|
1.0
|
Review all renovate PRs after fixing the CI known errors. - I first need to fix the errors described in #142 before being able to know for certain that renovate's dependency updates won't be a problem.
|
process
|
review all renovate prs after fixing the ci known errors i first need to fix the errors described in before being able to know for certain that renovate s dependency updates won t be a problem
| 1
|
395,472
| 27,070,541,042
|
IssuesEvent
|
2023-02-14 06:19:59
|
earloc/TypealizR
|
https://api.github.com/repos/earloc/TypealizR
|
closed
|
clarify the usage and intent of `code-first`
|
documentation
|
As kindly pointet out [here](https://github.com/earloc/TypealizR/discussions/78#discussioncomment-4961646), the current samples explaining the `code-first`-approach can be slightly optimized.
Thx @ivanhofer for your effort, very very appreciated!
|
1.0
|
clarify the usage and intent of `code-first` - As kindly pointet out [here](https://github.com/earloc/TypealizR/discussions/78#discussioncomment-4961646), the current samples explaining the `code-first`-approach can be slightly optimized.
Thx @ivanhofer for your effort, very very appreciated!
|
non_process
|
clarify the usage and intent of code first as kindly pointet out the current samples explaining the code first approach can be slightly optimized thx ivanhofer for your effort very very appreciated
| 0
|
280,996
| 21,315,319,521
|
IssuesEvent
|
2022-04-16 07:01:32
|
j4ck990/pe
|
https://api.github.com/repos/j4ck990/pe
|
opened
|
Inconsistent message(viewtask)
|
type.DocumentationBug severity.Low
|

The order of tasks from viewtask does not match the order shown by the student groups tab.
<!--session: 1650086616556-abe55de6-f984-418c-aad7-44e60f192af6-->
<!--Version: Web v3.4.2-->
|
1.0
|
Inconsistent message(viewtask) - 
The order of tasks from viewtask does not match the order shown by the student groups tab.
<!--session: 1650086616556-abe55de6-f984-418c-aad7-44e60f192af6-->
<!--Version: Web v3.4.2-->
|
non_process
|
inconsistent message viewtask the order of tasks from viewtask does not match the order shown by the student groups tab
| 0
|
8,881
| 11,981,357,968
|
IssuesEvent
|
2020-04-07 10:58:18
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
spindle pole body separation
|
Other term-related request PomBase cell cycle and DNA processes
|
GO:0000073
Name spindle pole body separation
is defined
A largely uncharacterized process involving the release of duplicated spindle pole bodies (SPBs) and their migration away from each other within the nuclear membrane. Duplicated SPBs are connected by a bridge structure that may be severed in order to release the SPBs from one another. Following liberation, SPBs diffuse through the nuclear membrane until they are across from each other. SPB separation must take place in order for a **bipolar mitotic spindle** to assemble
(i.e it is defined as the mitotic process)
Could this be renamed to
Name mitotic spindle pole body separation
and positioned under
GO:1905047
Name mitotic spindle pole body organization
|
1.0
|
spindle pole body separation - GO:0000073
Name spindle pole body separation
is defined
A largely uncharacterized process involving the release of duplicated spindle pole bodies (SPBs) and their migration away from each other within the nuclear membrane. Duplicated SPBs are connected by a bridge structure that may be severed in order to release the SPBs from one another. Following liberation, SPBs diffuse through the nuclear membrane until they are across from each other. SPB separation must take place in order for a **bipolar mitotic spindle** to assemble
(i.e it is defined as the mitotic process)
Could this be renamed to
Name mitotic spindle pole body separation
and positioned under
GO:1905047
Name mitotic spindle pole body organization
|
process
|
spindle pole body separation go name spindle pole body separation is defined a largely uncharacterized process involving the release of duplicated spindle pole bodies spbs and their migration away from each other within the nuclear membrane duplicated spbs are connected by a bridge structure that may be severed in order to release the spbs from one another following liberation spbs diffuse through the nuclear membrane until they are across from each other spb separation must take place in order for a bipolar mitotic spindle to assemble i e it is defined as the mitotic process could this be renamed to name mitotic spindle pole body separation and positioned under go name mitotic spindle pole body organization
| 1
|
16,684
| 21,786,516,707
|
IssuesEvent
|
2022-05-14 08:00:56
|
ymgw55/MLPapers
|
https://api.github.com/repos/ymgw55/MLPapers
|
closed
|
Effect of Post-processing on Contextualized Word Representations
|
BERT dynamic embedding post-process layer-wise
|
# Information
### URL
https://arxiv.org/abs/2104.07456
### Authors
[Hassan Sajjad](https://arxiv.org/search/cs?searchtype=author&query=Sajjad%2C+H), [Firoj Alam](https://arxiv.org/search/cs?searchtype=author&query=Alam%2C+F), [Fahim Dalvi](https://arxiv.org/search/cs?searchtype=author&query=Dalvi%2C+F), [Nadir Durrani](https://arxiv.org/search/cs?searchtype=author&query=Durrani%2C+N)
<!--- ### Published at --->
### Date of issue
15 Apr 2021
<!---
### Related Websites
### Package
### Code
--->
# Reading Motivation
BERTのPPについて.
# Summary
static embedding の PP は, 単語レベル・文レベルのタスクで性能が向上することが知られている. しかし, contextualized embedding の PP の研究はあまりされていない. この論文では, pre-train された言語モデルの各層から得られる contextualized embedding の PP の有用性について調べた. PP の種類として, z-score, min-max を用いた分散表現の各次元を正規化, all-the-but-top を用いた主要成分の除去, unit length normalization を採用. 様々なpre-trained model で実験した結果, PP によって, 分散表現中に存在する, (word similarity, analogy などの) lexical タスク・文分類のための重要な情報をとりだせることが分かった. この研究により, contextulized embedding を用いた研究はより興味深いものとなり, contextulized embbeding に PPを使うときに z-score を用いた正規化が重要になると分かる.
# Background
* pre-train され言語モデルの layer ごとに分散が異なる.

<!---
# Contribution
--->
# Proposed Method
Unit Length 以外, 次元ごとに PP. D をデータセット, Z を D 中の分散表現全体の集合とする.
z_{w_i, s_{w_i}} は, 文s_{w_i} 中の単語 w_i の分散表現.
* z-score Normalization

* Min-max Normalization

* Unit Length
L2 ノルムによる正規化.
* All-but-the-top
All-but-the-top は static embedding に対する操作として提案されたが, contextualized embedding でも計算することはできる.
k = d / 100, d = 768.

# Experiment
* Lexical-level Tasks
Unit Length は raw データとほとんど変わらす, MinMax はむしろ raw より性能が下がった.
zscore, All-but-the-top (abtt) は性能向上, 単体では abbt が良く, アンサンブルした zscore + abtt が大体一番良い.
XLNet では, 3, 4 層目にピーク, 3つのモデルで最終層付近での性能向上が顕著.

* Sequence classification Tasks
Unit Length, MinMax は性能が乏しいため無視.
zscore が一番良い. abbt は raw と似た結果. 一貫して zscore より zscore + abtt が良いというわけではない.

* Linguistic Correlation Analysis (LCA)
POS tag を決める際の, 層ごとの特徴量の寄与率を出している?
BERT の最終層では, raw では min, zscore 後では, max
LXNet の第0層では, raw では 0 だが, zscore 後ごでは, max
#185 を読んだ後に追記.

<!--
# Consideration
# Question
# Relation
--->
# Before Papers & Books
#183
# Next Papers & Books
#185
<!--
# Memo
--->
|
1.0
|
Effect of Post-processing on Contextualized Word Representations - # Information
### URL
https://arxiv.org/abs/2104.07456
### Authors
[Hassan Sajjad](https://arxiv.org/search/cs?searchtype=author&query=Sajjad%2C+H), [Firoj Alam](https://arxiv.org/search/cs?searchtype=author&query=Alam%2C+F), [Fahim Dalvi](https://arxiv.org/search/cs?searchtype=author&query=Dalvi%2C+F), [Nadir Durrani](https://arxiv.org/search/cs?searchtype=author&query=Durrani%2C+N)
<!--- ### Published at --->
### Date of issue
15 Apr 2021
<!---
### Related Websites
### Package
### Code
--->
# Reading Motivation
BERTのPPについて.
# Summary
static embedding の PP は, 単語レベル・文レベルのタスクで性能が向上することが知られている. しかし, contextualized embedding の PP の研究はあまりされていない. この論文では, pre-train された言語モデルの各層から得られる contextualized embedding の PP の有用性について調べた. PP の種類として, z-score, min-max を用いた分散表現の各次元を正規化, all-the-but-top を用いた主要成分の除去, unit length normalization を採用. 様々なpre-trained model で実験した結果, PP によって, 分散表現中に存在する, (word similarity, analogy などの) lexical タスク・文分類のための重要な情報をとりだせることが分かった. この研究により, contextulized embedding を用いた研究はより興味深いものとなり, contextulized embbeding に PPを使うときに z-score を用いた正規化が重要になると分かる.
# Background
* pre-train され言語モデルの layer ごとに分散が異なる.

<!---
# Contribution
--->
# Proposed Method
Unit Length 以外, 次元ごとに PP. D をデータセット, Z を D 中の分散表現全体の集合とする.
z_{w_i, s_{w_i}} は, 文s_{w_i} 中の単語 w_i の分散表現.
* z-score Normalization

* Min-max Normalization

* Unit Length
L2 ノルムによる正規化.
* All-but-the-top
All-but-the-top は static embedding に対する操作として提案されたが, contextualized embedding でも計算することはできる.
k = d / 100, d = 768.

# Experiment
* Lexical-level Tasks
Unit Length は raw データとほとんど変わらす, MinMax はむしろ raw より性能が下がった.
zscore, All-but-the-top (abtt) は性能向上, 単体では abbt が良く, アンサンブルした zscore + abtt が大体一番良い.
XLNet では, 3, 4 層目にピーク, 3つのモデルで最終層付近での性能向上が顕著.

* Sequence classification Tasks
Unit Length, MinMax は性能が乏しいため無視.
zscore が一番良い. abbt は raw と似た結果. 一貫して zscore より zscore + abtt が良いというわけではない.

* Linguistic Correlation Analysis (LCA)
POS tag を決める際の, 層ごとの特徴量の寄与率を出している?
BERT の最終層では, raw では min, zscore 後では, max
LXNet の第0層では, raw では 0 だが, zscore 後ごでは, max
#185 を読んだ後に追記.

<!--
# Consideration
# Question
# Relation
--->
# Before Papers & Books
#183
# Next Papers & Books
#185
<!--
# Memo
--->
|
process
|
effect of post processing on contextualized word representations information url authors date of issue apr related websites package code reading motivation bertのppについて summary static embedding の pp は 単語レベル・文レベルのタスクで性能が向上することが知られている しかし contextualized embedding の pp の研究はあまりされていない この論文では pre train された言語モデルの各層から得られる contextualized embedding の pp の有用性について調べた pp の種類として z score min max を用いた分散表現の各次元を正規化 all the but top を用いた主要成分の除去 unit length normalization を採用 様々なpre trained model で実験した結果 pp によって 分散表現中に存在する word similarity analogy などの lexical タスク・文分類のための重要な情報をとりだせることが分かった この研究により contextulized embedding を用いた研究はより興味深いものとなり contextulized embbeding に ppを使うときに z score を用いた正規化が重要になると分かる background pre train され言語モデルの layer ごとに分散が異なる contribution proposed method unit length 以外 次元ごとに pp d をデータセット z を d 中の分散表現全体の集合とする z w i s w i は 文s w i 中の単語 w i の分散表現 z score normalization min max normalization unit length ノルムによる正規化 all but the top all but the top は static embedding に対する操作として提案されたが contextualized embedding でも計算することはできる k d d experiment lexical level tasks unit length は raw データとほとんど変わらす minmax はむしろ raw より性能が下がった zscore all but the top abtt は性能向上 単体では abbt が良く アンサンブルした zscore abtt が大体一番良い xlnet では 層目にピーク sequence classification tasks unit length minmax は性能が乏しいため無視 zscore が一番良い abbt は raw と似た結果 一貫して zscore より zscore abtt が良いというわけではない linguistic correlation analysis lca pos tag を決める際の 層ごとの特徴量の寄与率を出している bert の最終層では raw では min zscore 後では max lxnet raw では だが zscore 後ごでは max を読んだ後に追記 consideration question relation before papers books next papers books memo
| 1
|
2,954
| 5,953,725,156
|
IssuesEvent
|
2017-05-27 10:37:34
|
kangarko/ChatControl-Pro
|
https://api.github.com/repos/kangarko/ChatControl-Pro
|
closed
|
Error on console
|
bug [queued / Your issue will be processed shortly]
|
13:16:19] [Async Chat Thread - #4/WARN]: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure
The last packet successfully received from the server was 29,748,105 milliseconds ago. The last packet sent successfully to the server was 0 milliseconds ago.
[13:16:19] [Async Chat Thread - #4/WARN]: at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
[13:16:19] [Async Chat Thread - #4/WARN]: at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
[13:16:19] [Async Chat Thread - #4/WARN]: at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.Util.handleNewInstance(Util.java:425)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:989)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.reuseAndReadPacket(MysqlIO.java:3556)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.reuseAndReadPacket(MysqlIO.java:3456)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:3897)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:2524)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2677)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2545)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2503)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.StatementImpl.executeQuery(StatementImpl.java:1369)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Com2.protected(SourceFile:62)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......lPT3.protected(SourceFile:139)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......lPT3.protected(SourceFile:47)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......LpT7.protected(SourceFile:116)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......LpT7.protected(SourceFile:99)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:526)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:420)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:273)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Com8.protected(SourceFile:173)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Com8.execute(SourceFile:43)
[13:16:19] [Async Chat Thread - #4/WARN]: at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:62)
[13:16:19] [Async Chat Thread - #4/WARN]: at org.bukkit.plugin.SimplePluginManager.fireEvent(SimplePluginManager.java:502)
[13:16:19] [Async Chat Thread - #4/WARN]: at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:484)
[13:16:19] [Async Chat Thread - #4/WARN]: at net.minecraft.server.v1_11_R1.PlayerConnection.chat(PlayerConnection.java:1276)
[13:16:19] [Async Chat Thread - #4/WARN]: at net.minecraft.server.v1_11_R1.PlayerConnection.a(PlayerConnection.java:1214)
[13:16:19] [Async Chat Thread - #4/WARN]: at net.minecraft.server.v1_11_R1.PacketPlayInChat$1.run(PacketPlayInChat.java:39)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.util.concurrent.FutureTask.run(FutureTask.java:266)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.lang.Thread.run(Thread.java:745)
[13:16:19] [Async Chat Thread - #4/WARN]: Caused by: java.io.EOFException: Can not read response from server. Expected to read 4 bytes, read 0 bytes before connection was unexpectedly lost.
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.readFully(MysqlIO.java:3008)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.reuseAndReadPacket(MysqlIO.java:3466)
[13:16:19] [Async Chat Thread - #4/WARN]: ... 28 more
[13:16:19] [Async Chat Thread - #4/ERROR]: Could not pass event AsyncPlayerChatEvent to ChatControl v7.3.15
java.lang.NullPointerException
at me.kangarko.chatcontrol. ......lPT3.protected(SourceFile:141) ~[?:?]
at me.kangarko.chatcontrol. ......lPT3.protected(SourceFile:47) ~[?:?]
at me.kangarko.chatcontrol. ......LpT7.protected(SourceFile:116) ~[?:?]
at me.kangarko.chatcontrol. ......LpT7.protected(SourceFile:99) ~[?:?]
at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:526) ~[?:?]
at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:420) ~[?:?]
at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:273) ~[?:?]
at me.kangarko.chatcontrol. ......Com8.protected(SourceFile:173) ~[?:?]
at me.kangarko.chatcontrol. ......Com8.execute(SourceFile:43) ~[?:?]
at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:62) ~[custom.jar:git-Spigot-38ddb34-c19c293]
at org.bukkit.plugin.SimplePluginManager.fireEvent(SimplePluginManager.java:502) [custom.jar:git-Spigot-38ddb34-c19c293]
at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:484) [custom.jar:git-Spigot-38ddb34-c19c293]
at net.minecraft.server.v1_11_R1.PlayerConnection.chat(PlayerConnection.java:1276) [custom.jar:git-Spigot-38ddb34-c19c293]
at net.minecraft.server.v1_11_R1.PlayerConnection.a(PlayerConnection.java:1214) [custom.jar:git-Spigot-38ddb34-c19c293]
at net.minecraft.server.v1_11_R1.PacketPlayInChat$1.run(PacketPlayInChat.java:39) [custom.jar:git-Spigot-38ddb34-c19c293]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) [?:1.8.0_121]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_121]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_121]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_121]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_121]
MC v 1.11.2 / ChatControl v 7.3.15
|
1.0
|
Error on console - 13:16:19] [Async Chat Thread - #4/WARN]: com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure
The last packet successfully received from the server was 29,748,105 milliseconds ago. The last packet sent successfully to the server was 0 milliseconds ago.
[13:16:19] [Async Chat Thread - #4/WARN]: at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
[13:16:19] [Async Chat Thread - #4/WARN]: at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
[13:16:19] [Async Chat Thread - #4/WARN]: at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.Util.handleNewInstance(Util.java:425)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.SQLError.createCommunicationsException(SQLError.java:989)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.reuseAndReadPacket(MysqlIO.java:3556)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.reuseAndReadPacket(MysqlIO.java:3456)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:3897)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:2524)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2677)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2545)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2503)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.StatementImpl.executeQuery(StatementImpl.java:1369)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Com2.protected(SourceFile:62)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......lPT3.protected(SourceFile:139)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......lPT3.protected(SourceFile:47)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......LpT7.protected(SourceFile:116)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......LpT7.protected(SourceFile:99)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:526)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:420)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:273)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Com8.protected(SourceFile:173)
[13:16:19] [Async Chat Thread - #4/WARN]: at me.kangarko.chatcontrol. ......Com8.execute(SourceFile:43)
[13:16:19] [Async Chat Thread - #4/WARN]: at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:62)
[13:16:19] [Async Chat Thread - #4/WARN]: at org.bukkit.plugin.SimplePluginManager.fireEvent(SimplePluginManager.java:502)
[13:16:19] [Async Chat Thread - #4/WARN]: at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:484)
[13:16:19] [Async Chat Thread - #4/WARN]: at net.minecraft.server.v1_11_R1.PlayerConnection.chat(PlayerConnection.java:1276)
[13:16:19] [Async Chat Thread - #4/WARN]: at net.minecraft.server.v1_11_R1.PlayerConnection.a(PlayerConnection.java:1214)
[13:16:19] [Async Chat Thread - #4/WARN]: at net.minecraft.server.v1_11_R1.PacketPlayInChat$1.run(PacketPlayInChat.java:39)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.util.concurrent.FutureTask.run(FutureTask.java:266)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
[13:16:19] [Async Chat Thread - #4/WARN]: at java.lang.Thread.run(Thread.java:745)
[13:16:19] [Async Chat Thread - #4/WARN]: Caused by: java.io.EOFException: Can not read response from server. Expected to read 4 bytes, read 0 bytes before connection was unexpectedly lost.
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.readFully(MysqlIO.java:3008)
[13:16:19] [Async Chat Thread - #4/WARN]: at com.mysql.jdbc.MysqlIO.reuseAndReadPacket(MysqlIO.java:3466)
[13:16:19] [Async Chat Thread - #4/WARN]: ... 28 more
[13:16:19] [Async Chat Thread - #4/ERROR]: Could not pass event AsyncPlayerChatEvent to ChatControl v7.3.15
java.lang.NullPointerException
at me.kangarko.chatcontrol. ......lPT3.protected(SourceFile:141) ~[?:?]
at me.kangarko.chatcontrol. ......lPT3.protected(SourceFile:47) ~[?:?]
at me.kangarko.chatcontrol. ......LpT7.protected(SourceFile:116) ~[?:?]
at me.kangarko.chatcontrol. ......LpT7.protected(SourceFile:99) ~[?:?]
at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:526) ~[?:?]
at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:420) ~[?:?]
at me.kangarko.chatcontrol. ......Lpt1.protected(SourceFile:273) ~[?:?]
at me.kangarko.chatcontrol. ......Com8.protected(SourceFile:173) ~[?:?]
at me.kangarko.chatcontrol. ......Com8.execute(SourceFile:43) ~[?:?]
at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:62) ~[custom.jar:git-Spigot-38ddb34-c19c293]
at org.bukkit.plugin.SimplePluginManager.fireEvent(SimplePluginManager.java:502) [custom.jar:git-Spigot-38ddb34-c19c293]
at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:484) [custom.jar:git-Spigot-38ddb34-c19c293]
at net.minecraft.server.v1_11_R1.PlayerConnection.chat(PlayerConnection.java:1276) [custom.jar:git-Spigot-38ddb34-c19c293]
at net.minecraft.server.v1_11_R1.PlayerConnection.a(PlayerConnection.java:1214) [custom.jar:git-Spigot-38ddb34-c19c293]
at net.minecraft.server.v1_11_R1.PacketPlayInChat$1.run(PacketPlayInChat.java:39) [custom.jar:git-Spigot-38ddb34-c19c293]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) [?:1.8.0_121]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_121]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_121]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_121]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_121]
MC v 1.11.2 / ChatControl v 7.3.15
|
process
|
error on console com mysql jdbc exceptions communicationsexception communications link failure the last packet successfully received from the server was milliseconds ago the last packet sent successfully to the server was milliseconds ago at sun reflect nativeconstructoraccessorimpl native method at sun reflect nativeconstructoraccessorimpl newinstance nativeconstructoraccessorimpl java at sun reflect delegatingconstructoraccessorimpl newinstance delegatingconstructoraccessorimpl java at java lang reflect constructor newinstance constructor java at com mysql jdbc util handlenewinstance util java at com mysql jdbc sqlerror createcommunicationsexception sqlerror java at com mysql jdbc mysqlio reuseandreadpacket mysqlio java at com mysql jdbc mysqlio reuseandreadpacket mysqlio java at com mysql jdbc mysqlio checkerrorpacket mysqlio java at com mysql jdbc mysqlio sendcommand mysqlio java at com mysql jdbc mysqlio sqlquerydirect mysqlio java at com mysql jdbc connectionimpl execsql connectionimpl java at com mysql jdbc connectionimpl execsql connectionimpl java at com mysql jdbc statementimpl executequery statementimpl java at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol execute sourcefile at org bukkit plugin registeredlistener callevent registeredlistener java at org bukkit plugin simplepluginmanager fireevent simplepluginmanager java at org bukkit plugin simplepluginmanager callevent simplepluginmanager java at net minecraft server playerconnection chat playerconnection java at net minecraft server playerconnection a playerconnection java at net minecraft server packetplayinchat run packetplayinchat java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by java io eofexception can not read response from server expected to read bytes read bytes before connection was unexpectedly lost at com mysql jdbc mysqlio readfully mysqlio java at com mysql jdbc mysqlio reuseandreadpacket mysqlio java more could not pass event asyncplayerchatevent to chatcontrol java lang nullpointerexception at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol protected sourcefile at me kangarko chatcontrol execute sourcefile at org bukkit plugin registeredlistener callevent registeredlistener java at org bukkit plugin simplepluginmanager fireevent simplepluginmanager java at org bukkit plugin simplepluginmanager callevent simplepluginmanager java at net minecraft server playerconnection chat playerconnection java at net minecraft server playerconnection a playerconnection java at net minecraft server packetplayinchat run packetplayinchat java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java mc v chatcontrol v
| 1
|
366,545
| 25,591,189,620
|
IssuesEvent
|
2022-12-01 13:07:50
|
spatial-data-lab/knime-geospatial-extension
|
https://api.github.com/repos/spatial-data-lab/knime-geospatial-extension
|
closed
|
US Census and Tiger maps node with links in option description
|
documentation
|
The new node dialogs have a little icon next to each dialog option. To make it easier for people to find the right MIPS etc numbers it would be great if we could add a link to a lookup page to the corresponding dialog option so that the link is shown directly in the node dialog when clicking on the question mark of the option.
|
1.0
|
US Census and Tiger maps node with links in option description - The new node dialogs have a little icon next to each dialog option. To make it easier for people to find the right MIPS etc numbers it would be great if we could add a link to a lookup page to the corresponding dialog option so that the link is shown directly in the node dialog when clicking on the question mark of the option.
|
non_process
|
us census and tiger maps node with links in option description the new node dialogs have a little icon next to each dialog option to make it easier for people to find the right mips etc numbers it would be great if we could add a link to a lookup page to the corresponding dialog option so that the link is shown directly in the node dialog when clicking on the question mark of the option
| 0
|
105,749
| 13,212,681,978
|
IssuesEvent
|
2020-08-16 08:32:48
|
oppia/oppia
|
https://api.github.com/repos/oppia/oppia
|
closed
|
MathExpressionInput: make rules use the mathematical interaction datatype
|
full-stack important needs design doc talk-to: @seanlip
|
For the mathematical expression interaction there is a single rule that compares the student's expression to an expression provided by the teacher. Currently the teacher enters their expression as text, but it would make sense to use the expression editor for that as well.
|
1.0
|
MathExpressionInput: make rules use the mathematical interaction datatype - For the mathematical expression interaction there is a single rule that compares the student's expression to an expression provided by the teacher. Currently the teacher enters their expression as text, but it would make sense to use the expression editor for that as well.
|
non_process
|
mathexpressioninput make rules use the mathematical interaction datatype for the mathematical expression interaction there is a single rule that compares the student s expression to an expression provided by the teacher currently the teacher enters their expression as text but it would make sense to use the expression editor for that as well
| 0
|
1,427
| 2,544,957,992
|
IssuesEvent
|
2015-01-29 14:17:02
|
AAndharia/ZIMS-School-Mgmt
|
https://api.github.com/repos/AAndharia/ZIMS-School-Mgmt
|
closed
|
Fields To Filter
|
Enhancement Ready For Testing
|
For Student Uniform Issue and Student Insurance Cover list pages, you have mentioned "Grade" as fields to filter. But those tables does not have "Grade" field.
I have added search by chapter and student name there for now. Please let me know which fields should be used to filter the list.
|
1.0
|
Fields To Filter - For Student Uniform Issue and Student Insurance Cover list pages, you have mentioned "Grade" as fields to filter. But those tables does not have "Grade" field.
I have added search by chapter and student name there for now. Please let me know which fields should be used to filter the list.
|
non_process
|
fields to filter for student uniform issue and student insurance cover list pages you have mentioned grade as fields to filter but those tables does not have grade field i have added search by chapter and student name there for now please let me know which fields should be used to filter the list
| 0
|
8,571
| 11,739,869,630
|
IssuesEvent
|
2020-03-11 18:30:18
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
child_process: spawnSync crashes trying to terminate setuid child because of maxBuffer exceeded
|
child_process confirmed-bug
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
-->
* **Version**: v13.8.0, v12.16.0
* **Platform**: macOS 10.15.3, Ubuntu 16.04.6 (4.15.0-70-generic)
* **Subsystem**: child_process, src/spawn_sync.c
### What steps will reproduce the bug?
<!--
Enter details about your bug, preferably a simple code snippet that can be
run using `node` directly without installing third-party dependencies.
-->
```
$ cat test.c
#include <unistd.h>
#include <stdio.h>
int main() { setuid(0); while (1) printf("hello"); }
$ gcc -o test test.c; sudo chown root test; sudo chmod 4755 test
$ node -e "require('child_process').spawnSync('./test')"
```
OR
```
$ node -e "require('child_process').spawnSync('sudo', ['bash', '-c', 'ls -lR /'])"
```
### How often does it reproduce? Is there a required condition?
Always
### What is the expected behavior?
<!--
If possible please provide textual output instead of screenshots.
-->
I expect to get `ENOBUF` (because of maxBuffer reached) and/or `EPERM` (because child process cannot be killed), so I can handle it somehow - but not crash.
### What do you see instead?
<!--
If possible please provide textual output instead of screenshots.
-->
```
node[68984]: ../src/spawn_sync.cc:611:void node::SyncProcessRunner::Kill(): Assertion `r >= 0 || r == UV_ESRCH' failed.
1: 0x100ba0c4a node::Abort() (.cold.1) [/usr/local/bin/node]
2: 0x100084961 node::FatalError(char const*, char const*) [/usr/local/bin/node]
3: 0x100084719 node::AppendExceptionLine(node::Environment*, v8::Local<v8::Value>, v8::Local<v8::Message>, node::ErrorHandlingMode) [/usr/local/bin/node]
4: 0x10010657d node::SyncProcessRunner::Kill() [/usr/local/bin/node]
5: 0x1006cb116 uv__stream_io [/usr/local/bin/node]
6: 0x1006d23a8 uv__io_poll [/usr/local/bin/node]
7: 0x1006c2fa2 uv_run [/usr/local/bin/node]
8: 0x10010598f node::SyncProcessRunner::TryInitializeAndRunLoop(v8::Local<v8::Value>) [/usr/local/bin/node]
9: 0x100105674 node::SyncProcessRunner::Run(v8::Local<v8::Value>) [/usr/local/bin/node]
10: 0x100105513 node::SyncProcessRunner::Spawn(v8::FunctionCallbackInfo<v8::Value> const&) [/usr/local/bin/node]
11: 0x1001cb578 v8::internal::FunctionCallbackArguments::Call(v8::internal::CallHandlerInfo) [/usr/local/bin/node]
12: 0x1001cac02 v8::internal::MaybeHandle<v8::internal::Object> v8::internal::(anonymous namespace)::HandleApiCallHelper<false>(v8::internal::Isolate*, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::FunctionTemplateInfo>, v8::internal::Handle<v8::internal::Object>, v8::internal::BuiltinArguments) [/usr/local/bin/node]
13: 0x1001ca40e v8::internal::Builtin_Impl_HandleApiCall(v8::internal::BuiltinArguments, v8::internal::Isolate*) [/usr/local/bin/node]
14: 0x1007503d9 Builtins_CEntry_Return1_DontSaveFPRegs_ArgvOnStack_BuiltinExit [/usr/local/bin/node]
zsh: abort node -e "require('child_process').spawnSync('./test')"
```
### Additional information
<!--
Tell us anything else you think we should know.
-->
According to documentation, `child_process.spawnSync()` terminates child process if it's output is larger than `maxBuffer`. If child process can't be killed (because of setuid call), node crashes in `CHECK()` below https://github.com/nodejs/node/blob/master/src/spawn_sync.cc#L611:
```
// If uv_kill failed with an error that isn't ESRCH, the user probably
// specified an invalid or unsupported signal. Signal this to the user as
// and error and kill the process with SIGKILL instead.
if (r < 0 && r != UV_ESRCH) {
SetError(r);
r = uv_process_kill(&uv_process_, SIGKILL);
CHECK(r >= 0 || r == UV_ESRCH);
}
```
Shouldn't we also check for UV_EPERM?
|
1.0
|
child_process: spawnSync crashes trying to terminate setuid child because of maxBuffer exceeded - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
-->
* **Version**: v13.8.0, v12.16.0
* **Platform**: macOS 10.15.3, Ubuntu 16.04.6 (4.15.0-70-generic)
* **Subsystem**: child_process, src/spawn_sync.c
### What steps will reproduce the bug?
<!--
Enter details about your bug, preferably a simple code snippet that can be
run using `node` directly without installing third-party dependencies.
-->
```
$ cat test.c
#include <unistd.h>
#include <stdio.h>
int main() { setuid(0); while (1) printf("hello"); }
$ gcc -o test test.c; sudo chown root test; sudo chmod 4755 test
$ node -e "require('child_process').spawnSync('./test')"
```
OR
```
$ node -e "require('child_process').spawnSync('sudo', ['bash', '-c', 'ls -lR /'])"
```
### How often does it reproduce? Is there a required condition?
Always
### What is the expected behavior?
<!--
If possible please provide textual output instead of screenshots.
-->
I expect to get `ENOBUF` (because of maxBuffer reached) and/or `EPERM` (because child process cannot be killed), so I can handle it somehow - but not crash.
### What do you see instead?
<!--
If possible please provide textual output instead of screenshots.
-->
```
node[68984]: ../src/spawn_sync.cc:611:void node::SyncProcessRunner::Kill(): Assertion `r >= 0 || r == UV_ESRCH' failed.
1: 0x100ba0c4a node::Abort() (.cold.1) [/usr/local/bin/node]
2: 0x100084961 node::FatalError(char const*, char const*) [/usr/local/bin/node]
3: 0x100084719 node::AppendExceptionLine(node::Environment*, v8::Local<v8::Value>, v8::Local<v8::Message>, node::ErrorHandlingMode) [/usr/local/bin/node]
4: 0x10010657d node::SyncProcessRunner::Kill() [/usr/local/bin/node]
5: 0x1006cb116 uv__stream_io [/usr/local/bin/node]
6: 0x1006d23a8 uv__io_poll [/usr/local/bin/node]
7: 0x1006c2fa2 uv_run [/usr/local/bin/node]
8: 0x10010598f node::SyncProcessRunner::TryInitializeAndRunLoop(v8::Local<v8::Value>) [/usr/local/bin/node]
9: 0x100105674 node::SyncProcessRunner::Run(v8::Local<v8::Value>) [/usr/local/bin/node]
10: 0x100105513 node::SyncProcessRunner::Spawn(v8::FunctionCallbackInfo<v8::Value> const&) [/usr/local/bin/node]
11: 0x1001cb578 v8::internal::FunctionCallbackArguments::Call(v8::internal::CallHandlerInfo) [/usr/local/bin/node]
12: 0x1001cac02 v8::internal::MaybeHandle<v8::internal::Object> v8::internal::(anonymous namespace)::HandleApiCallHelper<false>(v8::internal::Isolate*, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::FunctionTemplateInfo>, v8::internal::Handle<v8::internal::Object>, v8::internal::BuiltinArguments) [/usr/local/bin/node]
13: 0x1001ca40e v8::internal::Builtin_Impl_HandleApiCall(v8::internal::BuiltinArguments, v8::internal::Isolate*) [/usr/local/bin/node]
14: 0x1007503d9 Builtins_CEntry_Return1_DontSaveFPRegs_ArgvOnStack_BuiltinExit [/usr/local/bin/node]
zsh: abort node -e "require('child_process').spawnSync('./test')"
```
### Additional information
<!--
Tell us anything else you think we should know.
-->
According to documentation, `child_process.spawnSync()` terminates child process if it's output is larger than `maxBuffer`. If child process can't be killed (because of setuid call), node crashes in `CHECK()` below https://github.com/nodejs/node/blob/master/src/spawn_sync.cc#L611:
```
// If uv_kill failed with an error that isn't ESRCH, the user probably
// specified an invalid or unsupported signal. Signal this to the user as
// and error and kill the process with SIGKILL instead.
if (r < 0 && r != UV_ESRCH) {
SetError(r);
r = uv_process_kill(&uv_process_, SIGKILL);
CHECK(r >= 0 || r == UV_ESRCH);
}
```
Shouldn't we also check for UV_EPERM?
|
process
|
child process spawnsync crashes trying to terminate setuid child because of maxbuffer exceeded thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name version platform macos ubuntu generic subsystem child process src spawn sync c what steps will reproduce the bug enter details about your bug preferably a simple code snippet that can be run using node directly without installing third party dependencies cat test c include include int main setuid while printf hello gcc o test test c sudo chown root test sudo chmod test node e require child process spawnsync test or node e require child process spawnsync sudo how often does it reproduce is there a required condition always what is the expected behavior if possible please provide textual output instead of screenshots i expect to get enobuf because of maxbuffer reached and or eperm because child process cannot be killed so i can handle it somehow but not crash what do you see instead if possible please provide textual output instead of screenshots node src spawn sync cc void node syncprocessrunner kill assertion r r uv esrch failed node abort cold node fatalerror char const char const node appendexceptionline node environment local local node errorhandlingmode node syncprocessrunner kill uv stream io uv io poll uv run node syncprocessrunner tryinitializeandrunloop local node syncprocessrunner run local node syncprocessrunner spawn functioncallbackinfo const internal functioncallbackarguments call internal callhandlerinfo internal maybehandle internal anonymous namespace handleapicallhelper internal isolate internal handle internal handle internal handle internal handle internal builtinarguments internal builtin impl handleapicall internal builtinarguments internal isolate builtins centry dontsavefpregs argvonstack builtinexit zsh abort node e require child process spawnsync test additional information tell us anything else you think we should know according to documentation child process spawnsync terminates child process if it s output is larger than maxbuffer if child process can t be killed because of setuid call node crashes in check below if uv kill failed with an error that isn t esrch the user probably specified an invalid or unsupported signal signal this to the user as and error and kill the process with sigkill instead if r r uv esrch seterror r r uv process kill uv process sigkill check r r uv esrch shouldn t we also check for uv eperm
| 1
|
73,128
| 31,877,742,627
|
IssuesEvent
|
2023-09-16 02:30:50
|
jongfeel/BookReview
|
https://api.github.com/repos/jongfeel/BookReview
|
closed
|
10-1 모니터링의 이해
|
Hands-on microservices with kotlin 2023
|
### 모니터링의 이해
마이크로서비스를 운영 환경으로 전환하는데 있어서 얼마나 잘 만들었는지에 상관 없이 장애는 발생한다.
장애는 언젠가 발생하기 때문에 그 시점이 언제인지를 알고 상황을 처리할 수 있는 도구가 필요하다.
#### 모니터링은 무엇인가
운영 시스템에서 장애를 이해하고 대응하는 가장 좋은 방법은 실제로 무엇이 일어나고 있는지 파악하는 것이다. 이를 위해서는 모니터링해야 한다.
모니터링은 서비스에서 무슨 일이 일어나고 있는지 알 수 있고 측정 가능한 데이터를 제공한다.
모니터링 데이터의 변경, 추세 또는 패턴을 이해하기 위해 특정 시간이나 이벤트의 값 정보의 흐름으로 다른 분석 시스템 또는 MI시스템Management Information System에 제공할 수 있다.
고급 분석 기능을 사용해 정보를 탐색할 수 있는 시스템에 모니터링 피드를 통합하는 것이 일반적이다. 이런 시스템에서는 데이터를 탐색하고 분석하거나 기계 학습 알고리즘을 사용해 시스템에서 발생할 수 있는 것과 어떻게 처리해야 하는지를 예측할 수 있다.
> 모니터링 피드 통합은 업계에서 성공적인 관행으로 입증된 트렌드이다. 지금 바로 사용하지 않아도 정보를 저장하는 건 나중에 매우 유용하게 쓸 수 있으므로 이런 피드를 만드는 것을 고려해야 한다.
모니터링은 아키텍처와 시스템의 모든 요소가 애플리케이션의 현재 상태를 이해하는 데 사용할 수 있는 의미 있는 모니터링을 제공해야 한다. 모니터링은 데이터를 많이 가지고 있지만 실제로 볼 수 없다면 쓸모 없을 수 있다. 이런 정보를 지속적으로 제공할 수 있는 대시보드나 제어판을 갖는 것이 중요하다.
#### 경고
장애나 문제 발생 시 위험을 알릴 수 있는 정보를 가지고 있으면 상황에 대처할 수 있다. 경고는 제어판이나 대시보드에서 시각화하거나 개인 또는 그룹에게 알리거나 자동으로 처리되는 시스템으로 보낼 수 있다.
경고를 정의할 때는 다음 요소에 대해 고려해 보면 좋다.
- 의미 있는 정보: 단순 장애 메시지 말고 문제 상황이나 처리 방법, 힌트를 제공해야 한다.
- 경고를 촉발시키는 조건
- 맥락 정보: 실제 문제가 무엇이고 어디에서 발생했는가?
- 요약과 세부 정보 분류를 제공하려는 노력: 추가 정보를 볼 수 있어야 한다.
- 분류 가능하게 만들기: 어떻게 분류할지 고려
- 의미 있는 심각성: 심각성 등급을 다 높게 설정하는 건 의미 없음
- 추적 기능: 경고 발생 시간은 매우 중요하다
- 진정성: 진짜 경고인지 판별
> 경고는 분석 시스템에 다른 피드로 사용할 수 있고, 볼륨 추세, 예측 문제의 세부 정보를 생성하는 데도 사용할 수 있다. 이런 피드를 만드는 노력이 성과를 거둘 것이다.
#### 복구
장애가 발생하면 복구하기 위한 메커니즘을 제공해야 한다. 일부는 자동 처리 및 복구가 되지만 일부는 수동 작업이 필요하다.
> 하드 드라이브 고장의 경우는 자동 처리가 아니라 누군가 직접 하드 드라이브를 수동으로 교체해야 한다.
예로 데이터베이스 연결이 끊어졌을 때 문제 해결 후 다시 연결할 수 있어야 한다. 이런 경우에 서비스를 관리하는 데 도움이 되는 도구를 제공해야 한다.
도구의 경우 자체 내장이거나 외부에 별도로 만들어야 할 수도 있고, 기존 도구를 사용해 복구를 처리할 수도 있다.
#### 장애 발생 마이크로서비스 만들기
https://start.spring.io 에서 마이크로서비스 생성
<img width="1000" alt="image" src="https://github.com/jongfeel/BookReview/assets/17442457/706107d5-8cf3-43d5-bbd8-9e7ae069a3be">
고객 검색을 담당하는 서비스의 인터페이스 생성
``` kotlin
package com.microservice.chapter10
interface GreetingsService {
fun getGreetings(): String
}
```
GreetingsServiceImpl 클래스 구현
``` kotlin
package com.microservice.chapter10
import org.springframework.stereotype.Service
import java.util.*
@Service
class GreetingsServiceImpl : GreetingsService {
companion object {
private val greetingsMessages = arrayOf("Hello", "Ola", "Namaste", "Hola")
}
fun random(max: Int): Int = Random().nextInt(max) + 1
override fun getGreetings(): String = greetingsMessages[random(4)]
}
```
서비스가 호출될 때마다 임의의 메시지를 반환하지만 확률적으로 오류가 발생하는 간단한 구현이다.
즉 random 함수는 1에서 4 사이의 숫자를 반환하는데, 배열의 인덱스는 0에서 3사이의 숫자로 접근해야 한다.
무작위로 ArrayIndexOutOfBoundsException 장애가 발생하는 예제인 것이다.
테스트 단계에서 걸러질 수도 있지만 테스트에서는 통과했지만 운영에서 원치 않는 결과를 얻을 수도 있다. 여기서는 시연 목적으로 만든 것이다.
GreetingsService를 사용해 GET 요청의 메시지를 반환하는 GreetingsController를 만든다.
``` kotlin
package com.microservice.chapter10
import org.springframework.beans.factory.annotation.Autowired
import org.springframework.web.bind.annotation.GetMapping
import org.springframework.web.bind.annotation.RestController
@RestController
class GreetingsController {
@Autowired
lateinit var greetingsService: GreetingsService
@GetMapping("/hello")
fun message() = greetingsService.getGreetings()
}
```
이 컨트롤러는 http://localhost:8080/hello 를 요청하면 임의의 메시지를 반환한다.
실행한 후에는 메시지 혹은 오류가 발생하는 걸 확인할 수 있다.
무작위로 에러가 발생하는 상황
<img width="496" alt="image" src="https://github.com/jongfeel/BookReview/assets/17442457/7282f9f3-065b-46ef-b6a7-10d1dd63efb4">
|
1.0
|
10-1 모니터링의 이해 - ### 모니터링의 이해
마이크로서비스를 운영 환경으로 전환하는데 있어서 얼마나 잘 만들었는지에 상관 없이 장애는 발생한다.
장애는 언젠가 발생하기 때문에 그 시점이 언제인지를 알고 상황을 처리할 수 있는 도구가 필요하다.
#### 모니터링은 무엇인가
운영 시스템에서 장애를 이해하고 대응하는 가장 좋은 방법은 실제로 무엇이 일어나고 있는지 파악하는 것이다. 이를 위해서는 모니터링해야 한다.
모니터링은 서비스에서 무슨 일이 일어나고 있는지 알 수 있고 측정 가능한 데이터를 제공한다.
모니터링 데이터의 변경, 추세 또는 패턴을 이해하기 위해 특정 시간이나 이벤트의 값 정보의 흐름으로 다른 분석 시스템 또는 MI시스템Management Information System에 제공할 수 있다.
고급 분석 기능을 사용해 정보를 탐색할 수 있는 시스템에 모니터링 피드를 통합하는 것이 일반적이다. 이런 시스템에서는 데이터를 탐색하고 분석하거나 기계 학습 알고리즘을 사용해 시스템에서 발생할 수 있는 것과 어떻게 처리해야 하는지를 예측할 수 있다.
> 모니터링 피드 통합은 업계에서 성공적인 관행으로 입증된 트렌드이다. 지금 바로 사용하지 않아도 정보를 저장하는 건 나중에 매우 유용하게 쓸 수 있으므로 이런 피드를 만드는 것을 고려해야 한다.
모니터링은 아키텍처와 시스템의 모든 요소가 애플리케이션의 현재 상태를 이해하는 데 사용할 수 있는 의미 있는 모니터링을 제공해야 한다. 모니터링은 데이터를 많이 가지고 있지만 실제로 볼 수 없다면 쓸모 없을 수 있다. 이런 정보를 지속적으로 제공할 수 있는 대시보드나 제어판을 갖는 것이 중요하다.
#### 경고
장애나 문제 발생 시 위험을 알릴 수 있는 정보를 가지고 있으면 상황에 대처할 수 있다. 경고는 제어판이나 대시보드에서 시각화하거나 개인 또는 그룹에게 알리거나 자동으로 처리되는 시스템으로 보낼 수 있다.
경고를 정의할 때는 다음 요소에 대해 고려해 보면 좋다.
- 의미 있는 정보: 단순 장애 메시지 말고 문제 상황이나 처리 방법, 힌트를 제공해야 한다.
- 경고를 촉발시키는 조건
- 맥락 정보: 실제 문제가 무엇이고 어디에서 발생했는가?
- 요약과 세부 정보 분류를 제공하려는 노력: 추가 정보를 볼 수 있어야 한다.
- 분류 가능하게 만들기: 어떻게 분류할지 고려
- 의미 있는 심각성: 심각성 등급을 다 높게 설정하는 건 의미 없음
- 추적 기능: 경고 발생 시간은 매우 중요하다
- 진정성: 진짜 경고인지 판별
> 경고는 분석 시스템에 다른 피드로 사용할 수 있고, 볼륨 추세, 예측 문제의 세부 정보를 생성하는 데도 사용할 수 있다. 이런 피드를 만드는 노력이 성과를 거둘 것이다.
#### 복구
장애가 발생하면 복구하기 위한 메커니즘을 제공해야 한다. 일부는 자동 처리 및 복구가 되지만 일부는 수동 작업이 필요하다.
> 하드 드라이브 고장의 경우는 자동 처리가 아니라 누군가 직접 하드 드라이브를 수동으로 교체해야 한다.
예로 데이터베이스 연결이 끊어졌을 때 문제 해결 후 다시 연결할 수 있어야 한다. 이런 경우에 서비스를 관리하는 데 도움이 되는 도구를 제공해야 한다.
도구의 경우 자체 내장이거나 외부에 별도로 만들어야 할 수도 있고, 기존 도구를 사용해 복구를 처리할 수도 있다.
#### 장애 발생 마이크로서비스 만들기
https://start.spring.io 에서 마이크로서비스 생성
<img width="1000" alt="image" src="https://github.com/jongfeel/BookReview/assets/17442457/706107d5-8cf3-43d5-bbd8-9e7ae069a3be">
고객 검색을 담당하는 서비스의 인터페이스 생성
``` kotlin
package com.microservice.chapter10
interface GreetingsService {
fun getGreetings(): String
}
```
GreetingsServiceImpl 클래스 구현
``` kotlin
package com.microservice.chapter10
import org.springframework.stereotype.Service
import java.util.*
@Service
class GreetingsServiceImpl : GreetingsService {
companion object {
private val greetingsMessages = arrayOf("Hello", "Ola", "Namaste", "Hola")
}
fun random(max: Int): Int = Random().nextInt(max) + 1
override fun getGreetings(): String = greetingsMessages[random(4)]
}
```
서비스가 호출될 때마다 임의의 메시지를 반환하지만 확률적으로 오류가 발생하는 간단한 구현이다.
즉 random 함수는 1에서 4 사이의 숫자를 반환하는데, 배열의 인덱스는 0에서 3사이의 숫자로 접근해야 한다.
무작위로 ArrayIndexOutOfBoundsException 장애가 발생하는 예제인 것이다.
테스트 단계에서 걸러질 수도 있지만 테스트에서는 통과했지만 운영에서 원치 않는 결과를 얻을 수도 있다. 여기서는 시연 목적으로 만든 것이다.
GreetingsService를 사용해 GET 요청의 메시지를 반환하는 GreetingsController를 만든다.
``` kotlin
package com.microservice.chapter10
import org.springframework.beans.factory.annotation.Autowired
import org.springframework.web.bind.annotation.GetMapping
import org.springframework.web.bind.annotation.RestController
@RestController
class GreetingsController {
@Autowired
lateinit var greetingsService: GreetingsService
@GetMapping("/hello")
fun message() = greetingsService.getGreetings()
}
```
이 컨트롤러는 http://localhost:8080/hello 를 요청하면 임의의 메시지를 반환한다.
실행한 후에는 메시지 혹은 오류가 발생하는 걸 확인할 수 있다.
무작위로 에러가 발생하는 상황
<img width="496" alt="image" src="https://github.com/jongfeel/BookReview/assets/17442457/7282f9f3-065b-46ef-b6a7-10d1dd63efb4">
|
non_process
|
모니터링의 이해 모니터링의 이해 마이크로서비스를 운영 환경으로 전환하는데 있어서 얼마나 잘 만들었는지에 상관 없이 장애는 발생한다 장애는 언젠가 발생하기 때문에 그 시점이 언제인지를 알고 상황을 처리할 수 있는 도구가 필요하다 모니터링은 무엇인가 운영 시스템에서 장애를 이해하고 대응하는 가장 좋은 방법은 실제로 무엇이 일어나고 있는지 파악하는 것이다 이를 위해서는 모니터링해야 한다 모니터링은 서비스에서 무슨 일이 일어나고 있는지 알 수 있고 측정 가능한 데이터를 제공한다 모니터링 데이터의 변경 추세 또는 패턴을 이해하기 위해 특정 시간이나 이벤트의 값 정보의 흐름으로 다른 분석 시스템 또는 mi시스템management information system에 제공할 수 있다 고급 분석 기능을 사용해 정보를 탐색할 수 있는 시스템에 모니터링 피드를 통합하는 것이 일반적이다 이런 시스템에서는 데이터를 탐색하고 분석하거나 기계 학습 알고리즘을 사용해 시스템에서 발생할 수 있는 것과 어떻게 처리해야 하는지를 예측할 수 있다 모니터링 피드 통합은 업계에서 성공적인 관행으로 입증된 트렌드이다 지금 바로 사용하지 않아도 정보를 저장하는 건 나중에 매우 유용하게 쓸 수 있으므로 이런 피드를 만드는 것을 고려해야 한다 모니터링은 아키텍처와 시스템의 모든 요소가 애플리케이션의 현재 상태를 이해하는 데 사용할 수 있는 의미 있는 모니터링을 제공해야 한다 모니터링은 데이터를 많이 가지고 있지만 실제로 볼 수 없다면 쓸모 없을 수 있다 이런 정보를 지속적으로 제공할 수 있는 대시보드나 제어판을 갖는 것이 중요하다 경고 장애나 문제 발생 시 위험을 알릴 수 있는 정보를 가지고 있으면 상황에 대처할 수 있다 경고는 제어판이나 대시보드에서 시각화하거나 개인 또는 그룹에게 알리거나 자동으로 처리되는 시스템으로 보낼 수 있다 경고를 정의할 때는 다음 요소에 대해 고려해 보면 좋다 의미 있는 정보 단순 장애 메시지 말고 문제 상황이나 처리 방법 힌트를 제공해야 한다 경고를 촉발시키는 조건 맥락 정보 실제 문제가 무엇이고 어디에서 발생했는가 요약과 세부 정보 분류를 제공하려는 노력 추가 정보를 볼 수 있어야 한다 분류 가능하게 만들기 어떻게 분류할지 고려 의미 있는 심각성 심각성 등급을 다 높게 설정하는 건 의미 없음 추적 기능 경고 발생 시간은 매우 중요하다 진정성 진짜 경고인지 판별 경고는 분석 시스템에 다른 피드로 사용할 수 있고 볼륨 추세 예측 문제의 세부 정보를 생성하는 데도 사용할 수 있다 이런 피드를 만드는 노력이 성과를 거둘 것이다 복구 장애가 발생하면 복구하기 위한 메커니즘을 제공해야 한다 일부는 자동 처리 및 복구가 되지만 일부는 수동 작업이 필요하다 하드 드라이브 고장의 경우는 자동 처리가 아니라 누군가 직접 하드 드라이브를 수동으로 교체해야 한다 예로 데이터베이스 연결이 끊어졌을 때 문제 해결 후 다시 연결할 수 있어야 한다 이런 경우에 서비스를 관리하는 데 도움이 되는 도구를 제공해야 한다 도구의 경우 자체 내장이거나 외부에 별도로 만들어야 할 수도 있고 기존 도구를 사용해 복구를 처리할 수도 있다 장애 발생 마이크로서비스 만들기 에서 마이크로서비스 생성 img width alt image src 고객 검색을 담당하는 서비스의 인터페이스 생성 kotlin package com microservice interface greetingsservice fun getgreetings string greetingsserviceimpl 클래스 구현 kotlin package com microservice import org springframework stereotype service import java util service class greetingsserviceimpl greetingsservice companion object private val greetingsmessages arrayof hello ola namaste hola fun random max int int random nextint max override fun getgreetings string greetingsmessages 서비스가 호출될 때마다 임의의 메시지를 반환하지만 확률적으로 오류가 발생하는 간단한 구현이다 즉 random 함수는 사이의 숫자를 반환하는데 배열의 인덱스는 숫자로 접근해야 한다 무작위로 arrayindexoutofboundsexception 장애가 발생하는 예제인 것이다 테스트 단계에서 걸러질 수도 있지만 테스트에서는 통과했지만 운영에서 원치 않는 결과를 얻을 수도 있다 여기서는 시연 목적으로 만든 것이다 greetingsservice를 사용해 get 요청의 메시지를 반환하는 greetingscontroller를 만든다 kotlin package com microservice import org springframework beans factory annotation autowired import org springframework web bind annotation getmapping import org springframework web bind annotation restcontroller restcontroller class greetingscontroller autowired lateinit var greetingsservice greetingsservice getmapping hello fun message greetingsservice getgreetings 이 컨트롤러는 를 요청하면 임의의 메시지를 반환한다 실행한 후에는 메시지 혹은 오류가 발생하는 걸 확인할 수 있다 무작위로 에러가 발생하는 상황 img width alt image src
| 0
|
12,882
| 15,278,654,468
|
IssuesEvent
|
2021-02-23 02:00:49
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
closed
|
Hide hammerhead processing comments (style, script, processing header, etc) for client code
|
AREA: client STATE: Stale SYSTEM: client side processing TYPE: enhancement
|
In some very special cases we have different results with and without processing:
``` javascript
var e = document.createElement('div');
e.innerHTML = "<style><!--</style>--></style>";
e.firstChild && e.firstChild.firstChild && e.firstChild.firstChild.nodeValue == "<!--";
```
After processing `e.firstChild.firstChild.nodeValue =="/*hammerhead|stylesheet|start*/↵<!--↵/*hammerhead|stylesheet|end*/"`
|
1.0
|
Hide hammerhead processing comments (style, script, processing header, etc) for client code - In some very special cases we have different results with and without processing:
``` javascript
var e = document.createElement('div');
e.innerHTML = "<style><!--</style>--></style>";
e.firstChild && e.firstChild.firstChild && e.firstChild.firstChild.nodeValue == "<!--";
```
After processing `e.firstChild.firstChild.nodeValue =="/*hammerhead|stylesheet|start*/↵<!--↵/*hammerhead|stylesheet|end*/"`
|
process
|
hide hammerhead processing comments style script processing header etc for client code in some very special cases we have different results with and without processing javascript var e document createelement div e innerhtml e firstchild e firstchild firstchild e firstchild firstchild nodevalue after processing e firstchild firstchild nodevalue hammerhead stylesheet start ↵ ↵ hammerhead stylesheet end
| 1
|
10,983
| 4,862,409,469
|
IssuesEvent
|
2016-11-14 12:18:13
|
JabRef/jabref
|
https://api.github.com/repos/JabRef/jabref
|
closed
|
Entry editor steals focus from search bar
|
fixed-in-devBuilds waiting-for-feedback
|
# Summary
When having an entry in edit mode, and using the quick search to find another entry (to edit), then while typing into search bar, focus jumps to editor, and you begin editing some entry instead.
## Version
- JabRef 3.6
- linux 4.4.0-45-generic amd64
- Java 1.8.0_101
## Steps to reproduce:
1. select an entry, to have it in edit mode
2. begin typing in the quick search bar
3. observe how focus jumps, usually after 1-2 characters, to the editor where you instantly begin to change an entry. The edited entry is the first found entry using the typed characters so far.
## Solution
Focus should stay in search bar, at least until enter is hit, or something sane.
|
1.0
|
Entry editor steals focus from search bar - # Summary
When having an entry in edit mode, and using the quick search to find another entry (to edit), then while typing into search bar, focus jumps to editor, and you begin editing some entry instead.
## Version
- JabRef 3.6
- linux 4.4.0-45-generic amd64
- Java 1.8.0_101
## Steps to reproduce:
1. select an entry, to have it in edit mode
2. begin typing in the quick search bar
3. observe how focus jumps, usually after 1-2 characters, to the editor where you instantly begin to change an entry. The edited entry is the first found entry using the typed characters so far.
## Solution
Focus should stay in search bar, at least until enter is hit, or something sane.
|
non_process
|
entry editor steals focus from search bar summary when having an entry in edit mode and using the quick search to find another entry to edit then while typing into search bar focus jumps to editor and you begin editing some entry instead version jabref linux generic java steps to reproduce select an entry to have it in edit mode begin typing in the quick search bar observe how focus jumps usually after characters to the editor where you instantly begin to change an entry the edited entry is the first found entry using the typed characters so far solution focus should stay in search bar at least until enter is hit or something sane
| 0
|
287,997
| 31,856,885,599
|
IssuesEvent
|
2023-09-15 08:08:00
|
nidhi7598/linux-4.19.72_CVE-2022-3564
|
https://api.github.com/repos/nidhi7598/linux-4.19.72_CVE-2022-3564
|
closed
|
CVE-2022-1852 (Medium) detected in linuxlinux-4.19.294 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2022-1852 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A NULL pointer dereference flaw was found in the Linux kernel’s KVM module, which can lead to a denial of service in the x86_emulate_insn in arch/x86/kvm/emulate.c. This flaw occurs while executing an illegal instruction in guest in the Intel CPU.
<p>Publish Date: 2022-06-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1852>CVE-2022-1852</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1852">https://www.linuxkernelcves.com/cves/CVE-2022-1852</a></p>
<p>Release Date: 2022-05-24</p>
<p>Fix Resolution: v5.10.120,v5.15.45,v5.17.13,v5.18.2,v5.19-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-1852 (Medium) detected in linuxlinux-4.19.294 - autoclosed - ## CVE-2022-1852 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A NULL pointer dereference flaw was found in the Linux kernel’s KVM module, which can lead to a denial of service in the x86_emulate_insn in arch/x86/kvm/emulate.c. This flaw occurs while executing an illegal instruction in guest in the Intel CPU.
<p>Publish Date: 2022-06-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1852>CVE-2022-1852</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1852">https://www.linuxkernelcves.com/cves/CVE-2022-1852</a></p>
<p>Release Date: 2022-05-24</p>
<p>Fix Resolution: v5.10.120,v5.15.45,v5.17.13,v5.18.2,v5.19-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files vulnerability details a null pointer dereference flaw was found in the linux kernel’s kvm module which can lead to a denial of service in the emulate insn in arch kvm emulate c this flaw occurs while executing an illegal instruction in guest in the intel cpu publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
8,214
| 11,405,413,518
|
IssuesEvent
|
2020-01-31 12:02:36
|
parcel-bundler/parcel
|
https://api.github.com/repos/parcel-bundler/parcel
|
closed
|
Cannot disable loose entities encoding
|
:grey_question: Question HTML Preprocessing Stale
|
Hi! I'm using out-of-the-box parcel bundler, trying to use it to do some Backbone app development (yes, we still exist!)
The problem is that if you have an HTML template with stuff like this inside it:
```
<template id="IngredientsPage">
<section class="ingredients page">
<p>This are all the ingredients that the app can handle:</p>
{% if(data.get("ingredients").length) { %}
<ul>
{% data.get("ingredients").forEach( ingredient => %}
<li>Stuff!</li>
{% } %}
</ul>
{% } %}
</section>
</template>
```
Then Parcel will autoescape the ">" bit into > and I cannot disable that.
## 🤔 Expected Behavior
There should be some way of disabling this autoescaping for situations like mine
<!--- Tell us what should happen -->
## 😯 Current Behavior
Currently there is no way of disabling this escaping, not even using some other hacks like ```<script type="text/template"></script>```.
<!--- Tell us what happens instead of the expected behavior -->
<!--- If you are seeing an error, please include the full error message and stack trace -->
|
1.0
|
Cannot disable loose entities encoding - Hi! I'm using out-of-the-box parcel bundler, trying to use it to do some Backbone app development (yes, we still exist!)
The problem is that if you have an HTML template with stuff like this inside it:
```
<template id="IngredientsPage">
<section class="ingredients page">
<p>This are all the ingredients that the app can handle:</p>
{% if(data.get("ingredients").length) { %}
<ul>
{% data.get("ingredients").forEach( ingredient => %}
<li>Stuff!</li>
{% } %}
</ul>
{% } %}
</section>
</template>
```
Then Parcel will autoescape the ">" bit into > and I cannot disable that.
## 🤔 Expected Behavior
There should be some way of disabling this autoescaping for situations like mine
<!--- Tell us what should happen -->
## 😯 Current Behavior
Currently there is no way of disabling this escaping, not even using some other hacks like ```<script type="text/template"></script>```.
<!--- Tell us what happens instead of the expected behavior -->
<!--- If you are seeing an error, please include the full error message and stack trace -->
|
process
|
cannot disable loose entities encoding hi i m using out of the box parcel bundler trying to use it to do some backbone app development yes we still exist the problem is that if you have an html template with stuff like this inside it this are all the ingredients that the app can handle if data get ingredients length data get ingredients foreach ingredient stuff then parcel will autoescape the bit into gt and i cannot disable that 🤔 expected behavior there should be some way of disabling this autoescaping for situations like mine 😯 current behavior currently there is no way of disabling this escaping not even using some other hacks like
| 1
|
3,321
| 6,437,655,685
|
IssuesEvent
|
2017-08-11 00:08:55
|
GoogleCloudPlatform/google-cloud-java
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-java
|
closed
|
Convert version update process to gradle
|
Priority: P2+ release process Type: Enhancement
|
The current version update script is in bash and is hard to maintain. Currently it has a few bugs:
- it changes the version of the grpc dependency
- it doesn't update the versions of core-http or core-grpc
- it doesn't update the versions of modules under google-cloud-testing
Instead of continuing to maintain this in bash, we should convert it to gradle as a partial implementation of https://github.com/GoogleCloudPlatform/google-cloud-java/issues/1587 . We should figure out a way to invoke gradle from maven for this so we don't have to convert everything in one go.
I think we should use gax-java as a model for version updating. This would mean using a template for the quickstart examples (e.g. showing the user how to depend on the package from maven/gradle).
|
1.0
|
Convert version update process to gradle - The current version update script is in bash and is hard to maintain. Currently it has a few bugs:
- it changes the version of the grpc dependency
- it doesn't update the versions of core-http or core-grpc
- it doesn't update the versions of modules under google-cloud-testing
Instead of continuing to maintain this in bash, we should convert it to gradle as a partial implementation of https://github.com/GoogleCloudPlatform/google-cloud-java/issues/1587 . We should figure out a way to invoke gradle from maven for this so we don't have to convert everything in one go.
I think we should use gax-java as a model for version updating. This would mean using a template for the quickstart examples (e.g. showing the user how to depend on the package from maven/gradle).
|
process
|
convert version update process to gradle the current version update script is in bash and is hard to maintain currently it has a few bugs it changes the version of the grpc dependency it doesn t update the versions of core http or core grpc it doesn t update the versions of modules under google cloud testing instead of continuing to maintain this in bash we should convert it to gradle as a partial implementation of we should figure out a way to invoke gradle from maven for this so we don t have to convert everything in one go i think we should use gax java as a model for version updating this would mean using a template for the quickstart examples e g showing the user how to depend on the package from maven gradle
| 1
|
247,344
| 26,698,402,391
|
IssuesEvent
|
2023-01-27 12:24:54
|
GrimeyPickle/pytorch
|
https://api.github.com/repos/GrimeyPickle/pytorch
|
opened
|
bootstrap-3.3.6.min.js: 6 vulnerabilities (highest severity is: 6.1)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (bootstrap version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2019-8331](https://www.mend.io/vulnerability-database/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1 | ❌ |
| [CVE-2018-14040](https://www.mend.io/vulnerability-database/CVE-2018-14040) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0 | ❌ |
| [CVE-2018-20677](https://www.mend.io/vulnerability-database/CVE-2018-20677) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0 | ❌ |
| [CVE-2018-20676](https://www.mend.io/vulnerability-database/CVE-2018-20676) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | bootstrap - 3.4.0 | ❌ |
| [CVE-2018-14042](https://www.mend.io/vulnerability-database/CVE-2018-14042) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0 | ❌ |
| [CVE-2016-10735](https://www.mend.io/vulnerability-database/CVE-2016-10735) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | bootstrap - 3.4.0, 4.0.0-beta.2 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-8331</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14040</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-14040>CVE-2018-14040</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20677</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-20677>CVE-2018-20677</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20676</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14042</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-14042>CVE-2018-14042</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2016-10735</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-10735>CVE-2016-10735</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0, 4.0.0-beta.2</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
bootstrap-3.3.6.min.js: 6 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (bootstrap version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2019-8331](https://www.mend.io/vulnerability-database/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1 | ❌ |
| [CVE-2018-14040](https://www.mend.io/vulnerability-database/CVE-2018-14040) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0 | ❌ |
| [CVE-2018-20677](https://www.mend.io/vulnerability-database/CVE-2018-20677) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0 | ❌ |
| [CVE-2018-20676](https://www.mend.io/vulnerability-database/CVE-2018-20676) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | bootstrap - 3.4.0 | ❌ |
| [CVE-2018-14042](https://www.mend.io/vulnerability-database/CVE-2018-14042) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0 | ❌ |
| [CVE-2016-10735](https://www.mend.io/vulnerability-database/CVE-2016-10735) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.6.min.js | Direct | bootstrap - 3.4.0, 4.0.0-beta.2 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-8331</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14040</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-14040>CVE-2018-14040</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20677</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-20677>CVE-2018-20677</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20676</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14042</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-14042>CVE-2018-14042</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2016-10735</summary>
### Vulnerable Library - <b>bootstrap-3.3.6.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /third_party/NNPACK/web/nnpack.html</p>
<p>Path to vulnerable library: /third_party/NNPACK/web/nnpack.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GrimeyPickle/pytorch/commit/0f7d8523785b8ad829637ff5b0c48deea2452771">0f7d8523785b8ad829637ff5b0c48deea2452771</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-10735>CVE-2016-10735</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0, 4.0.0-beta.2</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_process
|
bootstrap min js vulnerabilities highest severity is vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file third party nnpack web nnpack html path to vulnerable library third party nnpack web nnpack html found in head commit a href vulnerabilities cve severity cvss dependency type fixed in bootstrap version remediation available medium bootstrap min js direct bootstrap bootstrap sass medium bootstrap min js direct org webjars npm bootstrap org webjars bootstrap medium bootstrap min js direct bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter medium bootstrap min js direct bootstrap medium bootstrap min js direct org webjars npm bootstrap org webjars bootstrap medium bootstrap min js direct bootstrap beta details cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file third party nnpack web nnpack html path to vulnerable library third party nnpack web nnpack html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution bootstrap bootstrap sass step up your open source security game with mend cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file third party nnpack web nnpack html path to vulnerable library third party nnpack web nnpack html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the collapse data parent attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution org webjars npm bootstrap org webjars bootstrap step up your open source security game with mend cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file third party nnpack web nnpack html path to vulnerable library third party nnpack web nnpack html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the affix configuration target property publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter step up your open source security game with mend cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file third party nnpack web nnpack html path to vulnerable library third party nnpack web nnpack html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap step up your open source security game with mend cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file third party nnpack web nnpack html path to vulnerable library third party nnpack web nnpack html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the data container property of tooltip publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution org webjars npm bootstrap org webjars bootstrap step up your open source security game with mend cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file third party nnpack web nnpack html path to vulnerable library third party nnpack web nnpack html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap beta step up your open source security game with mend
| 0
|
8,749
| 11,873,002,633
|
IssuesEvent
|
2020-03-26 16:38:19
|
sjbarag/brs
|
https://api.github.com/repos/sjbarag/brs
|
opened
|
Negation in conditional compilations is not supported
|
bug preprocessor
|
It appears syntax like `#if NOT ENABLE_FEATURE` is valid during conditional compilation in the reference BrightScript implementation, but not in `brs`. Fixing that will require the conditional compilation processor to get a bit more intelligent, but hopefully it's not a major change.
Originally reported by @triwav !
|
1.0
|
Negation in conditional compilations is not supported - It appears syntax like `#if NOT ENABLE_FEATURE` is valid during conditional compilation in the reference BrightScript implementation, but not in `brs`. Fixing that will require the conditional compilation processor to get a bit more intelligent, but hopefully it's not a major change.
Originally reported by @triwav !
|
process
|
negation in conditional compilations is not supported it appears syntax like if not enable feature is valid during conditional compilation in the reference brightscript implementation but not in brs fixing that will require the conditional compilation processor to get a bit more intelligent but hopefully it s not a major change originally reported by triwav
| 1
|
497,055
| 14,361,392,581
|
IssuesEvent
|
2020-11-30 18:11:53
|
dotCMS/core
|
https://api.github.com/repos/dotCMS/core
|
opened
|
Push publishing a page fails due to invalid urlMap detail page
|
Severity : Support Priority Severity : Support Requested Type : Bug
|
This is easy to reproduce.
1. Clean starter
2. Set up a PP endpoint (does not need to work)
3. Try to push publish the `/destinations/costa-rica` page.
You get an error b/c the structure of a content object on the page points to a non- existent URLMap detail page.
```
[30/11/20 18:08:44:922 GMT] INFO util.DependencyManager: PublisherFilter { excludeClassesSet = [] , excludeDependencyClassesSet = [] , excludeQueryIds = [] , excludeDependencyQueryIds = [] , relationships = true , dependencies = true}
[30/11/20 18:08:44:939 GMT] INFO caffine.CaffineCache: *** Building Cache : pushedassetscache, size:5000,Concurrency:32
[30/11/20 18:08:45:791 GMT] ERROR publishing.PublisherAPIImpl: null
com.dotcms.publishing.DotBundleException: null
at com.dotcms.enterprise.publishing.remote.bundler.DependencyBundler.generate(SourceFile:71) ~[ee-20.11.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish_aroundBody0(PublisherAPIImpl.java:123) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl$AjcClosure1.run(PublisherAPIImpl.java:1) ~[dotcms_20.11_999999.jar:?]
at org.aspectj.runtime.reflect.JoinPointImpl.proceed(JoinPointImpl.java:149) ~[aspectjrt-1.8.10.jar:?]
at com.dotcms.aspects.aspectj.AspectJDelegateMethodInvocation.proceed(AspectJDelegateMethodInvocation.java:42) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.interceptors.CloseDBIfOpenedMethodInterceptor.invoke(CloseDBIfOpenedMethodInterceptor.java:29) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.aspectj.CloseDBIfOpenedAspect.invoke(CloseDBIfOpenedAspect.java:41) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish(PublisherAPIImpl.java:43) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish(PublisherAPIImpl.java:38) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob.execute_aroundBody0(PublisherQueueJob.java:186) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob$AjcClosure1.run(PublisherQueueJob.java:1) ~[dotcms_20.11_999999.jar:?]
at org.aspectj.runtime.reflect.JoinPointImpl.proceed(JoinPointImpl.java:149) ~[aspectjrt-1.8.10.jar:?]
at com.dotcms.aspects.aspectj.AspectJDelegateMethodInvocation.proceed(AspectJDelegateMethodInvocation.java:42) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.interceptors.CloseDBIfOpenedMethodInterceptor.invoke(CloseDBIfOpenedMethodInterceptor.java:29) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.aspectj.CloseDBIfOpenedAspect.invoke(CloseDBIfOpenedAspect.java:41) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob.execute(PublisherQueueJob.java:112) ~[dotcms_20.11_999999.jar:?]
at org.quartz.core.JobRunShell.run(JobRunShell.java:223) ~[dot.quartz-all-1.8.6_2.jar:?]
at org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:549) ~[dot.quartz-all-1.8.6_2.jar:?]
Caused by: java.lang.NullPointerException
at com.dotcms.publisher.util.DependencyManager.setHTMLPagesDependencies(DependencyManager.java:692) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.structureDependencyHelper(DependencyManager.java:1165) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.setStructureDependencies(DependencyManager.java:1070) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.setDependencies(DependencyManager.java:343) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.enterprise.publishing.remote.bundler.DependencyBundler.generate(SourceFile:68) ~[ee-20.11.jar:?]
... 17 more
[30/11/20 18:08:45:791 GMT] ERROR business.PublisherQueueJob: Unable to publish Bundle '01ERD5FBSCF1KX0NC03QAB6V6T': null
com.dotcms.publishing.DotPublishingException: null
at com.dotcms.publishing.PublisherAPIImpl.publish_aroundBody0(PublisherAPIImpl.java:151) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl$AjcClosure1.run(PublisherAPIImpl.java:1) ~[dotcms_20.11_999999.jar:?]
at org.aspectj.runtime.reflect.JoinPointImpl.proceed(JoinPointImpl.java:149) ~[aspectjrt-1.8.10.jar:?]
at com.dotcms.aspects.aspectj.AspectJDelegateMethodInvocation.proceed(AspectJDelegateMethodInvocation.java:42) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.interceptors.CloseDBIfOpenedMethodInterceptor.invoke(CloseDBIfOpenedMethodInterceptor.java:29) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.aspectj.CloseDBIfOpenedAspect.invoke(CloseDBIfOpenedAspect.java:41) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish(PublisherAPIImpl.java:43) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish(PublisherAPIImpl.java:38) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob.execute_aroundBody0(PublisherQueueJob.java:186) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob$AjcClosure1.run(PublisherQueueJob.java:1) ~[dotcms_20.11_999999.jar:?]
at org.aspectj.runtime.reflect.JoinPointImpl.proceed(JoinPointImpl.java:149) ~[aspectjrt-1.8.10.jar:?]
at com.dotcms.aspects.aspectj.AspectJDelegateMethodInvocation.proceed(AspectJDelegateMethodInvocation.java:42) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.interceptors.CloseDBIfOpenedMethodInterceptor.invoke(CloseDBIfOpenedMethodInterceptor.java:29) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.aspectj.CloseDBIfOpenedAspect.invoke(CloseDBIfOpenedAspect.java:41) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob.execute(PublisherQueueJob.java:112) ~[dotcms_20.11_999999.jar:?]
at org.quartz.core.JobRunShell.run(JobRunShell.java:223) ~[dot.quartz-all-1.8.6_2.jar:?]
at org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:549) ~[dot.quartz-all-1.8.6_2.jar:?]
Caused by: com.dotcms.publishing.DotBundleException
at com.dotcms.enterprise.publishing.remote.bundler.DependencyBundler.generate(SourceFile:71) ~[ee-20.11.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish_aroundBody0(PublisherAPIImpl.java:123) ~[dotcms_20.11_999999.jar:?]
... 16 more
Caused by: java.lang.NullPointerException
at com.dotcms.publisher.util.DependencyManager.setHTMLPagesDependencies(DependencyManager.java:692) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.structureDependencyHelper(DependencyManager.java:1165) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.setStructureDependencies(DependencyManager.java:1070) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.setDependencies(DependencyManager.java:343) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.enterprise.publishing.remote.bundler.DependencyBundler.generate(SourceFile:68) ~[ee-20.11.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish_aroundBody0(PublisherAPIImpl.java:123) ~[dotcms_20.11_999999.jar:?]
... 16 more
[30/11/20 18:08:45:791 GMT] INFO util.PushPublishLogger: class com.dotcms.publisher.business.PublisherQueueJob : Status Update: Failed to bundle '01ERD5FBSCF1KX0NC03QAB6V6T'
[30/11/20 18:08:47:983 GMT] WARN auth.PrincipalThreadLocal: getName null
[30/11/20 18:08:47:986 GMT] INFO util.SecurityLogger: class com.dotcms.cms.login.LoginServiceAPIFactory$LoginServiceImpl : User dotcms.org.1 has successfully login from IP: 178.149.167.17 -- ip:178.149.167.17,user:Admin User [ID: dotcms.org.1][email:admin@dotcms.com]
[30/11/20 18:09:03:652 GMT] INFO delegate.SystemEventsJobDelegate: The event: 57d2813c-8199-409a-abe3-bbd35390f86a, has been skipped on the server: 525d0962-b246-42cf-8b44-2694f84673ae
```
|
1.0
|
Push publishing a page fails due to invalid urlMap detail page - This is easy to reproduce.
1. Clean starter
2. Set up a PP endpoint (does not need to work)
3. Try to push publish the `/destinations/costa-rica` page.
You get an error b/c the structure of a content object on the page points to a non- existent URLMap detail page.
```
[30/11/20 18:08:44:922 GMT] INFO util.DependencyManager: PublisherFilter { excludeClassesSet = [] , excludeDependencyClassesSet = [] , excludeQueryIds = [] , excludeDependencyQueryIds = [] , relationships = true , dependencies = true}
[30/11/20 18:08:44:939 GMT] INFO caffine.CaffineCache: *** Building Cache : pushedassetscache, size:5000,Concurrency:32
[30/11/20 18:08:45:791 GMT] ERROR publishing.PublisherAPIImpl: null
com.dotcms.publishing.DotBundleException: null
at com.dotcms.enterprise.publishing.remote.bundler.DependencyBundler.generate(SourceFile:71) ~[ee-20.11.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish_aroundBody0(PublisherAPIImpl.java:123) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl$AjcClosure1.run(PublisherAPIImpl.java:1) ~[dotcms_20.11_999999.jar:?]
at org.aspectj.runtime.reflect.JoinPointImpl.proceed(JoinPointImpl.java:149) ~[aspectjrt-1.8.10.jar:?]
at com.dotcms.aspects.aspectj.AspectJDelegateMethodInvocation.proceed(AspectJDelegateMethodInvocation.java:42) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.interceptors.CloseDBIfOpenedMethodInterceptor.invoke(CloseDBIfOpenedMethodInterceptor.java:29) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.aspectj.CloseDBIfOpenedAspect.invoke(CloseDBIfOpenedAspect.java:41) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish(PublisherAPIImpl.java:43) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish(PublisherAPIImpl.java:38) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob.execute_aroundBody0(PublisherQueueJob.java:186) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob$AjcClosure1.run(PublisherQueueJob.java:1) ~[dotcms_20.11_999999.jar:?]
at org.aspectj.runtime.reflect.JoinPointImpl.proceed(JoinPointImpl.java:149) ~[aspectjrt-1.8.10.jar:?]
at com.dotcms.aspects.aspectj.AspectJDelegateMethodInvocation.proceed(AspectJDelegateMethodInvocation.java:42) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.interceptors.CloseDBIfOpenedMethodInterceptor.invoke(CloseDBIfOpenedMethodInterceptor.java:29) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.aspectj.CloseDBIfOpenedAspect.invoke(CloseDBIfOpenedAspect.java:41) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob.execute(PublisherQueueJob.java:112) ~[dotcms_20.11_999999.jar:?]
at org.quartz.core.JobRunShell.run(JobRunShell.java:223) ~[dot.quartz-all-1.8.6_2.jar:?]
at org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:549) ~[dot.quartz-all-1.8.6_2.jar:?]
Caused by: java.lang.NullPointerException
at com.dotcms.publisher.util.DependencyManager.setHTMLPagesDependencies(DependencyManager.java:692) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.structureDependencyHelper(DependencyManager.java:1165) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.setStructureDependencies(DependencyManager.java:1070) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.setDependencies(DependencyManager.java:343) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.enterprise.publishing.remote.bundler.DependencyBundler.generate(SourceFile:68) ~[ee-20.11.jar:?]
... 17 more
[30/11/20 18:08:45:791 GMT] ERROR business.PublisherQueueJob: Unable to publish Bundle '01ERD5FBSCF1KX0NC03QAB6V6T': null
com.dotcms.publishing.DotPublishingException: null
at com.dotcms.publishing.PublisherAPIImpl.publish_aroundBody0(PublisherAPIImpl.java:151) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl$AjcClosure1.run(PublisherAPIImpl.java:1) ~[dotcms_20.11_999999.jar:?]
at org.aspectj.runtime.reflect.JoinPointImpl.proceed(JoinPointImpl.java:149) ~[aspectjrt-1.8.10.jar:?]
at com.dotcms.aspects.aspectj.AspectJDelegateMethodInvocation.proceed(AspectJDelegateMethodInvocation.java:42) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.interceptors.CloseDBIfOpenedMethodInterceptor.invoke(CloseDBIfOpenedMethodInterceptor.java:29) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.aspectj.CloseDBIfOpenedAspect.invoke(CloseDBIfOpenedAspect.java:41) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish(PublisherAPIImpl.java:43) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish(PublisherAPIImpl.java:38) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob.execute_aroundBody0(PublisherQueueJob.java:186) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob$AjcClosure1.run(PublisherQueueJob.java:1) ~[dotcms_20.11_999999.jar:?]
at org.aspectj.runtime.reflect.JoinPointImpl.proceed(JoinPointImpl.java:149) ~[aspectjrt-1.8.10.jar:?]
at com.dotcms.aspects.aspectj.AspectJDelegateMethodInvocation.proceed(AspectJDelegateMethodInvocation.java:42) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.interceptors.CloseDBIfOpenedMethodInterceptor.invoke(CloseDBIfOpenedMethodInterceptor.java:29) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.aspects.aspectj.CloseDBIfOpenedAspect.invoke(CloseDBIfOpenedAspect.java:41) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.business.PublisherQueueJob.execute(PublisherQueueJob.java:112) ~[dotcms_20.11_999999.jar:?]
at org.quartz.core.JobRunShell.run(JobRunShell.java:223) ~[dot.quartz-all-1.8.6_2.jar:?]
at org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:549) ~[dot.quartz-all-1.8.6_2.jar:?]
Caused by: com.dotcms.publishing.DotBundleException
at com.dotcms.enterprise.publishing.remote.bundler.DependencyBundler.generate(SourceFile:71) ~[ee-20.11.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish_aroundBody0(PublisherAPIImpl.java:123) ~[dotcms_20.11_999999.jar:?]
... 16 more
Caused by: java.lang.NullPointerException
at com.dotcms.publisher.util.DependencyManager.setHTMLPagesDependencies(DependencyManager.java:692) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.structureDependencyHelper(DependencyManager.java:1165) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.setStructureDependencies(DependencyManager.java:1070) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.publisher.util.DependencyManager.setDependencies(DependencyManager.java:343) ~[dotcms_20.11_999999.jar:?]
at com.dotcms.enterprise.publishing.remote.bundler.DependencyBundler.generate(SourceFile:68) ~[ee-20.11.jar:?]
at com.dotcms.publishing.PublisherAPIImpl.publish_aroundBody0(PublisherAPIImpl.java:123) ~[dotcms_20.11_999999.jar:?]
... 16 more
[30/11/20 18:08:45:791 GMT] INFO util.PushPublishLogger: class com.dotcms.publisher.business.PublisherQueueJob : Status Update: Failed to bundle '01ERD5FBSCF1KX0NC03QAB6V6T'
[30/11/20 18:08:47:983 GMT] WARN auth.PrincipalThreadLocal: getName null
[30/11/20 18:08:47:986 GMT] INFO util.SecurityLogger: class com.dotcms.cms.login.LoginServiceAPIFactory$LoginServiceImpl : User dotcms.org.1 has successfully login from IP: 178.149.167.17 -- ip:178.149.167.17,user:Admin User [ID: dotcms.org.1][email:admin@dotcms.com]
[30/11/20 18:09:03:652 GMT] INFO delegate.SystemEventsJobDelegate: The event: 57d2813c-8199-409a-abe3-bbd35390f86a, has been skipped on the server: 525d0962-b246-42cf-8b44-2694f84673ae
```
|
non_process
|
push publishing a page fails due to invalid urlmap detail page this is easy to reproduce clean starter set up a pp endpoint does not need to work try to push publish the destinations costa rica page you get an error b c the structure of a content object on the page points to a non existent urlmap detail page info util dependencymanager publisherfilter excludeclassesset excludedependencyclassesset excludequeryids excludedependencyqueryids relationships true dependencies true info caffine caffinecache building cache pushedassetscache size concurrency error publishing publisherapiimpl null com dotcms publishing dotbundleexception null at com dotcms enterprise publishing remote bundler dependencybundler generate sourcefile at com dotcms publishing publisherapiimpl publish publisherapiimpl java at com dotcms publishing publisherapiimpl run publisherapiimpl java at org aspectj runtime reflect joinpointimpl proceed joinpointimpl java at com dotcms aspects aspectj aspectjdelegatemethodinvocation proceed aspectjdelegatemethodinvocation java at com dotcms aspects interceptors closedbifopenedmethodinterceptor invoke closedbifopenedmethodinterceptor java at com dotcms aspects aspectj closedbifopenedaspect invoke closedbifopenedaspect java at com dotcms publishing publisherapiimpl publish publisherapiimpl java at com dotcms publishing publisherapiimpl publish publisherapiimpl java at com dotcms publisher business publisherqueuejob execute publisherqueuejob java at com dotcms publisher business publisherqueuejob run publisherqueuejob java at org aspectj runtime reflect joinpointimpl proceed joinpointimpl java at com dotcms aspects aspectj aspectjdelegatemethodinvocation proceed aspectjdelegatemethodinvocation java at com dotcms aspects interceptors closedbifopenedmethodinterceptor invoke closedbifopenedmethodinterceptor java at com dotcms aspects aspectj closedbifopenedaspect invoke closedbifopenedaspect java at com dotcms publisher business publisherqueuejob execute publisherqueuejob java at org quartz core jobrunshell run jobrunshell java at org quartz simpl simplethreadpool workerthread run simplethreadpool java caused by java lang nullpointerexception at com dotcms publisher util dependencymanager sethtmlpagesdependencies dependencymanager java at com dotcms publisher util dependencymanager structuredependencyhelper dependencymanager java at com dotcms publisher util dependencymanager setstructuredependencies dependencymanager java at com dotcms publisher util dependencymanager setdependencies dependencymanager java at com dotcms enterprise publishing remote bundler dependencybundler generate sourcefile more error business publisherqueuejob unable to publish bundle null com dotcms publishing dotpublishingexception null at com dotcms publishing publisherapiimpl publish publisherapiimpl java at com dotcms publishing publisherapiimpl run publisherapiimpl java at org aspectj runtime reflect joinpointimpl proceed joinpointimpl java at com dotcms aspects aspectj aspectjdelegatemethodinvocation proceed aspectjdelegatemethodinvocation java at com dotcms aspects interceptors closedbifopenedmethodinterceptor invoke closedbifopenedmethodinterceptor java at com dotcms aspects aspectj closedbifopenedaspect invoke closedbifopenedaspect java at com dotcms publishing publisherapiimpl publish publisherapiimpl java at com dotcms publishing publisherapiimpl publish publisherapiimpl java at com dotcms publisher business publisherqueuejob execute publisherqueuejob java at com dotcms publisher business publisherqueuejob run publisherqueuejob java at org aspectj runtime reflect joinpointimpl proceed joinpointimpl java at com dotcms aspects aspectj aspectjdelegatemethodinvocation proceed aspectjdelegatemethodinvocation java at com dotcms aspects interceptors closedbifopenedmethodinterceptor invoke closedbifopenedmethodinterceptor java at com dotcms aspects aspectj closedbifopenedaspect invoke closedbifopenedaspect java at com dotcms publisher business publisherqueuejob execute publisherqueuejob java at org quartz core jobrunshell run jobrunshell java at org quartz simpl simplethreadpool workerthread run simplethreadpool java caused by com dotcms publishing dotbundleexception at com dotcms enterprise publishing remote bundler dependencybundler generate sourcefile at com dotcms publishing publisherapiimpl publish publisherapiimpl java more caused by java lang nullpointerexception at com dotcms publisher util dependencymanager sethtmlpagesdependencies dependencymanager java at com dotcms publisher util dependencymanager structuredependencyhelper dependencymanager java at com dotcms publisher util dependencymanager setstructuredependencies dependencymanager java at com dotcms publisher util dependencymanager setdependencies dependencymanager java at com dotcms enterprise publishing remote bundler dependencybundler generate sourcefile at com dotcms publishing publisherapiimpl publish publisherapiimpl java more info util pushpublishlogger class com dotcms publisher business publisherqueuejob status update failed to bundle warn auth principalthreadlocal getname null info util securitylogger class com dotcms cms login loginserviceapifactory loginserviceimpl user dotcms org has successfully login from ip ip user admin user info delegate systemeventsjobdelegate the event has been skipped on the server
| 0
|
16,650
| 21,714,945,637
|
IssuesEvent
|
2022-05-10 16:56:52
|
quark-engine/quark-rules
|
https://api.github.com/repos/quark-engine/quark-rules
|
closed
|
Release new rules for call log and contact list accessing
|
issue-processing-state-06
|
Add new quark rules (number 00200 - 00204) for detecting call log and contact list accessing.
|
1.0
|
Release new rules for call log and contact list accessing - Add new quark rules (number 00200 - 00204) for detecting call log and contact list accessing.
|
process
|
release new rules for call log and contact list accessing add new quark rules number for detecting call log and contact list accessing
| 1
|
459,668
| 13,196,802,839
|
IssuesEvent
|
2020-08-13 21:25:52
|
kubernetes/website
|
https://api.github.com/repos/kubernetes/website
|
closed
|
modify layout of codenew shortcode
|
area/web-development kind/cleanup priority/backlog
|
**This is a Feature Request**
The `codenew` shortcode uses an HTML table to contain the "copy" icon, filename text,
and code block. Since the Docsy theme addition, the codenew shortcode generates code blocks
that appear squashed on the page.
**What would you like to be added**
Investigate whether the `codenew` shortcode should keep the table element for laying out the code block with
updated table styling or create a new layout using HTML `div` elements and styling.
**Why is this needed**
The copy-code-blocks should look like the other code blocks in the site.
**Comments**
Could test with page, https://kubernetes.io/docs/tasks/administer-cluster/manage-resources/memory-constraint-namespace/
|
1.0
|
modify layout of codenew shortcode - **This is a Feature Request**
The `codenew` shortcode uses an HTML table to contain the "copy" icon, filename text,
and code block. Since the Docsy theme addition, the codenew shortcode generates code blocks
that appear squashed on the page.
**What would you like to be added**
Investigate whether the `codenew` shortcode should keep the table element for laying out the code block with
updated table styling or create a new layout using HTML `div` elements and styling.
**Why is this needed**
The copy-code-blocks should look like the other code blocks in the site.
**Comments**
Could test with page, https://kubernetes.io/docs/tasks/administer-cluster/manage-resources/memory-constraint-namespace/
|
non_process
|
modify layout of codenew shortcode this is a feature request the codenew shortcode uses an html table to contain the copy icon filename text and code block since the docsy theme addition the codenew shortcode generates code blocks that appear squashed on the page what would you like to be added investigate whether the codenew shortcode should keep the table element for laying out the code block with updated table styling or create a new layout using html div elements and styling why is this needed the copy code blocks should look like the other code blocks in the site comments could test with page
| 0
|
18,500
| 3,068,135,303
|
IssuesEvent
|
2015-08-18 14:22:12
|
theAsgar/mdb-sqlite
|
https://api.github.com/repos/theAsgar/mdb-sqlite
|
closed
|
Converting from .db sqlite file to a .mdb file
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. java -jar dist/mdb-sqlite.jar Stock.db StockO.mdb
2.
3.
What is the expected output? What do you see instead?
Caused by: java.io.IOException: Unsupported version: 0
at
com.healthmarketscience.jackcess.JetFormat.getFormat(JetFormat.java:143)
at com.healthmarketscience.jackcess.Database.<init>(Database.java:347)
at com.healthmarketscience.jackcess.Database.open(Database.java:294)
at com.healthmarketscience.jackcess.Database.open(Database.java:270)
at com.plausiblelabs.mdb.Main.main(Unknown Source)
... 6 more
What version of the product are you using? On what operating system?
1.0.2 MS XP
Please provide any additional information below.
```
Original issue reported on code.google.com by `st...@branchadventures.org` on 9 Mar 2010 at 6:07
|
1.0
|
Converting from .db sqlite file to a .mdb file - ```
What steps will reproduce the problem?
1. java -jar dist/mdb-sqlite.jar Stock.db StockO.mdb
2.
3.
What is the expected output? What do you see instead?
Caused by: java.io.IOException: Unsupported version: 0
at
com.healthmarketscience.jackcess.JetFormat.getFormat(JetFormat.java:143)
at com.healthmarketscience.jackcess.Database.<init>(Database.java:347)
at com.healthmarketscience.jackcess.Database.open(Database.java:294)
at com.healthmarketscience.jackcess.Database.open(Database.java:270)
at com.plausiblelabs.mdb.Main.main(Unknown Source)
... 6 more
What version of the product are you using? On what operating system?
1.0.2 MS XP
Please provide any additional information below.
```
Original issue reported on code.google.com by `st...@branchadventures.org` on 9 Mar 2010 at 6:07
|
non_process
|
converting from db sqlite file to a mdb file what steps will reproduce the problem java jar dist mdb sqlite jar stock db stocko mdb what is the expected output what do you see instead caused by java io ioexception unsupported version at com healthmarketscience jackcess jetformat getformat jetformat java at com healthmarketscience jackcess database database java at com healthmarketscience jackcess database open database java at com healthmarketscience jackcess database open database java at com plausiblelabs mdb main main unknown source more what version of the product are you using on what operating system ms xp please provide any additional information below original issue reported on code google com by st branchadventures org on mar at
| 0
|
164,251
| 13,938,798,859
|
IssuesEvent
|
2020-10-22 15:41:24
|
janpf/ctt
|
https://api.github.com/repos/janpf/ctt
|
closed
|
Do declining valid keys indicate too-late notification?
|
documentation question
|
I hope this question is appropriate for the project. I have read that keys are only valid for a given day. Looking at the stats, I notice increasing published keys, but from 9/23 an strongly decreasing number of valid keys. Does that mean keys have been delivered to late (aka "over-stress at health agencies/labs")? I think this discrepancy is important to understand the value of CoronaWarnApp.
Thx!
|
1.0
|
Do declining valid keys indicate too-late notification? - I hope this question is appropriate for the project. I have read that keys are only valid for a given day. Looking at the stats, I notice increasing published keys, but from 9/23 an strongly decreasing number of valid keys. Does that mean keys have been delivered to late (aka "over-stress at health agencies/labs")? I think this discrepancy is important to understand the value of CoronaWarnApp.
Thx!
|
non_process
|
do declining valid keys indicate too late notification i hope this question is appropriate for the project i have read that keys are only valid for a given day looking at the stats i notice increasing published keys but from an strongly decreasing number of valid keys does that mean keys have been delivered to late aka over stress at health agencies labs i think this discrepancy is important to understand the value of coronawarnapp thx
| 0
|
22,643
| 31,895,826,712
|
IssuesEvent
|
2023-09-18 01:31:53
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - bed
|
Term - change Class - GeologicalContext normative Task Group - Material Sample Process - complete
|
## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_bed
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): bed
* Term label (English, not normative): Bed
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context
* Definition of the term (normative): The full name of the lithostratigraphic bed from which the ~~cataloged item~~**dwc:MaterialEntity** was collected.
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative): Harlem coal
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
|
1.0
|
Change term - bed - ## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_bed
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): bed
* Term label (English, not normative): Bed
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context
* Definition of the term (normative): The full name of the lithostratigraphic bed from which the ~~cataloged item~~**dwc:MaterialEntity** was collected.
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative): Harlem coal
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
|
process
|
change term bed term change submitter efficacy justification why is this change necessary create consistency of terms for material in darwin core demand justification if the change is semantic in nature name at least two organizations that independently need this term which includes representatives of over organizations stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version no current term definition proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes bed term label english not normative bed organized in class e g occurrence event location taxon geological context definition of the term normative the full name of the lithostratigraphic bed from which the cataloged item dwc materialentity was collected usage comments recommendations regarding content etc not normative examples not normative harlem coal refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd
| 1
|
8,777
| 6,652,360,988
|
IssuesEvent
|
2017-09-29 00:51:41
|
mozilla/lightbeam
|
https://api.github.com/repos/mozilla/lightbeam
|
closed
|
Lightbeam seems to be triggered every few milliseconds, even when I have no content at all
|
Performance
|
I'm currently investigating [why Lightbeam triggers the Slow Add-on Alert](https://bugzilla.mozilla.org/show_bug.cgi?id=1188507) in Firefox Nightly. For some reason, Lightbeam seems to be triggered every few milliseconds (or perhaps seconds, it's hard to tell with my heavily-instrumented Nightly), with the following stack:
```
0 ToolbarIconColor.inferFromText() ["chrome://browser/content/browser.js":18578]
this = [object Object]
this.init = [function]
this.uninit = [function]
this.handleEvent = [function]
this.observe = [function]
this.inferFromText = [function]
this._initialized = true
1 ToolbarIconColor.handleEvent(event = [object Event]) ["chrome://browser/content/browser.js":18543]
this = [object Object]
this.init = [function]
this.uninit = [function]
this.handleEvent = [function]
this.observe = [function]
this.inferFromText = [function]
this._initialized = true
```
|
True
|
Lightbeam seems to be triggered every few milliseconds, even when I have no content at all - I'm currently investigating [why Lightbeam triggers the Slow Add-on Alert](https://bugzilla.mozilla.org/show_bug.cgi?id=1188507) in Firefox Nightly. For some reason, Lightbeam seems to be triggered every few milliseconds (or perhaps seconds, it's hard to tell with my heavily-instrumented Nightly), with the following stack:
```
0 ToolbarIconColor.inferFromText() ["chrome://browser/content/browser.js":18578]
this = [object Object]
this.init = [function]
this.uninit = [function]
this.handleEvent = [function]
this.observe = [function]
this.inferFromText = [function]
this._initialized = true
1 ToolbarIconColor.handleEvent(event = [object Event]) ["chrome://browser/content/browser.js":18543]
this = [object Object]
this.init = [function]
this.uninit = [function]
this.handleEvent = [function]
this.observe = [function]
this.inferFromText = [function]
this._initialized = true
```
|
non_process
|
lightbeam seems to be triggered every few milliseconds even when i have no content at all i m currently investigating in firefox nightly for some reason lightbeam seems to be triggered every few milliseconds or perhaps seconds it s hard to tell with my heavily instrumented nightly with the following stack toolbariconcolor inferfromtext this this init this uninit this handleevent this observe this inferfromtext this initialized true toolbariconcolor handleevent event this this init this uninit this handleevent this observe this inferfromtext this initialized true
| 0
|
232,082
| 18,843,913,199
|
IssuesEvent
|
2021-11-11 12:53:34
|
eclipse-openj9/openj9
|
https://api.github.com/repos/eclipse-openj9/openj9
|
opened
|
JDK8 [zOS S390] 80 Load_Level_2.abbs.5mins.Mode103.5 - j9mm.141 * ** ASSERTION FAILED ** at ./MarkingScheme.hpp:71: ((false)) & j9mm.107 * ** ASSERTION FAILED **
|
test failure
|
Failure link
------------
From an internal build `job_output.php?id=19746935` (`fyrec517`):
```
java version "1.8.0_311"
Java(TM) SE Runtime Environment (build 8.0.7.0 - pmz3180sr7-20211011_01(SR7))
IBM J9 VM (build 2.9, JRE 1.8.0 z/OS s390-31-Bit 20211111_16455 (JIT enabled, AOT enabled)
OpenJ9 - 58668d5
OMR - 3804658
IBM - 3c151c1)
JCL - 20210930_01 based on Oracle jdk8u311-b11
```
Optional info
-------------
Failure output (captured from console output)
---------------------------------------------
```
j> 03:31:14 20211111 03:31:13 Runtime State Reporter IMPORTANT: 3373 tests complete, 262 currently running
j> 03:31:16 Object 00000600 not in heap range [1C500000,3C500000)
j> 03:31:16 11:31:15.992 0x3d753500 j9mm.141 * ** ASSERTION FAILED ** at ./MarkingScheme.hpp:71: ((false))
j> 03:31:16 11:31:16.024 0x464beb00 j9mm.107 * ** ASSERTION FAILED ** at ../../../gc_glue_java/MarkingDelegate.hpp:121: ((false && ((UDATA)0x99669966 == clazz->eyecatcher)))
j> 03:31:16 JVMDUMP039I Processing dump event "traceassert", detail "" at 2021/11/11 03:31:15 - please wait.
j> 03:31:16 JVMDUMP039I Processing dump event "traceassert", detail "" at 2021/11/11 03:31:16 - please wait.
j> 03:31:16 JVMDUMP032I JVM requested System dump using 'J9BUILD.JVM.TDUMP.J9BUILD6.D211111.T033115' in response to an event
```
fyi @dmitripivkine
|
1.0
|
JDK8 [zOS S390] 80 Load_Level_2.abbs.5mins.Mode103.5 - j9mm.141 * ** ASSERTION FAILED ** at ./MarkingScheme.hpp:71: ((false)) & j9mm.107 * ** ASSERTION FAILED ** - Failure link
------------
From an internal build `job_output.php?id=19746935` (`fyrec517`):
```
java version "1.8.0_311"
Java(TM) SE Runtime Environment (build 8.0.7.0 - pmz3180sr7-20211011_01(SR7))
IBM J9 VM (build 2.9, JRE 1.8.0 z/OS s390-31-Bit 20211111_16455 (JIT enabled, AOT enabled)
OpenJ9 - 58668d5
OMR - 3804658
IBM - 3c151c1)
JCL - 20210930_01 based on Oracle jdk8u311-b11
```
Optional info
-------------
Failure output (captured from console output)
---------------------------------------------
```
j> 03:31:14 20211111 03:31:13 Runtime State Reporter IMPORTANT: 3373 tests complete, 262 currently running
j> 03:31:16 Object 00000600 not in heap range [1C500000,3C500000)
j> 03:31:16 11:31:15.992 0x3d753500 j9mm.141 * ** ASSERTION FAILED ** at ./MarkingScheme.hpp:71: ((false))
j> 03:31:16 11:31:16.024 0x464beb00 j9mm.107 * ** ASSERTION FAILED ** at ../../../gc_glue_java/MarkingDelegate.hpp:121: ((false && ((UDATA)0x99669966 == clazz->eyecatcher)))
j> 03:31:16 JVMDUMP039I Processing dump event "traceassert", detail "" at 2021/11/11 03:31:15 - please wait.
j> 03:31:16 JVMDUMP039I Processing dump event "traceassert", detail "" at 2021/11/11 03:31:16 - please wait.
j> 03:31:16 JVMDUMP032I JVM requested System dump using 'J9BUILD.JVM.TDUMP.J9BUILD6.D211111.T033115' in response to an event
```
fyi @dmitripivkine
|
non_process
|
load level abbs assertion failed at markingscheme hpp false assertion failed failure link from an internal build job output php id java version java tm se runtime environment build ibm vm build jre z os bit jit enabled aot enabled omr ibm jcl based on oracle optional info failure output captured from console output j runtime state reporter important tests complete currently running j object not in heap range j assertion failed at markingscheme hpp false j assertion failed at gc glue java markingdelegate hpp false udata clazz eyecatcher j processing dump event traceassert detail at please wait j processing dump event traceassert detail at please wait j jvm requested system dump using jvm tdump in response to an event fyi dmitripivkine
| 0
|
40,897
| 5,282,170,718
|
IssuesEvent
|
2017-02-07 18:10:53
|
mozilla/network
|
https://api.github.com/repos/mozilla/network
|
opened
|
Analytics: identify events and goals
|
design engineering
|
A primary goal of the site is to increase membership of people with leadership potential. Let's plan our analytics to help us track efforts toward that.
First step: identify key events to track
Second step: gather advice from @vojtechmzl
|
1.0
|
Analytics: identify events and goals - A primary goal of the site is to increase membership of people with leadership potential. Let's plan our analytics to help us track efforts toward that.
First step: identify key events to track
Second step: gather advice from @vojtechmzl
|
non_process
|
analytics identify events and goals a primary goal of the site is to increase membership of people with leadership potential let s plan our analytics to help us track efforts toward that first step identify key events to track second step gather advice from vojtechmzl
| 0
|
16,732
| 21,893,853,329
|
IssuesEvent
|
2022-05-20 06:29:11
|
googleapis/google-cloud-dotnet
|
https://api.github.com/repos/googleapis/google-cloud-dotnet
|
closed
|
Remove redundant package references
|
type: process
|
Directory.Build.targets contains:
```xml
<PackageReference Include="Microsoft.SourceLink.GitHub" Version="1.1.1" PrivateAssets="All"/>
<PackageReference Include="Microsoft.DotNet.Analyzers.Compatibility" Version="0.2.12-alpha" PrivateAssets="All" />
<PackageReference Include="Microsoft.NETFramework.ReferenceAssemblies" Version="1.0.2" PrivateAssets="All" />
```
... but we still have the latter two references in our project files. We should make sure that each package reference is only in one place for any build - and ideally put the ConfigureAwaitChecker in Directory.Build.targets too
|
1.0
|
Remove redundant package references - Directory.Build.targets contains:
```xml
<PackageReference Include="Microsoft.SourceLink.GitHub" Version="1.1.1" PrivateAssets="All"/>
<PackageReference Include="Microsoft.DotNet.Analyzers.Compatibility" Version="0.2.12-alpha" PrivateAssets="All" />
<PackageReference Include="Microsoft.NETFramework.ReferenceAssemblies" Version="1.0.2" PrivateAssets="All" />
```
... but we still have the latter two references in our project files. We should make sure that each package reference is only in one place for any build - and ideally put the ConfigureAwaitChecker in Directory.Build.targets too
|
process
|
remove redundant package references directory build targets contains xml but we still have the latter two references in our project files we should make sure that each package reference is only in one place for any build and ideally put the configureawaitchecker in directory build targets too
| 1
|
212,936
| 23,963,109,006
|
IssuesEvent
|
2022-09-12 21:07:26
|
ManageIQ/manageiq-pods
|
https://api.github.com/repos/ManageIQ/manageiq-pods
|
closed
|
CVE-2018-17846 (High) detected in github.com/openshift/api-v3.9.0 - autoclosed
|
security vulnerability
|
## CVE-2018-17846 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/openshift/api-v3.9.0</b></p></summary>
<p>Canonical location of the OpenShift API definition.</p>
<p>
Dependency Hierarchy:
- :x: **github.com/openshift/api-v3.9.0** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ManageIQ/manageiq-pods/commit/85570f60339b4525836567309a9d6161d910b449">85570f60339b4525836567309a9d6161d910b449</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The html package (aka x/net/html) through 2018-09-25 in Go mishandles <table><math><select><mi><select></table>, leading to an infinite loop during an html.Parse call because inSelectIM and inSelectInTableIM do not comply with a specification.
<p>Publish Date: 2018-10-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-17846>CVE-2018-17846</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-17846">https://nvd.nist.gov/vuln/detail/CVE-2018-17846</a></p>
<p>Release Date: 2018-10-01</p>
<p>Fix Resolution: golang-golang-x-net-dev - 1:0.0+git20181201.351d144+dfsg-3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-17846 (High) detected in github.com/openshift/api-v3.9.0 - autoclosed - ## CVE-2018-17846 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/openshift/api-v3.9.0</b></p></summary>
<p>Canonical location of the OpenShift API definition.</p>
<p>
Dependency Hierarchy:
- :x: **github.com/openshift/api-v3.9.0** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ManageIQ/manageiq-pods/commit/85570f60339b4525836567309a9d6161d910b449">85570f60339b4525836567309a9d6161d910b449</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The html package (aka x/net/html) through 2018-09-25 in Go mishandles <table><math><select><mi><select></table>, leading to an infinite loop during an html.Parse call because inSelectIM and inSelectInTableIM do not comply with a specification.
<p>Publish Date: 2018-10-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-17846>CVE-2018-17846</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-17846">https://nvd.nist.gov/vuln/detail/CVE-2018-17846</a></p>
<p>Release Date: 2018-10-01</p>
<p>Fix Resolution: golang-golang-x-net-dev - 1:0.0+git20181201.351d144+dfsg-3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in github com openshift api autoclosed cve high severity vulnerability vulnerable library github com openshift api canonical location of the openshift api definition dependency hierarchy x github com openshift api vulnerable library found in head commit a href found in base branch master vulnerability details the html package aka x net html through in go mishandles leading to an infinite loop during an html parse call because inselectim and inselectintableim do not comply with a specification publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution golang golang x net dev dfsg step up your open source security game with mend
| 0
|
19,917
| 26,378,775,645
|
IssuesEvent
|
2023-01-12 06:28:49
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
env: python: No such file or directory
|
more data needed type: support / not a bug (process) team-OSS
|
### Description of the bug:
Hello all, I'm using macbook pro m1 macos 12.
I compiling tensorflow lite for ios but got this error.
Please help me, I really appreciat every your support!
Thanks you so much!
env: python: No such file or directory
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
_No response_
### Which operating system are you running Bazel on?
_No response_
### What is the output of `bazel info release`?
_No response_
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
_No response_
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
_No response_
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?
_No response_
|
1.0
|
env: python: No such file or directory - ### Description of the bug:
Hello all, I'm using macbook pro m1 macos 12.
I compiling tensorflow lite for ios but got this error.
Please help me, I really appreciat every your support!
Thanks you so much!
env: python: No such file or directory
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
_No response_
### Which operating system are you running Bazel on?
_No response_
### What is the output of `bazel info release`?
_No response_
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
_No response_
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
_No response_
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?
_No response_
|
process
|
env python no such file or directory description of the bug hello all i m using macbook pro macos i compiling tensorflow lite for ios but got this error please help me i really appreciat every your support thanks you so much env python no such file or directory what s the simplest easiest way to reproduce this bug please provide a minimal example if possible no response which operating system are you running bazel on no response what is the output of bazel info release no response if bazel info release returns development version or non git tell us how you built bazel no response what s the output of git remote get url origin git rev parse master git rev parse head no response have you found anything relevant by searching the web no response any other information logs or outputs that you want to share no response
| 1
|
333,117
| 29,508,928,051
|
IssuesEvent
|
2023-06-03 17:17:08
|
JuliaPluto/Malt.jl
|
https://api.github.com/repos/JuliaPluto/Malt.jl
|
closed
|
Test for nesting
|
test
|
We already know that this will work, but it is good to add a test:
Malt.jl should be usable inside a Malt.jl worker without any change in functionality.
|
1.0
|
Test for nesting - We already know that this will work, but it is good to add a test:
Malt.jl should be usable inside a Malt.jl worker without any change in functionality.
|
non_process
|
test for nesting we already know that this will work but it is good to add a test malt jl should be usable inside a malt jl worker without any change in functionality
| 0
|
268,581
| 20,356,906,974
|
IssuesEvent
|
2022-02-20 04:16:50
|
apache/buildstream
|
https://api.github.com/repos/apache/buildstream
|
closed
|
Document changes of the `format-version` in a central place
|
documentation
|
[See original issue on GitLab](https://gitlab.com/BuildStream/buildstream/-/issues/382)
In GitLab by [[Gitlab user @jjardon]](https://gitlab.com/jjardon) on Apr 30, 2018, 16:36
We currently have:
- When a new feature is added, a new string ``The <feature> configuration is available since format-version X```
- Buildstream projects need to specify the format-version in their projects
Still it would be good to have a central place to document what new features are being added in every bump, maybe a good location would be http://buildstream.gitlab.io/buildstream/projectconf.html#format-version ?
|
1.0
|
Document changes of the `format-version` in a central place - [See original issue on GitLab](https://gitlab.com/BuildStream/buildstream/-/issues/382)
In GitLab by [[Gitlab user @jjardon]](https://gitlab.com/jjardon) on Apr 30, 2018, 16:36
We currently have:
- When a new feature is added, a new string ``The <feature> configuration is available since format-version X```
- Buildstream projects need to specify the format-version in their projects
Still it would be good to have a central place to document what new features are being added in every bump, maybe a good location would be http://buildstream.gitlab.io/buildstream/projectconf.html#format-version ?
|
non_process
|
document changes of the format version in a central place in gitlab by on apr we currently have when a new feature is added a new string the configuration is available since format version x buildstream projects need to specify the format version in their projects still it would be good to have a central place to document what new features are being added in every bump maybe a good location would be
| 0
|
22,449
| 31,168,837,103
|
IssuesEvent
|
2023-08-16 22:21:37
|
NCAR/ucomp-pipeline
|
https://api.github.com/repos/NCAR/ucomp-pipeline
|
opened
|
Reprocess with version 0.5.x
|
process
|
Reprocess the mission.
- What date should I start with?
- Only quicklooks still?
- Reprocess 2022 first.
- [ ] clear dates before start date
- [ ] reprocess
|
1.0
|
Reprocess with version 0.5.x - Reprocess the mission.
- What date should I start with?
- Only quicklooks still?
- Reprocess 2022 first.
- [ ] clear dates before start date
- [ ] reprocess
|
process
|
reprocess with version x reprocess the mission what date should i start with only quicklooks still reprocess first clear dates before start date reprocess
| 1
|
8,711
| 11,851,337,826
|
IssuesEvent
|
2020-03-24 17:57:58
|
googleapis/python-pubsub
|
https://api.github.com/repos/googleapis/python-pubsub
|
closed
|
chore: group RBAC system tests into their own suite
|
api: pubsub testing type: process
|
Currently all RBAC system tests are placed under the `TestStreamingPull` class, even if they don't use streaming pull at all. These tests should be moved to their own class.
|
1.0
|
chore: group RBAC system tests into their own suite - Currently all RBAC system tests are placed under the `TestStreamingPull` class, even if they don't use streaming pull at all. These tests should be moved to their own class.
|
process
|
chore group rbac system tests into their own suite currently all rbac system tests are placed under the teststreamingpull class even if they don t use streaming pull at all these tests should be moved to their own class
| 1
|
256,536
| 22,060,088,557
|
IssuesEvent
|
2022-05-30 16:34:44
|
ukaea/Indica
|
https://api.github.com/repos/ukaea/Indica
|
opened
|
Improve test performance
|
enhancement testing
|
The tests take about 45 minutes on the GitHub runners currently. We could try to bring this down by:
- profiling and optimising.
- parallelising using `pytest-xdist`.
For parallelisation we have to be careful that the tests are independent, i.e. have no global internal or external mutable state (including files from e.g. caching, etc).
|
1.0
|
Improve test performance - The tests take about 45 minutes on the GitHub runners currently. We could try to bring this down by:
- profiling and optimising.
- parallelising using `pytest-xdist`.
For parallelisation we have to be careful that the tests are independent, i.e. have no global internal or external mutable state (including files from e.g. caching, etc).
|
non_process
|
improve test performance the tests take about minutes on the github runners currently we could try to bring this down by profiling and optimising parallelising using pytest xdist for parallelisation we have to be careful that the tests are independent i e have no global internal or external mutable state including files from e g caching etc
| 0
|
992
| 3,460,648,518
|
IssuesEvent
|
2015-12-19 10:23:43
|
rg3/youtube-dl
|
https://api.github.com/repos/rg3/youtube-dl
|
closed
|
Subtitles not embedded as expected
|
postprocessors subtitles
|
youtube-dl -c -f hls-2364 --all-subs --embed-subs https://tv.nrk.no/serie/tara-og-vegard/MYNT16000115/sesong-1/episode-1
Subtitles are not embedded into the `.mp4` file as expected. The following vague error message is given, though no obvious errors can be seen when examining the `.ttml` file.
ERROR: file:Bestegirlza-MYNT16000115.no.ttml: Invalid data found when processing input
|
1.0
|
Subtitles not embedded as expected - youtube-dl -c -f hls-2364 --all-subs --embed-subs https://tv.nrk.no/serie/tara-og-vegard/MYNT16000115/sesong-1/episode-1
Subtitles are not embedded into the `.mp4` file as expected. The following vague error message is given, though no obvious errors can be seen when examining the `.ttml` file.
ERROR: file:Bestegirlza-MYNT16000115.no.ttml: Invalid data found when processing input
|
process
|
subtitles not embedded as expected youtube dl c f hls all subs embed subs subtitles are not embedded into the file as expected the following vague error message is given though no obvious errors can be seen when examining the ttml file error file bestegirlza no ttml invalid data found when processing input
| 1
|
18,313
| 24,428,958,120
|
IssuesEvent
|
2022-10-06 06:32:29
|
maticnetwork/miden
|
https://api.github.com/repos/maticnetwork/miden
|
closed
|
Improvements to simplify exponentiation
|
assembly instruction set processor v0.3
|
To be able to verify FRI proofs efficiently within the VM, we need perform some number of exponentiations in the field. Currently, this will be relatively expensive - i.e., probably a couple of hundred cycles, if not more. So, it would be nice if the assembly provided something like an `exp` instructions. This instruction would work as follows:
```
[b, a, ... ] -> [a^b, ... ]
```
This would be exponentiation in the field. Specifically, `a` would be treated as a field element, while `b` would be treated like an integer in the range [0, 2^64 - 2^32 + 1).
Implementing this instruction via a native VM operation would probably add too much complexity, but we could provide an operation which would simplify this task greatly. This operation would be similar to the `BINACC` operation we discussed in #200 and even started to implement in #308. However, it would be more general and would support exponentiation for any base and exponent. We can call this operation `EXPACC`. We can use this operation then to make `pow2` instructions more efficient, or we may even decide to get rid of `pow2` instructions altogether if using `exp` instruction can be used relatively easily instead.
To summarize, I think we should do the following:
* Add `EXPACC` operation to the VM.
* Use it to implement `exp` assembly instruction.
* Either rewrite `pow2` instructions to use `EXPACC` or remove them and replace their usage with `exp`.
|
1.0
|
Improvements to simplify exponentiation - To be able to verify FRI proofs efficiently within the VM, we need perform some number of exponentiations in the field. Currently, this will be relatively expensive - i.e., probably a couple of hundred cycles, if not more. So, it would be nice if the assembly provided something like an `exp` instructions. This instruction would work as follows:
```
[b, a, ... ] -> [a^b, ... ]
```
This would be exponentiation in the field. Specifically, `a` would be treated as a field element, while `b` would be treated like an integer in the range [0, 2^64 - 2^32 + 1).
Implementing this instruction via a native VM operation would probably add too much complexity, but we could provide an operation which would simplify this task greatly. This operation would be similar to the `BINACC` operation we discussed in #200 and even started to implement in #308. However, it would be more general and would support exponentiation for any base and exponent. We can call this operation `EXPACC`. We can use this operation then to make `pow2` instructions more efficient, or we may even decide to get rid of `pow2` instructions altogether if using `exp` instruction can be used relatively easily instead.
To summarize, I think we should do the following:
* Add `EXPACC` operation to the VM.
* Use it to implement `exp` assembly instruction.
* Either rewrite `pow2` instructions to use `EXPACC` or remove them and replace their usage with `exp`.
|
process
|
improvements to simplify exponentiation to be able to verify fri proofs efficiently within the vm we need perform some number of exponentiations in the field currently this will be relatively expensive i e probably a couple of hundred cycles if not more so it would be nice if the assembly provided something like an exp instructions this instruction would work as follows this would be exponentiation in the field specifically a would be treated as a field element while b would be treated like an integer in the range implementing this instruction via a native vm operation would probably add too much complexity but we could provide an operation which would simplify this task greatly this operation would be similar to the binacc operation we discussed in and even started to implement in however it would be more general and would support exponentiation for any base and exponent we can call this operation expacc we can use this operation then to make instructions more efficient or we may even decide to get rid of instructions altogether if using exp instruction can be used relatively easily instead to summarize i think we should do the following add expacc operation to the vm use it to implement exp assembly instruction either rewrite instructions to use expacc or remove them and replace their usage with exp
| 1
|
22,022
| 30,533,920,594
|
IssuesEvent
|
2023-07-19 15:56:16
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
reopened
|
Edits to 'Modification of host cellular component' branch
|
ready multi-species process
|
Usually viruses/symbionts *dirsupt*, not only modify host cellular components. Here are all the 'annotations to modification' terms:
- [x] GO:0098865 modification by symbiont of host bicellular tight junctions -> 1 paper, disruption: change label
----
- [x] GO:0098864 modification by symbiont of host tight cell-cell junction
** PMID:24287273 disruption
** PMID:25838979 CACAO: "Our findings have uncovered a novel role of the bacterial protein AvrA in suppressing the inflammatory response of the host through JNK-regulated blockage of epithelial cell barrier function. " -> ok leave as 'modification by symbiont of host tight cell-cell junction'
----
- [x] GO:0052111 modification by symbiont of host structure-> used to mean 'induction of host (phagosome) membrane tubulation' annotated by Antonia [PMID:18996344](http://amigo.geneontology.org/amigo/reference/PMID:18996344) -> changed to modification by symbiont of host cell membrane
- [ ] AI Create new term modulation of endomembrane organization in another organism
----
- [x] GO:0052025 modification by symbiont of host cell membrane -> ok no more annotations
* * PMID:16085396 MTB -> inhibits macrophage phagocytosis -> changed to "GO:0052067 antiphagocytosis"
----
* GO:0052043 modification by symbiont of host cellular component -> 5 CACAO to 3 papers:
* * PMID:24304930 -> formation of replication-associated inclusion bodies
- [ ] NTR modification of host intracellular structure for assembly of symbiont replication compartment
- [ ] NTR viral factory assembly
- [ ] Check definition: GO:0046809 replication compartment - too specific (VZ term)
----
- [ ] PMID:10913092 -> Vibrio fischeri lux genes play an important role in colonization and development of the host light organ. -> probably need new terms
- [ ] NTR light-emitting organ development (for host)
is_a animal organ development
- [ ] NTR induction of host light-emitting organ development
- is_a modification of host intracellular structure for assembly of symbiont replication compartment
----
- [ ] PMID:24188660
-> missing MF 'toxin activity' (binary toxins consist of binding component and enzymatic component. The binding component recognizes a cell surface receptor and allows the internalization of the enzymatic component into the cytosol, and the enzymatic component catalyzes the reaction and induces the toxicity)
-> BP: killing of cells of another organism ("Binary toxin causes insect midgut CF-203 cells death via necrosis", so, not cytolysis)
-> "Confocal microscopy revealed that Plu1961/Plu1962 caused a notable decrease in cellular tubulin of CF-203 cells. " -> disruption by symbiont of host cytoskeleton (or, do we need 'host actin' and 'host microtubule' cytoskeleton?)
@dsiegele @genegodbold for a future discussion
|
1.0
|
Edits to 'Modification of host cellular component' branch - Usually viruses/symbionts *dirsupt*, not only modify host cellular components. Here are all the 'annotations to modification' terms:
- [x] GO:0098865 modification by symbiont of host bicellular tight junctions -> 1 paper, disruption: change label
----
- [x] GO:0098864 modification by symbiont of host tight cell-cell junction
** PMID:24287273 disruption
** PMID:25838979 CACAO: "Our findings have uncovered a novel role of the bacterial protein AvrA in suppressing the inflammatory response of the host through JNK-regulated blockage of epithelial cell barrier function. " -> ok leave as 'modification by symbiont of host tight cell-cell junction'
----
- [x] GO:0052111 modification by symbiont of host structure-> used to mean 'induction of host (phagosome) membrane tubulation' annotated by Antonia [PMID:18996344](http://amigo.geneontology.org/amigo/reference/PMID:18996344) -> changed to modification by symbiont of host cell membrane
- [ ] AI Create new term modulation of endomembrane organization in another organism
----
- [x] GO:0052025 modification by symbiont of host cell membrane -> ok no more annotations
* * PMID:16085396 MTB -> inhibits macrophage phagocytosis -> changed to "GO:0052067 antiphagocytosis"
----
* GO:0052043 modification by symbiont of host cellular component -> 5 CACAO to 3 papers:
* * PMID:24304930 -> formation of replication-associated inclusion bodies
- [ ] NTR modification of host intracellular structure for assembly of symbiont replication compartment
- [ ] NTR viral factory assembly
- [ ] Check definition: GO:0046809 replication compartment - too specific (VZ term)
----
- [ ] PMID:10913092 -> Vibrio fischeri lux genes play an important role in colonization and development of the host light organ. -> probably need new terms
- [ ] NTR light-emitting organ development (for host)
is_a animal organ development
- [ ] NTR induction of host light-emitting organ development
- is_a modification of host intracellular structure for assembly of symbiont replication compartment
----
- [ ] PMID:24188660
-> missing MF 'toxin activity' (binary toxins consist of binding component and enzymatic component. The binding component recognizes a cell surface receptor and allows the internalization of the enzymatic component into the cytosol, and the enzymatic component catalyzes the reaction and induces the toxicity)
-> BP: killing of cells of another organism ("Binary toxin causes insect midgut CF-203 cells death via necrosis", so, not cytolysis)
-> "Confocal microscopy revealed that Plu1961/Plu1962 caused a notable decrease in cellular tubulin of CF-203 cells. " -> disruption by symbiont of host cytoskeleton (or, do we need 'host actin' and 'host microtubule' cytoskeleton?)
@dsiegele @genegodbold for a future discussion
|
process
|
edits to modification of host cellular component branch usually viruses symbionts dirsupt not only modify host cellular components here are all the annotations to modification terms go modification by symbiont of host bicellular tight junctions paper disruption change label go modification by symbiont of host tight cell cell junction pmid disruption pmid cacao our findings have uncovered a novel role of the bacterial protein avra in suppressing the inflammatory response of the host through jnk regulated blockage of epithelial cell barrier function ok leave as modification by symbiont of host tight cell cell junction go modification by symbiont of host structure used to mean induction of host phagosome membrane tubulation annotated by antonia changed to modification by symbiont of host cell membrane ai create new term modulation of endomembrane organization in another organism go modification by symbiont of host cell membrane ok no more annotations pmid mtb inhibits macrophage phagocytosis changed to go antiphagocytosis go modification by symbiont of host cellular component cacao to papers pmid formation of replication associated inclusion bodies ntr modification of host intracellular structure for assembly of symbiont replication compartment ntr viral factory assembly check definition go replication compartment too specific vz term pmid vibrio fischeri lux genes play an important role in colonization and development of the host light organ probably need new terms ntr light emitting organ development for host is a animal organ development ntr induction of host light emitting organ development is a modification of host intracellular structure for assembly of symbiont replication compartment pmid missing mf toxin activity binary toxins consist of binding component and enzymatic component the binding component recognizes a cell surface receptor and allows the internalization of the enzymatic component into the cytosol and the enzymatic component catalyzes the reaction and induces the toxicity bp killing of cells of another organism binary toxin causes insect midgut cf cells death via necrosis so not cytolysis confocal microscopy revealed that caused a notable decrease in cellular tubulin of cf cells disruption by symbiont of host cytoskeleton or do we need host actin and host microtubule cytoskeleton dsiegele genegodbold for a future discussion
| 1
|
39,394
| 5,079,373,106
|
IssuesEvent
|
2016-12-28 19:46:00
|
18F/cg-dashboard
|
https://api.github.com/repos/18F/cg-dashboard
|
closed
|
Tier-1 priority design refinements
|
design Navigator
|
Based on usability session documented in #716:
Address the following issues that came up in testing:
- [ ] Did not observe an intentional decision to clicking org name as opposed to clicking in the box to expand. Consider a more clear interaction here between jump to space vs expand accordion. [3]
- [ ] Users tripped on distinction between apps and services. Consider A) a different visual style here for services or B) hiding services by default. [3]
- [ ] Breadcrumbs: one user expressed surprise when clicking on the breadcrumb resulted in a dropdown rather than being moved to the point in the experience. *This one is very open to numerous solutions.* Consider an interaction that allows a single click to move to parent org or space and multiple clicks to jump to non-parent org or space. [2]
Acceptance criteria:
- Prototype designs that address the above issues.
|
1.0
|
Tier-1 priority design refinements - Based on usability session documented in #716:
Address the following issues that came up in testing:
- [ ] Did not observe an intentional decision to clicking org name as opposed to clicking in the box to expand. Consider a more clear interaction here between jump to space vs expand accordion. [3]
- [ ] Users tripped on distinction between apps and services. Consider A) a different visual style here for services or B) hiding services by default. [3]
- [ ] Breadcrumbs: one user expressed surprise when clicking on the breadcrumb resulted in a dropdown rather than being moved to the point in the experience. *This one is very open to numerous solutions.* Consider an interaction that allows a single click to move to parent org or space and multiple clicks to jump to non-parent org or space. [2]
Acceptance criteria:
- Prototype designs that address the above issues.
|
non_process
|
tier priority design refinements based on usability session documented in address the following issues that came up in testing did not observe an intentional decision to clicking org name as opposed to clicking in the box to expand consider a more clear interaction here between jump to space vs expand accordion users tripped on distinction between apps and services consider a a different visual style here for services or b hiding services by default breadcrumbs one user expressed surprise when clicking on the breadcrumb resulted in a dropdown rather than being moved to the point in the experience this one is very open to numerous solutions consider an interaction that allows a single click to move to parent org or space and multiple clicks to jump to non parent org or space acceptance criteria prototype designs that address the above issues
| 0
|
78,909
| 3,519,231,913
|
IssuesEvent
|
2016-01-12 16:08:08
|
uwdub/web-dub
|
https://api.github.com/repos/uwdub/web-dub
|
opened
|
Cleaning TBD Titles on Upcoming Talks
|
priority-high
|
A common scenario is that we have a speaker scheduled, but no title yet. Currently, this just appears as TBD in several places (the front page, the calendar). Let's instead emphasize the speaker.
This starts in part from an @andyjko observation:
> On the calendar, there are many entries with "TBD" headers that nevertheless have names next to them. After digging around, I figured out that "TBD" is the title of the talk. I think it would be less confusing if we just skipped the title altogether if it's empty.
I think it requires:
- Updating the main page to show speaker name (i.e., always).
- Updating the main page to not show TBD titles.
- Updating the calendar page to not show TBD titles.
- This may require changing our TBD syntax, having a few field-specific flags instead of one global.
|
1.0
|
Cleaning TBD Titles on Upcoming Talks - A common scenario is that we have a speaker scheduled, but no title yet. Currently, this just appears as TBD in several places (the front page, the calendar). Let's instead emphasize the speaker.
This starts in part from an @andyjko observation:
> On the calendar, there are many entries with "TBD" headers that nevertheless have names next to them. After digging around, I figured out that "TBD" is the title of the talk. I think it would be less confusing if we just skipped the title altogether if it's empty.
I think it requires:
- Updating the main page to show speaker name (i.e., always).
- Updating the main page to not show TBD titles.
- Updating the calendar page to not show TBD titles.
- This may require changing our TBD syntax, having a few field-specific flags instead of one global.
|
non_process
|
cleaning tbd titles on upcoming talks a common scenario is that we have a speaker scheduled but no title yet currently this just appears as tbd in several places the front page the calendar let s instead emphasize the speaker this starts in part from an andyjko observation on the calendar there are many entries with tbd headers that nevertheless have names next to them after digging around i figured out that tbd is the title of the talk i think it would be less confusing if we just skipped the title altogether if it s empty i think it requires updating the main page to show speaker name i e always updating the main page to not show tbd titles updating the calendar page to not show tbd titles this may require changing our tbd syntax having a few field specific flags instead of one global
| 0
|
13,628
| 16,239,653,139
|
IssuesEvent
|
2021-05-07 07:53:53
|
googleapis/nodejs-logging
|
https://api.github.com/repos/googleapis/nodejs-logging
|
closed
|
Fix broken source map in package
|
api: logging priority: p2 type: bug type: process
|
I think sourcemaps never worked in this repo.
Source maps are generated as: `{"version":3,"file":"common.js","sourceRoot":"","sources":["../../src/common.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;`
But package only ships with files `build/src` and `build/protos`. There's no file to be accessed by the path `../../src/common.ts`
Solution(s): ship package with ts files, or point sourceRoot to this repo url
|
1.0
|
Fix broken source map in package - I think sourcemaps never worked in this repo.
Source maps are generated as: `{"version":3,"file":"common.js","sourceRoot":"","sources":["../../src/common.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;`
But package only ships with files `build/src` and `build/protos`. There's no file to be accessed by the path `../../src/common.ts`
Solution(s): ship package with ts files, or point sourceRoot to this repo url
|
process
|
fix broken source map in package i think sourcemaps never worked in this repo source maps are generated as version file common js sourceroot sources names mappings aaaa but package only ships with files build src and build protos there s no file to be accessed by the path src common ts solution s ship package with ts files or point sourceroot to this repo url
| 1
|
8,338
| 11,497,798,420
|
IssuesEvent
|
2020-02-12 10:42:11
|
18F/tts-tech-portfolio
|
https://api.github.com/repos/18F/tts-tech-portfolio
|
closed
|
document how we prioritize within the board
|
Jan2020-inperson epic: internal workflow/procedures groomed - accepted workflow: process
|
## Background information
https://github.com/18F/tts-tech-portfolio/issues/181#issuecomment-557642484
Ran out of time to do this, but some general principles to use for prioritization:
- Close out things that are in-progress before picking up something new. If there's too much or part of it can wait, break out a new issue for the Icebox.
- Keep an eye on things that have due dates
- Look for the tasks that are force-multipliers, i.e. where we can spend a little time now to save a lot of time in the long run
- Solve the active pain points before going to nice-to-haves
Would like to think about other/additional criteria can be added to the principle
- Is this something that people within TTS are asking for?
- Is this an action item from another meeting (Working Group, Guild, council, etc)?
- Is this a customer service task?
- Is this something that is specifically called out in the Handbook that the Tech Portfolio does?
- Has this been lingering for a while?
- Will this task make another task on the board easier?
- Is there a related card in progress?
## User stories
As someone working on issues, I would like to know the criteria for prioritizing.
## Implementation
- [ ] Pull request for the Project Board on how to prioritize
- [ ] Team members review pull request
- [ ] Pull request is approved
## Acceptance criteria
- [ ] demo at Review
- [ ] pull request is linked to issue
|
1.0
|
document how we prioritize within the board - ## Background information
https://github.com/18F/tts-tech-portfolio/issues/181#issuecomment-557642484
Ran out of time to do this, but some general principles to use for prioritization:
- Close out things that are in-progress before picking up something new. If there's too much or part of it can wait, break out a new issue for the Icebox.
- Keep an eye on things that have due dates
- Look for the tasks that are force-multipliers, i.e. where we can spend a little time now to save a lot of time in the long run
- Solve the active pain points before going to nice-to-haves
Would like to think about other/additional criteria can be added to the principle
- Is this something that people within TTS are asking for?
- Is this an action item from another meeting (Working Group, Guild, council, etc)?
- Is this a customer service task?
- Is this something that is specifically called out in the Handbook that the Tech Portfolio does?
- Has this been lingering for a while?
- Will this task make another task on the board easier?
- Is there a related card in progress?
## User stories
As someone working on issues, I would like to know the criteria for prioritizing.
## Implementation
- [ ] Pull request for the Project Board on how to prioritize
- [ ] Team members review pull request
- [ ] Pull request is approved
## Acceptance criteria
- [ ] demo at Review
- [ ] pull request is linked to issue
|
process
|
document how we prioritize within the board background information ran out of time to do this but some general principles to use for prioritization close out things that are in progress before picking up something new if there s too much or part of it can wait break out a new issue for the icebox keep an eye on things that have due dates look for the tasks that are force multipliers i e where we can spend a little time now to save a lot of time in the long run solve the active pain points before going to nice to haves would like to think about other additional criteria can be added to the principle is this something that people within tts are asking for is this an action item from another meeting working group guild council etc is this a customer service task is this something that is specifically called out in the handbook that the tech portfolio does has this been lingering for a while will this task make another task on the board easier is there a related card in progress user stories as someone working on issues i would like to know the criteria for prioritizing implementation pull request for the project board on how to prioritize team members review pull request pull request is approved acceptance criteria demo at review pull request is linked to issue
| 1
|
7,456
| 10,561,203,723
|
IssuesEvent
|
2019-10-04 15:23:15
|
ESMValGroup/ESMValCore
|
https://api.github.com/repos/ESMValGroup/ESMValCore
|
closed
|
Apply cmor checks (and fixes) to fx-files
|
cmor paper preprocessor
|
Going through the fx data in CMIP5, I've noticed that there are several issues with the mask variables `sftlf` and `sftof` which needs to be fixed (wrong units, wrong coordinates, wrong fillvalues, etc.), but this is currently not implemented, although some fixes are available (see for example [here](https://github.com/ESMValGroup/ESMValTool/blob/fc376aded0d2092b05e3d1f74132592f0a1dc2a6/esmvaltool/cmor/_fixes/CMIP5/EC_EARTH.py#L29)).
This affects both, the use of fx variables in the diagnostics (via `fx-files` key in the variable dict) and the use of fx variables for the masking (via `mask_landsea` key in the preprocessor dict).
In principle, whenever a fx-variable is needed, it should be read, cmor-checked, (fixed) and saved in the `work_dir` / `preproc_dir`, as for any other variable. The corresponding path should be then made available to the diagnostics if the `fx-files` key is specified in the recipe (as it happens now with the original input fx-file).
|
1.0
|
Apply cmor checks (and fixes) to fx-files - Going through the fx data in CMIP5, I've noticed that there are several issues with the mask variables `sftlf` and `sftof` which needs to be fixed (wrong units, wrong coordinates, wrong fillvalues, etc.), but this is currently not implemented, although some fixes are available (see for example [here](https://github.com/ESMValGroup/ESMValTool/blob/fc376aded0d2092b05e3d1f74132592f0a1dc2a6/esmvaltool/cmor/_fixes/CMIP5/EC_EARTH.py#L29)).
This affects both, the use of fx variables in the diagnostics (via `fx-files` key in the variable dict) and the use of fx variables for the masking (via `mask_landsea` key in the preprocessor dict).
In principle, whenever a fx-variable is needed, it should be read, cmor-checked, (fixed) and saved in the `work_dir` / `preproc_dir`, as for any other variable. The corresponding path should be then made available to the diagnostics if the `fx-files` key is specified in the recipe (as it happens now with the original input fx-file).
|
process
|
apply cmor checks and fixes to fx files going through the fx data in i ve noticed that there are several issues with the mask variables sftlf and sftof which needs to be fixed wrong units wrong coordinates wrong fillvalues etc but this is currently not implemented although some fixes are available see for example this affects both the use of fx variables in the diagnostics via fx files key in the variable dict and the use of fx variables for the masking via mask landsea key in the preprocessor dict in principle whenever a fx variable is needed it should be read cmor checked fixed and saved in the work dir preproc dir as for any other variable the corresponding path should be then made available to the diagnostics if the fx files key is specified in the recipe as it happens now with the original input fx file
| 1
|
568,894
| 16,989,970,508
|
IssuesEvent
|
2021-06-30 19:03:30
|
hochschule-darmstadt/openartbrowser
|
https://api.github.com/repos/hochschule-darmstadt/openartbrowser
|
opened
|
Dropdown list missing for search result page
|
User Interface bug high priority mobile
|
**Describe the bug**
The search result page should have a dropdown list for selection between tabs (as used for the mobile view on all other pages).
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'openartbrowser' in mobile view (use chrome inspection tools).
2. Enter a term in the search bar and perform a search
4. See error (no dropdown list for tabs, just a list of all of them)
**Expected behavior**
A dropdown list should be provided on the search result page in mobile view instead of a list of all items. The default entry should be the "All" tab.
**Screenshots**

|
1.0
|
Dropdown list missing for search result page - **Describe the bug**
The search result page should have a dropdown list for selection between tabs (as used for the mobile view on all other pages).
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'openartbrowser' in mobile view (use chrome inspection tools).
2. Enter a term in the search bar and perform a search
4. See error (no dropdown list for tabs, just a list of all of them)
**Expected behavior**
A dropdown list should be provided on the search result page in mobile view instead of a list of all items. The default entry should be the "All" tab.
**Screenshots**

|
non_process
|
dropdown list missing for search result page describe the bug the search result page should have a dropdown list for selection between tabs as used for the mobile view on all other pages to reproduce steps to reproduce the behavior go to openartbrowser in mobile view use chrome inspection tools enter a term in the search bar and perform a search see error no dropdown list for tabs just a list of all of them expected behavior a dropdown list should be provided on the search result page in mobile view instead of a list of all items the default entry should be the all tab screenshots
| 0
|
16,271
| 20,865,022,058
|
IssuesEvent
|
2022-03-22 05:48:54
|
google/gvisor
|
https://api.github.com/repos/google/gvisor
|
closed
|
gVisor Networking Features
|
type: enhancement area: networking area: performance type: process
|
### Description
We are trying to gather requirements from gVisor's Open source community to prioritize Networking features that we should be working on.
Feel free to leave comments on this issue with features requests. Please use the following template to help us prioritize these features.
Description: <Please describe the feature in the best way possible>
Use-case: Why do you need it?
Blocker: Is it a blocker to your gVisor adoption?
|
1.0
|
gVisor Networking Features - ### Description
We are trying to gather requirements from gVisor's Open source community to prioritize Networking features that we should be working on.
Feel free to leave comments on this issue with features requests. Please use the following template to help us prioritize these features.
Description: <Please describe the feature in the best way possible>
Use-case: Why do you need it?
Blocker: Is it a blocker to your gVisor adoption?
|
process
|
gvisor networking features description we are trying to gather requirements from gvisor s open source community to prioritize networking features that we should be working on feel free to leave comments on this issue with features requests please use the following template to help us prioritize these features description use case why do you need it blocker is it a blocker to your gvisor adoption
| 1
|
267,506
| 8,389,787,370
|
IssuesEvent
|
2018-10-09 10:36:11
|
CS2103-AY1819S1-T12-3/main
|
https://api.github.com/repos/CS2103-AY1819S1-T12-3/main
|
opened
|
As a student I can edit my tasks
|
priority.High type.Story
|
so that I can adjust/change my tasks according to my schedule.
|
1.0
|
As a student I can edit my tasks - so that I can adjust/change my tasks according to my schedule.
|
non_process
|
as a student i can edit my tasks so that i can adjust change my tasks according to my schedule
| 0
|
11,513
| 14,396,830,019
|
IssuesEvent
|
2020-12-03 07:05:17
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
--experimental_remote_grpc_log and --build_event_binary_file together result in bazel hanging
|
more data needed team-Remote-Exec type: support / not a bug (process) untriaged
|
This one's bizarre. I've been able to get it to happen with bazel 3.4.1 and 3.5, but haven't narrowed down a reproduction case yet; filing the bug now in case it's more obvious to somebody else what is going on. I also have remote cache enabled for this particular build.
The build will complete normally, and then bazel will print "Waiting for build events: BinaryFormatFileTransport" forever and never exit.
|
1.0
|
--experimental_remote_grpc_log and --build_event_binary_file together result in bazel hanging - This one's bizarre. I've been able to get it to happen with bazel 3.4.1 and 3.5, but haven't narrowed down a reproduction case yet; filing the bug now in case it's more obvious to somebody else what is going on. I also have remote cache enabled for this particular build.
The build will complete normally, and then bazel will print "Waiting for build events: BinaryFormatFileTransport" forever and never exit.
|
process
|
experimental remote grpc log and build event binary file together result in bazel hanging this one s bizarre i ve been able to get it to happen with bazel and but haven t narrowed down a reproduction case yet filing the bug now in case it s more obvious to somebody else what is going on i also have remote cache enabled for this particular build the build will complete normally and then bazel will print waiting for build events binaryformatfiletransport forever and never exit
| 1
|
11,620
| 14,483,948,257
|
IssuesEvent
|
2020-12-10 15:45:20
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Enable strict mode across the codebase
|
process/candidate team/client tech/typescript
|
<img width="280" alt="Screen_Shot_2020-12-10_at_14 49 33" src="https://user-images.githubusercontent.com/1094804/101794305-0636da00-3b07-11eb-99a6-581618e19ed0.png">
|
1.0
|
Enable strict mode across the codebase - <img width="280" alt="Screen_Shot_2020-12-10_at_14 49 33" src="https://user-images.githubusercontent.com/1094804/101794305-0636da00-3b07-11eb-99a6-581618e19ed0.png">
|
process
|
enable strict mode across the codebase img width alt screen shot at src
| 1
|
15,823
| 20,018,007,460
|
IssuesEvent
|
2022-02-01 13:59:56
|
threefoldtech/tfchain
|
https://api.github.com/repos/threefoldtech/tfchain
|
closed
|
Use versioned tfchain_pallets dependencies
|
process_wontfix type_feature
|
currently the dependency version is "development", which is a moving target.
|
1.0
|
Use versioned tfchain_pallets dependencies - currently the dependency version is "development", which is a moving target.
|
process
|
use versioned tfchain pallets dependencies currently the dependency version is development which is a moving target
| 1
|
22,696
| 32,006,392,701
|
IssuesEvent
|
2023-09-21 15:04:40
|
googleapis/sdk-platform-java
|
https://api.github.com/repos/googleapis/sdk-platform-java
|
closed
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* must have required property 'release_level' in gax-java/.repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* must have required property 'release_level' in gax-java/.repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 must have required property release level in gax java repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
| 1
|
13,371
| 15,834,316,582
|
IssuesEvent
|
2021-04-06 16:38:07
|
DSpace/DSpace
|
https://api.github.com/repos/DSpace/DSpace
|
closed
|
Create a script which can reset `dspace.agreements.end-user` metadata field whenever the User Agreement changes.
|
e/4 new feature tools: processes
|
**Is your feature request related to a problem? Please describe.**
Based on discussion in https://github.com/DSpace/dspace-angular/pull/862#issuecomment-694173128
If a site changes/modifies the "User Agreement" text, the administrator would need a way to "reset" the `dspace.agreements.end-user` for all EPersons. Currently, when an EPerson agrees to the "User Agreement", this metadata field is set to `true`. They no longer see the "User Agreement" after that point.
**Describe the solution you'd like**
A Script should be created which can delete this `dspace.agreements.end-user` field (or set to `false`) for all EPersons. This will force EPersons to then read the Agreement again and accept it again. This script would need to be triggered manually whenever an Administrator knows they've made major modifications to the User Agreement which require users to agree to it again.
This Script should ideally support the Scripts & Processes framework. That would allow it to be run from *either* the CLI or from the Admin UI.
**Describe alternatives or workarounds you've considered**
For the time being, a valid workaround is to remove this `dspace.agreements.end-user` field (or set them all to `false`) using SQL queries on database.
|
1.0
|
Create a script which can reset `dspace.agreements.end-user` metadata field whenever the User Agreement changes. - **Is your feature request related to a problem? Please describe.**
Based on discussion in https://github.com/DSpace/dspace-angular/pull/862#issuecomment-694173128
If a site changes/modifies the "User Agreement" text, the administrator would need a way to "reset" the `dspace.agreements.end-user` for all EPersons. Currently, when an EPerson agrees to the "User Agreement", this metadata field is set to `true`. They no longer see the "User Agreement" after that point.
**Describe the solution you'd like**
A Script should be created which can delete this `dspace.agreements.end-user` field (or set to `false`) for all EPersons. This will force EPersons to then read the Agreement again and accept it again. This script would need to be triggered manually whenever an Administrator knows they've made major modifications to the User Agreement which require users to agree to it again.
This Script should ideally support the Scripts & Processes framework. That would allow it to be run from *either* the CLI or from the Admin UI.
**Describe alternatives or workarounds you've considered**
For the time being, a valid workaround is to remove this `dspace.agreements.end-user` field (or set them all to `false`) using SQL queries on database.
|
process
|
create a script which can reset dspace agreements end user metadata field whenever the user agreement changes is your feature request related to a problem please describe based on discussion in if a site changes modifies the user agreement text the administrator would need a way to reset the dspace agreements end user for all epersons currently when an eperson agrees to the user agreement this metadata field is set to true they no longer see the user agreement after that point describe the solution you d like a script should be created which can delete this dspace agreements end user field or set to false for all epersons this will force epersons to then read the agreement again and accept it again this script would need to be triggered manually whenever an administrator knows they ve made major modifications to the user agreement which require users to agree to it again this script should ideally support the scripts processes framework that would allow it to be run from either the cli or from the admin ui describe alternatives or workarounds you ve considered for the time being a valid workaround is to remove this dspace agreements end user field or set them all to false using sql queries on database
| 1
|
121,686
| 4,820,191,312
|
IssuesEvent
|
2016-11-04 21:49:45
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
RFC: Allow perm-failed deployments
|
area/app-config-deployment kind/enhancement priority/P2 team/ux
|
There are times when the deployment system can infer that the latest deployment state has no reasonable chance of being realized (e.g. a bad or unpullable image). The current deployment controller design will continue to try reconciling indefinitely regardless of the possibly of success. If based on inference or user constraints (e.g. timeout conditions specified in an enhancement to the deployment API #1743) the system is ready to give up its best-effort attempt, the deployment could be somehow marked as "permananently failed" for a given spec hash so that the system won't continue thrashing on a doomed deployment.
There is a bit of functional overlap with inert deployments (#14516) in that both concepts result in the deployment controller "ignoring" a deployment whose state still needs realized, but inert deployments as described don't seem to capture all the context users would want in this case (i.e., it's not enough to just mark a doomed deployment intert without more context about why, and without UX safety nets to distinguish the repurcussions of re-activating the suddenly inert deployment vs. a permafailed deployment.)
|
1.0
|
RFC: Allow perm-failed deployments - There are times when the deployment system can infer that the latest deployment state has no reasonable chance of being realized (e.g. a bad or unpullable image). The current deployment controller design will continue to try reconciling indefinitely regardless of the possibly of success. If based on inference or user constraints (e.g. timeout conditions specified in an enhancement to the deployment API #1743) the system is ready to give up its best-effort attempt, the deployment could be somehow marked as "permananently failed" for a given spec hash so that the system won't continue thrashing on a doomed deployment.
There is a bit of functional overlap with inert deployments (#14516) in that both concepts result in the deployment controller "ignoring" a deployment whose state still needs realized, but inert deployments as described don't seem to capture all the context users would want in this case (i.e., it's not enough to just mark a doomed deployment intert without more context about why, and without UX safety nets to distinguish the repurcussions of re-activating the suddenly inert deployment vs. a permafailed deployment.)
|
non_process
|
rfc allow perm failed deployments there are times when the deployment system can infer that the latest deployment state has no reasonable chance of being realized e g a bad or unpullable image the current deployment controller design will continue to try reconciling indefinitely regardless of the possibly of success if based on inference or user constraints e g timeout conditions specified in an enhancement to the deployment api the system is ready to give up its best effort attempt the deployment could be somehow marked as permananently failed for a given spec hash so that the system won t continue thrashing on a doomed deployment there is a bit of functional overlap with inert deployments in that both concepts result in the deployment controller ignoring a deployment whose state still needs realized but inert deployments as described don t seem to capture all the context users would want in this case i e it s not enough to just mark a doomed deployment intert without more context about why and without ux safety nets to distinguish the repurcussions of re activating the suddenly inert deployment vs a permafailed deployment
| 0
|
2,265
| 5,096,222,237
|
IssuesEvent
|
2017-01-03 17:30:10
|
Alfresco/alfresco-ng2-components
|
https://api.github.com/repos/Alfresco/alfresco-ng2-components
|
opened
|
Error is logged when trying to complete the Risk Process
|
bug comp: activiti-processList comp: activiti-taskList
|
<!--
PLEASE FILL OUT THE FOLLOWING INFORMATION, THIS WILL HELP US TO RESOLVE YOUR PROBLEM FASTER.
REMEMBER FOR SUPPORT REQUESTS YOU CAN ALSO ASK ON OUR GITTER CHAT:
Please ask before on our gitter channel https://gitter.im/Alfresco/alfresco-ng2-components
-->
**Type of issue:** (check with "[x]")
```
- [ ] New feature request
- [X] Bug
- [ ] Support request
```
**Current behavior:**
Start the Risk Process (attached to this issue)
Go on Task Approver Review Request
Select Approve and complete
Go on Task Implement Request
Select Approve and then click Implemented.
An error is triggered and the process is stucked.
**Expected behavior:**
User should be able to complete the process
**Steps to reproduce the issue:**
<!-- Describe the steps to reproduce the issue. -->
**Component name and version:**
<!-- Example: ng2-alfresco-login. Check before if this issue is still present in the most recent version -->
<img width="1430" alt="screen shot 2017-01-03 at 17 28 36" src="https://cloud.githubusercontent.com/assets/7974125/21616631/149ad5a0-d1da-11e6-8ee9-bb8ff0417d39.png">
[Risk Limit Management.zip](https://github.com/Alfresco/alfresco-ng2-components/files/682863/Risk.Limit.Management.zip)
|
1.0
|
Error is logged when trying to complete the Risk Process - <!--
PLEASE FILL OUT THE FOLLOWING INFORMATION, THIS WILL HELP US TO RESOLVE YOUR PROBLEM FASTER.
REMEMBER FOR SUPPORT REQUESTS YOU CAN ALSO ASK ON OUR GITTER CHAT:
Please ask before on our gitter channel https://gitter.im/Alfresco/alfresco-ng2-components
-->
**Type of issue:** (check with "[x]")
```
- [ ] New feature request
- [X] Bug
- [ ] Support request
```
**Current behavior:**
Start the Risk Process (attached to this issue)
Go on Task Approver Review Request
Select Approve and complete
Go on Task Implement Request
Select Approve and then click Implemented.
An error is triggered and the process is stucked.
**Expected behavior:**
User should be able to complete the process
**Steps to reproduce the issue:**
<!-- Describe the steps to reproduce the issue. -->
**Component name and version:**
<!-- Example: ng2-alfresco-login. Check before if this issue is still present in the most recent version -->
<img width="1430" alt="screen shot 2017-01-03 at 17 28 36" src="https://cloud.githubusercontent.com/assets/7974125/21616631/149ad5a0-d1da-11e6-8ee9-bb8ff0417d39.png">
[Risk Limit Management.zip](https://github.com/Alfresco/alfresco-ng2-components/files/682863/Risk.Limit.Management.zip)
|
process
|
error is logged when trying to complete the risk process please fill out the following information this will help us to resolve your problem faster remember for support requests you can also ask on our gitter chat please ask before on our gitter channel type of issue check with new feature request bug support request current behavior start the risk process attached to this issue go on task approver review request select approve and complete go on task implement request select approve and then click implemented an error is triggered and the process is stucked expected behavior user should be able to complete the process steps to reproduce the issue component name and version img width alt screen shot at src
| 1
|
20,718
| 27,417,156,647
|
IssuesEvent
|
2023-03-01 14:30:11
|
camunda/issues
|
https://api.github.com/repos/camunda/issues
|
opened
|
Support Elastic Search 8
|
component:distribution component:operate component:optimize component:sre component:tasklist component:zeebe component:zeebe-process-automation public target:8.2 potential:8.2
|
> This is an epic internal-docs issue. It bundles all activities we conduct around a certain initiative. It will typically links to various child issues from various repositories and can be spread across multiple teams.
### Value Proposition Statement
Use Elasticsearch 8.x with Camunda Plattform 8.
### User Problem
- Elastic 8 is out for many months already.
- Elastic 7 runs out of support 8/2023.
- Currently C8 SM only supports Elastic 7.x
### User Stories
* SM: I can make use of Elasticsearch 8.x when installing Camunda Platform 8 Self-Managed.
* SaaS: When creating a new cluster or updating to the latest minor version on SaaS we are using Elastic 8 underneath.
Consider https://github.com/camunda/product-hub/issues/256
### Implementation Notes
* We should support the latest two minors of Elastic 8.
* We need to consider what way we can roll-out Elastic 8 to our managed service (SaaS). It should be part of this Epic, but we need to define still how we do it in best way.
* We need to consider update paths for customers and our SaaS customers running on ES7. Since ES7 is out for a long time and ES7 will go out of maintenance soon, we can only support ES8.
### Links
<!-- Additional links, for example:
* Support Case X
* Bug on team board Y
* Concept document Z
* Topic Slack Channel
* Value Proposition Slide
* Initiative in OKR sheet
* -->
* Support
* Optimize - https://jira.camunda.com/browse/OPT-6404
* C8 - https://jira.camunda.com/browse/SUPPORT-15270
* https://jira.camunda.com/browse/SUPPORT-15365
* https://jira.camunda.com/browse/SUPPORT-12775
- https://github.com/camunda/camunda-platform/issues/83
#### Design Tickets
Reviewed: Jan 30, 2023
No design needed
#### Implementation Tickets
* Operate - https://github.com/camunda/operate/issues/3931
* Tasklist - https://github.com/camunda/operate/issues/4112
Tested from Operate side using:
- Elasticsearch 8 image: docker.elastic.co/elasticsearch/elasticsearch:8.6.0
- Elasticsearch 7 image: docker.elastic.co/elasticsearch/elasticsearch:7.17.0
* Optimize - https://github.com/camunda/camunda-optimize/pull/5860
#### Doc Tickets
|
1.0
|
Support Elastic Search 8 - > This is an epic internal-docs issue. It bundles all activities we conduct around a certain initiative. It will typically links to various child issues from various repositories and can be spread across multiple teams.
### Value Proposition Statement
Use Elasticsearch 8.x with Camunda Plattform 8.
### User Problem
- Elastic 8 is out for many months already.
- Elastic 7 runs out of support 8/2023.
- Currently C8 SM only supports Elastic 7.x
### User Stories
* SM: I can make use of Elasticsearch 8.x when installing Camunda Platform 8 Self-Managed.
* SaaS: When creating a new cluster or updating to the latest minor version on SaaS we are using Elastic 8 underneath.
Consider https://github.com/camunda/product-hub/issues/256
### Implementation Notes
* We should support the latest two minors of Elastic 8.
* We need to consider what way we can roll-out Elastic 8 to our managed service (SaaS). It should be part of this Epic, but we need to define still how we do it in best way.
* We need to consider update paths for customers and our SaaS customers running on ES7. Since ES7 is out for a long time and ES7 will go out of maintenance soon, we can only support ES8.
### Links
<!-- Additional links, for example:
* Support Case X
* Bug on team board Y
* Concept document Z
* Topic Slack Channel
* Value Proposition Slide
* Initiative in OKR sheet
* -->
* Support
* Optimize - https://jira.camunda.com/browse/OPT-6404
* C8 - https://jira.camunda.com/browse/SUPPORT-15270
* https://jira.camunda.com/browse/SUPPORT-15365
* https://jira.camunda.com/browse/SUPPORT-12775
- https://github.com/camunda/camunda-platform/issues/83
#### Design Tickets
Reviewed: Jan 30, 2023
No design needed
#### Implementation Tickets
* Operate - https://github.com/camunda/operate/issues/3931
* Tasklist - https://github.com/camunda/operate/issues/4112
Tested from Operate side using:
- Elasticsearch 8 image: docker.elastic.co/elasticsearch/elasticsearch:8.6.0
- Elasticsearch 7 image: docker.elastic.co/elasticsearch/elasticsearch:7.17.0
* Optimize - https://github.com/camunda/camunda-optimize/pull/5860
#### Doc Tickets
|
process
|
support elastic search this is an epic internal docs issue it bundles all activities we conduct around a certain initiative it will typically links to various child issues from various repositories and can be spread across multiple teams value proposition statement use elasticsearch x with camunda plattform user problem elastic is out for many months already elastic runs out of support currently sm only supports elastic x user stories sm i can make use of elasticsearch x when installing camunda platform self managed saas when creating a new cluster or updating to the latest minor version on saas we are using elastic underneath consider implementation notes we should support the latest two minors of elastic we need to consider what way we can roll out elastic to our managed service saas it should be part of this epic but we need to define still how we do it in best way we need to consider update paths for customers and our saas customers running on since is out for a long time and will go out of maintenance soon we can only support links additional links for example support case x bug on team board y concept document z topic slack channel value proposition slide initiative in okr sheet support optimize design tickets reviewed jan no design needed implementation tickets operate tasklist tested from operate side using elasticsearch image docker elastic co elasticsearch elasticsearch elasticsearch image docker elastic co elasticsearch elasticsearch optimize doc tickets
| 1
|
6,090
| 8,951,001,976
|
IssuesEvent
|
2019-01-25 12:36:44
|
enKryptIO/ethvm
|
https://api.github.com/repos/enKryptIO/ethvm
|
closed
|
Exception on Kafka Processing
|
bug priority:low project:processing
|
From time to time, if I destroy the docker environment with everything (including volumes), whenever I process again a new downloaded chain, I obtain this issue:
'''
Aldo Borrero, [22.01.19 21:37]
16:13:41.185 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 0
16:13:42.169 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 1
16:13:42.204 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 2
16:13:42.218 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 3
16:13:42.341 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 4
16:13:42.364 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 5
16:13:42.377 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 6
16:13:42.380 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 7
16:13:42.415 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 8
16:13:42.417 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 9
16:13:42.421 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 10
16:13:42.423 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 11
16:13:42.428 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 12
16:13:42.432 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 13
16:13:42.443 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 14
16:13:42.452 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 15
16:13:42.459 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 16
16:13:42.460 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 17
16:13:42.462 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 18
16:13:42.467 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 19
16:13:42.469 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 20
16:13:42.471 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 21
16:13:42.474 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 22
16:13:42.476 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 23
16:13:42.479 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 24
16:13:42.491 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 25
16:13:42.509 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 26
16:13:42.511 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 27
16:13:42.512 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 28
16:13:42.514 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 29
16:13:42.516 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 30
16:13:42.517 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 31
16:13:42.519 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 32
16:13:42.520 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 33
16:13:42.524 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 34
16:13:42.525 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 35
16:13:42.530 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 36
16:13:42.532 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 37
16:13:42.535 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 38
16:13:42.538 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 39
16:13:42.540 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 40
16:13:42.542 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 41
16:13:42.544 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 42
16:13:42.552 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 43
16:13:42.553 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 44
16:13:42.553 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 45
16:13:42.563 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 46
16:13:42.565 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 47
Aldo Borrero, [22.01.19 21:37]
16:13:42.566 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 48
16:13:42.568 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 49
16:13:42.569 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 50
16:13:42.570 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 51
16:13:42.571 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 52
16:13:42.573 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 53
16:13:42.575 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 54
16:13:42.576 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 55
16:13:42.577 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 56
16:13:42.578 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 57
16:13:42.579 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 58
16:13:42.580 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 59
16:13:42.581 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 60
16:13:42.583 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 61
16:13:42.585 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 62
16:13:42.587 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 63
16:13:42.589 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 64
16:13:42.591 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 65
16:13:42.599 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 66
16:13:42.601 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 67
16:13:42.619 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 68
16:13:42.625 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 69
16:13:42.641 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 70
16:13:42.643 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 71
16:13:42.648 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 72
16:13:42.650 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 73
16:13:42.653 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 74
16:13:42.655 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 75
16:13:42.676 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 76
16:13:42.706 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 77
16:13:42.721 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 78
16:13:42.731 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 79
16:13:42.734 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 80
16:13:42.736 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 81
16:13:42.742 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 82
16:13:42.745 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 83
16:13:42.760 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 84
16:13:42.766 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 85
16:13:42.768 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 86
16:13:42.769 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 87
16:13:42.771 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 88
16:13:42.772 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 89
16:13:42.774 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 90
16:13:42.775 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 91
16:13:42.776 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 92
16:13:42.778 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 93
16:13:42.779 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 94
16:13:42.780 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 95
Aldo Borrero, [22.01.19 21:37]
16:13:42.781 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 96
16:13:42.786 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 97
16:13:42.788 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 98
16:13:42.790 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 99
16:13:42.813 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 100
16:13:42.820 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 101
16:13:42.826 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 102
16:13:42.828 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 103
16:13:42.831 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 104
16:13:42.833 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 105
16:13:42.835 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 106
16:13:42.836 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 107
16:13:42.841 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 108
16:13:42.845 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 109
16:13:42.859 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 110
16:13:42.868 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 111
16:13:42.871 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 112
16:13:42.875 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 113
16:13:42.877 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 114
16:13:42.881 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 115
16:13:42.885 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 116
16:13:42.887 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 117
16:13:42.891 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 118
16:13:42.899 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 119
16:13:42.912 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 120
16:13:42.923 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 121
16:13:42.927 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 122
16:13:42.929 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 123
16:13:42.932 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 124
16:13:42.936 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 125
16:13:42.939 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 126
16:13:42.944 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 127
16:13:42.972 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 128
16:13:42.977 WARN i.e.k.s.p.b.ChainEventsTransformer - Ignoring block 128, already processed
16:13:42.978 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 129
16:13:42.982 ERROR o.a.k.s.p.i.AssignedStreamsTasks - stream-thread [block-processor-1fe1a9ba-e648-4bc0-a02d-78e567ba37aa-StreamThread-1] Failed to process stream task 0_0 due to the following error:
java.lang.IllegalStateException: Block out of sequence. Expected = 128, received = 129
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.ensureSequentialProcessing(ChainEventsTransformer.kt:103)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:88)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:25)
at org.apache.kafka.streams.kstream.internals.KStreamTransform$KStreamTransformProcessor.process(KStreamTransform.java:56)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:132)
Aldo Borrero, [22.01.19 21:37]
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
at org.apache.kafka.streams.kstream.internals.KStreamPeek$KStreamPeekProcessor.process(KStreamPeek.java:44)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:129)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
at org.apache.kafka.streams.processor.internals.SourceNode.process(SourceNode.java:84)
at org.apache.kafka.streams.processor.internals.StreamTask.process(StreamTask.java:351)
at org.apache.kafka.streams.processor.internals.AssignedStreamsTasks.process(AssignedStreamsTasks.java:104)
at org.apache.kafka.streams.processor.internals.TaskManager.process(TaskManager.java:413)
at org.apache.kafka.streams.processor.internals.StreamThread.runOnce(StreamThread.java:862)
at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:777)
at org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:747)
16:13:42.982 ERROR o.a.k.s.p.internals.StreamThread - stream-thread [block-processor-1fe1a9ba-e648-4bc0-a02d-78e567ba37aa-StreamThread-1] Encountered the following error during processing:
java.lang.IllegalStateException: Block out of sequence. Expected = 128, received = 129
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.ensureSequentialProcessing(ChainEventsTransformer.kt:103)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:88)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:25)
at org.apache.kafka.streams.kstream.internals.KStreamTransform$KStreamTransformProcessor.process(KStreamTransform.java:56)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:132)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
at org.apache.kafka.streams.kstream.internals.KStreamPeek$KStreamPeekProcessor.process(KStreamPeek.java:44)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:129)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
at org.apache.kafka.streams.processor.internals.SourceNode.process(SourceNode.java:84)
at org.apache.kafka.streams.processor.internals.StreamTask.process(StreamTask.java:351)
at org.apache.kafka.streams.processor.internals.AssignedStreamsTasks.process(AssignedStreamsTasks.java:104)
at org.apache.kafka.streams.processor.internals.TaskManager.process(TaskManager.java:413)
at org.apache.kafka.streams.processor.internals.StreamThread.runOnce(StreamThread.java:862)
at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:777)
at org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:747)
Aldo Borrero, [22.01.19 21:37]
16:13:43.024 ERROR o.a.k.s.p.i.RecordCollectorImpl - task [0_0] Error sending record (key \x00\x00\x00\x00\x05\x02p|\xF0\x06\xB7\x10Ih\xEC\xC9\x08\x94,\x03\xFF\xCA[\x13J+\xDA\xEE\x1C\x1E\x83\xA2\xA3\xA3\xAA\x85\x91\xED\x00\x00\x00\x00\x00 value [0, 0, 0, 0, 6, 2, 2, 0, 2, 27, 112, 71, -76, 51, -118, -49, 101, -66, -108, -63, -93, -24, -59, -55, 51, -118, -41, -42, 124, 16, 69, 99, -111, -126, 68, -12, 0, 0, 0, 0] timestamp 1548202124660) to topic block-processor-chain-events-changelog due to org.apache.kafka.common.KafkaException: Failing batch since transaction was aborted; No more records will be sent and no more offsets will be recorded for this task.
16:13:43.056 ERROR o.a.k.s.p.internals.StreamTask - task [0_0] Failed to close producer due to the following error:
org.apache.kafka.streams.errors.StreamsException: task [0_0] Abort sending since an error caught with a previous record (key \x00\x00\x00\x00\x05\x02p|\xF0\x06\xB7\x10Ih\xEC\xC9\x08\x94,\x03\xFF\xCA[\x13J+\xDA\xEE\x1C\x1E\x83\xA2\xA3\xA3\xAA\x85\x91\xED\x00\x00\x00\x00\x00 value [B@22a4ed8 timestamp 1548202124660) to topic block-processor-chain-events-changelog due to org.apache.kafka.common.KafkaException: Failing batch since transaction was aborted
at org.apache.kafka.streams.processor.internals.RecordCollectorImpl.recordSendError(RecordCollectorImpl.java:133)
at org.apache.kafka.streams.processor.internals.RecordCollectorImpl.access$500(RecordCollectorImpl.java:50)
at org.apache.kafka.streams.processor.internals.RecordCollectorImpl$1.onCompletion(RecordCollectorImpl.java:192)
at org.apache.kafka.clients.producer.KafkaProducer$InterceptorCallback.onCompletion(KafkaProducer.java:1288)
at org.apache.kafka.clients.producer.internals.ProducerBatch.completeFutureAndFireCallbacks(ProducerBatch.java:230)
at org.apache.kafka.clients.producer.internals.ProducerBatch.abort(ProducerBatch.java:158)
at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortUndrainedBatches(RecordAccumulator.java:762)
at org.apache.kafka.clients.producer.internals.Sender.maybeSendTransactionalRequest(Sender.java:396)
at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:284)
at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:233)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.kafka.common.KafkaException: Failing batch since transaction was aborted
... 4 common frames omitted
Exception in thread "block-processor-1fe1a9ba-e648-4bc0-a02d-78e567ba37aa-StreamThread-1" java.lang.IllegalStateException: Block out of sequence. Expected = 128, received = 129
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.ensureSequentialProcessing(ChainEventsTransformer.kt:103)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:88)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:25)
at org.apache.kafka.streams.kstream.internals.KStreamTransform$KStreamTransformProcessor.process(KStreamTransform.java:56)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:132)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
at org.apache.kafka.streams.kstream.internals.KStreamPeek$KStreamPeekProcessor.process(KStreamPeek.java:44)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:129)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
Aldo Borrero, [22.01.19 21:37]
at org.apache.kafka.streams.processor.internals.SourceNode.process(SourceNode.java:84)
at org.apache.kafka.streams.processor.internals.StreamTask.process(StreamTask.java:351)
at org.apache.kafka.streams.processor.internals.AssignedStreamsTasks.process(AssignedStreamsTasks.java:104)
at org.apache.kafka.streams.processor.internals.TaskManager.process(TaskManager.java:413)
at org.apache.kafka.streams.processor.internals.StreamThread.runOnce(StreamThread.java:862)
at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:777)
at org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:747)
16:13:43.109 WARN o.apache.kafka.streams.KafkaStreams - stream-client [block-processor-1fe1a9ba-e648-4bc0-a02d-78e567ba37aa] All stream threads have died.
'''
|
1.0
|
Exception on Kafka Processing - From time to time, if I destroy the docker environment with everything (including volumes), whenever I process again a new downloaded chain, I obtain this issue:
'''
Aldo Borrero, [22.01.19 21:37]
16:13:41.185 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 0
16:13:42.169 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 1
16:13:42.204 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 2
16:13:42.218 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 3
16:13:42.341 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 4
16:13:42.364 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 5
16:13:42.377 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 6
16:13:42.380 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 7
16:13:42.415 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 8
16:13:42.417 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 9
16:13:42.421 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 10
16:13:42.423 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 11
16:13:42.428 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 12
16:13:42.432 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 13
16:13:42.443 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 14
16:13:42.452 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 15
16:13:42.459 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 16
16:13:42.460 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 17
16:13:42.462 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 18
16:13:42.467 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 19
16:13:42.469 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 20
16:13:42.471 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 21
16:13:42.474 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 22
16:13:42.476 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 23
16:13:42.479 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 24
16:13:42.491 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 25
16:13:42.509 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 26
16:13:42.511 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 27
16:13:42.512 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 28
16:13:42.514 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 29
16:13:42.516 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 30
16:13:42.517 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 31
16:13:42.519 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 32
16:13:42.520 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 33
16:13:42.524 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 34
16:13:42.525 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 35
16:13:42.530 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 36
16:13:42.532 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 37
16:13:42.535 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 38
16:13:42.538 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 39
16:13:42.540 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 40
16:13:42.542 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 41
16:13:42.544 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 42
16:13:42.552 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 43
16:13:42.553 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 44
16:13:42.553 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 45
16:13:42.563 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 46
16:13:42.565 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 47
Aldo Borrero, [22.01.19 21:37]
16:13:42.566 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 48
16:13:42.568 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 49
16:13:42.569 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 50
16:13:42.570 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 51
16:13:42.571 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 52
16:13:42.573 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 53
16:13:42.575 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 54
16:13:42.576 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 55
16:13:42.577 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 56
16:13:42.578 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 57
16:13:42.579 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 58
16:13:42.580 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 59
16:13:42.581 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 60
16:13:42.583 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 61
16:13:42.585 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 62
16:13:42.587 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 63
16:13:42.589 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 64
16:13:42.591 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 65
16:13:42.599 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 66
16:13:42.601 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 67
16:13:42.619 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 68
16:13:42.625 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 69
16:13:42.641 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 70
16:13:42.643 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 71
16:13:42.648 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 72
16:13:42.650 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 73
16:13:42.653 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 74
16:13:42.655 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 75
16:13:42.676 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 76
16:13:42.706 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 77
16:13:42.721 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 78
16:13:42.731 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 79
16:13:42.734 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 80
16:13:42.736 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 81
16:13:42.742 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 82
16:13:42.745 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 83
16:13:42.760 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 84
16:13:42.766 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 85
16:13:42.768 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 86
16:13:42.769 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 87
16:13:42.771 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 88
16:13:42.772 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 89
16:13:42.774 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 90
16:13:42.775 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 91
16:13:42.776 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 92
16:13:42.778 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 93
16:13:42.779 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 94
16:13:42.780 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 95
Aldo Borrero, [22.01.19 21:37]
16:13:42.781 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 96
16:13:42.786 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 97
16:13:42.788 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 98
16:13:42.790 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 99
16:13:42.813 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 100
16:13:42.820 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 101
16:13:42.826 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 102
16:13:42.828 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 103
16:13:42.831 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 104
16:13:42.833 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 105
16:13:42.835 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 106
16:13:42.836 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 107
16:13:42.841 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 108
16:13:42.845 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 109
16:13:42.859 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 110
16:13:42.868 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 111
16:13:42.871 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 112
16:13:42.875 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 113
16:13:42.877 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 114
16:13:42.881 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 115
16:13:42.885 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 116
16:13:42.887 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 117
16:13:42.891 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 118
16:13:42.899 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 119
16:13:42.912 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 120
16:13:42.923 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 121
16:13:42.927 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 122
16:13:42.929 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 123
16:13:42.932 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 124
16:13:42.936 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 125
16:13:42.939 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 126
16:13:42.944 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 127
16:13:42.972 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 128
16:13:42.977 WARN i.e.k.s.p.b.ChainEventsTransformer - Ignoring block 128, already processed
16:13:42.978 INFO i.e.k.s.processors.BlockProcessor - Processing block number = 129
16:13:42.982 ERROR o.a.k.s.p.i.AssignedStreamsTasks - stream-thread [block-processor-1fe1a9ba-e648-4bc0-a02d-78e567ba37aa-StreamThread-1] Failed to process stream task 0_0 due to the following error:
java.lang.IllegalStateException: Block out of sequence. Expected = 128, received = 129
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.ensureSequentialProcessing(ChainEventsTransformer.kt:103)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:88)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:25)
at org.apache.kafka.streams.kstream.internals.KStreamTransform$KStreamTransformProcessor.process(KStreamTransform.java:56)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:132)
Aldo Borrero, [22.01.19 21:37]
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
at org.apache.kafka.streams.kstream.internals.KStreamPeek$KStreamPeekProcessor.process(KStreamPeek.java:44)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:129)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
at org.apache.kafka.streams.processor.internals.SourceNode.process(SourceNode.java:84)
at org.apache.kafka.streams.processor.internals.StreamTask.process(StreamTask.java:351)
at org.apache.kafka.streams.processor.internals.AssignedStreamsTasks.process(AssignedStreamsTasks.java:104)
at org.apache.kafka.streams.processor.internals.TaskManager.process(TaskManager.java:413)
at org.apache.kafka.streams.processor.internals.StreamThread.runOnce(StreamThread.java:862)
at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:777)
at org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:747)
16:13:42.982 ERROR o.a.k.s.p.internals.StreamThread - stream-thread [block-processor-1fe1a9ba-e648-4bc0-a02d-78e567ba37aa-StreamThread-1] Encountered the following error during processing:
java.lang.IllegalStateException: Block out of sequence. Expected = 128, received = 129
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.ensureSequentialProcessing(ChainEventsTransformer.kt:103)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:88)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:25)
at org.apache.kafka.streams.kstream.internals.KStreamTransform$KStreamTransformProcessor.process(KStreamTransform.java:56)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:132)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
at org.apache.kafka.streams.kstream.internals.KStreamPeek$KStreamPeekProcessor.process(KStreamPeek.java:44)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:129)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
at org.apache.kafka.streams.processor.internals.SourceNode.process(SourceNode.java:84)
at org.apache.kafka.streams.processor.internals.StreamTask.process(StreamTask.java:351)
at org.apache.kafka.streams.processor.internals.AssignedStreamsTasks.process(AssignedStreamsTasks.java:104)
at org.apache.kafka.streams.processor.internals.TaskManager.process(TaskManager.java:413)
at org.apache.kafka.streams.processor.internals.StreamThread.runOnce(StreamThread.java:862)
at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:777)
at org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:747)
Aldo Borrero, [22.01.19 21:37]
16:13:43.024 ERROR o.a.k.s.p.i.RecordCollectorImpl - task [0_0] Error sending record (key \x00\x00\x00\x00\x05\x02p|\xF0\x06\xB7\x10Ih\xEC\xC9\x08\x94,\x03\xFF\xCA[\x13J+\xDA\xEE\x1C\x1E\x83\xA2\xA3\xA3\xAA\x85\x91\xED\x00\x00\x00\x00\x00 value [0, 0, 0, 0, 6, 2, 2, 0, 2, 27, 112, 71, -76, 51, -118, -49, 101, -66, -108, -63, -93, -24, -59, -55, 51, -118, -41, -42, 124, 16, 69, 99, -111, -126, 68, -12, 0, 0, 0, 0] timestamp 1548202124660) to topic block-processor-chain-events-changelog due to org.apache.kafka.common.KafkaException: Failing batch since transaction was aborted; No more records will be sent and no more offsets will be recorded for this task.
16:13:43.056 ERROR o.a.k.s.p.internals.StreamTask - task [0_0] Failed to close producer due to the following error:
org.apache.kafka.streams.errors.StreamsException: task [0_0] Abort sending since an error caught with a previous record (key \x00\x00\x00\x00\x05\x02p|\xF0\x06\xB7\x10Ih\xEC\xC9\x08\x94,\x03\xFF\xCA[\x13J+\xDA\xEE\x1C\x1E\x83\xA2\xA3\xA3\xAA\x85\x91\xED\x00\x00\x00\x00\x00 value [B@22a4ed8 timestamp 1548202124660) to topic block-processor-chain-events-changelog due to org.apache.kafka.common.KafkaException: Failing batch since transaction was aborted
at org.apache.kafka.streams.processor.internals.RecordCollectorImpl.recordSendError(RecordCollectorImpl.java:133)
at org.apache.kafka.streams.processor.internals.RecordCollectorImpl.access$500(RecordCollectorImpl.java:50)
at org.apache.kafka.streams.processor.internals.RecordCollectorImpl$1.onCompletion(RecordCollectorImpl.java:192)
at org.apache.kafka.clients.producer.KafkaProducer$InterceptorCallback.onCompletion(KafkaProducer.java:1288)
at org.apache.kafka.clients.producer.internals.ProducerBatch.completeFutureAndFireCallbacks(ProducerBatch.java:230)
at org.apache.kafka.clients.producer.internals.ProducerBatch.abort(ProducerBatch.java:158)
at org.apache.kafka.clients.producer.internals.RecordAccumulator.abortUndrainedBatches(RecordAccumulator.java:762)
at org.apache.kafka.clients.producer.internals.Sender.maybeSendTransactionalRequest(Sender.java:396)
at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:284)
at org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:233)
at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.kafka.common.KafkaException: Failing batch since transaction was aborted
... 4 common frames omitted
Exception in thread "block-processor-1fe1a9ba-e648-4bc0-a02d-78e567ba37aa-StreamThread-1" java.lang.IllegalStateException: Block out of sequence. Expected = 128, received = 129
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.ensureSequentialProcessing(ChainEventsTransformer.kt:103)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:88)
at io.enkrypt.kafka.streams.processors.block.ChainEventsTransformer.transform(ChainEventsTransformer.kt:25)
at org.apache.kafka.streams.kstream.internals.KStreamTransform$KStreamTransformProcessor.process(KStreamTransform.java:56)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:132)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
at org.apache.kafka.streams.kstream.internals.KStreamPeek$KStreamPeekProcessor.process(KStreamPeek.java:44)
at org.apache.kafka.streams.processor.internals.ProcessorNode.process(ProcessorNode.java:115)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:146)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:129)
at org.apache.kafka.streams.processor.internals.ProcessorContextImpl.forward(ProcessorContextImpl.java:93)
Aldo Borrero, [22.01.19 21:37]
at org.apache.kafka.streams.processor.internals.SourceNode.process(SourceNode.java:84)
at org.apache.kafka.streams.processor.internals.StreamTask.process(StreamTask.java:351)
at org.apache.kafka.streams.processor.internals.AssignedStreamsTasks.process(AssignedStreamsTasks.java:104)
at org.apache.kafka.streams.processor.internals.TaskManager.process(TaskManager.java:413)
at org.apache.kafka.streams.processor.internals.StreamThread.runOnce(StreamThread.java:862)
at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:777)
at org.apache.kafka.streams.processor.internals.StreamThread.run(StreamThread.java:747)
16:13:43.109 WARN o.apache.kafka.streams.KafkaStreams - stream-client [block-processor-1fe1a9ba-e648-4bc0-a02d-78e567ba37aa] All stream threads have died.
'''
|
process
|
exception on kafka processing from time to time if i destroy the docker environment with everything including volumes whenever i process again a new downloaded chain i obtain this issue aldo borrero info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number aldo borrero info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number aldo borrero info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number info i e k s processors blockprocessor processing block number warn i e k s p b chaineventstransformer ignoring block already processed info i e k s processors blockprocessor processing block number error o a k s p i assignedstreamstasks stream thread failed to process stream task due to the following error java lang illegalstateexception block out of sequence expected received at io enkrypt kafka streams processors block chaineventstransformer ensuresequentialprocessing chaineventstransformer kt at io enkrypt kafka streams processors block chaineventstransformer transform chaineventstransformer kt at io enkrypt kafka streams processors block chaineventstransformer transform chaineventstransformer kt at org apache kafka streams kstream internals kstreamtransform kstreamtransformprocessor process kstreamtransform java at org apache kafka streams processor internals processornode process processornode java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java aldo borrero at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams kstream internals kstreampeek kstreampeekprocessor process kstreampeek java at org apache kafka streams processor internals processornode process processornode java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals sourcenode process sourcenode java at org apache kafka streams processor internals streamtask process streamtask java at org apache kafka streams processor internals assignedstreamstasks process assignedstreamstasks java at org apache kafka streams processor internals taskmanager process taskmanager java at org apache kafka streams processor internals streamthread runonce streamthread java at org apache kafka streams processor internals streamthread runloop streamthread java at org apache kafka streams processor internals streamthread run streamthread java error o a k s p internals streamthread stream thread encountered the following error during processing java lang illegalstateexception block out of sequence expected received at io enkrypt kafka streams processors block chaineventstransformer ensuresequentialprocessing chaineventstransformer kt at io enkrypt kafka streams processors block chaineventstransformer transform chaineventstransformer kt at io enkrypt kafka streams processors block chaineventstransformer transform chaineventstransformer kt at org apache kafka streams kstream internals kstreamtransform kstreamtransformprocessor process kstreamtransform java at org apache kafka streams processor internals processornode process processornode java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams kstream internals kstreampeek kstreampeekprocessor process kstreampeek java at org apache kafka streams processor internals processornode process processornode java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals sourcenode process sourcenode java at org apache kafka streams processor internals streamtask process streamtask java at org apache kafka streams processor internals assignedstreamstasks process assignedstreamstasks java at org apache kafka streams processor internals taskmanager process taskmanager java at org apache kafka streams processor internals streamthread runonce streamthread java at org apache kafka streams processor internals streamthread runloop streamthread java at org apache kafka streams processor internals streamthread run streamthread java aldo borrero error o a k s p i recordcollectorimpl task error sending record key xec xff xca timestamp to topic block processor chain events changelog due to org apache kafka common kafkaexception failing batch since transaction was aborted no more records will be sent and no more offsets will be recorded for this task error o a k s p internals streamtask task failed to close producer due to the following error org apache kafka streams errors streamsexception task abort sending since an error caught with a previous record key xec xff xca xda xee xaa xed value b timestamp to topic block processor chain events changelog due to org apache kafka common kafkaexception failing batch since transaction was aborted at org apache kafka streams processor internals recordcollectorimpl recordsenderror recordcollectorimpl java at org apache kafka streams processor internals recordcollectorimpl access recordcollectorimpl java at org apache kafka streams processor internals recordcollectorimpl oncompletion recordcollectorimpl java at org apache kafka clients producer kafkaproducer interceptorcallback oncompletion kafkaproducer java at org apache kafka clients producer internals producerbatch completefutureandfirecallbacks producerbatch java at org apache kafka clients producer internals producerbatch abort producerbatch java at org apache kafka clients producer internals recordaccumulator abortundrainedbatches recordaccumulator java at org apache kafka clients producer internals sender maybesendtransactionalrequest sender java at org apache kafka clients producer internals sender run sender java at org apache kafka clients producer internals sender run sender java at java lang thread run thread java caused by org apache kafka common kafkaexception failing batch since transaction was aborted common frames omitted exception in thread block processor streamthread java lang illegalstateexception block out of sequence expected received at io enkrypt kafka streams processors block chaineventstransformer ensuresequentialprocessing chaineventstransformer kt at io enkrypt kafka streams processors block chaineventstransformer transform chaineventstransformer kt at io enkrypt kafka streams processors block chaineventstransformer transform chaineventstransformer kt at org apache kafka streams kstream internals kstreamtransform kstreamtransformprocessor process kstreamtransform java at org apache kafka streams processor internals processornode process processornode java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams kstream internals kstreampeek kstreampeekprocessor process kstreampeek java at org apache kafka streams processor internals processornode process processornode java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java at org apache kafka streams processor internals processorcontextimpl forward processorcontextimpl java aldo borrero at org apache kafka streams processor internals sourcenode process sourcenode java at org apache kafka streams processor internals streamtask process streamtask java at org apache kafka streams processor internals assignedstreamstasks process assignedstreamstasks java at org apache kafka streams processor internals taskmanager process taskmanager java at org apache kafka streams processor internals streamthread runonce streamthread java at org apache kafka streams processor internals streamthread runloop streamthread java at org apache kafka streams processor internals streamthread run streamthread java warn o apache kafka streams kafkastreams stream client all stream threads have died
| 1
|
23,301
| 10,869,035,410
|
IssuesEvent
|
2019-11-15 06:14:16
|
nycbeardo/nycbeardo.github.io
|
https://api.github.com/repos/nycbeardo/nycbeardo.github.io
|
closed
|
CVE-2018-20677 (Medium) detected in bootstrap-3.3.4-3.3.4.min.js, bootstrap-3.3.4-3.3.4.js
|
security vulnerability
|
## CVE-2018-20677 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bootstrap-3.3.4-3.3.4.min.js</b>, <b>bootstrap-3.3.4-3.3.4.js</b></p></summary>
<p>
<details><summary><b>bootstrap-3.3.4-3.3.4.min.js</b></p></summary>
<p>Google-styled theme for Bootstrap.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/todc-bootstrap/3.3.4-3.3.4/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/todc-bootstrap/3.3.4-3.3.4/js/bootstrap.min.js</a></p>
<p>Path to vulnerable library: /nycbeardo.github.io/js/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.4-3.3.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>bootstrap-3.3.4-3.3.4.js</b></p></summary>
<p>Google-styled theme for Bootstrap.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/todc-bootstrap/3.3.4-3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/todc-bootstrap/3.3.4-3.3.4/js/bootstrap.js</a></p>
<p>Path to vulnerable library: /nycbeardo.github.io/js/bootstrap.js,/nycbeardo.github.io/js/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.4-3.3.4.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/nycbeardo/nycbeardo.github.io/commit/3d3541f32330e18b0a9a64e75d16d80a330780b3">3d3541f32330e18b0a9a64e75d16d80a330780b3</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677>CVE-2018-20677</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-20677 (Medium) detected in bootstrap-3.3.4-3.3.4.min.js, bootstrap-3.3.4-3.3.4.js - ## CVE-2018-20677 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bootstrap-3.3.4-3.3.4.min.js</b>, <b>bootstrap-3.3.4-3.3.4.js</b></p></summary>
<p>
<details><summary><b>bootstrap-3.3.4-3.3.4.min.js</b></p></summary>
<p>Google-styled theme for Bootstrap.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/todc-bootstrap/3.3.4-3.3.4/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/todc-bootstrap/3.3.4-3.3.4/js/bootstrap.min.js</a></p>
<p>Path to vulnerable library: /nycbeardo.github.io/js/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.4-3.3.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>bootstrap-3.3.4-3.3.4.js</b></p></summary>
<p>Google-styled theme for Bootstrap.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/todc-bootstrap/3.3.4-3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/todc-bootstrap/3.3.4-3.3.4/js/bootstrap.js</a></p>
<p>Path to vulnerable library: /nycbeardo.github.io/js/bootstrap.js,/nycbeardo.github.io/js/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.4-3.3.4.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/nycbeardo/nycbeardo.github.io/commit/3d3541f32330e18b0a9a64e75d16d80a330780b3">3d3541f32330e18b0a9a64e75d16d80a330780b3</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677>CVE-2018-20677</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in bootstrap min js bootstrap js cve medium severity vulnerability vulnerable libraries bootstrap min js bootstrap js bootstrap min js google styled theme for bootstrap library home page a href path to vulnerable library nycbeardo github io js bootstrap min js dependency hierarchy x bootstrap min js vulnerable library bootstrap js google styled theme for bootstrap library home page a href path to vulnerable library nycbeardo github io js bootstrap js nycbeardo github io js bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href vulnerability details in bootstrap before xss is possible in the affix configuration target property publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter step up your open source security game with whitesource
| 0
|
612,341
| 19,010,284,998
|
IssuesEvent
|
2021-11-23 08:30:10
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
Use a PackageRepository instance to represent the repository in PackageDescriptor
|
Type/Improvement Priority/Low Team/DevTools Area/ProjectAPI
|
**Description:**
https://github.com/ballerina-platform/ballerina-lang/pull/28857#discussion_r586005165
|
1.0
|
Use a PackageRepository instance to represent the repository in PackageDescriptor - **Description:**
https://github.com/ballerina-platform/ballerina-lang/pull/28857#discussion_r586005165
|
non_process
|
use a packagerepository instance to represent the repository in packagedescriptor description
| 0
|
10,666
| 13,455,325,010
|
IssuesEvent
|
2020-09-09 05:59:00
|
LighteningZero/LighteningSlide
|
https://api.github.com/repos/LighteningZero/LighteningSlide
|
reopened
|
[FEATURE]Markdown List
|
Extension High Priority Need Process enhancement
|
**Is your feature request related to a problem? Please describe.**
A List is required, e.g.
1. Hi
1. Hello
1. Top
- Hight
- Width
- [ ] Unmarked
- [x] marked
- [ ] Add a list
**Describe the solution you'd like**
add a new ext like `default.list.js`
**Additional context**
This must be solved before 1st Sept. 2020
|
1.0
|
[FEATURE]Markdown List - **Is your feature request related to a problem? Please describe.**
A List is required, e.g.
1. Hi
1. Hello
1. Top
- Hight
- Width
- [ ] Unmarked
- [x] marked
- [ ] Add a list
**Describe the solution you'd like**
add a new ext like `default.list.js`
**Additional context**
This must be solved before 1st Sept. 2020
|
process
|
markdown list is your feature request related to a problem please describe a list is required e g hi hello top hight width unmarked marked add a list describe the solution you d like add a new ext like default list js additional context this must be solved before sept
| 1
|
175,537
| 6,551,973,786
|
IssuesEvent
|
2017-09-05 16:30:01
|
UNIVALI-LITE/Portugol-Studio
|
https://api.github.com/repos/UNIVALI-LITE/Portugol-Studio
|
closed
|
Escolha dentro de escolha de tipos diferentes.
|
bug high-priority in-progress
|
_Bug relatado pelo Paulo Vieira e descoberto por um estudante dele durante a aula de algoritmos._
1. Você faz um primeiro escolha caso de um tipo (por exemplo inteiro)
2. Dentro do primeiro caso você bota outro escolha caso de outro tipo (por exemplo caracter), e bota quantos casos quiser.
3. Quando você volta para o primeiro escolha caso, e tenta botar outro caso. Ele da um erro falando que esperava o tipo do outro escolha caso.
Eu duvido que alguém tenha entendido a minha explicação, então como uma imagem vale mais que mil palavras segue a imagem do ocorrido:

Código:
http://pastebin.com/d4dGAhmG
Acho que vale investigar como o analisador sintático está empilhando a função escolha
|
1.0
|
Escolha dentro de escolha de tipos diferentes. - _Bug relatado pelo Paulo Vieira e descoberto por um estudante dele durante a aula de algoritmos._
1. Você faz um primeiro escolha caso de um tipo (por exemplo inteiro)
2. Dentro do primeiro caso você bota outro escolha caso de outro tipo (por exemplo caracter), e bota quantos casos quiser.
3. Quando você volta para o primeiro escolha caso, e tenta botar outro caso. Ele da um erro falando que esperava o tipo do outro escolha caso.
Eu duvido que alguém tenha entendido a minha explicação, então como uma imagem vale mais que mil palavras segue a imagem do ocorrido:

Código:
http://pastebin.com/d4dGAhmG
Acho que vale investigar como o analisador sintático está empilhando a função escolha
|
non_process
|
escolha dentro de escolha de tipos diferentes bug relatado pelo paulo vieira e descoberto por um estudante dele durante a aula de algoritmos você faz um primeiro escolha caso de um tipo por exemplo inteiro dentro do primeiro caso você bota outro escolha caso de outro tipo por exemplo caracter e bota quantos casos quiser quando você volta para o primeiro escolha caso e tenta botar outro caso ele da um erro falando que esperava o tipo do outro escolha caso eu duvido que alguém tenha entendido a minha explicação então como uma imagem vale mais que mil palavras segue a imagem do ocorrido código acho que vale investigar como o analisador sintático está empilhando a função escolha
| 0
|
4,914
| 7,787,377,955
|
IssuesEvent
|
2018-06-06 22:13:29
|
bazelbuild/rules_swift
|
https://api.github.com/repos/bazelbuild/rules_swift
|
opened
|
Configure CI
|
process
|
Set up CI to build all the examples on both Linux and Apple platforms. Once we have proper unit tests (#1), run those as well.
|
1.0
|
Configure CI - Set up CI to build all the examples on both Linux and Apple platforms. Once we have proper unit tests (#1), run those as well.
|
process
|
configure ci set up ci to build all the examples on both linux and apple platforms once we have proper unit tests run those as well
| 1
|
15,950
| 20,169,276,966
|
IssuesEvent
|
2022-02-10 08:55:43
|
Jeffail/benthos
|
https://api.github.com/repos/Jeffail/benthos
|
opened
|
Consider clearing errors at the beginning of a try block, or showing a linting error
|
enhancement processors v4
|
Right now nested `try` blocks within `catch` will never trigger:
```yaml
pipeline:
processors:
- try:
- resource: foo
- catch:
- try:
- resource: bar
- resource: baz
```
So if `foo` fails neither `bar` nor `baz` would execute. This is because the error context is carried into the processors of the catch block and aren't cleared until the end so that the error message can be referrenced. A work around is to add a nested catch before the try block:
```yaml
pipeline:
processors:
- try:
- resource: foo
- catch:
- catch: []
- try:
- resource: bar
- resource: baz
```
This clears the error before the `try` block but looks odd and ugly.
|
1.0
|
Consider clearing errors at the beginning of a try block, or showing a linting error - Right now nested `try` blocks within `catch` will never trigger:
```yaml
pipeline:
processors:
- try:
- resource: foo
- catch:
- try:
- resource: bar
- resource: baz
```
So if `foo` fails neither `bar` nor `baz` would execute. This is because the error context is carried into the processors of the catch block and aren't cleared until the end so that the error message can be referrenced. A work around is to add a nested catch before the try block:
```yaml
pipeline:
processors:
- try:
- resource: foo
- catch:
- catch: []
- try:
- resource: bar
- resource: baz
```
This clears the error before the `try` block but looks odd and ugly.
|
process
|
consider clearing errors at the beginning of a try block or showing a linting error right now nested try blocks within catch will never trigger yaml pipeline processors try resource foo catch try resource bar resource baz so if foo fails neither bar nor baz would execute this is because the error context is carried into the processors of the catch block and aren t cleared until the end so that the error message can be referrenced a work around is to add a nested catch before the try block yaml pipeline processors try resource foo catch catch try resource bar resource baz this clears the error before the try block but looks odd and ugly
| 1
|
15,713
| 19,848,804,290
|
IssuesEvent
|
2022-01-21 09:56:43
|
ooi-data/CE06ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample
|
https://api.github.com/repos/ooi-data/CE06ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T09:56:43.128543.
## Details
Flow name: `CE06ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T09:56:43.128543.
## Details
Flow name: `CE06ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered cspp flort sample task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
| 1
|
260,769
| 27,784,716,889
|
IssuesEvent
|
2023-03-17 01:31:18
|
tomdgl397/juice-shop
|
https://api.github.com/repos/tomdgl397/juice-shop
|
opened
|
CVE-2023-28155 (Medium) detected in request-2.88.2.tgz
|
Mend: dependency security vulnerability
|
## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.2.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/request/package.json,/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- :x: **request-2.88.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tomdgl397/juice-shop/commit/24e92478a2e956132cc96bf9e3bc8ca7fecf375d">24e92478a2e956132cc96bf9e3bc8ca7fecf375d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-28155 (Medium) detected in request-2.88.2.tgz - ## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.2.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/request/package.json,/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- :x: **request-2.88.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tomdgl397/juice-shop/commit/24e92478a2e956132cc96bf9e3bc8ca7fecf375d">24e92478a2e956132cc96bf9e3bc8ca7fecf375d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in request tgz cve medium severity vulnerability vulnerable library request tgz simplified http request client library home page a href path to dependency file package json path to vulnerable library node modules request package json node modules request package json dependency hierarchy x request tgz vulnerable library found in head commit a href vulnerability details unsupported when assigned the request package through for node js allows a bypass of ssrf mitigations via an attacker controller server that does a cross protocol redirect http to https or https to http note this vulnerability only affects products that are no longer supported by the maintainer publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
| 0
|
14,192
| 17,095,621,498
|
IssuesEvent
|
2021-07-09 01:47:49
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
processing extension error when starting qgis zurich on m1 mac os
|
Bug Feedback MacOS Processing stale
|
Here the error message when launching qgis 3.18.1 on a m1 mac (all is fine with 3.16 version):
UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 195: ordinal not in range(128)
Traceback (most recent call last):
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 335, in _startPlugin
plugins[packageName] = package.classFactory(iface)
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/plugins/processing/__init__.py", line 51, in classFactory
return ProcessingPlugin(iface)
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/plugins/processing/ProcessingPlugin.py", line 168, in __init__
self.initProcessing()
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/plugins/processing/ProcessingPlugin.py", line 173, in initProcessing
Processing.initialize()
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/plugins/processing/core/Processing.py", line 134, in initialize
RenderingStyles.loadStyles()
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/plugins/processing/gui/RenderingStyles.py", line 45, in loadStyles
line = lines.readline().strip('\n')
File "/Applications/QGIS.app/Contents/MacOS/lib/python3.8/encodings/ascii.py", line 26, in decode
return codecs.ascii_decode(input, self.errors)[0]
UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 195: ordinal not in range(128)
|
1.0
|
processing extension error when starting qgis zurich on m1 mac os - Here the error message when launching qgis 3.18.1 on a m1 mac (all is fine with 3.16 version):
UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 195: ordinal not in range(128)
Traceback (most recent call last):
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 335, in _startPlugin
plugins[packageName] = package.classFactory(iface)
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/plugins/processing/__init__.py", line 51, in classFactory
return ProcessingPlugin(iface)
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/plugins/processing/ProcessingPlugin.py", line 168, in __init__
self.initProcessing()
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/plugins/processing/ProcessingPlugin.py", line 173, in initProcessing
Processing.initialize()
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/plugins/processing/core/Processing.py", line 134, in initialize
RenderingStyles.loadStyles()
File "/Applications/QGIS.app/Contents/MacOS/../Resources/python/plugins/processing/gui/RenderingStyles.py", line 45, in loadStyles
line = lines.readline().strip('\n')
File "/Applications/QGIS.app/Contents/MacOS/lib/python3.8/encodings/ascii.py", line 26, in decode
return codecs.ascii_decode(input, self.errors)[0]
UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 195: ordinal not in range(128)
|
process
|
processing extension error when starting qgis zurich on mac os here the error message when launching qgis on a mac all is fine with version unicodedecodeerror ascii codec can t decode byte in position ordinal not in range traceback most recent call last file applications qgis app contents macos resources python qgis utils py line in startplugin plugins package classfactory iface file applications qgis app contents macos resources python plugins processing init py line in classfactory return processingplugin iface file applications qgis app contents macos resources python plugins processing processingplugin py line in init self initprocessing file applications qgis app contents macos resources python plugins processing processingplugin py line in initprocessing processing initialize file applications qgis app contents macos resources python plugins processing core processing py line in initialize renderingstyles loadstyles file applications qgis app contents macos resources python plugins processing gui renderingstyles py line in loadstyles line lines readline strip n file applications qgis app contents macos lib encodings ascii py line in decode return codecs ascii decode input self errors unicodedecodeerror ascii codec can t decode byte in position ordinal not in range
| 1
|
115
| 2,546,506,348
|
IssuesEvent
|
2015-01-30 00:38:40
|
GsDevKit/gsApplicationTools
|
https://api.github.com/repos/GsDevKit/gsApplicationTools
|
opened
|
stop/startGemServer relies upon env var GEMSTONE_LOGDIR
|
in process
|
makes debugging server errors pretty difficult:)
|
1.0
|
stop/startGemServer relies upon env var GEMSTONE_LOGDIR - makes debugging server errors pretty difficult:)
|
process
|
stop startgemserver relies upon env var gemstone logdir makes debugging server errors pretty difficult
| 1
|
4,152
| 7,103,404,847
|
IssuesEvent
|
2018-01-16 04:48:41
|
PaddlePaddle/models
|
https://api.github.com/repos/PaddlePaddle/models
|
closed
|
Implement Xception
|
model in process
|
Xception is a popular model for image classification and object detection.
https://arxiv.org/pdf/1610.02357.pdf
|
1.0
|
Implement Xception - Xception is a popular model for image classification and object detection.
https://arxiv.org/pdf/1610.02357.pdf
|
process
|
implement xception xception is a popular model for image classification and object detection
| 1
|
17,639
| 23,461,835,606
|
IssuesEvent
|
2022-08-16 13:42:40
|
hashgraph/hedera-json-rpc-relay
|
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
|
closed
|
Add k6 performance test suite
|
enhancement P3 process
|
### Problem
Currently the relay has UTs, some integration tests and now acceptances tests.
However, we don't have a dedicated performance tests suite to extract performance numbers
### Solution
Add support in line with our existing API testing frameworks used.
In the mirror Node we found success with k6 so a similar approach is recommend
- Add documentation for test run similar to https://github.com/hashgraph/hedera-mirror-node/tree/main/hedera-mirror-test/k6
- Add test scripts similar to https://github.com/hashgraph/hedera-mirror-node/tree/main/hedera-mirror-test/k6/src/web3
### Alternatives
_No response_
|
1.0
|
Add k6 performance test suite - ### Problem
Currently the relay has UTs, some integration tests and now acceptances tests.
However, we don't have a dedicated performance tests suite to extract performance numbers
### Solution
Add support in line with our existing API testing frameworks used.
In the mirror Node we found success with k6 so a similar approach is recommend
- Add documentation for test run similar to https://github.com/hashgraph/hedera-mirror-node/tree/main/hedera-mirror-test/k6
- Add test scripts similar to https://github.com/hashgraph/hedera-mirror-node/tree/main/hedera-mirror-test/k6/src/web3
### Alternatives
_No response_
|
process
|
add performance test suite problem currently the relay has uts some integration tests and now acceptances tests however we don t have a dedicated performance tests suite to extract performance numbers solution add support in line with our existing api testing frameworks used in the mirror node we found success with so a similar approach is recommend add documentation for test run similar to add test scripts similar to alternatives no response
| 1
|
21,754
| 30,272,923,844
|
IssuesEvent
|
2023-07-07 16:52:46
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Creating/editing database connection with SSH tunnel fails silently, when database port is not filled out
|
Type:Bug Priority:P1 .Backend First Experience/Setup .Regression Administration/Databases .Team/QueryProcessor :hammer_and_wrench:
|
**Describe the bug**
I'm attempting to edit a MySQL connection to use an SSH tunnel through a bastion server to Google Cloud SQL. The tunnel uses SSH key authentication. The changes fail silently, returning me to the edit screen with the tunnel options reset, while spitting out a 400 error to the troubleshooting logs.
The same SSH tunnel works ok for other, self-hosted databases connected to Metabase and also from my desktop machine. I manually cloned the details from another connection in the metabase_database table, adjusted them to suit, and the actual tunnelled connection appears to work fine.
**Logs**
I've blanked out the server and key details.
```
[62f0d42b-8177-4d38-8dc3-a4d7a5ba8e72] 2022-02-25T18:12:57-08:00 WARN org.apache.sshd.client.keyverifier.AcceptAllServerKeyVerifier ParameterizedMessage[messagePattern=Server at {} presented unverified {} key: {}, stringArgs=[XXX, EC, XXX], throwable=null]
[62f0d42b-8177-4d38-8dc3-a4d7a5ba8e72] 2022-02-25T18:12:57-08:00 ERROR metabase.driver.util Database connection error
java.lang.NullPointerException
at clojure.lang.RT.intCast(RT.java:1221)
at metabase.util.ssh$start_ssh_tunnel_BANG_.invokeStatic(ssh.clj:63)
at metabase.util.ssh$start_ssh_tunnel_BANG_.invoke(ssh.clj:49)
at metabase.util.ssh$include_ssh_tunnel_BANG_.invokeStatic(ssh.clj:89)
at metabase.util.ssh$include_ssh_tunnel_BANG_.invoke(ssh.clj:83)
at metabase.util.ssh$fn__33886.invokeStatic(ssh.clj:115)
at metabase.util.ssh$fn__33886.invoke(ssh.clj:105)
at clojure.lang.MultiFn.invoke(MultiFn.java:234)
at metabase.driver.sql_jdbc.connection$details__GT_connection_spec_for_testing_connection.invokeStatic(connection.clj:239)
at metabase.driver.sql_jdbc.connection$details__GT_connection_spec_for_testing_connection.invoke(connection.clj:235)
at metabase.driver.sql_jdbc.connection$can_connect_QMARK_.invokeStatic(connection.clj:253)
at metabase.driver.sql_jdbc.connection$can_connect_QMARK_.invoke(connection.clj:249)
at metabase.driver.sql_jdbc$fn__81966.invokeStatic(sql_jdbc.clj:35)
at metabase.driver.sql_jdbc$fn__81966.invoke(sql_jdbc.clj:33)
at metabase.driver.mysql$fn__79845.invokeStatic(mysql.clj:71)
at metabase.driver.mysql$fn__79845.invoke(mysql.clj:67)
at clojure.lang.MultiFn.invoke(MultiFn.java:234)
at metabase.driver.util$can_connect_with_details_QMARK_$fn__19263.invoke(util.clj:48)
at metabase.util$do_with_timeout$fn__6209.invoke(util.clj:352)
at clojure.core$binding_conveyor_fn$fn__5772.invoke(core.clj:2034)
at clojure.lang.AFn.call(AFn.java:18)
at java.base/java.util.concurrent.FutureTask.run(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.base/java.lang.Thread.run(Unknown Source)
[62f0d42b-8177-4d38-8dc3-a4d7a5ba8e72] 2022-02-25T18:12:57-08:00 ERROR metabase.api.database Cannot connect to Database
java.lang.NullPointerException
at java.base/java.util.regex.Matcher.getTextLength(Unknown Source)
at java.base/java.util.regex.Matcher.reset(Unknown Source)
at java.base/java.util.regex.Matcher.<init>(Unknown Source)
at java.base/java.util.regex.Pattern.matcher(Unknown Source)
at clojure.core$re_matcher.invokeStatic(core.clj:4845)
at clojure.core$re_matches.invokeStatic(core.clj:4875)
at clojure.core$re_matches.invoke(core.clj:4875)
at metabase.driver.mysql$fn__79855.invokeStatic(mysql.clj:117)
at metabase.driver.mysql$fn__79855.invoke(mysql.clj:115)
at clojure.lang.MultiFn.invoke(MultiFn.java:234)
at metabase.driver.util$can_connect_with_details_QMARK_.invokeStatic(util.clj:53)
at metabase.driver.util$can_connect_with_details_QMARK_.doInvoke(util.clj:36)
at clojure.lang.RestFn.invoke(RestFn.java:442)
at metabase.api.database$test_database_connection.invokeStatic(database.clj:431)
at metabase.api.database$test_database_connection.doInvoke(database.clj:421)
at clojure.lang.RestFn.invoke(RestFn.java:425)
at metabase.api.database$fn__73371.invokeStatic(database.clj:587)
at metabase.api.database$fn__73371.invoke(database.clj:566)
at compojure.core$wrap_response$fn__32271.invoke(core.clj:160)
at compojure.core$wrap_route_middleware$fn__32255.invoke(core.clj:132)
at compojure.core$wrap_route_info$fn__32260.invoke(core.clj:139)
at compojure.core$wrap_route_matches$fn__32264.invoke(core.clj:151)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$wrap_route_matches$fn__32264.invoke(core.clj:153)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283.invoke(core.clj:200)
at metabase.server.middleware.auth$enforce_authentication$fn__65343.invoke(auth.clj:14)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283.invoke(core.clj:200)
at compojure.core$make_context$handler__32311.invoke(core.clj:289)
at compojure.core$make_context$fn__32315.invoke(core.clj:299)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at metabase.api.routes$fn__76468$fn__76471.invoke(routes.clj:56)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283.invoke(core.clj:200)
at clojure.lang.AFn.applyToHelper(AFn.java:160)
at clojure.lang.AFn.applyTo(AFn.java:144)
at clojure.core$apply.invokeStatic(core.clj:667)
at clojure.core$apply.invoke(core.clj:662)
at metabase.server.routes$fn__76613$fn__76614.doInvoke(routes.clj:57)
at clojure.lang.RestFn.invoke(RestFn.java:436)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283.invoke(core.clj:200)
at compojure.core$make_context$handler__32311.invoke(core.clj:289)
at compojure.core$make_context$fn__32315.invoke(core.clj:299)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$wrap_route_matches$fn__32264.invoke(core.clj:153)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$wrap_route_matches$fn__32264.invoke(core.clj:153)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$wrap_route_matches$fn__32264.invoke(core.clj:153)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at metabase.server.routes$fn__76601$fn__76603.invoke(routes.clj:41)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283.invoke(core.clj:200)
at metabase.server.middleware.exceptions$catch_uncaught_exceptions$fn__73596.invoke(exceptions.clj:98)
at metabase.server.middleware.exceptions$catch_api_exceptions$fn__73593.invoke(exceptions.clj:86)
at metabase.server.middleware.log$log_api_call$fn__77079$fn__77080$fn__77081.invoke(log.clj:209)
at metabase.driver.sql_jdbc.execute.diagnostic$do_with_diagnostic_info.invokeStatic(diagnostic.clj:15)
at metabase.driver.sql_jdbc.execute.diagnostic$do_with_diagnostic_info.invoke(diagnostic.clj:9)
at metabase.server.middleware.log$log_api_call$fn__77079$fn__77080.invoke(log.clj:201)
at toucan.db$_do_with_call_counting.invokeStatic(db.clj:216)
at toucan.db$_do_with_call_counting.invoke(db.clj:209)
at metabase.server.middleware.log$log_api_call$fn__77079.invoke(log.clj:200)
at metabase.server.middleware.browser_cookie$ensure_browser_id_cookie$fn__80432.invoke(browser_cookie.clj:38)
at metabase.server.middleware.security$add_security_headers$fn__55203.invoke(security.clj:148)
at metabase.server.middleware.json$wrap_json_body$fn__79587.invoke(json.clj:62)
at metabase.server.middleware.json$wrap_streamed_json_response$fn__79605.invoke(json.clj:98)
at metabase.server.middleware.offset_paging$handle_paging$fn__55227.invoke(offset_paging.clj:42)
at ring.middleware.keyword_params$wrap_keyword_params$fn__80699.invoke(keyword_params.clj:55)
at ring.middleware.params$wrap_params$fn__80715.invoke(params.clj:69)
at metabase.server.middleware.misc$maybe_set_site_url$fn__34377.invoke(misc.clj:59)
at metabase.server.middleware.session$bind_current_user$fn__45720$fn__45721.invoke(session.clj:262)
at metabase.server.middleware.session$do_with_current_user.invokeStatic(session.clj:243)
at metabase.server.middleware.session$do_with_current_user.invoke(session.clj:235)
at metabase.server.middleware.session$bind_current_user$fn__45720.invoke(session.clj:261)
at metabase.server.middleware.session$wrap_current_user_info$fn__45707.invoke(session.clj:221)
at metabase.server.middleware.session$wrap_session_id$fn__45693.invoke(session.clj:167)
at metabase.server.middleware.auth$wrap_api_key$fn__65351.invoke(auth.clj:27)
at ring.middleware.cookies$wrap_cookies$fn__80619.invoke(cookies.clj:216)
at metabase.server.middleware.misc$add_content_type$fn__34360.invoke(misc.clj:27)
at metabase.server.middleware.misc$disable_streaming_buffering$fn__34385.invoke(misc.clj:76)
at ring.middleware.gzip$wrap_gzip$fn__80661.invoke(gzip.clj:86)
at metabase.server.middleware.misc$bind_request$fn__34388.invoke(misc.clj:93)
at metabase.server.middleware.ssl$redirect_to_https_middleware$fn__80448.invoke(ssl.clj:48)
at metabase.server$async_proxy_handler$fn__76852.invoke(server.clj:73)
at metabase.server.proxy$org.eclipse.jetty.server.handler.AbstractHandler$ff19274a.handle(Unknown Source)
at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127)
at org.eclipse.jetty.server.Server.handle(Server.java:516)
at org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:388)
at org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:633)
at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:380)
at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:277)
at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311)
at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105)
at org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104)
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:338)
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:315)
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173)
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:131)
at org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:386)
at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)
at org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)
at java.base/java.lang.Thread.run(Unknown Source)
[62f0d42b-8177-4d38-8dc3-a4d7a5ba8e72] 2022-02-25T18:12:57-08:00 DEBUG metabase.server.middleware.log PUT /api/database/12 400 437.5 ms (1 DB calls)
{:valid false, :dbname nil, :message nil}
```
**To Reproduce**
n/a
**Expected behavior**
Either:
- For the SSH tunnel to successfully connect, or;
- For a connection error to be reported in-page, retaining the details that were submitted.
**Screenshots**
n/a
**Information about your Metabase Installation:**
Running the Docker image on a Kubernetes cluster.
```
{
"browser-info": {
"language": "en-US",
"platform": "MacIntel",
"userAgent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/98.0.4758.109 Safari/537.36",
"vendor": "Google Inc."
},
"system-info": {
"file.encoding": "UTF-8",
"java.runtime.name": "OpenJDK Runtime Environment",
"java.runtime.version": "11.0.14.1+1",
"java.vendor": "Eclipse Adoptium",
"java.vendor.url": "https://adoptium.net/",
"java.version": "11.0.14.1",
"java.vm.name": "OpenJDK 64-Bit Server VM",
"java.vm.version": "11.0.14.1+1",
"os.name": "Linux",
"os.version": "5.4.144+",
"user.language": "en",
"user.timezone": "US/Pacific"
},
"metabase-info": {
"databases": [
"googleanalytics",
"mysql",
"postgres"
],
"hosting-env": "unknown",
"application-database": "postgres",
"application-database-details": {
"database": {
"name": "PostgreSQL",
"version": "9.6.24"
},
"jdbc-driver": {
"name": "PostgreSQL JDBC Driver",
"version": "42.2.23"
}
},
"run-mode": "prod",
"version": {
"date": "2022-02-17",
"tag": "v0.42.1",
"branch": "release-x.42.x",
"hash": "629f4de"
},
"settings": {
"report-timezone": "America/Vancouver"
}
}
}
```
**Severity**
Blocking usage, prior to direct database manipulation. Could not connect to significant service database.
**Additional context**
n/a
|
1.0
|
Creating/editing database connection with SSH tunnel fails silently, when database port is not filled out - **Describe the bug**
I'm attempting to edit a MySQL connection to use an SSH tunnel through a bastion server to Google Cloud SQL. The tunnel uses SSH key authentication. The changes fail silently, returning me to the edit screen with the tunnel options reset, while spitting out a 400 error to the troubleshooting logs.
The same SSH tunnel works ok for other, self-hosted databases connected to Metabase and also from my desktop machine. I manually cloned the details from another connection in the metabase_database table, adjusted them to suit, and the actual tunnelled connection appears to work fine.
**Logs**
I've blanked out the server and key details.
```
[62f0d42b-8177-4d38-8dc3-a4d7a5ba8e72] 2022-02-25T18:12:57-08:00 WARN org.apache.sshd.client.keyverifier.AcceptAllServerKeyVerifier ParameterizedMessage[messagePattern=Server at {} presented unverified {} key: {}, stringArgs=[XXX, EC, XXX], throwable=null]
[62f0d42b-8177-4d38-8dc3-a4d7a5ba8e72] 2022-02-25T18:12:57-08:00 ERROR metabase.driver.util Database connection error
java.lang.NullPointerException
at clojure.lang.RT.intCast(RT.java:1221)
at metabase.util.ssh$start_ssh_tunnel_BANG_.invokeStatic(ssh.clj:63)
at metabase.util.ssh$start_ssh_tunnel_BANG_.invoke(ssh.clj:49)
at metabase.util.ssh$include_ssh_tunnel_BANG_.invokeStatic(ssh.clj:89)
at metabase.util.ssh$include_ssh_tunnel_BANG_.invoke(ssh.clj:83)
at metabase.util.ssh$fn__33886.invokeStatic(ssh.clj:115)
at metabase.util.ssh$fn__33886.invoke(ssh.clj:105)
at clojure.lang.MultiFn.invoke(MultiFn.java:234)
at metabase.driver.sql_jdbc.connection$details__GT_connection_spec_for_testing_connection.invokeStatic(connection.clj:239)
at metabase.driver.sql_jdbc.connection$details__GT_connection_spec_for_testing_connection.invoke(connection.clj:235)
at metabase.driver.sql_jdbc.connection$can_connect_QMARK_.invokeStatic(connection.clj:253)
at metabase.driver.sql_jdbc.connection$can_connect_QMARK_.invoke(connection.clj:249)
at metabase.driver.sql_jdbc$fn__81966.invokeStatic(sql_jdbc.clj:35)
at metabase.driver.sql_jdbc$fn__81966.invoke(sql_jdbc.clj:33)
at metabase.driver.mysql$fn__79845.invokeStatic(mysql.clj:71)
at metabase.driver.mysql$fn__79845.invoke(mysql.clj:67)
at clojure.lang.MultiFn.invoke(MultiFn.java:234)
at metabase.driver.util$can_connect_with_details_QMARK_$fn__19263.invoke(util.clj:48)
at metabase.util$do_with_timeout$fn__6209.invoke(util.clj:352)
at clojure.core$binding_conveyor_fn$fn__5772.invoke(core.clj:2034)
at clojure.lang.AFn.call(AFn.java:18)
at java.base/java.util.concurrent.FutureTask.run(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
at java.base/java.lang.Thread.run(Unknown Source)
[62f0d42b-8177-4d38-8dc3-a4d7a5ba8e72] 2022-02-25T18:12:57-08:00 ERROR metabase.api.database Cannot connect to Database
java.lang.NullPointerException
at java.base/java.util.regex.Matcher.getTextLength(Unknown Source)
at java.base/java.util.regex.Matcher.reset(Unknown Source)
at java.base/java.util.regex.Matcher.<init>(Unknown Source)
at java.base/java.util.regex.Pattern.matcher(Unknown Source)
at clojure.core$re_matcher.invokeStatic(core.clj:4845)
at clojure.core$re_matches.invokeStatic(core.clj:4875)
at clojure.core$re_matches.invoke(core.clj:4875)
at metabase.driver.mysql$fn__79855.invokeStatic(mysql.clj:117)
at metabase.driver.mysql$fn__79855.invoke(mysql.clj:115)
at clojure.lang.MultiFn.invoke(MultiFn.java:234)
at metabase.driver.util$can_connect_with_details_QMARK_.invokeStatic(util.clj:53)
at metabase.driver.util$can_connect_with_details_QMARK_.doInvoke(util.clj:36)
at clojure.lang.RestFn.invoke(RestFn.java:442)
at metabase.api.database$test_database_connection.invokeStatic(database.clj:431)
at metabase.api.database$test_database_connection.doInvoke(database.clj:421)
at clojure.lang.RestFn.invoke(RestFn.java:425)
at metabase.api.database$fn__73371.invokeStatic(database.clj:587)
at metabase.api.database$fn__73371.invoke(database.clj:566)
at compojure.core$wrap_response$fn__32271.invoke(core.clj:160)
at compojure.core$wrap_route_middleware$fn__32255.invoke(core.clj:132)
at compojure.core$wrap_route_info$fn__32260.invoke(core.clj:139)
at compojure.core$wrap_route_matches$fn__32264.invoke(core.clj:151)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$wrap_route_matches$fn__32264.invoke(core.clj:153)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283.invoke(core.clj:200)
at metabase.server.middleware.auth$enforce_authentication$fn__65343.invoke(auth.clj:14)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283.invoke(core.clj:200)
at compojure.core$make_context$handler__32311.invoke(core.clj:289)
at compojure.core$make_context$fn__32315.invoke(core.clj:299)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$make_context$fn__32315.invoke(core.clj:300)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at metabase.api.routes$fn__76468$fn__76471.invoke(routes.clj:56)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283.invoke(core.clj:200)
at clojure.lang.AFn.applyToHelper(AFn.java:160)
at clojure.lang.AFn.applyTo(AFn.java:144)
at clojure.core$apply.invokeStatic(core.clj:667)
at clojure.core$apply.invoke(core.clj:662)
at metabase.server.routes$fn__76613$fn__76614.doInvoke(routes.clj:57)
at clojure.lang.RestFn.invoke(RestFn.java:436)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283.invoke(core.clj:200)
at compojure.core$make_context$handler__32311.invoke(core.clj:289)
at compojure.core$make_context$fn__32315.invoke(core.clj:299)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$wrap_route_matches$fn__32264.invoke(core.clj:153)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$wrap_route_matches$fn__32264.invoke(core.clj:153)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at compojure.core$wrap_route_matches$fn__32264.invoke(core.clj:153)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283$f__32284$respond_SINGLEQUOTE___32285.invoke(core.clj:197)
at metabase.server.routes$fn__76601$fn__76603.invoke(routes.clj:41)
at compojure.core$routes$fn__32283$f__32284.invoke(core.clj:198)
at compojure.core$routes$fn__32283.invoke(core.clj:200)
at metabase.server.middleware.exceptions$catch_uncaught_exceptions$fn__73596.invoke(exceptions.clj:98)
at metabase.server.middleware.exceptions$catch_api_exceptions$fn__73593.invoke(exceptions.clj:86)
at metabase.server.middleware.log$log_api_call$fn__77079$fn__77080$fn__77081.invoke(log.clj:209)
at metabase.driver.sql_jdbc.execute.diagnostic$do_with_diagnostic_info.invokeStatic(diagnostic.clj:15)
at metabase.driver.sql_jdbc.execute.diagnostic$do_with_diagnostic_info.invoke(diagnostic.clj:9)
at metabase.server.middleware.log$log_api_call$fn__77079$fn__77080.invoke(log.clj:201)
at toucan.db$_do_with_call_counting.invokeStatic(db.clj:216)
at toucan.db$_do_with_call_counting.invoke(db.clj:209)
at metabase.server.middleware.log$log_api_call$fn__77079.invoke(log.clj:200)
at metabase.server.middleware.browser_cookie$ensure_browser_id_cookie$fn__80432.invoke(browser_cookie.clj:38)
at metabase.server.middleware.security$add_security_headers$fn__55203.invoke(security.clj:148)
at metabase.server.middleware.json$wrap_json_body$fn__79587.invoke(json.clj:62)
at metabase.server.middleware.json$wrap_streamed_json_response$fn__79605.invoke(json.clj:98)
at metabase.server.middleware.offset_paging$handle_paging$fn__55227.invoke(offset_paging.clj:42)
at ring.middleware.keyword_params$wrap_keyword_params$fn__80699.invoke(keyword_params.clj:55)
at ring.middleware.params$wrap_params$fn__80715.invoke(params.clj:69)
at metabase.server.middleware.misc$maybe_set_site_url$fn__34377.invoke(misc.clj:59)
at metabase.server.middleware.session$bind_current_user$fn__45720$fn__45721.invoke(session.clj:262)
at metabase.server.middleware.session$do_with_current_user.invokeStatic(session.clj:243)
at metabase.server.middleware.session$do_with_current_user.invoke(session.clj:235)
at metabase.server.middleware.session$bind_current_user$fn__45720.invoke(session.clj:261)
at metabase.server.middleware.session$wrap_current_user_info$fn__45707.invoke(session.clj:221)
at metabase.server.middleware.session$wrap_session_id$fn__45693.invoke(session.clj:167)
at metabase.server.middleware.auth$wrap_api_key$fn__65351.invoke(auth.clj:27)
at ring.middleware.cookies$wrap_cookies$fn__80619.invoke(cookies.clj:216)
at metabase.server.middleware.misc$add_content_type$fn__34360.invoke(misc.clj:27)
at metabase.server.middleware.misc$disable_streaming_buffering$fn__34385.invoke(misc.clj:76)
at ring.middleware.gzip$wrap_gzip$fn__80661.invoke(gzip.clj:86)
at metabase.server.middleware.misc$bind_request$fn__34388.invoke(misc.clj:93)
at metabase.server.middleware.ssl$redirect_to_https_middleware$fn__80448.invoke(ssl.clj:48)
at metabase.server$async_proxy_handler$fn__76852.invoke(server.clj:73)
at metabase.server.proxy$org.eclipse.jetty.server.handler.AbstractHandler$ff19274a.handle(Unknown Source)
at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127)
at org.eclipse.jetty.server.Server.handle(Server.java:516)
at org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:388)
at org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:633)
at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:380)
at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:277)
at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311)
at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105)
at org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104)
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:338)
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:315)
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173)
at org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:131)
at org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:386)
at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)
at org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)
at java.base/java.lang.Thread.run(Unknown Source)
[62f0d42b-8177-4d38-8dc3-a4d7a5ba8e72] 2022-02-25T18:12:57-08:00 DEBUG metabase.server.middleware.log PUT /api/database/12 400 437.5 ms (1 DB calls)
{:valid false, :dbname nil, :message nil}
```
**To Reproduce**
n/a
**Expected behavior**
Either:
- For the SSH tunnel to successfully connect, or;
- For a connection error to be reported in-page, retaining the details that were submitted.
**Screenshots**
n/a
**Information about your Metabase Installation:**
Running the Docker image on a Kubernetes cluster.
```
{
"browser-info": {
"language": "en-US",
"platform": "MacIntel",
"userAgent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/98.0.4758.109 Safari/537.36",
"vendor": "Google Inc."
},
"system-info": {
"file.encoding": "UTF-8",
"java.runtime.name": "OpenJDK Runtime Environment",
"java.runtime.version": "11.0.14.1+1",
"java.vendor": "Eclipse Adoptium",
"java.vendor.url": "https://adoptium.net/",
"java.version": "11.0.14.1",
"java.vm.name": "OpenJDK 64-Bit Server VM",
"java.vm.version": "11.0.14.1+1",
"os.name": "Linux",
"os.version": "5.4.144+",
"user.language": "en",
"user.timezone": "US/Pacific"
},
"metabase-info": {
"databases": [
"googleanalytics",
"mysql",
"postgres"
],
"hosting-env": "unknown",
"application-database": "postgres",
"application-database-details": {
"database": {
"name": "PostgreSQL",
"version": "9.6.24"
},
"jdbc-driver": {
"name": "PostgreSQL JDBC Driver",
"version": "42.2.23"
}
},
"run-mode": "prod",
"version": {
"date": "2022-02-17",
"tag": "v0.42.1",
"branch": "release-x.42.x",
"hash": "629f4de"
},
"settings": {
"report-timezone": "America/Vancouver"
}
}
}
```
**Severity**
Blocking usage, prior to direct database manipulation. Could not connect to significant service database.
**Additional context**
n/a
|
process
|
creating editing database connection with ssh tunnel fails silently when database port is not filled out describe the bug i m attempting to edit a mysql connection to use an ssh tunnel through a bastion server to google cloud sql the tunnel uses ssh key authentication the changes fail silently returning me to the edit screen with the tunnel options reset while spitting out a error to the troubleshooting logs the same ssh tunnel works ok for other self hosted databases connected to metabase and also from my desktop machine i manually cloned the details from another connection in the metabase database table adjusted them to suit and the actual tunnelled connection appears to work fine logs i ve blanked out the server and key details warn org apache sshd client keyverifier acceptallserverkeyverifier parameterizedmessage throwable null error metabase driver util database connection error java lang nullpointerexception at clojure lang rt intcast rt java at metabase util ssh start ssh tunnel bang invokestatic ssh clj at metabase util ssh start ssh tunnel bang invoke ssh clj at metabase util ssh include ssh tunnel bang invokestatic ssh clj at metabase util ssh include ssh tunnel bang invoke ssh clj at metabase util ssh fn invokestatic ssh clj at metabase util ssh fn invoke ssh clj at clojure lang multifn invoke multifn java at metabase driver sql jdbc connection details gt connection spec for testing connection invokestatic connection clj at metabase driver sql jdbc connection details gt connection spec for testing connection invoke connection clj at metabase driver sql jdbc connection can connect qmark invokestatic connection clj at metabase driver sql jdbc connection can connect qmark invoke connection clj at metabase driver sql jdbc fn invokestatic sql jdbc clj at metabase driver sql jdbc fn invoke sql jdbc clj at metabase driver mysql fn invokestatic mysql clj at metabase driver mysql fn invoke mysql clj at clojure lang multifn invoke multifn java at metabase driver util can connect with details qmark fn invoke util clj at metabase util do with timeout fn invoke util clj at clojure core binding conveyor fn fn invoke core clj at clojure lang afn call afn java at java base java util concurrent futuretask run unknown source at java base java util concurrent threadpoolexecutor runworker unknown source at java base java util concurrent threadpoolexecutor worker run unknown source at java base java lang thread run unknown source error metabase api database cannot connect to database java lang nullpointerexception at java base java util regex matcher gettextlength unknown source at java base java util regex matcher reset unknown source at java base java util regex matcher unknown source at java base java util regex pattern matcher unknown source at clojure core re matcher invokestatic core clj at clojure core re matches invokestatic core clj at clojure core re matches invoke core clj at metabase driver mysql fn invokestatic mysql clj at metabase driver mysql fn invoke mysql clj at clojure lang multifn invoke multifn java at metabase driver util can connect with details qmark invokestatic util clj at metabase driver util can connect with details qmark doinvoke util clj at clojure lang restfn invoke restfn java at metabase api database test database connection invokestatic database clj at metabase api database test database connection doinvoke database clj at clojure lang restfn invoke restfn java at metabase api database fn invokestatic database clj at metabase api database fn invoke database clj at compojure core wrap response fn invoke core clj at compojure core wrap route middleware fn invoke core clj at compojure core wrap route info fn invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at metabase server middleware auth enforce authentication fn invoke auth clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at compojure core make context handler invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at metabase api routes fn fn invoke routes clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at clojure lang afn applytohelper afn java at clojure lang afn applyto afn java at clojure core apply invokestatic core clj at clojure core apply invoke core clj at metabase server routes fn fn doinvoke routes clj at clojure lang restfn invoke restfn java at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at compojure core make context handler invoke core clj at compojure core make context fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at compojure core wrap route matches fn invoke core clj at compojure core routes fn f invoke core clj at compojure core routes fn f respond singlequote invoke core clj at metabase server routes fn fn invoke routes clj at compojure core routes fn f invoke core clj at compojure core routes fn invoke core clj at metabase server middleware exceptions catch uncaught exceptions fn invoke exceptions clj at metabase server middleware exceptions catch api exceptions fn invoke exceptions clj at metabase server middleware log log api call fn fn fn invoke log clj at metabase driver sql jdbc execute diagnostic do with diagnostic info invokestatic diagnostic clj at metabase driver sql jdbc execute diagnostic do with diagnostic info invoke diagnostic clj at metabase server middleware log log api call fn fn invoke log clj at toucan db do with call counting invokestatic db clj at toucan db do with call counting invoke db clj at metabase server middleware log log api call fn invoke log clj at metabase server middleware browser cookie ensure browser id cookie fn invoke browser cookie clj at metabase server middleware security add security headers fn invoke security clj at metabase server middleware json wrap json body fn invoke json clj at metabase server middleware json wrap streamed json response fn invoke json clj at metabase server middleware offset paging handle paging fn invoke offset paging clj at ring middleware keyword params wrap keyword params fn invoke keyword params clj at ring middleware params wrap params fn invoke params clj at metabase server middleware misc maybe set site url fn invoke misc clj at metabase server middleware session bind current user fn fn invoke session clj at metabase server middleware session do with current user invokestatic session clj at metabase server middleware session do with current user invoke session clj at metabase server middleware session bind current user fn invoke session clj at metabase server middleware session wrap current user info fn invoke session clj at metabase server middleware session wrap session id fn invoke session clj at metabase server middleware auth wrap api key fn invoke auth clj at ring middleware cookies wrap cookies fn invoke cookies clj at metabase server middleware misc add content type fn invoke misc clj at metabase server middleware misc disable streaming buffering fn invoke misc clj at ring middleware gzip wrap gzip fn invoke gzip clj at metabase server middleware misc bind request fn invoke misc clj at metabase server middleware ssl redirect to https middleware fn invoke ssl clj at metabase server async proxy handler fn invoke server clj at metabase server proxy org eclipse jetty server handler abstracthandler handle unknown source at org eclipse jetty server handler handlerwrapper handle handlerwrapper java at org eclipse jetty server server handle server java at org eclipse jetty server httpchannel lambda handle httpchannel java at org eclipse jetty server httpchannel dispatch httpchannel java at org eclipse jetty server httpchannel handle httpchannel java at org eclipse jetty server httpconnection onfillable httpconnection java at org eclipse jetty io abstractconnection readcallback succeeded abstractconnection java at org eclipse jetty io fillinterest fillable fillinterest java at org eclipse jetty io channelendpoint run channelendpoint java at org eclipse jetty util thread strategy eatwhatyoukill runtask eatwhatyoukill java at org eclipse jetty util thread strategy eatwhatyoukill doproduce eatwhatyoukill java at org eclipse jetty util thread strategy eatwhatyoukill tryproduce eatwhatyoukill java at org eclipse jetty util thread strategy eatwhatyoukill run eatwhatyoukill java at org eclipse jetty util thread reservedthreadexecutor reservedthread run reservedthreadexecutor java at org eclipse jetty util thread queuedthreadpool runjob queuedthreadpool java at org eclipse jetty util thread queuedthreadpool runner run queuedthreadpool java at java base java lang thread run unknown source debug metabase server middleware log put api database ms db calls valid false dbname nil message nil to reproduce n a expected behavior either for the ssh tunnel to successfully connect or for a connection error to be reported in page retaining the details that were submitted screenshots n a information about your metabase installation running the docker image on a kubernetes cluster browser info language en us platform macintel useragent mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari vendor google inc system info file encoding utf java runtime name openjdk runtime environment java runtime version java vendor eclipse adoptium java vendor url java version java vm name openjdk bit server vm java vm version os name linux os version user language en user timezone us pacific metabase info databases googleanalytics mysql postgres hosting env unknown application database postgres application database details database name postgresql version jdbc driver name postgresql jdbc driver version run mode prod version date tag branch release x x hash settings report timezone america vancouver severity blocking usage prior to direct database manipulation could not connect to significant service database additional context n a
| 1
|
16,230
| 20,767,048,292
|
IssuesEvent
|
2022-03-15 21:51:49
|
crim-ca/weaver
|
https://api.github.com/repos/crim-ca/weaver
|
closed
|
[Feature] Builtin process marked with WPS-REST isStatic field
|
good first issue triage/feature process/wps3 triage/conformance process/builtin
|
WPS-REST reference introduced parameter `isStatic = true|false` to indicate a pre-deployed vs dynamically deployed processes after WPS app startup.
This basically corresponds to our `builtin` processes that are marked using CWL `BuiltinRequirement`.
We should add this new parameter accordingly to indicate this behaviour if it remains in the standard.
This can also help filtering processes that should not be allowed to be undeployed as they come with Weaver and are expected to be available at all time.
references:
- https://github.com/opengeospatial/wps-rest-binding/blob/e6e5c02dd963205f339822b946076a4dbf2daa87/extensions/transactions/standard/requirements/transactions/static/REQ_indicator.adoc
- https://github.com/opengeospatial/wps-rest-binding/blob/9b302c59a669115f34116a574d1e0d5db733bfa7/extensions/transactions/standard/clause_6_transactions.adoc
Corresponding requirements must be added as supported in conformance route once integrated.
|
2.0
|
[Feature] Builtin process marked with WPS-REST isStatic field - WPS-REST reference introduced parameter `isStatic = true|false` to indicate a pre-deployed vs dynamically deployed processes after WPS app startup.
This basically corresponds to our `builtin` processes that are marked using CWL `BuiltinRequirement`.
We should add this new parameter accordingly to indicate this behaviour if it remains in the standard.
This can also help filtering processes that should not be allowed to be undeployed as they come with Weaver and are expected to be available at all time.
references:
- https://github.com/opengeospatial/wps-rest-binding/blob/e6e5c02dd963205f339822b946076a4dbf2daa87/extensions/transactions/standard/requirements/transactions/static/REQ_indicator.adoc
- https://github.com/opengeospatial/wps-rest-binding/blob/9b302c59a669115f34116a574d1e0d5db733bfa7/extensions/transactions/standard/clause_6_transactions.adoc
Corresponding requirements must be added as supported in conformance route once integrated.
|
process
|
builtin process marked with wps rest isstatic field wps rest reference introduced parameter isstatic true false to indicate a pre deployed vs dynamically deployed processes after wps app startup this basically corresponds to our builtin processes that are marked using cwl builtinrequirement we should add this new parameter accordingly to indicate this behaviour if it remains in the standard this can also help filtering processes that should not be allowed to be undeployed as they come with weaver and are expected to be available at all time references corresponding requirements must be added as supported in conformance route once integrated
| 1
|
4,775
| 7,642,111,885
|
IssuesEvent
|
2018-05-08 08:09:44
|
Bw2801/environment
|
https://api.github.com/repos/Bw2801/environment
|
opened
|
Add secure websocket support
|
enhancement processor
|
Add support for certificates to allow secure websocket connections.
|
1.0
|
Add secure websocket support - Add support for certificates to allow secure websocket connections.
|
process
|
add secure websocket support add support for certificates to allow secure websocket connections
| 1
|
15,463
| 19,680,335,777
|
IssuesEvent
|
2022-01-11 16:11:21
|
plazi/community
|
https://api.github.com/repos/plazi/community
|
closed
|
to be processed: science 371.941
|
process request
|
I would appreciate if this article could be processed. It has not treatments, but figures etc.
Please let @millerjeremya have the UUID
t
[science.371.941-944.pdf](https://github.com/plazi/community/files/7846427/science.371.941-944.pdf)
x
|
1.0
|
to be processed: science 371.941 - I would appreciate if this article could be processed. It has not treatments, but figures etc.
Please let @millerjeremya have the UUID
t
[science.371.941-944.pdf](https://github.com/plazi/community/files/7846427/science.371.941-944.pdf)
x
|
process
|
to be processed science i would appreciate if this article could be processed it has not treatments but figures etc please let millerjeremya have the uuid t x
| 1
|
2,159
| 5,006,374,395
|
IssuesEvent
|
2016-12-12 13:58:06
|
openvstorage/framework
|
https://api.github.com/repos/openvstorage/framework
|
closed
|
Correctly process 'Ensure safety' for vdisks
|
priority_normal process_cantreproduce type_bug
|
This tasks should complete correctly and just report that it did not process it correctly for that vdisk.
Params `ovs.vdisk.migrate_from_voldrv`:
`{'new_owner_id': u'globaljlKIkrmtTHLINeSg', 'volume_id': u'2b45a75c-6a40-4bf3-93c8-087f949f411f'}`
As seen here in the `ovs.vdisk.migrate_from_voldrv` task:
```
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 240, in trace_task
R = retval = fun(*args, **kwargs)
File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 438, in __protected_call__
return self.run(*args, **kwargs)
File "/opt/OpenvStorage/ovs/lib/helpers/decorators.py", line 69, in new_function
return function(*args, **kwargs)
File "/opt/OpenvStorage/ovs/lib/vdisk.py", line 220, in migrate_from_voldrv
MDSServiceController.mds_checkup()
File "/usr/lib/python2.7/dist-packages/celery/local.py", line 188, in __call__
return self._get_current_object()(*a, **kw)
File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 439, in __protected_call__
return orig(self, *args, **kwargs)
File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__
return self.run(*args, **kwargs)
File "/opt/OpenvStorage/ovs/lib/helpers/decorators.py", line 305, in new_function
output = function(*args, **kwargs)
File "/opt/OpenvStorage/ovs/lib/mdsservice.py", line 822, in mds_checkup
raise Exception('\n - ' + '\n - '.join(failures))
Exception:
- Ensure safety for vDisk KVM13-fio-rou-d1 with guid ff16eae1-5f25-49a8-8465-77be71e9b983 failed
```
|
1.0
|
Correctly process 'Ensure safety' for vdisks - This tasks should complete correctly and just report that it did not process it correctly for that vdisk.
Params `ovs.vdisk.migrate_from_voldrv`:
`{'new_owner_id': u'globaljlKIkrmtTHLINeSg', 'volume_id': u'2b45a75c-6a40-4bf3-93c8-087f949f411f'}`
As seen here in the `ovs.vdisk.migrate_from_voldrv` task:
```
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 240, in trace_task
R = retval = fun(*args, **kwargs)
File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 438, in __protected_call__
return self.run(*args, **kwargs)
File "/opt/OpenvStorage/ovs/lib/helpers/decorators.py", line 69, in new_function
return function(*args, **kwargs)
File "/opt/OpenvStorage/ovs/lib/vdisk.py", line 220, in migrate_from_voldrv
MDSServiceController.mds_checkup()
File "/usr/lib/python2.7/dist-packages/celery/local.py", line 188, in __call__
return self._get_current_object()(*a, **kw)
File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 439, in __protected_call__
return orig(self, *args, **kwargs)
File "/usr/lib/python2.7/dist-packages/celery/app/task.py", line 420, in __call__
return self.run(*args, **kwargs)
File "/opt/OpenvStorage/ovs/lib/helpers/decorators.py", line 305, in new_function
output = function(*args, **kwargs)
File "/opt/OpenvStorage/ovs/lib/mdsservice.py", line 822, in mds_checkup
raise Exception('\n - ' + '\n - '.join(failures))
Exception:
- Ensure safety for vDisk KVM13-fio-rou-d1 with guid ff16eae1-5f25-49a8-8465-77be71e9b983 failed
```
|
process
|
correctly process ensure safety for vdisks this tasks should complete correctly and just report that it did not process it correctly for that vdisk params ovs vdisk migrate from voldrv new owner id u globaljlkikrmtthlinesg volume id u as seen here in the ovs vdisk migrate from voldrv task traceback most recent call last file usr lib dist packages celery app trace py line in trace task r retval fun args kwargs file usr lib dist packages celery app trace py line in protected call return self run args kwargs file opt openvstorage ovs lib helpers decorators py line in new function return function args kwargs file opt openvstorage ovs lib vdisk py line in migrate from voldrv mdsservicecontroller mds checkup file usr lib dist packages celery local py line in call return self get current object a kw file usr lib dist packages celery app trace py line in protected call return orig self args kwargs file usr lib dist packages celery app task py line in call return self run args kwargs file opt openvstorage ovs lib helpers decorators py line in new function output function args kwargs file opt openvstorage ovs lib mdsservice py line in mds checkup raise exception n n join failures exception ensure safety for vdisk fio rou with guid failed
| 1
|
270,446
| 23,509,650,988
|
IssuesEvent
|
2022-08-18 15:23:12
|
Mamr96insatbug/test
|
https://api.github.com/repos/Mamr96insatbug/test
|
opened
|
Error playing
|
Forwarded-to-Test
|
# :clipboard: Bug Details
>Error playing
key | value
--|--
Reported At | 2022-08-18 15:22:57 UTC
Email | aelbashary@instabug.com
Categories | Report a bug, Usability Issue
Tags | Forwarded-to-Test
App Version | 2.0.8 (6)
Session Duration | 1162
Device | arm64, iOS 15.5
Display | 390x844 (@3x)
Location | Cairo, Egypt (en)
## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/test123/beta/bugs/401?utm_source=github&utm_medium=integrations) :point_left:
___
# :chart_with_downwards_trend: Session Profiler
Here is what the app was doing right before the bug was reported:
Key | Value
--|--
CPU Load | 0.8%
Used Memory | 100.0% - 0.03/0.03 GB
Used Storage | 40.6% - 92.68/228.27 GB
Connectivity | WiFi - Simulator WiFi
Battery | 100% - unplugged
Orientation | portrait
Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/test123/beta/bugs/401?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :mag_right: Logs
### User Steps
Here are the last 10 steps done by the user right before the bug was reported:
```
15:22:13 Top View: SwiftRadio.StationsViewController
15:22:13 Tap in _UIButtonBarButton in SwiftRadio.NowPlayingViewController
15:22:12 Tap in UIButton in SwiftRadio.NowPlayingViewController
15:22:09 Top View: SwiftRadio.NowPlayingViewController
15:22:09 Tap in _UIParallaxDimmingView in SwiftRadio.NowPlayingViewController
15:22:09 Tap in UIStackView in SwiftRadio.StationsViewController
15:03:10 Tap in Dismiss of type UIButton in SwiftRadio.StationsViewController
15:03:08 Top View: SwiftRadio.StationsViewController
15:03:08 Application: DidBecomeActive
15:03:08 Application: SceneDidActivate
```
Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/test123/beta/bugs/401?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left:
___
# :camera: Images
[](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/18748412/672c31e1ddbe61d2e1ce532a556f421e_original/26937922/2022081805225541421474.jpg?Expires=4816509791&Signature=q27ktNgxrkk2kHXzXnR1AZ1k6-rOk8wGLdn88sh5UVeNSa31NjUeIE1kUXCKDxn4NgO6IcpvVH48F9mGCNy5DJNVp42tgmBq15D4uAWETm01aYNg6lfOYm5Id6WN~ILFotXiNkAcLu8YpUXrntXgHXof4GpO-tNgCB4CMR9EEmE0pCXoiU0vWqMgqrKT8T35qfs0b8cMKGb4tURbtsHT5EO1J1uj~7w2AlRVpyhlEITmh~hl4dAgvNnqmpEScUWiAUGZG1McMTNVmiwdXM0cxpt~2k7u2G8GNKINgYnU6RwCuWApk~BHaEmVZT4ajsj6YbA5Zlr9~ZnKWHdmauwPew__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)
___
# :warning: Looking for More Details?
1. **Network Log**: we are unable to capture your network requests automatically. If you are using AFNetworking or Alamofire, [**check the details mentioned here**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations#section-requests-not-appearing-in-logs).
2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations).
3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations).
4. **Console Log**: when enabled you will see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations).
|
1.0
|
Error playing - # :clipboard: Bug Details
>Error playing
key | value
--|--
Reported At | 2022-08-18 15:22:57 UTC
Email | aelbashary@instabug.com
Categories | Report a bug, Usability Issue
Tags | Forwarded-to-Test
App Version | 2.0.8 (6)
Session Duration | 1162
Device | arm64, iOS 15.5
Display | 390x844 (@3x)
Location | Cairo, Egypt (en)
## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/test123/beta/bugs/401?utm_source=github&utm_medium=integrations) :point_left:
___
# :chart_with_downwards_trend: Session Profiler
Here is what the app was doing right before the bug was reported:
Key | Value
--|--
CPU Load | 0.8%
Used Memory | 100.0% - 0.03/0.03 GB
Used Storage | 40.6% - 92.68/228.27 GB
Connectivity | WiFi - Simulator WiFi
Battery | 100% - unplugged
Orientation | portrait
Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/test123/beta/bugs/401?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :mag_right: Logs
### User Steps
Here are the last 10 steps done by the user right before the bug was reported:
```
15:22:13 Top View: SwiftRadio.StationsViewController
15:22:13 Tap in _UIButtonBarButton in SwiftRadio.NowPlayingViewController
15:22:12 Tap in UIButton in SwiftRadio.NowPlayingViewController
15:22:09 Top View: SwiftRadio.NowPlayingViewController
15:22:09 Tap in _UIParallaxDimmingView in SwiftRadio.NowPlayingViewController
15:22:09 Tap in UIStackView in SwiftRadio.StationsViewController
15:03:10 Tap in Dismiss of type UIButton in SwiftRadio.StationsViewController
15:03:08 Top View: SwiftRadio.StationsViewController
15:03:08 Application: DidBecomeActive
15:03:08 Application: SceneDidActivate
```
Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/test123/beta/bugs/401?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left:
___
# :camera: Images
[](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/18748412/672c31e1ddbe61d2e1ce532a556f421e_original/26937922/2022081805225541421474.jpg?Expires=4816509791&Signature=q27ktNgxrkk2kHXzXnR1AZ1k6-rOk8wGLdn88sh5UVeNSa31NjUeIE1kUXCKDxn4NgO6IcpvVH48F9mGCNy5DJNVp42tgmBq15D4uAWETm01aYNg6lfOYm5Id6WN~ILFotXiNkAcLu8YpUXrntXgHXof4GpO-tNgCB4CMR9EEmE0pCXoiU0vWqMgqrKT8T35qfs0b8cMKGb4tURbtsHT5EO1J1uj~7w2AlRVpyhlEITmh~hl4dAgvNnqmpEScUWiAUGZG1McMTNVmiwdXM0cxpt~2k7u2G8GNKINgYnU6RwCuWApk~BHaEmVZT4ajsj6YbA5Zlr9~ZnKWHdmauwPew__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)
___
# :warning: Looking for More Details?
1. **Network Log**: we are unable to capture your network requests automatically. If you are using AFNetworking or Alamofire, [**check the details mentioned here**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations#section-requests-not-appearing-in-logs).
2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations).
3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations).
4. **Console Log**: when enabled you will see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations).
|
non_process
|
error playing clipboard bug details error playing key value reported at utc email aelbashary instabug com categories report a bug usability issue tags forwarded to test app version session duration device ios display location cairo egypt en point right point left chart with downwards trend session profiler here is what the app was doing right before the bug was reported key value cpu load used memory gb used storage gb connectivity wifi simulator wifi battery unplugged orientation portrait find all the changes that happened in the parameters mentioned above during the last seconds before the bug was reported here point right point left mag right logs user steps here are the last steps done by the user right before the bug was reported top view swiftradio stationsviewcontroller tap in uibuttonbarbutton in swiftradio nowplayingviewcontroller tap in uibutton in swiftradio nowplayingviewcontroller top view swiftradio nowplayingviewcontroller tap in uiparallaxdimmingview in swiftradio nowplayingviewcontroller tap in uistackview in swiftradio stationsviewcontroller tap in dismiss of type uibutton in swiftradio stationsviewcontroller top view swiftradio stationsviewcontroller application didbecomeactive application scenedidactivate find all the user steps done by the user throughout the session here point right point left camera images warning looking for more details network log we are unable to capture your network requests automatically if you are using afnetworking or alamofire user events start capturing custom user events to send them along with each report instabug log start adding instabug logs to see them right inside each report you receive console log when enabled you will see them right inside each report you receive
| 0
|
22,408
| 31,142,292,407
|
IssuesEvent
|
2023-08-16 01:44:47
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Flaky test: AssertionError: expected 'POST /runs/00748421-e035-4a3d-8604-8468cc48bdb5/instances' to equal 'POST /instances/e9e81b5e-cc58-4026-b2ff-8ae3161435a6/results'
|
OS: linux process: flaky test topic: flake ❄️ stage: flake stale
|
### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41370/workflows/b5842bda-87dd-488c-ac98-4855e959a900/jobs/1713324
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/system-tests/test/record_spec.js#L395
### Analysis
<img width="1300" alt="Screen Shot 2022-08-05 at 12 26 27 PM" src="https://user-images.githubusercontent.com/26726429/183147285-63d8f397-13f7-4701-a14d-c501824c7493.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
|
1.0
|
Flaky test: AssertionError: expected 'POST /runs/00748421-e035-4a3d-8604-8468cc48bdb5/instances' to equal 'POST /instances/e9e81b5e-cc58-4026-b2ff-8ae3161435a6/results' - ### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41370/workflows/b5842bda-87dd-488c-ac98-4855e959a900/jobs/1713324
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/system-tests/test/record_spec.js#L395
### Analysis
<img width="1300" alt="Screen Shot 2022-08-05 at 12 26 27 PM" src="https://user-images.githubusercontent.com/26726429/183147285-63d8f397-13f7-4701-a14d-c501824c7493.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
|
process
|
flaky test assertionerror expected post runs instances to equal post instances results link to dashboard or circleci failure link to failing test in github analysis img width alt screen shot at pm src cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed
| 1
|
120,795
| 15,803,666,503
|
IssuesEvent
|
2021-04-03 15:14:35
|
dyl10s/TimeTracker
|
https://api.github.com/repos/dyl10s/TimeTracker
|
opened
|
Remove User
|
api design web app
|
A teacher should be able to remove a specific user from a project. Probably a button on the project details page next to each member with a confirmation popup. When a user is removed all the data should be deleted for that project as well.
|
1.0
|
Remove User - A teacher should be able to remove a specific user from a project. Probably a button on the project details page next to each member with a confirmation popup. When a user is removed all the data should be deleted for that project as well.
|
non_process
|
remove user a teacher should be able to remove a specific user from a project probably a button on the project details page next to each member with a confirmation popup when a user is removed all the data should be deleted for that project as well
| 0
|
7,917
| 11,096,907,168
|
IssuesEvent
|
2019-12-16 12:13:23
|
googleapis/google-cloud-dotnet
|
https://api.github.com/repos/googleapis/google-cloud-dotnet
|
closed
|
Add better integration tests for Google.Cloud.Diagnostics.AspNet
|
api: clouderrorreporting api: cloudtrace status: blocked type: process
|
Currently it has very few integration tests and should have better coverage.
|
1.0
|
Add better integration tests for Google.Cloud.Diagnostics.AspNet - Currently it has very few integration tests and should have better coverage.
|
process
|
add better integration tests for google cloud diagnostics aspnet currently it has very few integration tests and should have better coverage
| 1
|
9,501
| 12,489,593,214
|
IssuesEvent
|
2020-05-31 19:34:56
|
Arch666Angel/mods
|
https://api.github.com/repos/Arch666Angel/mods
|
closed
|
[BUG] Crystal Powder --> Slurry has no fluid input?
|
Angels Bio Processing Impact: Enhancement
|
Every other recipe making crystal slurry (from dust or geodes) requires sulphuric acid as an input. Crystal powder to crystal slurry requires no fluid input at all? Seems like it should take something...
|
1.0
|
[BUG] Crystal Powder --> Slurry has no fluid input? - Every other recipe making crystal slurry (from dust or geodes) requires sulphuric acid as an input. Crystal powder to crystal slurry requires no fluid input at all? Seems like it should take something...
|
process
|
crystal powder slurry has no fluid input every other recipe making crystal slurry from dust or geodes requires sulphuric acid as an input crystal powder to crystal slurry requires no fluid input at all seems like it should take something
| 1
|
768,617
| 26,973,065,015
|
IssuesEvent
|
2023-02-09 07:16:27
|
WavesHQ/bridge
|
https://api.github.com/repos/WavesHQ/bridge
|
closed
|
`Contracts`: Adding name and version in contract as constants
|
needs/area needs/triage kind/feature needs/priority
|
<!-- Please only use this template for submitting enhancement/feature requests -->
#### What would you like to be added:
With the current implementation, there are no mentioned of name and version of the contract.
For the backend, devs have to manually change the `env` var if upgrading the contract later.
#### Why is this needed:
|
1.0
|
`Contracts`: Adding name and version in contract as constants - <!-- Please only use this template for submitting enhancement/feature requests -->
#### What would you like to be added:
With the current implementation, there are no mentioned of name and version of the contract.
For the backend, devs have to manually change the `env` var if upgrading the contract later.
#### Why is this needed:
|
non_process
|
contracts adding name and version in contract as constants what would you like to be added with the current implementation there are no mentioned of name and version of the contract for the backend devs have to manually change the env var if upgrading the contract later why is this needed
| 0
|
89,908
| 25,921,262,879
|
IssuesEvent
|
2022-12-15 22:13:24
|
LLNL/serac
|
https://api.github.com/repos/LLNL/serac
|
opened
|
Turn on position independent code for MFEM builds
|
CS build
|
This has come up in several niche cases and has no side effects that I know of. We should turn it on by default in our MFEM builds.
|
1.0
|
Turn on position independent code for MFEM builds - This has come up in several niche cases and has no side effects that I know of. We should turn it on by default in our MFEM builds.
|
non_process
|
turn on position independent code for mfem builds this has come up in several niche cases and has no side effects that i know of we should turn it on by default in our mfem builds
| 0
|
12,207
| 14,742,778,466
|
IssuesEvent
|
2021-01-07 12:53:04
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Emails -
|
anc-process anc-ui anp-2.5 ant-enhancement grt-ui processes
|
In GitLab by @kdjstudios on Jun 5, 2019, 15:26
Hello Team,
I just recently noticed that all of the emails that are going out from SA Billing seem to have "Hi Support," as the first line in the body. I think I may not have noticed this before as most of these emails were originally directly to only the support team. However, now we are sending these emails to operations and clients. I think we will need to update that line or completely remove it.
Also I noticed, that when we send the emails it goes to both the user who attempted to perform the step and the site's email or the Manager email. Sometimes this is the same person, is there a way we can determine if they are the same and only include that user once, as to not send duplicate emails?
|
2.0
|
Emails - - In GitLab by @kdjstudios on Jun 5, 2019, 15:26
Hello Team,
I just recently noticed that all of the emails that are going out from SA Billing seem to have "Hi Support," as the first line in the body. I think I may not have noticed this before as most of these emails were originally directly to only the support team. However, now we are sending these emails to operations and clients. I think we will need to update that line or completely remove it.
Also I noticed, that when we send the emails it goes to both the user who attempted to perform the step and the site's email or the Manager email. Sometimes this is the same person, is there a way we can determine if they are the same and only include that user once, as to not send duplicate emails?
|
process
|
emails in gitlab by kdjstudios on jun hello team i just recently noticed that all of the emails that are going out from sa billing seem to have hi support as the first line in the body i think i may not have noticed this before as most of these emails were originally directly to only the support team however now we are sending these emails to operations and clients i think we will need to update that line or completely remove it also i noticed that when we send the emails it goes to both the user who attempted to perform the step and the site s email or the manager email sometimes this is the same person is there a way we can determine if they are the same and only include that user once as to not send duplicate emails
| 1
|
216,480
| 16,766,364,870
|
IssuesEvent
|
2021-06-14 09:18:12
|
Evywell/codeR
|
https://api.github.com/repos/Evywell/codeR
|
closed
|
A better way to - handle exceptions in functional tests
|
draft test
|
In order to:
1. throw assertion errors instead of exception
2. throw the exceptions in the main thread
3. create test with exception catches
we need to use a common thread manager which will catch exceptions and throw them in the main thread (maybe using a synchronized method)
|
1.0
|
A better way to - handle exceptions in functional tests - In order to:
1. throw assertion errors instead of exception
2. throw the exceptions in the main thread
3. create test with exception catches
we need to use a common thread manager which will catch exceptions and throw them in the main thread (maybe using a synchronized method)
|
non_process
|
a better way to handle exceptions in functional tests in order to throw assertion errors instead of exception throw the exceptions in the main thread create test with exception catches we need to use a common thread manager which will catch exceptions and throw them in the main thread maybe using a synchronized method
| 0
|
492,205
| 14,190,239,519
|
IssuesEvent
|
2020-11-14 05:12:48
|
ArkEcosystem/core
|
https://api.github.com/repos/ArkEcosystem/core
|
closed
|
Burn balance transaction
|
Priority: Low Type: Feature
|
As discussed in #3442 there is need to simply burn balance. Calculating supply is prone to errors without dedicated burn transaction. Creating new transaction type is clean solution to the problem.
|
1.0
|
Burn balance transaction - As discussed in #3442 there is need to simply burn balance. Calculating supply is prone to errors without dedicated burn transaction. Creating new transaction type is clean solution to the problem.
|
non_process
|
burn balance transaction as discussed in there is need to simply burn balance calculating supply is prone to errors without dedicated burn transaction creating new transaction type is clean solution to the problem
| 0
|
95,656
| 10,884,911,518
|
IssuesEvent
|
2019-11-18 09:21:17
|
WeakAuras/WeakAuras2
|
https://api.github.com/repos/WeakAuras/WeakAuras2
|
closed
|
Add wiki links to edit boxes
|
🎨 Enhancement 📚Documentation
|
All our custom code edit boxes should have a small text + wiki link below them so people have a easier time finding info about them
I guesst this format:
`(white)Learn more about custom code that you can use here on our (yellow) wiki(wikiurl).`
|
1.0
|
Add wiki links to edit boxes - All our custom code edit boxes should have a small text + wiki link below them so people have a easier time finding info about them
I guesst this format:
`(white)Learn more about custom code that you can use here on our (yellow) wiki(wikiurl).`
|
non_process
|
add wiki links to edit boxes all our custom code edit boxes should have a small text wiki link below them so people have a easier time finding info about them i guesst this format white learn more about custom code that you can use here on our yellow wiki wikiurl
| 0
|
18,828
| 24,730,430,613
|
IssuesEvent
|
2022-10-20 17:05:58
|
Sunbird-cQube/community
|
https://api.github.com/repos/Sunbird-cQube/community
|
closed
|
Test pre-aggregated files by filling in the required sections in the current configuration file
|
Backlog Processing From-v4.0-Alpha
|
To test the aggregated data processing through the existing configuration file by filling all the sections in the configuration file
|
1.0
|
Test pre-aggregated files by filling in the required sections in the current configuration file - To test the aggregated data processing through the existing configuration file by filling all the sections in the configuration file
|
process
|
test pre aggregated files by filling in the required sections in the current configuration file to test the aggregated data processing through the existing configuration file by filling all the sections in the configuration file
| 1
|
599,370
| 18,272,218,720
|
IssuesEvent
|
2021-10-04 14:52:37
|
GoogleContainerTools/skaffold
|
https://api.github.com/repos/GoogleContainerTools/skaffold
|
closed
|
Deployment status-check is misleading on our pod-only getting-started
|
kind/bug priority/p2 area/status-check
|
We have only a **pod** in our getting-started example. This can cause some oddness with status-check which only looks at **Deployments**: the deployment seems to succeed even if the pod is pending! It looks like logging is broken, until user runs `kubectl get pods`.
Repro in minikube
1. `kubectl taint nodes minikube key=value:NoSchedule`
1. `skaffold dev` on the getting-started example
The output is from Tejal's PR https://github.com/GoogleContainerTools/skaffold/pull/4223 which is better than master but still not sufficient:
```
...
Successfully tagged gcr.io/balintp-gcp-lab/skaffold-example:v1.10.0-47-g0bec3d0e7
Tags used in deployment:
- skaffold-example -> gcr.io/balintp-gcp-lab/skaffold-example:2d2c5bd5243ac8274bf2d6eccb52208c0a08136990eb69735e2886b8a447deee
Starting deploy...
- pod/getting-started created
Waiting for deployments to stabilize...
Deployments stabilized in 15.01709ms
Press Ctrl+C to exit
Watching for changes...
^CCleaning up...
- pod "getting-started" deleted
There is a new version (1.10.1) of Skaffold available. Download it from:
```
While `kubectl describe pod` shows the events
```
...
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute for 300s
node.kubernetes.io/unreachable:NoExecute for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning FailedScheduling <unknown> default-scheduler 0/1 nodes are available: 1 node(s) had taint {key: value}, that the pod didn't tolerate.
Warning FailedScheduling <unknown> default-scheduler 0/1 nodes are available: 1 node(s) had taint {key: value}, that the pod didn't tolerate.
```
|
1.0
|
Deployment status-check is misleading on our pod-only getting-started - We have only a **pod** in our getting-started example. This can cause some oddness with status-check which only looks at **Deployments**: the deployment seems to succeed even if the pod is pending! It looks like logging is broken, until user runs `kubectl get pods`.
Repro in minikube
1. `kubectl taint nodes minikube key=value:NoSchedule`
1. `skaffold dev` on the getting-started example
The output is from Tejal's PR https://github.com/GoogleContainerTools/skaffold/pull/4223 which is better than master but still not sufficient:
```
...
Successfully tagged gcr.io/balintp-gcp-lab/skaffold-example:v1.10.0-47-g0bec3d0e7
Tags used in deployment:
- skaffold-example -> gcr.io/balintp-gcp-lab/skaffold-example:2d2c5bd5243ac8274bf2d6eccb52208c0a08136990eb69735e2886b8a447deee
Starting deploy...
- pod/getting-started created
Waiting for deployments to stabilize...
Deployments stabilized in 15.01709ms
Press Ctrl+C to exit
Watching for changes...
^CCleaning up...
- pod "getting-started" deleted
There is a new version (1.10.1) of Skaffold available. Download it from:
```
While `kubectl describe pod` shows the events
```
...
Node-Selectors: <none>
Tolerations: node.kubernetes.io/not-ready:NoExecute for 300s
node.kubernetes.io/unreachable:NoExecute for 300s
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning FailedScheduling <unknown> default-scheduler 0/1 nodes are available: 1 node(s) had taint {key: value}, that the pod didn't tolerate.
Warning FailedScheduling <unknown> default-scheduler 0/1 nodes are available: 1 node(s) had taint {key: value}, that the pod didn't tolerate.
```
|
non_process
|
deployment status check is misleading on our pod only getting started we have only a pod in our getting started example this can cause some oddness with status check which only looks at deployments the deployment seems to succeed even if the pod is pending it looks like logging is broken until user runs kubectl get pods repro in minikube kubectl taint nodes minikube key value noschedule skaffold dev on the getting started example the output is from tejal s pr which is better than master but still not sufficient successfully tagged gcr io balintp gcp lab skaffold example tags used in deployment skaffold example gcr io balintp gcp lab skaffold example starting deploy pod getting started created waiting for deployments to stabilize deployments stabilized in press ctrl c to exit watching for changes ccleaning up pod getting started deleted there is a new version of skaffold available download it from while kubectl describe pod shows the events node selectors tolerations node kubernetes io not ready noexecute for node kubernetes io unreachable noexecute for events type reason age from message warning failedscheduling default scheduler nodes are available node s had taint key value that the pod didn t tolerate warning failedscheduling default scheduler nodes are available node s had taint key value that the pod didn t tolerate
| 0
|
299,716
| 9,205,788,457
|
IssuesEvent
|
2019-03-08 11:39:38
|
qissue-bot/QGIS
|
https://api.github.com/repos/qissue-bot/QGIS
|
closed
|
ECW takes a long time to display
|
Category: Rasters Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report
|
---
Author Name: **William Kyngesburye** (William Kyngesburye)
Original Redmine Issue: 1084, https://issues.qgis.org/issues/1084
Original Assignee: nobody -
---
In 0.9.1 ECW display is instantaneous, no matter how large the ECW was, probably because it used the inherent overviews. In 0.10.0 ECW takes a lot longer to display.
As a example, any of the Blue Marble ECW images takes about a minute before anything is displayed, during this time there is a lot of disk read activity. I gave up waiting for the Landsat742 ECW to display. I hadn't tried until now, but it's the same with 0.9.2rc1. I wonder if Qgis is not reading the overviews now?
I don't have a large JP2 to try, but I wonder if they have the same problem, since both ECW and JP2 have inherent overviews.
OSX 10.5.2, GDAL 1.5.1.
|
1.0
|
ECW takes a long time to display - ---
Author Name: **William Kyngesburye** (William Kyngesburye)
Original Redmine Issue: 1084, https://issues.qgis.org/issues/1084
Original Assignee: nobody -
---
In 0.9.1 ECW display is instantaneous, no matter how large the ECW was, probably because it used the inherent overviews. In 0.10.0 ECW takes a lot longer to display.
As a example, any of the Blue Marble ECW images takes about a minute before anything is displayed, during this time there is a lot of disk read activity. I gave up waiting for the Landsat742 ECW to display. I hadn't tried until now, but it's the same with 0.9.2rc1. I wonder if Qgis is not reading the overviews now?
I don't have a large JP2 to try, but I wonder if they have the same problem, since both ECW and JP2 have inherent overviews.
OSX 10.5.2, GDAL 1.5.1.
|
non_process
|
ecw takes a long time to display author name william kyngesburye william kyngesburye original redmine issue original assignee nobody in ecw display is instantaneous no matter how large the ecw was probably because it used the inherent overviews in ecw takes a lot longer to display as a example any of the blue marble ecw images takes about a minute before anything is displayed during this time there is a lot of disk read activity i gave up waiting for the ecw to display i hadn t tried until now but it s the same with i wonder if qgis is not reading the overviews now i don t have a large to try but i wonder if they have the same problem since both ecw and have inherent overviews osx gdal
| 0
|
57,343
| 7,050,133,893
|
IssuesEvent
|
2018-01-03 03:20:20
|
simonbw/commander-periscope
|
https://api.github.com/repos/simonbw/commander-periscope
|
opened
|
Colorblindness Audit
|
Design
|
Make sure the game is playable by colorblind people. We should never rely on color being the only characteristic conveying some information.
|
1.0
|
Colorblindness Audit - Make sure the game is playable by colorblind people. We should never rely on color being the only characteristic conveying some information.
|
non_process
|
colorblindness audit make sure the game is playable by colorblind people we should never rely on color being the only characteristic conveying some information
| 0
|
218,092
| 7,330,393,223
|
IssuesEvent
|
2018-03-05 09:47:00
|
NCEAS/metacat
|
https://api.github.com/repos/NCEAS/metacat
|
closed
|
add function to submit data with registry entities
|
Category: registry Component: Bugzilla-Id Priority: Immediate Status: Resolved Tracker: Bug
|
---
Author Name: **Matt Jones** (Matt Jones)
Original Redmine Issue: 1982, https://projects.ecoinformatics.org/ecoinfo/issues/1982
Original Date: 2005-02-17
Original Assignee: Shaun Walbridge
---
A common request is to be able to submit data to the metacat along with a
registry entry as it is submitted. We've not done this because of the
additional metadata burden that is typically required to be EML compliant
(entity and attribute). However, there is some utility to being able to archive
the data even without the detailed entity and attribute metadata. Once strategy
is this:
1) Add a form upload field to allo a user to point at a data file on disk for upload
2) Upon submit, upload the metadata and data, parse out the data file and submit
it to metacat as its own object.
3) Use the ID for the submitted dataset to create a metacat URL, and write that
metacat url into the distribution section for the overall data package, and then
submit the metadata.
This is not the traditional approach to uploading EML data, as data are
typically uploaded in the dataTable/distribution section. A couple of problems
arise. First, because the distribution url is in an atypical place in EML,
metacat doesn't know how to set access control rules on the data object
properly, so this will need to be specially handled. Second, because the
distribution url is in an atypical place in EML, morpho won't know how to deal
with the data package and will need to be modified to detect the presence of
these data tables and 'move' the reference to the dataTable subtree to allow the
user to provide entity/attribute info for the table and to view the data. This
will be entered as a separate bug. So, an additional feature to contemplate is:
4) Add a form field for specifying access restrictions on the data. This could
be a simple radio button like:
"Make data 1) publicly available 2) private"
In some special cases, such as the NCEAS data registry, we might want a third
choice like "Group available" where the group consists of the people that are
members of the associated working group. THis will require being able to get
the list of ldap accounts associated with a group from the registry script, but
this could be presumably added to the admindb access script in perl.
|
1.0
|
add function to submit data with registry entities - ---
Author Name: **Matt Jones** (Matt Jones)
Original Redmine Issue: 1982, https://projects.ecoinformatics.org/ecoinfo/issues/1982
Original Date: 2005-02-17
Original Assignee: Shaun Walbridge
---
A common request is to be able to submit data to the metacat along with a
registry entry as it is submitted. We've not done this because of the
additional metadata burden that is typically required to be EML compliant
(entity and attribute). However, there is some utility to being able to archive
the data even without the detailed entity and attribute metadata. Once strategy
is this:
1) Add a form upload field to allo a user to point at a data file on disk for upload
2) Upon submit, upload the metadata and data, parse out the data file and submit
it to metacat as its own object.
3) Use the ID for the submitted dataset to create a metacat URL, and write that
metacat url into the distribution section for the overall data package, and then
submit the metadata.
This is not the traditional approach to uploading EML data, as data are
typically uploaded in the dataTable/distribution section. A couple of problems
arise. First, because the distribution url is in an atypical place in EML,
metacat doesn't know how to set access control rules on the data object
properly, so this will need to be specially handled. Second, because the
distribution url is in an atypical place in EML, morpho won't know how to deal
with the data package and will need to be modified to detect the presence of
these data tables and 'move' the reference to the dataTable subtree to allow the
user to provide entity/attribute info for the table and to view the data. This
will be entered as a separate bug. So, an additional feature to contemplate is:
4) Add a form field for specifying access restrictions on the data. This could
be a simple radio button like:
"Make data 1) publicly available 2) private"
In some special cases, such as the NCEAS data registry, we might want a third
choice like "Group available" where the group consists of the people that are
members of the associated working group. THis will require being able to get
the list of ldap accounts associated with a group from the registry script, but
this could be presumably added to the admindb access script in perl.
|
non_process
|
add function to submit data with registry entities author name matt jones matt jones original redmine issue original date original assignee shaun walbridge a common request is to be able to submit data to the metacat along with a registry entry as it is submitted we ve not done this because of the additional metadata burden that is typically required to be eml compliant entity and attribute however there is some utility to being able to archive the data even without the detailed entity and attribute metadata once strategy is this add a form upload field to allo a user to point at a data file on disk for upload upon submit upload the metadata and data parse out the data file and submit it to metacat as its own object use the id for the submitted dataset to create a metacat url and write that metacat url into the distribution section for the overall data package and then submit the metadata this is not the traditional approach to uploading eml data as data are typically uploaded in the datatable distribution section a couple of problems arise first because the distribution url is in an atypical place in eml metacat doesn t know how to set access control rules on the data object properly so this will need to be specially handled second because the distribution url is in an atypical place in eml morpho won t know how to deal with the data package and will need to be modified to detect the presence of these data tables and move the reference to the datatable subtree to allow the user to provide entity attribute info for the table and to view the data this will be entered as a separate bug so an additional feature to contemplate is add a form field for specifying access restrictions on the data this could be a simple radio button like make data publicly available private in some special cases such as the nceas data registry we might want a third choice like group available where the group consists of the people that are members of the associated working group this will require being able to get the list of ldap accounts associated with a group from the registry script but this could be presumably added to the admindb access script in perl
| 0
|
9,308
| 12,322,419,246
|
IssuesEvent
|
2020-05-13 10:18:32
|
threebotserver/publishingtools
|
https://api.github.com/repos/threebotserver/publishingtools
|
closed
|
Blog: Can we standardize the image header size?
|
process_wontfix
|
It's bigger in this one – http://134.122.109.244:3000/blog/vonsubblog/posts/hewlett-packard-enterprise--threefold-sign-mou – all the other ones seem to be the same.
|
1.0
|
Blog: Can we standardize the image header size? - It's bigger in this one – http://134.122.109.244:3000/blog/vonsubblog/posts/hewlett-packard-enterprise--threefold-sign-mou – all the other ones seem to be the same.
|
process
|
blog can we standardize the image header size it s bigger in this one – – all the other ones seem to be the same
| 1
|
8,070
| 11,251,352,366
|
IssuesEvent
|
2020-01-11 00:03:21
|
googleapis/java-phishingprotection
|
https://api.github.com/repos/googleapis/java-phishingprotection
|
opened
|
Promote to GA
|
type: process
|
Package name: **google-cloud-phishingprotection**
Current release: **beta**
Proposed release: **GA**
## Instructions
Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue.
## Required
- [ ] 28 days elapsed since last beta release with new API surface
- [ ] Server API is GA
- [ ] Package API is stable, and we can commit to backward compatibility
- [ ] All dependencies are GA
## Optional
- [ ] Most common / important scenarios have descriptive samples
- [ ] Public manual methods have at least one usage sample each (excluding overloads)
- [ ] Per-API README includes a full description of the API
- [ ] Per-API README contains at least one “getting started” sample using the most common API scenario
- [ ] Manual code has been reviewed by API producer
- [ ] Manual code has been reviewed by a DPE responsible for samples
- [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
|
1.0
|
Promote to GA - Package name: **google-cloud-phishingprotection**
Current release: **beta**
Proposed release: **GA**
## Instructions
Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue.
## Required
- [ ] 28 days elapsed since last beta release with new API surface
- [ ] Server API is GA
- [ ] Package API is stable, and we can commit to backward compatibility
- [ ] All dependencies are GA
## Optional
- [ ] Most common / important scenarios have descriptive samples
- [ ] Public manual methods have at least one usage sample each (excluding overloads)
- [ ] Per-API README includes a full description of the API
- [ ] Per-API README contains at least one “getting started” sample using the most common API scenario
- [ ] Manual code has been reviewed by API producer
- [ ] Manual code has been reviewed by a DPE responsible for samples
- [ ] 'Client Libraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
|
process
|
promote to ga package name google cloud phishingprotection current release beta proposed release ga instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required days elapsed since last beta release with new api surface server api is ga package api is stable and we can commit to backward compatibility all dependencies are ga optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
| 1
|
382,252
| 11,302,986,526
|
IssuesEvent
|
2020-01-17 18:59:45
|
kubernetes-sigs/cluster-api
|
https://api.github.com/repos/kubernetes-sigs/cluster-api
|
closed
|
Write conversion unit tests
|
kind/cleanup priority/important-soon
|
As a follow-up to #1992, we need to write some unit tests in a `conversion_test.go` file that goes alongside our `conversion.go` files that unit-tests conversions from v1alpha2 -> v1alpha3 and vice-versa.
/kind cleanup
/milestone v0.3.0
/priority important-soon
|
1.0
|
Write conversion unit tests - As a follow-up to #1992, we need to write some unit tests in a `conversion_test.go` file that goes alongside our `conversion.go` files that unit-tests conversions from v1alpha2 -> v1alpha3 and vice-versa.
/kind cleanup
/milestone v0.3.0
/priority important-soon
|
non_process
|
write conversion unit tests as a follow up to we need to write some unit tests in a conversion test go file that goes alongside our conversion go files that unit tests conversions from and vice versa kind cleanup milestone priority important soon
| 0
|
69,354
| 30,248,501,645
|
IssuesEvent
|
2023-07-06 18:27:11
|
hashgraph/hedera-services
|
https://api.github.com/repos/hashgraph/hedera-services
|
closed
|
Do zero sums on all transfer lists and change NFT owners
|
Hedera Token Service Modularization
|
- [x] Implement ensuring token associations for recepients
- [x] Do zero sum of hbar balance changes
- [x] Do zero sum of fungible token balance changes
- [x] Change NFT owners
|
1.0
|
Do zero sums on all transfer lists and change NFT owners - - [x] Implement ensuring token associations for recepients
- [x] Do zero sum of hbar balance changes
- [x] Do zero sum of fungible token balance changes
- [x] Change NFT owners
|
non_process
|
do zero sums on all transfer lists and change nft owners implement ensuring token associations for recepients do zero sum of hbar balance changes do zero sum of fungible token balance changes change nft owners
| 0
|
20,032
| 26,517,128,872
|
IssuesEvent
|
2023-01-18 21:50:41
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
A better name for "Manually entered demands" is something like "Agent Demands"
|
doc-enhancement devops/prod Pri2 devops-cicd-process/tech
|
Considering that the first subtitle in the page is "Task Demands" (task is the requester), "Agent Demands" makes more sense as the provider.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: e7541ee6-d2bb-84c0-fead-1aa8ee7d2372
* Version Independent ID: 5cf7c51e-37e1-6c67-e6c6-80262c4eb662
* Content: [Demands - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/demands?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/demands.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/demands.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @steved0x
* Microsoft Alias: **sdanie**
|
1.0
|
A better name for "Manually entered demands" is something like "Agent Demands" - Considering that the first subtitle in the page is "Task Demands" (task is the requester), "Agent Demands" makes more sense as the provider.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: e7541ee6-d2bb-84c0-fead-1aa8ee7d2372
* Version Independent ID: 5cf7c51e-37e1-6c67-e6c6-80262c4eb662
* Content: [Demands - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/demands?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/demands.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/demands.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @steved0x
* Microsoft Alias: **sdanie**
|
process
|
a better name for manually entered demands is something like agent demands considering that the first subtitle in the page is task demands task is the requester agent demands makes more sense as the provider document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id fead version independent id content content source product devops technology devops cicd process github login microsoft alias sdanie
| 1
|
73,560
| 9,670,583,942
|
IssuesEvent
|
2019-05-21 20:17:27
|
microsoft/msbuild
|
https://api.github.com/repos/microsoft/msbuild
|
closed
|
Add GetFileHash and VerifyFileHash tasks to XSD (for IntelliSense)
|
Documentation
|
_Reported by @japj in https://github.com/Microsoft/msbuild/pull/3999#issuecomment-458141311._
#3398 exposed two new tasks, but they're not exposed to Visual Studio's autocompletion because they're not mentioned in the XSD used for that.
They need an entry like
https://github.com/Microsoft/msbuild/blob/966cdf2ac602e359530ad63f608e6cbba9c9fd0f/src/MSBuild/Microsoft.Build.CommonTypes.xsd#L2198-L2211
but with the right inputs/outputs.
|
1.0
|
Add GetFileHash and VerifyFileHash tasks to XSD (for IntelliSense) - _Reported by @japj in https://github.com/Microsoft/msbuild/pull/3999#issuecomment-458141311._
#3398 exposed two new tasks, but they're not exposed to Visual Studio's autocompletion because they're not mentioned in the XSD used for that.
They need an entry like
https://github.com/Microsoft/msbuild/blob/966cdf2ac602e359530ad63f608e6cbba9c9fd0f/src/MSBuild/Microsoft.Build.CommonTypes.xsd#L2198-L2211
but with the right inputs/outputs.
|
non_process
|
add getfilehash and verifyfilehash tasks to xsd for intellisense reported by japj in exposed two new tasks but they re not exposed to visual studio s autocompletion because they re not mentioned in the xsd used for that they need an entry like but with the right inputs outputs
| 0
|
69,600
| 8,435,231,562
|
IssuesEvent
|
2018-10-17 12:37:37
|
chapel-lang/chapel
|
https://api.github.com/repos/chapel-lang/chapel
|
closed
|
rename domain.member?
|
area: Libraries type: Design
|
The current method name for querying if a value is in a domain is `domain.member()`.
E.g.
var D:domain(int);
D += 4;
assert(D.member(4));
assert(!D.member(3));
This method name bothers me for English/Programming Style reasons. `member` is a noun and I really want the method name to be a verb in this case.
What should we do?
1) add domain.contains()
a) and deprecate domain.member()
b) and keep domain.member() as a synonym
2) enable parser/build to translate `x in someDomain` into this query (#5034)
a) and deprecate domain.member()
b) and keep domain.member() as a synonym
3) do nothing
|
1.0
|
rename domain.member? - The current method name for querying if a value is in a domain is `domain.member()`.
E.g.
var D:domain(int);
D += 4;
assert(D.member(4));
assert(!D.member(3));
This method name bothers me for English/Programming Style reasons. `member` is a noun and I really want the method name to be a verb in this case.
What should we do?
1) add domain.contains()
a) and deprecate domain.member()
b) and keep domain.member() as a synonym
2) enable parser/build to translate `x in someDomain` into this query (#5034)
a) and deprecate domain.member()
b) and keep domain.member() as a synonym
3) do nothing
|
non_process
|
rename domain member the current method name for querying if a value is in a domain is domain member e g var d domain int d assert d member assert d member this method name bothers me for english programming style reasons member is a noun and i really want the method name to be a verb in this case what should we do add domain contains a and deprecate domain member b and keep domain member as a synonym enable parser build to translate x in somedomain into this query a and deprecate domain member b and keep domain member as a synonym do nothing
| 0
|
22,237
| 30,785,825,491
|
IssuesEvent
|
2023-07-31 13:11:19
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
@appsignal/cli 1.2.5 has 1 guarddog issues
|
npm-silent-process-execution
|
```{"npm-silent-process-execution":[{"code":" (0, child_process_1.spawn)(\"node\", [__filename], {\n cwd: process.cwd(),\n detached: true,\n env: __assign(__assign({}, process.env), env),\n stdio: \"ignore\"\n }).unref();","location":"package/dist/commands/demo.js:38","message":"This package is silently executing another executable"}]}```
|
1.0
|
@appsignal/cli 1.2.5 has 1 guarddog issues - ```{"npm-silent-process-execution":[{"code":" (0, child_process_1.spawn)(\"node\", [__filename], {\n cwd: process.cwd(),\n detached: true,\n env: __assign(__assign({}, process.env), env),\n stdio: \"ignore\"\n }).unref();","location":"package/dist/commands/demo.js:38","message":"This package is silently executing another executable"}]}```
|
process
|
appsignal cli has guarddog issues npm silent process execution n cwd process cwd n detached true n env assign assign process env env n stdio ignore n unref location package dist commands demo js message this package is silently executing another executable
| 1
|
175,766
| 21,330,054,781
|
IssuesEvent
|
2022-04-18 07:03:24
|
LaudateCorpus1/JQuery-Mobile
|
https://api.github.com/repos/LaudateCorpus1/JQuery-Mobile
|
opened
|
CVE-2018-14042 (Medium) detected in bootstrap-3.3.1.min.js
|
security vulnerability
|
## CVE-2018-14042 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.1.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.1/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.1/js/bootstrap.min.js</a></p>
<p>Path to vulnerable library: /js/boostrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LaudateCorpus1/JQuery-Mobile/commit/a0d9ec452c9a1910d7f9dcf8bd62ed007ae52d13">a0d9ec452c9a1910d7f9dcf8bd62ed007ae52d13</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-14042 (Medium) detected in bootstrap-3.3.1.min.js - ## CVE-2018-14042 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.1.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.1/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.1/js/bootstrap.min.js</a></p>
<p>Path to vulnerable library: /js/boostrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LaudateCorpus1/JQuery-Mobile/commit/a0d9ec452c9a1910d7f9dcf8bd62ed007ae52d13">a0d9ec452c9a1910d7f9dcf8bd62ed007ae52d13</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library js boostrap min js dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the data container property of tooltip publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org webjars npm bootstrap org webjars bootstrap step up your open source security game with whitesource
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.