Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
20,475
| 27,132,324,529
|
IssuesEvent
|
2023-02-16 10:39:34
|
gobuffalo/middleware
|
https://api.github.com/repos/gobuffalo/middleware
|
closed
|
process: consolidate the default middlewares into a single module
|
process
|
As of today, we have 7 official middleware modules and 4 of them are used in a new app by default. Now, these middleware are stable and have no much update recently except dependency updates. To reduce the maintenance effort on these modules, and also to simplify the module dependency structure, I would like to consolidate the default middleware (which are `mw-csrf`, `mw-forcessl`, `mw-i18n`, and `mw-paramlogger`) into a single module. Each of them will be a separate package in the module but will share the same module version and other stuff.
The process will be:
* [x] prepare the new module by merging them all together into this repository
* [x] update and clean up all the old modules before deprecating them, and releasing them as the last versions.
* [x] improve the CLI to use the new module for a new app
* [x] improve the CLI fixer to update existing apps, but this is optional
Hey @paganotoni, as we discussed before, I am going to deprecate the old default middleware modules (4 modules) and merge them into this module. Please take a look at the rough process above and let me know if you have any concern or comment on it. ETA for this work is the end of this month roughly.
Related tasks:
* [x] gobuffalo/mw-paramlogger/pull/12
* [x] gobuffalo/mw-i18n/pull/36
* [x] gobuffalo/mw-forcessl/pull/4
* [x] gobuffalo/mw-csrf/pull/15
* [x] gobuffalo/mw-contenttype/pull/3
|
1.0
|
process: consolidate the default middlewares into a single module - As of today, we have 7 official middleware modules and 4 of them are used in a new app by default. Now, these middleware are stable and have no much update recently except dependency updates. To reduce the maintenance effort on these modules, and also to simplify the module dependency structure, I would like to consolidate the default middleware (which are `mw-csrf`, `mw-forcessl`, `mw-i18n`, and `mw-paramlogger`) into a single module. Each of them will be a separate package in the module but will share the same module version and other stuff.
The process will be:
* [x] prepare the new module by merging them all together into this repository
* [x] update and clean up all the old modules before deprecating them, and releasing them as the last versions.
* [x] improve the CLI to use the new module for a new app
* [x] improve the CLI fixer to update existing apps, but this is optional
Hey @paganotoni, as we discussed before, I am going to deprecate the old default middleware modules (4 modules) and merge them into this module. Please take a look at the rough process above and let me know if you have any concern or comment on it. ETA for this work is the end of this month roughly.
Related tasks:
* [x] gobuffalo/mw-paramlogger/pull/12
* [x] gobuffalo/mw-i18n/pull/36
* [x] gobuffalo/mw-forcessl/pull/4
* [x] gobuffalo/mw-csrf/pull/15
* [x] gobuffalo/mw-contenttype/pull/3
|
process
|
process consolidate the default middlewares into a single module as of today we have official middleware modules and of them are used in a new app by default now these middleware are stable and have no much update recently except dependency updates to reduce the maintenance effort on these modules and also to simplify the module dependency structure i would like to consolidate the default middleware which are mw csrf mw forcessl mw and mw paramlogger into a single module each of them will be a separate package in the module but will share the same module version and other stuff the process will be prepare the new module by merging them all together into this repository update and clean up all the old modules before deprecating them and releasing them as the last versions improve the cli to use the new module for a new app improve the cli fixer to update existing apps but this is optional hey paganotoni as we discussed before i am going to deprecate the old default middleware modules modules and merge them into this module please take a look at the rough process above and let me know if you have any concern or comment on it eta for this work is the end of this month roughly related tasks gobuffalo mw paramlogger pull gobuffalo mw pull gobuffalo mw forcessl pull gobuffalo mw csrf pull gobuffalo mw contenttype pull
| 1
|
37,745
| 8,513,817,439
|
IssuesEvent
|
2018-10-31 16:57:09
|
robolectric/robolectric
|
https://api.github.com/repos/robolectric/robolectric
|
closed
|
android.support.v7.app.AlertDialog crash when try to open
|
defect
|
Using the new support library with the material alert dialog, when the dialog try to open the test crash with this exception:
Robolectric version: 3.0-SNAPSHOT
Android Studio 1.2 RC3
emulateSDK = 21
Code:
```
android.support.v7.app.AlertDialog.show();
```
Stacktrace:
```
java.lang.IllegalStateException: Could not execute method of the activity
...
Caused by: java.lang.reflect.InvocationTargetException
..
Caused by: java.lang.StringIndexOutOfBoundsException: String index out of range: -9
at java.lang.String.substring(String.java:1954)
at org.robolectric.res.ResName.qualifyResName(ResName.java:51)
at org.robolectric.res.Attribute.getStyleReference(Attribute.java:147)
at org.robolectric.res.builder.XmlFileBuilder$XmlResourceParserImpl.getResourceId(XmlFileBuilder.java:786)
at org.robolectric.res.builder.XmlFileBuilder$XmlResourceParserImpl.getStyleAttribute(XmlFileBuilder.java:769)
at org.robolectric.shadows.ShadowResources.attrsToTypedArray(ShadowResources.java:152)
at org.robolectric.shadows.ShadowResources.access$000(ShadowResources.java:51)
at org.robolectric.shadows.ShadowResources$ShadowTheme.obtainStyledAttributes(ShadowResources.java:489)
at android.content.res.Resources$Theme.obtainStyledAttributes(Resources.java)
at android.content.Context.obtainStyledAttributes(Context.java:460)
at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:708)
at android.view.LayoutInflater.rInflate(LayoutInflater.java:806)
at android.view.LayoutInflater.rInflate(LayoutInflater.java:809)
at android.view.LayoutInflater.rInflate(LayoutInflater.java:809)
at android.view.LayoutInflater.inflate(LayoutInflater.java:504)
at android.view.LayoutInflater.inflate(LayoutInflater.java:414)
at android.view.LayoutInflater.inflate(LayoutInflater.java:365)
at android.support.v7.app.AppCompatDelegateImplV7.setContentView(AppCompatDelegateImplV7.java:249)
at android.support.v7.app.AppCompatDialog.setContentView(AppCompatDialog.java:75)
at android.support.v7.app.AlertController.installContent(AlertController.java:216)
at android.support.v7.app.AlertDialog.onCreate(AlertDialog.java:240)
at android.app.Dialog.dispatchOnCreate(Dialog.java:373)
at android.app.Dialog.show(Dialog.java:274)
at org.robolectric.shadows.ShadowDialog.show(ShadowDialog.java:65)
at android.app.Dialog.show(Dialog.java)
```
Dependency:
```
compile 'com.android.support:appcompat-v7:22.1.0'
```
|
1.0
|
android.support.v7.app.AlertDialog crash when try to open - Using the new support library with the material alert dialog, when the dialog try to open the test crash with this exception:
Robolectric version: 3.0-SNAPSHOT
Android Studio 1.2 RC3
emulateSDK = 21
Code:
```
android.support.v7.app.AlertDialog.show();
```
Stacktrace:
```
java.lang.IllegalStateException: Could not execute method of the activity
...
Caused by: java.lang.reflect.InvocationTargetException
..
Caused by: java.lang.StringIndexOutOfBoundsException: String index out of range: -9
at java.lang.String.substring(String.java:1954)
at org.robolectric.res.ResName.qualifyResName(ResName.java:51)
at org.robolectric.res.Attribute.getStyleReference(Attribute.java:147)
at org.robolectric.res.builder.XmlFileBuilder$XmlResourceParserImpl.getResourceId(XmlFileBuilder.java:786)
at org.robolectric.res.builder.XmlFileBuilder$XmlResourceParserImpl.getStyleAttribute(XmlFileBuilder.java:769)
at org.robolectric.shadows.ShadowResources.attrsToTypedArray(ShadowResources.java:152)
at org.robolectric.shadows.ShadowResources.access$000(ShadowResources.java:51)
at org.robolectric.shadows.ShadowResources$ShadowTheme.obtainStyledAttributes(ShadowResources.java:489)
at android.content.res.Resources$Theme.obtainStyledAttributes(Resources.java)
at android.content.Context.obtainStyledAttributes(Context.java:460)
at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:708)
at android.view.LayoutInflater.rInflate(LayoutInflater.java:806)
at android.view.LayoutInflater.rInflate(LayoutInflater.java:809)
at android.view.LayoutInflater.rInflate(LayoutInflater.java:809)
at android.view.LayoutInflater.inflate(LayoutInflater.java:504)
at android.view.LayoutInflater.inflate(LayoutInflater.java:414)
at android.view.LayoutInflater.inflate(LayoutInflater.java:365)
at android.support.v7.app.AppCompatDelegateImplV7.setContentView(AppCompatDelegateImplV7.java:249)
at android.support.v7.app.AppCompatDialog.setContentView(AppCompatDialog.java:75)
at android.support.v7.app.AlertController.installContent(AlertController.java:216)
at android.support.v7.app.AlertDialog.onCreate(AlertDialog.java:240)
at android.app.Dialog.dispatchOnCreate(Dialog.java:373)
at android.app.Dialog.show(Dialog.java:274)
at org.robolectric.shadows.ShadowDialog.show(ShadowDialog.java:65)
at android.app.Dialog.show(Dialog.java)
```
Dependency:
```
compile 'com.android.support:appcompat-v7:22.1.0'
```
|
non_process
|
android support app alertdialog crash when try to open using the new support library with the material alert dialog when the dialog try to open the test crash with this exception robolectric version snapshot android studio emulatesdk code android support app alertdialog show stacktrace java lang illegalstateexception could not execute method of the activity caused by java lang reflect invocationtargetexception caused by java lang stringindexoutofboundsexception string index out of range at java lang string substring string java at org robolectric res resname qualifyresname resname java at org robolectric res attribute getstylereference attribute java at org robolectric res builder xmlfilebuilder xmlresourceparserimpl getresourceid xmlfilebuilder java at org robolectric res builder xmlfilebuilder xmlresourceparserimpl getstyleattribute xmlfilebuilder java at org robolectric shadows shadowresources attrstotypedarray shadowresources java at org robolectric shadows shadowresources access shadowresources java at org robolectric shadows shadowresources shadowtheme obtainstyledattributes shadowresources java at android content res resources theme obtainstyledattributes resources java at android content context obtainstyledattributes context java at android view layoutinflater createviewfromtag layoutinflater java at android view layoutinflater rinflate layoutinflater java at android view layoutinflater rinflate layoutinflater java at android view layoutinflater rinflate layoutinflater java at android view layoutinflater inflate layoutinflater java at android view layoutinflater inflate layoutinflater java at android view layoutinflater inflate layoutinflater java at android support app setcontentview java at android support app appcompatdialog setcontentview appcompatdialog java at android support app alertcontroller installcontent alertcontroller java at android support app alertdialog oncreate alertdialog java at android app dialog dispatchoncreate dialog java at android app dialog show dialog java at org robolectric shadows shadowdialog show shadowdialog java at android app dialog show dialog java dependency compile com android support appcompat
| 0
|
16,881
| 22,160,959,955
|
IssuesEvent
|
2022-06-04 13:59:14
|
tornadoweb/tornado
|
https://api.github.com/repos/tornadoweb/tornado
|
closed
|
process, tcpserver: Soft-deprecate multi-process mode
|
tcpserver process
|
Multi-process mode has a lot of subtleties; it is responsible for a large fraction of our questions on stack overflow. And now that we have SO_REUSEPORT, I see little reason to use it instead of multiple independent processes (I've always viewed multiple independent processes as desirable for "serious" production use, but the need for an external load balancer is a significant operational overhead that justifies the shared-listener multiprocess mode. But if you can share a port without fork and its associates subtlety, that's even better)
We should at least update the docs to steer people towards reusing ports instead of forking multiple-processes. We might want to go as far as deprecating the process argument to `server.start()`, although I'm in no rush to actually remove it (I don't think I'd deprecate the `process.fork_processes()` function - empirically the people who get in trouble with multi-process mode are all using `start(N)`; explicit forking is much more explicit and either the people who try it are able to get it working without difficulty or no one's trying it).
|
1.0
|
process, tcpserver: Soft-deprecate multi-process mode - Multi-process mode has a lot of subtleties; it is responsible for a large fraction of our questions on stack overflow. And now that we have SO_REUSEPORT, I see little reason to use it instead of multiple independent processes (I've always viewed multiple independent processes as desirable for "serious" production use, but the need for an external load balancer is a significant operational overhead that justifies the shared-listener multiprocess mode. But if you can share a port without fork and its associates subtlety, that's even better)
We should at least update the docs to steer people towards reusing ports instead of forking multiple-processes. We might want to go as far as deprecating the process argument to `server.start()`, although I'm in no rush to actually remove it (I don't think I'd deprecate the `process.fork_processes()` function - empirically the people who get in trouble with multi-process mode are all using `start(N)`; explicit forking is much more explicit and either the people who try it are able to get it working without difficulty or no one's trying it).
|
process
|
process tcpserver soft deprecate multi process mode multi process mode has a lot of subtleties it is responsible for a large fraction of our questions on stack overflow and now that we have so reuseport i see little reason to use it instead of multiple independent processes i ve always viewed multiple independent processes as desirable for serious production use but the need for an external load balancer is a significant operational overhead that justifies the shared listener multiprocess mode but if you can share a port without fork and its associates subtlety that s even better we should at least update the docs to steer people towards reusing ports instead of forking multiple processes we might want to go as far as deprecating the process argument to server start although i m in no rush to actually remove it i don t think i d deprecate the process fork processes function empirically the people who get in trouble with multi process mode are all using start n explicit forking is much more explicit and either the people who try it are able to get it working without difficulty or no one s trying it
| 1
|
15,593
| 19,719,133,569
|
IssuesEvent
|
2022-01-13 13:56:24
|
googleapis/docuploader
|
https://api.github.com/repos/googleapis/docuploader
|
closed
|
Update storage pin to v2
|
priority: p2 type: process
|
trying to update storage requirements on doc-templates repo: https://github.com/googleapis/doc-templates/pull/337, but cannot be done unless storage version is bumped here as well, if possible.
|
1.0
|
Update storage pin to v2 - trying to update storage requirements on doc-templates repo: https://github.com/googleapis/doc-templates/pull/337, but cannot be done unless storage version is bumped here as well, if possible.
|
process
|
update storage pin to trying to update storage requirements on doc templates repo but cannot be done unless storage version is bumped here as well if possible
| 1
|
58,452
| 16,541,803,477
|
IssuesEvent
|
2021-05-27 17:48:16
|
opencaching/opencaching-pl
|
https://api.github.com/repos/opencaching/opencaching-pl
|
opened
|
Google Analytics doesn't work
|
Component Core Priority Medium Type Defect
|
From (about) 1 april 2021 Google Analytics doesn't work on OC pages (I've checked OC PL and OC UK):

|
1.0
|
Google Analytics doesn't work - From (about) 1 april 2021 Google Analytics doesn't work on OC pages (I've checked OC PL and OC UK):

|
non_process
|
google analytics doesn t work from about april google analytics doesn t work on oc pages i ve checked oc pl and oc uk
| 0
|
16,696
| 21,795,485,690
|
IssuesEvent
|
2022-05-15 15:09:50
|
uncrustify/uncrustify
|
https://api.github.com/repos/uncrustify/uncrustify
|
closed
|
Incorrect spacing added after address-of operator in macro when macro has an argument.
|
C and C++11 Preprocessor
|
[debug.txt](https://github.com/uncrustify/uncrustify/files/8164688/debug.txt)
test.c:
```
#define my_thing \
&object1, &object2
#define my_thing(stuff) \
&object1, &object2
```
Expect no change, but get:
```
#define my_thing \
&object1, &object2
#define my_thing(stuff) \
& object1, &object2
```
(Notice the extra space after the ampersand)
```
➜ Downloads uncrustify -v
Uncrustify-0.73.0_f
```
|
1.0
|
Incorrect spacing added after address-of operator in macro when macro has an argument. - [debug.txt](https://github.com/uncrustify/uncrustify/files/8164688/debug.txt)
test.c:
```
#define my_thing \
&object1, &object2
#define my_thing(stuff) \
&object1, &object2
```
Expect no change, but get:
```
#define my_thing \
&object1, &object2
#define my_thing(stuff) \
& object1, &object2
```
(Notice the extra space after the ampersand)
```
➜ Downloads uncrustify -v
Uncrustify-0.73.0_f
```
|
process
|
incorrect spacing added after address of operator in macro when macro has an argument test c define my thing define my thing stuff expect no change but get define my thing define my thing stuff notice the extra space after the ampersand ➜ downloads uncrustify v uncrustify f
| 1
|
158,330
| 13,728,698,270
|
IssuesEvent
|
2020-10-04 12:52:53
|
teamcadi/android-tutorial
|
https://api.github.com/repos/teamcadi/android-tutorial
|
opened
|
3주차. AlertDialog와 ListView
|
documentation enhancement
|
- Dialog/AlertDialog 개념 정리
- 배열과 ArrayList를 통한 ListView 기초 다지기 및 개념 정리
- Lombok, Junit, log4J/SLF4j 실습
|
1.0
|
3주차. AlertDialog와 ListView - - Dialog/AlertDialog 개념 정리
- 배열과 ArrayList를 통한 ListView 기초 다지기 및 개념 정리
- Lombok, Junit, log4J/SLF4j 실습
|
non_process
|
alertdialog와 listview dialog alertdialog 개념 정리 배열과 arraylist를 통한 listview 기초 다지기 및 개념 정리 lombok junit 실습
| 0
|
4,824
| 7,719,783,552
|
IssuesEvent
|
2018-05-23 20:34:57
|
freesteph/peach-melpa.org
|
https://api.github.com/repos/freesteph/peach-melpa.org
|
opened
|
Ability to to refine by brightness
|
image processing
|
Some users have asked for a way to tell light and dark themes. Since that notion isn't exactly black or white (ha! most fun I've ever had on a GitHub issue) it would be nice to use ImageMagick to get the level of brightness and offer a toggle/slider depending on the accuracy/usefulness of these respective controls.
|
1.0
|
Ability to to refine by brightness - Some users have asked for a way to tell light and dark themes. Since that notion isn't exactly black or white (ha! most fun I've ever had on a GitHub issue) it would be nice to use ImageMagick to get the level of brightness and offer a toggle/slider depending on the accuracy/usefulness of these respective controls.
|
process
|
ability to to refine by brightness some users have asked for a way to tell light and dark themes since that notion isn t exactly black or white ha most fun i ve ever had on a github issue it would be nice to use imagemagick to get the level of brightness and offer a toggle slider depending on the accuracy usefulness of these respective controls
| 1
|
7,814
| 10,966,337,191
|
IssuesEvent
|
2019-11-28 06:41:22
|
hashgraph/hedera-mirror-node
|
https://api.github.com/repos/hashgraph/hedera-mirror-node
|
closed
|
HCS Setup
|
P1 enhancement hcs process
|
**Problem**
The first step to start HCS development is to create a Maven module to contain the GRPC and protobuf relate artifacts.
**Solution**
See `Setup` section in HCS design document (PR #365)
**Alternatives**
**Additional Context**
|
1.0
|
HCS Setup - **Problem**
The first step to start HCS development is to create a Maven module to contain the GRPC and protobuf relate artifacts.
**Solution**
See `Setup` section in HCS design document (PR #365)
**Alternatives**
**Additional Context**
|
process
|
hcs setup problem the first step to start hcs development is to create a maven module to contain the grpc and protobuf relate artifacts solution see setup section in hcs design document pr alternatives additional context
| 1
|
172,207
| 21,041,957,649
|
IssuesEvent
|
2022-03-31 13:06:39
|
SmartBear/ready-api-plugins
|
https://api.github.com/repos/SmartBear/ready-api-plugins
|
opened
|
CVE-2020-36518 (High) detected in jackson-databind-2.4.2.jar
|
security vulnerability
|
## CVE-2020-36518 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.4.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: /ready-api-plugin-template/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.4.2/jackson-databind-2.4.2.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-1.7.0.jar (Root Library)
- jackson-datatype-joda-2.4.2.jar
- :x: **jackson-databind-2.4.2.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.
WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.
<p>Publish Date: 2022-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2816">https://github.com/FasterXML/jackson-databind/issues/2816</a></p>
<p>Release Date: 2022-03-11</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.12.6.1,2.13.2.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.4.2","packageFilePaths":["/ready-api-plugin-template/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;com.fasterxml.jackson.datatype:jackson-datatype-joda:2.4.2;com.fasterxml.jackson.core:jackson-databind:2.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.12.6.1,2.13.2.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36518","vulnerabilityDetails":"jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.\n WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-36518 (High) detected in jackson-databind-2.4.2.jar - ## CVE-2020-36518 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.4.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: /ready-api-plugin-template/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.4.2/jackson-databind-2.4.2.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-1.7.0.jar (Root Library)
- jackson-datatype-joda-2.4.2.jar
- :x: **jackson-databind-2.4.2.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.
WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.
<p>Publish Date: 2022-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2816">https://github.com/FasterXML/jackson-databind/issues/2816</a></p>
<p>Release Date: 2022-03-11</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.12.6.1,2.13.2.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.4.2","packageFilePaths":["/ready-api-plugin-template/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;com.fasterxml.jackson.datatype:jackson-datatype-joda:2.4.2;com.fasterxml.jackson.core:jackson-databind:2.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.12.6.1,2.13.2.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36518","vulnerabilityDetails":"jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.\n WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to dependency file ready api plugin template pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy ready api soapui pro jar root library jackson datatype joda jar x jackson databind jar vulnerable library found in base branch master vulnerability details jackson databind before allows a java stackoverflow exception and denial of service via a large depth of nested objects whitesource note after conducting further research whitesource has determined that all versions of com fasterxml jackson core jackson databind up to version are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com smartbear ready api soapui pro com fasterxml jackson datatype jackson datatype joda com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails jackson databind before allows a java stackoverflow exception and denial of service via a large depth of nested objects n whitesource note after conducting further research whitesource has determined that all versions of com fasterxml jackson core jackson databind up to version are vulnerable to cve vulnerabilityurl
| 0
|
147,320
| 13,205,953,710
|
IssuesEvent
|
2020-08-14 19:05:07
|
onflow/flow-js-sdk
|
https://api.github.com/repos/onflow/flow-js-sdk
|
closed
|
Create build-limit
|
JS documentation
|
# build-limit
### Background
Moving builders out of the SDK package into their own single purpose package will enable better code health, smaller bundle size and generally decrease entropy between packages when things inevitably change and progress.
### Expected Outcomes
- [ ] @onflow/sdk-build-limit is created
|
1.0
|
Create build-limit - # build-limit
### Background
Moving builders out of the SDK package into their own single purpose package will enable better code health, smaller bundle size and generally decrease entropy between packages when things inevitably change and progress.
### Expected Outcomes
- [ ] @onflow/sdk-build-limit is created
|
non_process
|
create build limit build limit background moving builders out of the sdk package into their own single purpose package will enable better code health smaller bundle size and generally decrease entropy between packages when things inevitably change and progress expected outcomes onflow sdk build limit is created
| 0
|
6,685
| 9,805,502,574
|
IssuesEvent
|
2019-06-12 09:08:14
|
EthVM/EthVM
|
https://api.github.com/repos/EthVM/EthVM
|
closed
|
After some time, fetching blocks decreases to 0
|
bug project:processing
|
* **I'm submitting a ...**
- [ ] feature request
- [X] bug report
* **Bug Report**
I run twice on two different Ropsten envs and after sometime the fetching of new blocks decreases to 0 and never recovers:
```
[2019-06-07 22:55:08,033] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:08,033] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:08,046] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:08,046] DEBUG Next range. Current tail = 913882, current head = 5750432 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:08,046] DEBUG Next range from chain tracker: 913882..913882 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:08,046] DEBUG Range = 913882..913882, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:08,046] DEBUG Fetching range: 913882..913882. Next = 913882..913882 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:08,910] DEBUG Parity request: 913882, 913882, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:08,925] DEBUG Total blocks fetched = 913883, fetch count = 1. Elapsed = 879 ms, target fetch = 1000 ms, % of target fetch = 0.879, trace count = 835, avg trace count = 1 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:08,925] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:08,925] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:08,975] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:08,975] DEBUG Next range. Current tail = 913883, current head = 5750432 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:08,976] DEBUG Next range from chain tracker: 913883..913883 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:08,976] DEBUG Range = 913883..913883, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:08,976] DEBUG Fetching range: 913883..913883. Next = 913883..913883 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:10,174] DEBUG Parity request: 913883, 913883, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:10,188] DEBUG Total blocks fetched = 913884, fetch count = 1. Elapsed = 1212 ms, target fetch = 1000 ms, % of target fetch = 1.212, trace count = 836, avg trace count = 1 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:10,188] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:10,188] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:10,210] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:10,210] DEBUG Next range. Current tail = 913884, current head = 5750432 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:10,210] DEBUG Next range from chain tracker: 913884..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:10,210] DEBUG Range = 913884..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:10,210] DEBUG Fetching range: 913884..913884. Next = 913884..913884 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:12,265] DEBUG Parity request: 913884, 913884, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:12,281] DEBUG Total blocks fetched = 913885, fetch count = 1. Elapsed = 2071 ms, target fetch = 1000 ms, % of target fetch = 2.071, trace count = 838, avg trace count = 1 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:12,281] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:12,281] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:12,301] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:12,301] DEBUG Next range. Current tail = 913885, current head = 5750432 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:12,301] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:12,301] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:12,302] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:12,302] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:12,385] DEBUG Trying to reset tail to 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:12,385] DEBUG Current tail = 913885 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:12,385] DEBUG After reset attempt tail = 913885 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:13,302] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:13,302] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:13,302] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:13,302] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:13,302] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:13,302] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:14,302] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:14,303] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:14,303] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:14,303] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:14,303] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:14,303] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:15,303] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:15,303] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:15,303] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:15,303] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:15,303] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:15,303] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:16,304] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:16,304] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:16,304] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:16,304] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:16,304] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:16,304] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:17,304] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:17,304] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:17,304] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:17,304] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:17,305] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:17,305] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:18,305] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:18,305] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:18,305] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:18,305] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:18,305] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:18,305] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:19,305] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:19,306] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:19,306] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:19,306] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:19,306] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:19,306] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:20,306] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:20,307] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:20,307] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:20,307] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:20,307] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:20,307] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:21,307] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:21,307] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:21,307] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:21,307] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:21,307] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:21,307] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:22,308] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:22,308] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:22,309] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:22,309] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:22,309] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:22,309] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:23,309] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:23,309] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:23,309] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:23,309] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:23,309] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:23,309] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:24,309] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:24,310] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:24,310] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:24,310] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:24,310] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:24,310] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:24,538] WARN WorkerSinkTask{id=postgres-trace-sink-0} Commit of offsets timed out (org.apache.kafka.connect.runtime.WorkerSinkTask)
[2019-06-07 22:55:25,310] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:25,310] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:25,310] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[Another execution different from previous one]
[2019-06-08 13:01:48,342] DEBUG Total blocks fetched = 599402, fetch count = 1. Elapsed = 453 ms, target fetch = 1000 ms, % of target fetch = 0.453, trace count = 689, avg trace count = 1 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,342] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,342] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:48,352] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:48,352] DEBUG Next range. Current tail = 599402, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:48,352] DEBUG Next range from chain tracker: 599402..599403 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:48,352] DEBUG Range = 599402..599403, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,352] DEBUG Fetching range: 599402..599402. Next = 599402..599402 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,352] DEBUG Fetching range: 599403..599403. Next = 599403..599403 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,421] DEBUG Parity request: 599402, 599402, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,563] DEBUG Parity request: 599403, 599403, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,593] DEBUG Total blocks fetched = 599404, fetch count = 2. Elapsed = 241 ms, target fetch = 1000 ms, % of target fetch = 0.241, trace count = 340, avg trace count = 1 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,593] DEBUG Reorg size = 0, records size = 11 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,593] DEBUG Polled 11 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:48,598] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:48,598] DEBUG Next range. Current tail = 599404, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:48,598] DEBUG Next range from chain tracker: 599404..599407 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:48,598] DEBUG Range = 599404..599407, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,598] DEBUG Fetching range: 599404..599404. Next = 599404..599404 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,598] DEBUG Fetching range: 599405..599405. Next = 599405..599405 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,598] DEBUG Fetching range: 599406..599406. Next = 599406..599406 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,598] DEBUG Fetching range: 599407..599407. Next = 599407..599407 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,851] DEBUG Parity request: 599404, 599404, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,910] DEBUG Parity request: 599405, 599405, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,940] DEBUG Parity request: 599406, 599406, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:49,958] DEBUG Parity request: 599407, 599407, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:50,169] DEBUG Total blocks fetched = 599408, fetch count = 4. Elapsed = 1571 ms, target fetch = 1000 ms, % of target fetch = 1.571, trace count = 4587, avg trace count = 4 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:50,169] DEBUG Reorg size = 0, records size = 21 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:50,169] DEBUG Polled 21 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:50,320] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:50,322] DEBUG Next range. Current tail = 599408, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:50,322] DEBUG Next range from chain tracker: 599408..599409 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:50,322] DEBUG Range = 599408..599409, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:50,323] DEBUG Fetching range: 599408..599408. Next = 599408..599408 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:50,323] DEBUG Fetching range: 599409..599409. Next = 599409..599409 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:51,095] DEBUG Parity request: 599409, 599409, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:51,361] DEBUG Parity request: 599408, 599408, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:51,534] DEBUG Total blocks fetched = 599410, fetch count = 2. Elapsed = 1212 ms, target fetch = 1000 ms, % of target fetch = 1.212, trace count = 8935, avg trace count = 16 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:51,534] DEBUG Reorg size = 0, records size = 11 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:51,534] DEBUG Polled 11 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:51,666] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:51,666] DEBUG Next range. Current tail = 599410, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:51,666] DEBUG Next range from chain tracker: 599410..599411 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:51,666] DEBUG Range = 599410..599411, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:51,666] DEBUG Fetching range: 599411..599411. Next = 599411..599411 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:51,666] DEBUG Fetching range: 599410..599410. Next = 599410..599410 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:52,999] DEBUG Parity request: 599411, 599411, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:53,874] DEBUG Parity request: 599410, 599410, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:54,384] DEBUG Total blocks fetched = 599412, fetch count = 2. Elapsed = 2718 ms, target fetch = 1000 ms, % of target fetch = 2.718, trace count = 20348, avg trace count = 21 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:54,384] DEBUG Reorg size = 0, records size = 11 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:54,384] DEBUG Polled 11 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:54,640] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:54,640] DEBUG Next range. Current tail = 599412, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:54,640] DEBUG Next range from chain tracker: 599412..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:54,640] DEBUG Range = 599412..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:54,640] DEBUG Fetching range: 599412..599412. Next = 599412..599412 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:55,872] DEBUG Parity request: 599412, 599412, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:56,157] DEBUG Total blocks fetched = 599413, fetch count = 1. Elapsed = 1517 ms, target fetch = 1000 ms, % of target fetch = 1.517, trace count = 8474, avg trace count = 11 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:56,157] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:56,158] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:56,233] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:56,233] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:56,233] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:56,233] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:56,233] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:56,233] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:57,233] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:57,233] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:57,233] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:57,233] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:57,233] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:57,234] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:58,234] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:58,234] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:58,234] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:58,234] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:58,234] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:58,234] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:59,234] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:59,234] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:59,234] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:59,234] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:59,234] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:59,235] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:00,235] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:00,235] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:00,235] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:00,235] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:00,235] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:00,235] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:01,235] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:01,235] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:01,235] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:01,236] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:01,236] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:01,236] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:02,236] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:02,236] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:02,236] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:02,236] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:02,236] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:02,236] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:03,236] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:03,236] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:03,236] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:03,236] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:03,236] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:03,236] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:04,237] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:04,237] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:04,237] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:04,237] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:04,237] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:04,237] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:05,237] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:05,237] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:05,237] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:05,237] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:05,237] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:05,238] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:05,851] DEBUG Trying to reset tail to 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:05,852] DEBUG Current tail = 599413 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:05,852] DEBUG After reset attempt tail = 599413 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:06,238] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:06,238] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:06,238] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:06,238] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:06,238] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:06,238] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:07,241] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:07,242] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:07,242] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:07,242] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:07,242] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:07,242] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:08,242] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:08,242] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:08,242] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:08,242] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:08,242] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:08,242] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:09,242] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:09,242] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:09,242] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:09,242] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:09,242] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:09,242] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:10,243] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:10,243] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:10,243] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:10,243] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:10,243] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:10,243] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:11,243] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:11,243] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:11,243] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:11,243] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:11,243] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:11,243] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:12,243] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:12,243] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:12,244] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:12,244] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:12,244] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:12,244] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:13,244] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:13,244] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:13,244] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:13,244] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:13,244] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:13,244] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
```
|
1.0
|
After some time, fetching blocks decreases to 0 - * **I'm submitting a ...**
- [ ] feature request
- [X] bug report
* **Bug Report**
I run twice on two different Ropsten envs and after sometime the fetching of new blocks decreases to 0 and never recovers:
```
[2019-06-07 22:55:08,033] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:08,033] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:08,046] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:08,046] DEBUG Next range. Current tail = 913882, current head = 5750432 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:08,046] DEBUG Next range from chain tracker: 913882..913882 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:08,046] DEBUG Range = 913882..913882, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:08,046] DEBUG Fetching range: 913882..913882. Next = 913882..913882 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:08,910] DEBUG Parity request: 913882, 913882, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:08,925] DEBUG Total blocks fetched = 913883, fetch count = 1. Elapsed = 879 ms, target fetch = 1000 ms, % of target fetch = 0.879, trace count = 835, avg trace count = 1 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:08,925] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:08,925] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:08,975] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:08,975] DEBUG Next range. Current tail = 913883, current head = 5750432 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:08,976] DEBUG Next range from chain tracker: 913883..913883 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:08,976] DEBUG Range = 913883..913883, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:08,976] DEBUG Fetching range: 913883..913883. Next = 913883..913883 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:10,174] DEBUG Parity request: 913883, 913883, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:10,188] DEBUG Total blocks fetched = 913884, fetch count = 1. Elapsed = 1212 ms, target fetch = 1000 ms, % of target fetch = 1.212, trace count = 836, avg trace count = 1 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:10,188] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:10,188] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:10,210] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:10,210] DEBUG Next range. Current tail = 913884, current head = 5750432 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:10,210] DEBUG Next range from chain tracker: 913884..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:10,210] DEBUG Range = 913884..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:10,210] DEBUG Fetching range: 913884..913884. Next = 913884..913884 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:12,265] DEBUG Parity request: 913884, 913884, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-07 22:55:12,281] DEBUG Total blocks fetched = 913885, fetch count = 1. Elapsed = 2071 ms, target fetch = 1000 ms, % of target fetch = 2.071, trace count = 838, avg trace count = 1 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:12,281] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:12,281] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:12,301] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:12,301] DEBUG Next range. Current tail = 913885, current head = 5750432 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:12,301] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:12,301] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:12,302] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:12,302] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:12,385] DEBUG Trying to reset tail to 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:12,385] DEBUG Current tail = 913885 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:12,385] DEBUG After reset attempt tail = 913885 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:13,302] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:13,302] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:13,302] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:13,302] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:13,302] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:13,302] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:14,302] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:14,303] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:14,303] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:14,303] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:14,303] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:14,303] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:15,303] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:15,303] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:15,303] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:15,303] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:15,303] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:15,303] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:16,304] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:16,304] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:16,304] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:16,304] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:16,304] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:16,304] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:17,304] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:17,304] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:17,304] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:17,304] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:17,305] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:17,305] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:18,305] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:18,305] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:18,305] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:18,305] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:18,305] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:18,305] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:19,305] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:19,306] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:19,306] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:19,306] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:19,306] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:19,306] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:20,306] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:20,307] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:20,307] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:20,307] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:20,307] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:20,307] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:21,307] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:21,307] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:21,307] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:21,307] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:21,307] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:21,307] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:22,308] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:22,308] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:22,309] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:22,309] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:22,309] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:22,309] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:23,309] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:23,309] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:23,309] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:23,309] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:23,309] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:23,309] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:24,309] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:24,310] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:24,310] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:24,310] DEBUG Range = 913885..913884, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:24,310] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-07 22:55:24,310] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:24,538] WARN WorkerSinkTask{id=postgres-trace-sink-0} Commit of offsets timed out (org.apache.kafka.connect.runtime.WorkerSinkTask)
[2019-06-07 22:55:25,310] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-07 22:55:25,310] DEBUG Next range. Current tail = 913885, current head = 5750433 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-07 22:55:25,310] DEBUG Next range from chain tracker: 913885..913884 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[Another execution different from previous one]
[2019-06-08 13:01:48,342] DEBUG Total blocks fetched = 599402, fetch count = 1. Elapsed = 453 ms, target fetch = 1000 ms, % of target fetch = 0.453, trace count = 689, avg trace count = 1 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,342] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,342] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:48,352] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:48,352] DEBUG Next range. Current tail = 599402, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:48,352] DEBUG Next range from chain tracker: 599402..599403 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:48,352] DEBUG Range = 599402..599403, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,352] DEBUG Fetching range: 599402..599402. Next = 599402..599402 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,352] DEBUG Fetching range: 599403..599403. Next = 599403..599403 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,421] DEBUG Parity request: 599402, 599402, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,563] DEBUG Parity request: 599403, 599403, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,593] DEBUG Total blocks fetched = 599404, fetch count = 2. Elapsed = 241 ms, target fetch = 1000 ms, % of target fetch = 0.241, trace count = 340, avg trace count = 1 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,593] DEBUG Reorg size = 0, records size = 11 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,593] DEBUG Polled 11 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:48,598] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:48,598] DEBUG Next range. Current tail = 599404, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:48,598] DEBUG Next range from chain tracker: 599404..599407 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:48,598] DEBUG Range = 599404..599407, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:48,598] DEBUG Fetching range: 599404..599404. Next = 599404..599404 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,598] DEBUG Fetching range: 599405..599405. Next = 599405..599405 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,598] DEBUG Fetching range: 599406..599406. Next = 599406..599406 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,598] DEBUG Fetching range: 599407..599407. Next = 599407..599407 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,851] DEBUG Parity request: 599404, 599404, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,910] DEBUG Parity request: 599405, 599405, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:48,940] DEBUG Parity request: 599406, 599406, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:49,958] DEBUG Parity request: 599407, 599407, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:50,169] DEBUG Total blocks fetched = 599408, fetch count = 4. Elapsed = 1571 ms, target fetch = 1000 ms, % of target fetch = 1.571, trace count = 4587, avg trace count = 4 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:50,169] DEBUG Reorg size = 0, records size = 21 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:50,169] DEBUG Polled 21 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:50,320] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:50,322] DEBUG Next range. Current tail = 599408, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:50,322] DEBUG Next range from chain tracker: 599408..599409 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:50,322] DEBUG Range = 599408..599409, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:50,323] DEBUG Fetching range: 599408..599408. Next = 599408..599408 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:50,323] DEBUG Fetching range: 599409..599409. Next = 599409..599409 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:51,095] DEBUG Parity request: 599409, 599409, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:51,361] DEBUG Parity request: 599408, 599408, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:51,534] DEBUG Total blocks fetched = 599410, fetch count = 2. Elapsed = 1212 ms, target fetch = 1000 ms, % of target fetch = 1.212, trace count = 8935, avg trace count = 16 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:51,534] DEBUG Reorg size = 0, records size = 11 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:51,534] DEBUG Polled 11 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:51,666] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:51,666] DEBUG Next range. Current tail = 599410, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:51,666] DEBUG Next range from chain tracker: 599410..599411 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:51,666] DEBUG Range = 599410..599411, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:51,666] DEBUG Fetching range: 599411..599411. Next = 599411..599411 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:51,666] DEBUG Fetching range: 599410..599410. Next = 599410..599410 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:52,999] DEBUG Parity request: 599411, 599411, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:53,874] DEBUG Parity request: 599410, 599410, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:54,384] DEBUG Total blocks fetched = 599412, fetch count = 2. Elapsed = 2718 ms, target fetch = 1000 ms, % of target fetch = 2.718, trace count = 20348, avg trace count = 21 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:54,384] DEBUG Reorg size = 0, records size = 11 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:54,384] DEBUG Polled 11 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:54,640] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:54,640] DEBUG Next range. Current tail = 599412, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:54,640] DEBUG Next range from chain tracker: 599412..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:54,640] DEBUG Range = 599412..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:54,640] DEBUG Fetching range: 599412..599412. Next = 599412..599412 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:55,872] DEBUG Parity request: 599412, 599412, 1000 (com.ethvm.kafka.connect.sources.web3.sources.ParityFullBlockSource)
[2019-06-08 13:01:56,157] DEBUG Total blocks fetched = 599413, fetch count = 1. Elapsed = 1517 ms, target fetch = 1000 ms, % of target fetch = 1.517, trace count = 8474, avg trace count = 11 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:56,157] DEBUG Reorg size = 0, records size = 6 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:56,158] DEBUG Polled 6 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:56,233] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:56,233] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:56,233] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:56,233] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:56,233] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:56,233] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:57,233] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:57,233] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:57,233] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:57,233] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:57,233] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:57,234] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:58,234] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:58,234] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:58,234] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:58,234] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:58,234] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:58,234] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:59,234] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:01:59,234] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:59,234] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:01:59,234] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:59,234] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:01:59,235] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:00,235] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:00,235] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:00,235] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:00,235] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:00,235] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:00,235] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:01,235] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:01,235] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:01,235] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:01,236] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:01,236] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:01,236] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:02,236] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:02,236] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:02,236] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:02,236] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:02,236] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:02,236] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:03,236] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:03,236] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:03,236] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:03,236] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:03,236] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:03,236] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:04,237] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:04,237] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:04,237] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:04,237] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:04,237] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:04,237] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:05,237] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:05,237] DEBUG Next range. Current tail = 599413, current head = 5753124 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:05,237] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:05,237] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:05,237] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:05,238] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:05,851] DEBUG Trying to reset tail to 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:05,852] DEBUG Current tail = 599413 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:05,852] DEBUG After reset attempt tail = 599413 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:06,238] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:06,238] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:06,238] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:06,238] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:06,238] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:06,238] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:07,241] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:07,242] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:07,242] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:07,242] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:07,242] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:07,242] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:08,242] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:08,242] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:08,242] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:08,242] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:08,242] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:08,242] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:09,242] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:09,242] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:09,242] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:09,242] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:09,242] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:09,242] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:10,243] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:10,243] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:10,243] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:10,243] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:10,243] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:10,243] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:11,243] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:11,243] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:11,243] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:11,243] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:11,243] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:11,243] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:12,243] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:12,243] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:12,244] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:12,244] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:12,244] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:12,244] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:13,244] DEBUG Polling (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
[2019-06-08 13:02:13,244] DEBUG Next range. Current tail = 599413, current head = 5753125 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:13,244] DEBUG Next range from chain tracker: 599413..599412 (com.ethvm.kafka.connect.sources.web3.tracker.CanonicalChainTracker)
[2019-06-08 13:02:13,244] DEBUG Range = 599413..599412, reOrgs = [] (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:13,244] DEBUG Reorg size = 0, records size = 0 (com.ethvm.kafka.connect.sources.web3.sources.AbstractParityEntitySource)
[2019-06-08 13:02:13,244] DEBUG Polled 0 records (com.ethvm.kafka.connect.sources.web3.ParitySourceTask)
```
|
process
|
after some time fetching blocks decreases to i m submitting a feature request bug report bug report i run twice on two different ropsten envs and after sometime the fetching of new blocks decreases to and never recovers debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug total blocks fetched fetch count elapsed ms target fetch ms of target fetch trace count avg trace count com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug total blocks fetched fetch count elapsed ms target fetch ms of target fetch trace count avg trace count com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug total blocks fetched fetch count elapsed ms target fetch ms of target fetch trace count avg trace count com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug trying to reset tail to com ethvm kafka connect sources tracker canonicalchaintracker debug current tail com ethvm kafka connect sources tracker canonicalchaintracker debug after reset attempt tail com ethvm kafka connect sources tracker canonicalchaintracker debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask warn workersinktask id postgres trace sink commit of offsets timed out org apache kafka connect runtime workersinktask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug total blocks fetched fetch count elapsed ms target fetch ms of target fetch trace count avg trace count com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug total blocks fetched fetch count elapsed ms target fetch ms of target fetch trace count avg trace count com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug total blocks fetched fetch count elapsed ms target fetch ms of target fetch trace count avg trace count com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug total blocks fetched fetch count elapsed ms target fetch ms of target fetch trace count avg trace count com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug total blocks fetched fetch count elapsed ms target fetch ms of target fetch trace count avg trace count com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug fetching range next com ethvm kafka connect sources sources parityfullblocksource debug parity request com ethvm kafka connect sources sources parityfullblocksource debug total blocks fetched fetch count elapsed ms target fetch ms of target fetch trace count avg trace count com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug trying to reset tail to com ethvm kafka connect sources tracker canonicalchaintracker debug current tail com ethvm kafka connect sources tracker canonicalchaintracker debug after reset attempt tail com ethvm kafka connect sources tracker canonicalchaintracker debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask debug polling com ethvm kafka connect sources paritysourcetask debug next range current tail current head com ethvm kafka connect sources tracker canonicalchaintracker debug next range from chain tracker com ethvm kafka connect sources tracker canonicalchaintracker debug range reorgs com ethvm kafka connect sources sources abstractparityentitysource debug reorg size records size com ethvm kafka connect sources sources abstractparityentitysource debug polled records com ethvm kafka connect sources paritysourcetask
| 1
|
21,824
| 30,316,761,926
|
IssuesEvent
|
2023-07-10 16:05:12
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
New term - tribe
|
Term - add Class - Taxon normative Process - complete
|
Was https://code.google.com/p/darwincore/issues/detail?id=147
==New Term Recommendation==
Submitter: David Remsen
Justification: Many taxa belong to large taxonomic families, particularly among the insects but also among other animal and plant groups. Often the 'interesting' components of the classification are operating at the sub-familial-and-below level. In order to effectively capture this more refined classification information, I recommend the addition of tribe (and sub-tribe).
Definition: The full scientific name of the tribe in which the taxon is classified.
Comment: Examples "Ortaliini", "Arethuseae"
Refines:dwc:Family
Has Domain: http://rs.tdwg.org/dwc/terms/Taxon
Has Range:
Replaces:
ABCD 2.06:
Oct 3, 2013 comment #5 gtuco.btuco
I would like to promote the adoption of the concept mentioned in this issue. To do so, I will need a stronger proposal demonstrating the need to share this information - that is, that independent groups, organizations, projects have the same need and can reach a consensus proposal about how the term should be used. It might be a good idea to circulate the proposal on tdwg-content and see if a community can be built around and support the addition.
|
1.0
|
New term - tribe - Was https://code.google.com/p/darwincore/issues/detail?id=147
==New Term Recommendation==
Submitter: David Remsen
Justification: Many taxa belong to large taxonomic families, particularly among the insects but also among other animal and plant groups. Often the 'interesting' components of the classification are operating at the sub-familial-and-below level. In order to effectively capture this more refined classification information, I recommend the addition of tribe (and sub-tribe).
Definition: The full scientific name of the tribe in which the taxon is classified.
Comment: Examples "Ortaliini", "Arethuseae"
Refines:dwc:Family
Has Domain: http://rs.tdwg.org/dwc/terms/Taxon
Has Range:
Replaces:
ABCD 2.06:
Oct 3, 2013 comment #5 gtuco.btuco
I would like to promote the adoption of the concept mentioned in this issue. To do so, I will need a stronger proposal demonstrating the need to share this information - that is, that independent groups, organizations, projects have the same need and can reach a consensus proposal about how the term should be used. It might be a good idea to circulate the proposal on tdwg-content and see if a community can be built around and support the addition.
|
process
|
new term tribe was new term recommendation submitter david remsen justification many taxa belong to large taxonomic families particularly among the insects but also among other animal and plant groups often the interesting components of the classification are operating at the sub familial and below level in order to effectively capture this more refined classification information i recommend the addition of tribe and sub tribe definition the full scientific name of the tribe in which the taxon is classified comment examples ortaliini arethuseae refines dwc family has domain has range replaces abcd oct comment gtuco btuco i would like to promote the adoption of the concept mentioned in this issue to do so i will need a stronger proposal demonstrating the need to share this information that is that independent groups organizations projects have the same need and can reach a consensus proposal about how the term should be used it might be a good idea to circulate the proposal on tdwg content and see if a community can be built around and support the addition
| 1
|
16,764
| 21,938,146,803
|
IssuesEvent
|
2022-05-23 15:24:56
|
googleapis/google-cloud-go
|
https://api.github.com/repos/googleapis/google-cloud-go
|
closed
|
internal/apidiff: apidiff is suddenly failing broadly on all PRs
|
priority: p2 type: process
|
Suddenly, the `internal/apidiff` tool is failing (when run via internal/kokoro/check_incompat_changes.sh). We don't know why but since it is blocking every PR, we must disable it until we can figure it out.
[Example job result](https://source.cloud.google.com/results/invocations/f6800f4c-f577-4c39-9af1-b4bb17c014cf/targets/cloud-devrel%2Fclient-libraries%2Fgo%2Fgoogle-cloud-go%2Fpresubmit%2Flatest-version/log).
|
1.0
|
internal/apidiff: apidiff is suddenly failing broadly on all PRs - Suddenly, the `internal/apidiff` tool is failing (when run via internal/kokoro/check_incompat_changes.sh). We don't know why but since it is blocking every PR, we must disable it until we can figure it out.
[Example job result](https://source.cloud.google.com/results/invocations/f6800f4c-f577-4c39-9af1-b4bb17c014cf/targets/cloud-devrel%2Fclient-libraries%2Fgo%2Fgoogle-cloud-go%2Fpresubmit%2Flatest-version/log).
|
process
|
internal apidiff apidiff is suddenly failing broadly on all prs suddenly the internal apidiff tool is failing when run via internal kokoro check incompat changes sh we don t know why but since it is blocking every pr we must disable it until we can figure it out
| 1
|
144,872
| 5,547,199,214
|
IssuesEvent
|
2017-03-23 04:24:31
|
nossas/bonde-client
|
https://api.github.com/repos/nossas/bonde-client
|
closed
|
Replace process.env.BROWSER to exenv lib
|
Category: DevOps Priority: Highest Stage: Done Type: Improvement
|
# Introduction
It seems that the `process.env.BROWSER` variable is does not working.. In replacement of this node environment variable, we should use the [exenv](https://github.com/JedWatson/exenv) library
|
1.0
|
Replace process.env.BROWSER to exenv lib - # Introduction
It seems that the `process.env.BROWSER` variable is does not working.. In replacement of this node environment variable, we should use the [exenv](https://github.com/JedWatson/exenv) library
|
non_process
|
replace process env browser to exenv lib introduction it seems that the process env browser variable is does not working in replacement of this node environment variable we should use the library
| 0
|
6,464
| 9,546,602,166
|
IssuesEvent
|
2019-05-01 20:25:46
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
closed
|
Department of State - Progress Bar
|
Apply Process Approved Requirements Ready State Dept.
|
Who: Student Applicant
What: Progress bar in the Application process
Why: As a student applicant I want to see my progress and be able to see what else needs to be done
A/C
- There will be a progress bar at the top of each page (same design in USAJOBS application process)
- There will be 6 numbered steps
- Select Internships
- Experience & References
- Education & Transcripts
- Languages & Skills
- Statement of Interest
- Review Application (this is not in the design)
- The circle will be a solid green once the applicant has completed that page
- The circle will be outlined in green when the applicant is on the page
- The circle will be outlined in white for pages the applicant hasn't completed yet
- A student must complete a page prior to moving forward to the next page.
- Once a page has been completed, however, they can navigate backward and forward to the page by clicking on the number.
InVision Link: https://opm.invisionapp.com/d/main/#/console/15360465/319289320/preview
Public Link: https://opm.invisionapp.com/share/ZEPNZR09Q54
|
1.0
|
Department of State - Progress Bar - Who: Student Applicant
What: Progress bar in the Application process
Why: As a student applicant I want to see my progress and be able to see what else needs to be done
A/C
- There will be a progress bar at the top of each page (same design in USAJOBS application process)
- There will be 6 numbered steps
- Select Internships
- Experience & References
- Education & Transcripts
- Languages & Skills
- Statement of Interest
- Review Application (this is not in the design)
- The circle will be a solid green once the applicant has completed that page
- The circle will be outlined in green when the applicant is on the page
- The circle will be outlined in white for pages the applicant hasn't completed yet
- A student must complete a page prior to moving forward to the next page.
- Once a page has been completed, however, they can navigate backward and forward to the page by clicking on the number.
InVision Link: https://opm.invisionapp.com/d/main/#/console/15360465/319289320/preview
Public Link: https://opm.invisionapp.com/share/ZEPNZR09Q54
|
process
|
department of state progress bar who student applicant what progress bar in the application process why as a student applicant i want to see my progress and be able to see what else needs to be done a c there will be a progress bar at the top of each page same design in usajobs application process there will be numbered steps select internships experience references education transcripts languages skills statement of interest review application this is not in the design the circle will be a solid green once the applicant has completed that page the circle will be outlined in green when the applicant is on the page the circle will be outlined in white for pages the applicant hasn t completed yet a student must complete a page prior to moving forward to the next page once a page has been completed however they can navigate backward and forward to the page by clicking on the number invision link public link
| 1
|
13,851
| 8,380,494,299
|
IssuesEvent
|
2018-10-07 15:18:31
|
nudibranchrecords/hedron
|
https://api.github.com/repos/nudibranchrecords/hedron
|
opened
|
Macro React components have poor performance
|
bug performance
|
After some testing in the profiler, it appears somewhere on the macro page is causing a bit of jank
|
True
|
Macro React components have poor performance - After some testing in the profiler, it appears somewhere on the macro page is causing a bit of jank
|
non_process
|
macro react components have poor performance after some testing in the profiler it appears somewhere on the macro page is causing a bit of jank
| 0
|
14,241
| 17,172,627,525
|
IssuesEvent
|
2021-07-15 07:26:22
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[FEATURE] Allow running model files direct from standalone qgis_process tool
|
3.14 Automatic new feature Processing
|
Original commit: https://github.com/qgis/QGIS/commit/bfadf47d16eb5b297f84547bfaac7384b1b38d22 by nyalldawson
Unfortunately this naughty coder did not write a description... :-(
|
1.0
|
[FEATURE] Allow running model files direct from standalone qgis_process tool - Original commit: https://github.com/qgis/QGIS/commit/bfadf47d16eb5b297f84547bfaac7384b1b38d22 by nyalldawson
Unfortunately this naughty coder did not write a description... :-(
|
process
|
allow running model files direct from standalone qgis process tool original commit by nyalldawson unfortunately this naughty coder did not write a description
| 1
|
119,856
| 25,594,749,714
|
IssuesEvent
|
2022-12-01 15:25:09
|
CleverRaven/Cataclysm-DDA
|
https://api.github.com/repos/CleverRaven/Cataclysm-DDA
|
closed
|
Allow each different save files to choose different default tilesets for themselves (or default to the global default tileset).
|
<Enhancement / Feature> <Suggestion / Discussion> SDL: Tiles / Sound Character / World Generation stale Code: Debug
|
# Is your feature request related to a problem? Please describe.
Cataclysm is a game that can offer drastically different experiences in tone and game-feel depending on what kind of mod you chose to enable. For example, a basic, DDA-only game will differ a lot from an Aftershock enabled one, a Magiclysm one, a No Hope one, or a Dark Days of the Dead one. Sometimes I want to switch from one game with a particular tone to another with a vastly different tone, but the game won't switch to the right tileset by itself so I'd have to manually navigate the GFX menu and do the switch myself, which personally ruins the experience for me.
# Describe the solution you'd like
In the world creation menu, there could be an option to choose your preferred tileset that would only apply to the world that you are currently making, so that when you switch between save files, it would automatically load that tileset for you, instead of using only one default tileset. The game should of course give you the option to default to whatever tileset that the game is using as the global default.
# Describe alternatives you've considered
# Additional context
|
1.0
|
Allow each different save files to choose different default tilesets for themselves (or default to the global default tileset). - # Is your feature request related to a problem? Please describe.
Cataclysm is a game that can offer drastically different experiences in tone and game-feel depending on what kind of mod you chose to enable. For example, a basic, DDA-only game will differ a lot from an Aftershock enabled one, a Magiclysm one, a No Hope one, or a Dark Days of the Dead one. Sometimes I want to switch from one game with a particular tone to another with a vastly different tone, but the game won't switch to the right tileset by itself so I'd have to manually navigate the GFX menu and do the switch myself, which personally ruins the experience for me.
# Describe the solution you'd like
In the world creation menu, there could be an option to choose your preferred tileset that would only apply to the world that you are currently making, so that when you switch between save files, it would automatically load that tileset for you, instead of using only one default tileset. The game should of course give you the option to default to whatever tileset that the game is using as the global default.
# Describe alternatives you've considered
# Additional context
|
non_process
|
allow each different save files to choose different default tilesets for themselves or default to the global default tileset is your feature request related to a problem please describe cataclysm is a game that can offer drastically different experiences in tone and game feel depending on what kind of mod you chose to enable for example a basic dda only game will differ a lot from an aftershock enabled one a magiclysm one a no hope one or a dark days of the dead one sometimes i want to switch from one game with a particular tone to another with a vastly different tone but the game won t switch to the right tileset by itself so i d have to manually navigate the gfx menu and do the switch myself which personally ruins the experience for me describe the solution you d like in the world creation menu there could be an option to choose your preferred tileset that would only apply to the world that you are currently making so that when you switch between save files it would automatically load that tileset for you instead of using only one default tileset the game should of course give you the option to default to whatever tileset that the game is using as the global default describe alternatives you ve considered additional context
| 0
|
12,376
| 14,897,085,141
|
IssuesEvent
|
2021-01-21 11:15:29
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
Missing h in http panic error
|
bug/0-needs-info kind/bug process/candidate team/client
|
<!--
Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client.
Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports
-->
## Bug description
<!-- A clear and concise description of what the bug is. -->
Link is missing an h
```
PrismaClientUnknownRequestError2 [PrismaClientUnknownRequestError]:
Invalid `prisma.executeRaw()` invocation:
PANIC in /root/.cargo/git/checkouts/quaint-9f01e008b9a89c14/bc3aa8a/src/connector/postgres/error.rs:138:29
index out of bounds: the len is 1 but the index is 1
This is a non-recoverable error which probably happens when the Prisma Query Engine has a panic.
ttps://github.com/prisma/prisma/issues/new?body=Hi+Prisma+Team%21+My+Prisma+Client+just+crashed.+This+is+the+report%3A%0A%23%23+Versions%0A%0A%7C+Name++++++++++++%7C+Version++++++++++++%7C%0A%7C-----------------%7C--------------------%7C%0A%7C+Node++++++++++++%7C+v12.16.1+++++++++++%7C+%0A%7C+OS++++++++++++++%7C+darwin+++++++++++++%7C%0A%7C+Prisma+Client+++%7C+2.14.0+++++++++++++%7C%0A%7C+Query+Engine++++%7C+query-engine+5d491261d382a2a5ffdc71de17072b0e409f1cc1%7C%0A%7C+Database++++++++%7C+undefined%7C%0A%0A%0A%0A%23%23+Query%0A%60%60%60%0Amutation+%7B%0A++executeRaw%28%0A++++query%3A+%22X%22%0A++++parameters%3A+%22X%22%0A++%29%0A%7D%0A%60%60%60%0A%0A%23%23+Logs%0A%60%60%60%0A++tryLoadEnv+Environment+variables+loaded+from+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2F.env++%0A++tryLoadEnv+Environment+variables+loaded+from+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2F.env++%0A++prisma-client+%7B+clientVersion%3A+%272.14.0%27+%7D++%0A++engine+%7B+cwd%3A+%27%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fprisma%27+%7D++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++engine+%7B%0A++engine+++flags%3A+%5B%0A++engine+++++%27--enable-raw-queries%27%2C%0A++engine+++++%27--unix-path%27%2C%0A++engine+++++%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%5D%0A++engine+%7D++%0A++engine+stdout++Starting+a+postgresql+pool+with+13+connections.++%0A++engine+stdout++Started+http+server+on+http%2Bunix%3A%2F%2F%2Fprivate%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++prisma-client+prisma.%24executeRaw%28SET+jwt.claims.user_id+%3D+10%2C+%5B%5D%29++%0A++prisma-client+Prisma+Client+call%3A++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++prisma-client+prisma.%24executeRaw%28select+jwt.claims.user_id%2C+%5B%5D%29++%0A++prisma-client+Prisma+Client+call%3A++%0A++engine+stdout++PANIC+in+%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%3A138%3A29%0Aindex+out+of+bounds%3A+the+len+is+1+but+the+index+is+1++%0A++engine+TypeError%3A+this.currentRequestPromise.cancel+is+not+a+function%0A++engine+++++at+NodeEngine.handlePanic+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26344%3A36%29%0A++engine+++++at+NodeEngine.setError+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26290%3A16%29%0A++engine+++++at+LineStream.%3Canonymous%3E+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26585%3A24%29%0A++engine+++++at+LineStream.emit+%28events.js%3A311%3A20%29%0A++engine+++++at+addChunk+%28_stream_readable.js%3A294%3A12%29%0A++engine+++++at+readableAddChunk+%28_stream_readable.js%3A275%3A11%29%0A++engine+++++at+LineStream.Readable.push+%28_stream_readable.js%3A209%3A10%29%0A++engine+++++at+LineStream.Transform.push+%28_stream_transform.js%3A152%3A32%29%0A++engine+++++at+LineStream._pushBuffer+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A23721%3A19%29%0A++engine+++++at+LineStream._transform+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A23715%3A10%29++%7B%22timestamp%22%3A%22Jan+15+17%3A04%3A51.789%22%2C%22level%22%3A%22ERROR%22%2C%22fields%22%3A%7B%22message%22%3A%22PANIC%22%2C%22reason%22%3A%22index+out+of+bounds%3A+the+len+is+1+but+the+index+is+1%22%2C%22file%22%3A%22%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%22%2C%22line%22%3A138%2C%22column%22%3A29%7D%2C%22target%22%3A%22query_engine%22%7D++%0A++engine+%7B%0A++engine+++error%3A+SocketError%3A+closed%0A++engine+++++++at+Socket.onSocketClose+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A24792%3A33%29%0A++engine+++++++at+Socket.emit+%28events.js%3A311%3A20%29%0A++engine+++++++at+Pipe.%3Canonymous%3E+%28net.js%3A668%3A12%29+%7B%0A++engine+++++name%3A+%27SocketError%27%2C%0A++engine+++++code%3A+%27UND_ERR_SOCKET%27%0A++engine+++%7D%0A++engine+%7D++%0A++engine+%7B%0A++engine+++error%3A+Error%3A+connect+ECONNREFUSED+%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%0A++engine+++++++at+PipeConnectWrap.afterConnect+%5Bas+oncomplete%5D+%28net.js%3A1137%3A16%29+%7B%0A++engine+++++errno%3A+%27ECONNREFUSED%27%2C%0A++engine+++++code%3A+%27ECONNREFUSED%27%2C%0A++engine+++++syscall%3A+%27connect%27%2C%0A++engine+++++address%3A+%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%7D%0A++engine+%7D++%0A++engine+Client+Version%3A+2.14.0++%0A++engine+Engine+Version%3A+query-engine+5d491261d382a2a5ffdc71de17072b0e409f1cc1++%0A++engine+Active+provider%3A+postgresql++%0A++engine+%7B%0A++engine+++error%3A+Error%3A+connect+ECONNREFUSED+%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%0A++engine+++++++at+PipeConnectWrap.afterConnect+%5Bas+oncomplete%5D+%28net.js%3A1137%3A16%29+%7B%0A++engine+++++errno%3A+%27ECONNREFUSED%27%2C%0A++engine+++++code%3A+%27ECONNREFUSED%27%2C%0A++engine+++++syscall%3A+%27connect%27%2C%0A++engine+++++address%3A+%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%7D%0A++engine+%7D++%2B1s%0A%60%60%60%0A%0A%23%23+Client+Snippet%0A%60%60%60ts%0A%2F%2F+PLEASE+FILL+YOUR+CODE+SNIPPET+HERE%0A%60%60%60%0A%0A%23%23+Schema%0A%60%60%60prisma%0A%2F%2F+PLEASE+ADD+YOUR+SCHEMA+HERE+IF+POSSIBLE%0A%60%60%60%0A&title=PANIC+in+%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%3A138%3A29%0Aindex+out+of+bounds%3A+the+len+is+1+but+the+index+is+1&template=bug_report.mdhttps://github.com/prisma/prisma/issues/new?body=Hi+Prisma+Team%21+My+Prisma+Client+just+crashed.+This+is+the+report%3A%0A%23%23+Versions%0A%0A%7C+Name++++++++++++%7C+Version++++++++++++%7C%0A%7C-----------------%7C--------------------%7C%0A%7C+Node++++++++++++%7C+v12.16.1+++++++++++%7C+%0A%7C+OS++++++++++++++%7C+darwin+++++++++++++%7C%0A%7C+Prisma+Client+++%7C+2.14.0+++++++++++++%7C%0A%7C+Query+Engine++++%7C+query-engine+5d491261d382a2a5ffdc71de17072b0e409f1cc1%7C%0A%7C+Database++++++++%7C+undefined%7C%0A%0A%0A%0A%23%23+Query%0A%60%60%60%0Amutation+%7B%0A++executeRaw%28%0A++++query%3A+%22X%22%0A++++parameters%3A+%22X%22%0A++%29%0A%7D%0A%60%60%60%0A%0A%23%23+Logs%0A%60%60%60%0A++tryLoadEnv+Environment+variables+loaded+from+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2F.env++%0A++tryLoadEnv+Environment+variables+loaded+from+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2F.env++%0A++prisma-client+%7B+clientVersion%3A+%272.14.0%27+%7D++%0A++engine+%7B+cwd%3A+%27%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fprisma%27+%7D++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++engine+%7B%0A++engine+++flags%3A+%5B%0A++engine+++++%27--enable-raw-queries%27%2C%0A++engine+++++%27--unix-path%27%2C%0A++engine+++++%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%5D%0A++engine+%7D++%0A++engine+stdout++Starting+a+postgresql+pool+with+13+connections.++%0A++engine+stdout++Started+http+server+on+http%2Bunix%3A%2F%2F%2Fprivate%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++prisma-client+prisma.%24executeRaw%28SET+jwt.claims.user_id+%3D+10%2C+%5B%5D%29++%0A++prisma-client+Prisma+Client+call%3A++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++prisma-client+prisma.%24executeRaw%28select+jwt.claims.user_id%2C+%5B%5D%29++%0A++prisma-client+Prisma+Client+call%3A++%0A++engine+stdout++PANIC+in+%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%3A138%3A29%0Aindex+out+of+bounds%3A+the+len+is+1+but+the+index+is+1++%0A++engine+TypeError%3A+this.currentRequestPromise.cancel+is+not+a+function%0A++engine+++++at+NodeEngine.handlePanic+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26344%3A36%29%0A++engine+++++at+NodeEngine.setError+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26290%3A16%29%0A++engine+++++at+LineStream.%3Canonymous%3E+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26585%3A24%29%0A++engine+++++at+LineStream.emit+%28events.js%3A311%3A20%29%0A++engine+++++at+addChunk+%28_stream_readable.js%3A294%3A12%29%0A++engine+++++at+readableAddChunk+%28_stream_readable.js%3A275%3A11%29%0A++engine+++++at+LineStream.Readable.push+%28_stream_readable.js%3A209%3A10%29%0A++engine+++++at+LineStream.Transform.push+%28_stream_transform.js%3A152%3A32%29%0A++engine+++++at+LineStream._pushBuffer+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A23721%3A19%29%0A++engine+++++at+LineStream._transform+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A23715%3A10%29++%7B%22timestamp%22%3A%22Jan+15+17%3A04%3A51.789%22%2C%22level%22%3A%22ERROR%22%2C%22fields%22%3A%7B%22message%22%3A%22PANIC%22%2C%22reason%22%3A%22index+out+of+bounds%3A+the+len+is+1+but+the+index+is+1%22%2C%22file%22%3A%22%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%22%2C%22line%22%3A138%2C%22column%22%3A29%7D%2C%22target%22%3A%22query_engine%22%7D++%0A++engine+%7B%0A++engine+++error%3A+SocketError%3A+closed%0A++engine+++++++at+Socket.onSocketClose+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A24792%3A33%29%0A++engine+++++++at+Socket.emit+%28events.js%3A311%3A20%29%0A++engine+++++++at+Pipe.%3Canonymous%3E+%28net.js%3A668%3A12%29+%7B%0A++engine+++++name%3A+%27SocketError%27%2C%0A++engine+++++code%3A+%27UND_ERR_SOCKET%27%0A++engine+++%7D%0A++engine+%7D++%0A++engine+%7B%0A++engine+++error%3A+Error%3A+connect+ECONNREFUSED+%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%0A++engine+++++++at+PipeConnectWrap.afterConnect+%5Bas+oncomplete%5D+%28net.js%3A1137%3A16%29+%7B%0A++engine+++++errno%3A+%27ECONNREFUSED%27%2C%0A++engine+++++code%3A+%27ECONNREFUSED%27%2C%0A++engine+++++syscall%3A+%27connect%27%2C%0A++engine+++++address%3A+%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%7D%0A++engine+%7D++%0A++engine+Client+Version%3A+2.14.0++%0A++engine+Engine+Version%3A+query-engine+5d491261d382a2a5ffdc71de17072b0e409f1cc1++%0A++engine+Active+provider%3A+postgresql++%0A++engine+%7B%0A++engine+++error%3A+Error%3A+connect+ECONNREFUSED+%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%0A++engine+++++++at+PipeConnectWrap.afterConnect+%5Bas+oncomplete%5D+%28net.js%3A1137%3A16%29+%7B%0A++engine+++++errno%3A+%27ECONNREFUSED%27%2C%0A++engine+++++code%3A+%27ECONNREFUSED%27%2C%0A++engine+++++syscall%3A+%27connect%27%2C%0A++engine+++++address%3A+%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%7D%0A++engine+%7D++%2B1s%0A%60%60%60%0A%0A%23%23+Client+Snippet%0A%60%60%60ts%0A%2F%2F+PLEASE+FILL+YOUR+CODE+SNIPPET+HERE%0A%60%60%60%0A%0A%23%23+Schema%0A%60%60%60prisma%0A%2F%2F+PLEASE+ADD+YOUR+SCHEMA+HERE+IF+POSSIBLE%0A%60%60%60%0A&title=PANIC+in+%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%3A138%3A29%0Aindex+out+of+bounds%3A+the+len+is+1+but+the+index+is+1&template=bug_report.md
If you want the Prisma team to look into it, please open the link above 🙏
To increase the chance of success, please post your schema and a snippet of
how you used Prisma Client in the issue.
at PrismaClientFetcher.request (/Users/m/Go/src/github.com/prisma/issue-4303/node_modules/@prisma/client/runtime/index.js:78134:15)
at processTicksAndRejections (internal/process/task_queues.js:97:5) {
clientVersion: '2.14.0'
}
```
## How to reproduce
<!--
Steps to reproduce the behavior:
1. Go to '...'
2. Change '....'
3. Run '....'
4. See error
-->
```js
const { PrismaClient} = require('@prisma/client')
const client= new PrismaClient()
async function main() {
await client.$connect()
const id = 10
await client.$executeRaw(`SET jwt.claims.user_id = ${id}`)
const result = await client.$executeRaw(`select jwt.claims.user_id`)
console.log(result)
}
main().catch(console.error).finally(() => client.$disconnect())
```
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
Should be able to click on the issue to create an issue
## Prisma information
<!-- Your Prisma schema, Prisma Client queries, ...
Do not include your database credentials when sharing your Prisma schema! -->
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]-->
- Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]-->
- Node.js version: <!--[Run `node -v` to see your Node.js version]-->
- Prisma version:
<!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]-->
```
Environment variables loaded from .env
@prisma/cli : 2.14.0
@prisma/client : 2.14.0
Current platform : darwin
Query Engine : query-engine 5d491261d382a2a5ffdc71de17072b0e409f1cc1 (at node_modules/@prisma/engines/query-engine-darwin)
Migration Engine : migration-engine-cli 5d491261d382a2a5ffdc71de17072b0e409f1cc1 (at node_modules/@prisma/engines/migration-engine-darwin)
Introspection Engine : introspection-core 5d491261d382a2a5ffdc71de17072b0e409f1cc1 (at node_modules/@prisma/engines/introspection-engine-darwin)
Format Binary : prisma-fmt 5d491261d382a2a5ffdc71de17072b0e409f1cc1 (at node_modules/@prisma/engines/prisma-fmt-darwin)
Studio : 0.332.0
```
|
1.0
|
Missing h in http panic error - <!--
Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client.
Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports
-->
## Bug description
<!-- A clear and concise description of what the bug is. -->
Link is missing an h
```
PrismaClientUnknownRequestError2 [PrismaClientUnknownRequestError]:
Invalid `prisma.executeRaw()` invocation:
PANIC in /root/.cargo/git/checkouts/quaint-9f01e008b9a89c14/bc3aa8a/src/connector/postgres/error.rs:138:29
index out of bounds: the len is 1 but the index is 1
This is a non-recoverable error which probably happens when the Prisma Query Engine has a panic.
ttps://github.com/prisma/prisma/issues/new?body=Hi+Prisma+Team%21+My+Prisma+Client+just+crashed.+This+is+the+report%3A%0A%23%23+Versions%0A%0A%7C+Name++++++++++++%7C+Version++++++++++++%7C%0A%7C-----------------%7C--------------------%7C%0A%7C+Node++++++++++++%7C+v12.16.1+++++++++++%7C+%0A%7C+OS++++++++++++++%7C+darwin+++++++++++++%7C%0A%7C+Prisma+Client+++%7C+2.14.0+++++++++++++%7C%0A%7C+Query+Engine++++%7C+query-engine+5d491261d382a2a5ffdc71de17072b0e409f1cc1%7C%0A%7C+Database++++++++%7C+undefined%7C%0A%0A%0A%0A%23%23+Query%0A%60%60%60%0Amutation+%7B%0A++executeRaw%28%0A++++query%3A+%22X%22%0A++++parameters%3A+%22X%22%0A++%29%0A%7D%0A%60%60%60%0A%0A%23%23+Logs%0A%60%60%60%0A++tryLoadEnv+Environment+variables+loaded+from+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2F.env++%0A++tryLoadEnv+Environment+variables+loaded+from+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2F.env++%0A++prisma-client+%7B+clientVersion%3A+%272.14.0%27+%7D++%0A++engine+%7B+cwd%3A+%27%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fprisma%27+%7D++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++engine+%7B%0A++engine+++flags%3A+%5B%0A++engine+++++%27--enable-raw-queries%27%2C%0A++engine+++++%27--unix-path%27%2C%0A++engine+++++%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%5D%0A++engine+%7D++%0A++engine+stdout++Starting+a+postgresql+pool+with+13+connections.++%0A++engine+stdout++Started+http+server+on+http%2Bunix%3A%2F%2F%2Fprivate%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++prisma-client+prisma.%24executeRaw%28SET+jwt.claims.user_id+%3D+10%2C+%5B%5D%29++%0A++prisma-client+Prisma+Client+call%3A++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++prisma-client+prisma.%24executeRaw%28select+jwt.claims.user_id%2C+%5B%5D%29++%0A++prisma-client+Prisma+Client+call%3A++%0A++engine+stdout++PANIC+in+%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%3A138%3A29%0Aindex+out+of+bounds%3A+the+len+is+1+but+the+index+is+1++%0A++engine+TypeError%3A+this.currentRequestPromise.cancel+is+not+a+function%0A++engine+++++at+NodeEngine.handlePanic+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26344%3A36%29%0A++engine+++++at+NodeEngine.setError+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26290%3A16%29%0A++engine+++++at+LineStream.%3Canonymous%3E+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26585%3A24%29%0A++engine+++++at+LineStream.emit+%28events.js%3A311%3A20%29%0A++engine+++++at+addChunk+%28_stream_readable.js%3A294%3A12%29%0A++engine+++++at+readableAddChunk+%28_stream_readable.js%3A275%3A11%29%0A++engine+++++at+LineStream.Readable.push+%28_stream_readable.js%3A209%3A10%29%0A++engine+++++at+LineStream.Transform.push+%28_stream_transform.js%3A152%3A32%29%0A++engine+++++at+LineStream._pushBuffer+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A23721%3A19%29%0A++engine+++++at+LineStream._transform+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A23715%3A10%29++%7B%22timestamp%22%3A%22Jan+15+17%3A04%3A51.789%22%2C%22level%22%3A%22ERROR%22%2C%22fields%22%3A%7B%22message%22%3A%22PANIC%22%2C%22reason%22%3A%22index+out+of+bounds%3A+the+len+is+1+but+the+index+is+1%22%2C%22file%22%3A%22%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%22%2C%22line%22%3A138%2C%22column%22%3A29%7D%2C%22target%22%3A%22query_engine%22%7D++%0A++engine+%7B%0A++engine+++error%3A+SocketError%3A+closed%0A++engine+++++++at+Socket.onSocketClose+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A24792%3A33%29%0A++engine+++++++at+Socket.emit+%28events.js%3A311%3A20%29%0A++engine+++++++at+Pipe.%3Canonymous%3E+%28net.js%3A668%3A12%29+%7B%0A++engine+++++name%3A+%27SocketError%27%2C%0A++engine+++++code%3A+%27UND_ERR_SOCKET%27%0A++engine+++%7D%0A++engine+%7D++%0A++engine+%7B%0A++engine+++error%3A+Error%3A+connect+ECONNREFUSED+%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%0A++engine+++++++at+PipeConnectWrap.afterConnect+%5Bas+oncomplete%5D+%28net.js%3A1137%3A16%29+%7B%0A++engine+++++errno%3A+%27ECONNREFUSED%27%2C%0A++engine+++++code%3A+%27ECONNREFUSED%27%2C%0A++engine+++++syscall%3A+%27connect%27%2C%0A++engine+++++address%3A+%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%7D%0A++engine+%7D++%0A++engine+Client+Version%3A+2.14.0++%0A++engine+Engine+Version%3A+query-engine+5d491261d382a2a5ffdc71de17072b0e409f1cc1++%0A++engine+Active+provider%3A+postgresql++%0A++engine+%7B%0A++engine+++error%3A+Error%3A+connect+ECONNREFUSED+%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%0A++engine+++++++at+PipeConnectWrap.afterConnect+%5Bas+oncomplete%5D+%28net.js%3A1137%3A16%29+%7B%0A++engine+++++errno%3A+%27ECONNREFUSED%27%2C%0A++engine+++++code%3A+%27ECONNREFUSED%27%2C%0A++engine+++++syscall%3A+%27connect%27%2C%0A++engine+++++address%3A+%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%7D%0A++engine+%7D++%2B1s%0A%60%60%60%0A%0A%23%23+Client+Snippet%0A%60%60%60ts%0A%2F%2F+PLEASE+FILL+YOUR+CODE+SNIPPET+HERE%0A%60%60%60%0A%0A%23%23+Schema%0A%60%60%60prisma%0A%2F%2F+PLEASE+ADD+YOUR+SCHEMA+HERE+IF+POSSIBLE%0A%60%60%60%0A&title=PANIC+in+%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%3A138%3A29%0Aindex+out+of+bounds%3A+the+len+is+1+but+the+index+is+1&template=bug_report.mdhttps://github.com/prisma/prisma/issues/new?body=Hi+Prisma+Team%21+My+Prisma+Client+just+crashed.+This+is+the+report%3A%0A%23%23+Versions%0A%0A%7C+Name++++++++++++%7C+Version++++++++++++%7C%0A%7C-----------------%7C--------------------%7C%0A%7C+Node++++++++++++%7C+v12.16.1+++++++++++%7C+%0A%7C+OS++++++++++++++%7C+darwin+++++++++++++%7C%0A%7C+Prisma+Client+++%7C+2.14.0+++++++++++++%7C%0A%7C+Query+Engine++++%7C+query-engine+5d491261d382a2a5ffdc71de17072b0e409f1cc1%7C%0A%7C+Database++++++++%7C+undefined%7C%0A%0A%0A%0A%23%23+Query%0A%60%60%60%0Amutation+%7B%0A++executeRaw%28%0A++++query%3A+%22X%22%0A++++parameters%3A+%22X%22%0A++%29%0A%7D%0A%60%60%60%0A%0A%23%23+Logs%0A%60%60%60%0A++tryLoadEnv+Environment+variables+loaded+from+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2F.env++%0A++tryLoadEnv+Environment+variables+loaded+from+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2F.env++%0A++prisma-client+%7B+clientVersion%3A+%272.14.0%27+%7D++%0A++engine+%7B+cwd%3A+%27%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fprisma%27+%7D++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++engine+%7B%0A++engine+++flags%3A+%5B%0A++engine+++++%27--enable-raw-queries%27%2C%0A++engine+++++%27--unix-path%27%2C%0A++engine+++++%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%5D%0A++engine+%7D++%0A++engine+stdout++Starting+a+postgresql+pool+with+13+connections.++%0A++engine+stdout++Started+http+server+on+http%2Bunix%3A%2F%2F%2Fprivate%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock++%0A++engine+Search+for+Query+Engine+in+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient++%0A++prisma-client+prisma.%24executeRaw%28SET+jwt.claims.user_id+%3D+10%2C+%5B%5D%29++%0A++prisma-client+Prisma+Client+call%3A++%0A++plusX+Execution+permissions+of+%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F.prisma%2Fclient%2Fquery-engine-darwin+are+fine++%0A++prisma-client+prisma.%24executeRaw%28select+jwt.claims.user_id%2C+%5B%5D%29++%0A++prisma-client+Prisma+Client+call%3A++%0A++engine+stdout++PANIC+in+%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%3A138%3A29%0Aindex+out+of+bounds%3A+the+len+is+1+but+the+index+is+1++%0A++engine+TypeError%3A+this.currentRequestPromise.cancel+is+not+a+function%0A++engine+++++at+NodeEngine.handlePanic+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26344%3A36%29%0A++engine+++++at+NodeEngine.setError+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26290%3A16%29%0A++engine+++++at+LineStream.%3Canonymous%3E+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A26585%3A24%29%0A++engine+++++at+LineStream.emit+%28events.js%3A311%3A20%29%0A++engine+++++at+addChunk+%28_stream_readable.js%3A294%3A12%29%0A++engine+++++at+readableAddChunk+%28_stream_readable.js%3A275%3A11%29%0A++engine+++++at+LineStream.Readable.push+%28_stream_readable.js%3A209%3A10%29%0A++engine+++++at+LineStream.Transform.push+%28_stream_transform.js%3A152%3A32%29%0A++engine+++++at+LineStream._pushBuffer+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A23721%3A19%29%0A++engine+++++at+LineStream._transform+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A23715%3A10%29++%7B%22timestamp%22%3A%22Jan+15+17%3A04%3A51.789%22%2C%22level%22%3A%22ERROR%22%2C%22fields%22%3A%7B%22message%22%3A%22PANIC%22%2C%22reason%22%3A%22index+out+of+bounds%3A+the+len+is+1+but+the+index+is+1%22%2C%22file%22%3A%22%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%22%2C%22line%22%3A138%2C%22column%22%3A29%7D%2C%22target%22%3A%22query_engine%22%7D++%0A++engine+%7B%0A++engine+++error%3A+SocketError%3A+closed%0A++engine+++++++at+Socket.onSocketClose+%28%2FUsers%2Fm%2FGo%2Fsrc%2Fgithub.com%2Fprisma%2Fissue-4303%2Fnode_modules%2F%40prisma%2Fclient%2Fruntime%2Findex.js%3A24792%3A33%29%0A++engine+++++++at+Socket.emit+%28events.js%3A311%3A20%29%0A++engine+++++++at+Pipe.%3Canonymous%3E+%28net.js%3A668%3A12%29+%7B%0A++engine+++++name%3A+%27SocketError%27%2C%0A++engine+++++code%3A+%27UND_ERR_SOCKET%27%0A++engine+++%7D%0A++engine+%7D++%0A++engine+%7B%0A++engine+++error%3A+Error%3A+connect+ECONNREFUSED+%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%0A++engine+++++++at+PipeConnectWrap.afterConnect+%5Bas+oncomplete%5D+%28net.js%3A1137%3A16%29+%7B%0A++engine+++++errno%3A+%27ECONNREFUSED%27%2C%0A++engine+++++code%3A+%27ECONNREFUSED%27%2C%0A++engine+++++syscall%3A+%27connect%27%2C%0A++engine+++++address%3A+%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%7D%0A++engine+%7D++%0A++engine+Client+Version%3A+2.14.0++%0A++engine+Engine+Version%3A+query-engine+5d491261d382a2a5ffdc71de17072b0e409f1cc1++%0A++engine+Active+provider%3A+postgresql++%0A++engine+%7B%0A++engine+++error%3A+Error%3A+connect+ECONNREFUSED+%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%0A++engine+++++++at+PipeConnectWrap.afterConnect+%5Bas+oncomplete%5D+%28net.js%3A1137%3A16%29+%7B%0A++engine+++++errno%3A+%27ECONNREFUSED%27%2C%0A++engine+++++code%3A+%27ECONNREFUSED%27%2C%0A++engine+++++syscall%3A+%27connect%27%2C%0A++engine+++++address%3A+%27%2Ftmp%2Fprisma-309d86ec139dc8c202ae8134.sock%27%0A++engine+++%7D%0A++engine+%7D++%2B1s%0A%60%60%60%0A%0A%23%23+Client+Snippet%0A%60%60%60ts%0A%2F%2F+PLEASE+FILL+YOUR+CODE+SNIPPET+HERE%0A%60%60%60%0A%0A%23%23+Schema%0A%60%60%60prisma%0A%2F%2F+PLEASE+ADD+YOUR+SCHEMA+HERE+IF+POSSIBLE%0A%60%60%60%0A&title=PANIC+in+%2Froot%2F.cargo%2Fgit%2Fcheckouts%2Fquaint-9f01e008b9a89c14%2Fbc3aa8a%2Fsrc%2Fconnector%2Fpostgres%2Ferror.rs%3A138%3A29%0Aindex+out+of+bounds%3A+the+len+is+1+but+the+index+is+1&template=bug_report.md
If you want the Prisma team to look into it, please open the link above 🙏
To increase the chance of success, please post your schema and a snippet of
how you used Prisma Client in the issue.
at PrismaClientFetcher.request (/Users/m/Go/src/github.com/prisma/issue-4303/node_modules/@prisma/client/runtime/index.js:78134:15)
at processTicksAndRejections (internal/process/task_queues.js:97:5) {
clientVersion: '2.14.0'
}
```
## How to reproduce
<!--
Steps to reproduce the behavior:
1. Go to '...'
2. Change '....'
3. Run '....'
4. See error
-->
```js
const { PrismaClient} = require('@prisma/client')
const client= new PrismaClient()
async function main() {
await client.$connect()
const id = 10
await client.$executeRaw(`SET jwt.claims.user_id = ${id}`)
const result = await client.$executeRaw(`select jwt.claims.user_id`)
console.log(result)
}
main().catch(console.error).finally(() => client.$disconnect())
```
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
Should be able to click on the issue to create an issue
## Prisma information
<!-- Your Prisma schema, Prisma Client queries, ...
Do not include your database credentials when sharing your Prisma schema! -->
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]-->
- Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]-->
- Node.js version: <!--[Run `node -v` to see your Node.js version]-->
- Prisma version:
<!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]-->
```
Environment variables loaded from .env
@prisma/cli : 2.14.0
@prisma/client : 2.14.0
Current platform : darwin
Query Engine : query-engine 5d491261d382a2a5ffdc71de17072b0e409f1cc1 (at node_modules/@prisma/engines/query-engine-darwin)
Migration Engine : migration-engine-cli 5d491261d382a2a5ffdc71de17072b0e409f1cc1 (at node_modules/@prisma/engines/migration-engine-darwin)
Introspection Engine : introspection-core 5d491261d382a2a5ffdc71de17072b0e409f1cc1 (at node_modules/@prisma/engines/introspection-engine-darwin)
Format Binary : prisma-fmt 5d491261d382a2a5ffdc71de17072b0e409f1cc1 (at node_modules/@prisma/engines/prisma-fmt-darwin)
Studio : 0.332.0
```
|
process
|
missing h in http panic error thanks for helping us improve prisma 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by setting the debug environment variable and enabling additional logging output in prisma client learn more about writing proper bug reports here bug description link is missing an h invalid prisma executeraw invocation panic in root cargo git checkouts quaint src connector postgres error rs index out of bounds the len is but the index is this is a non recoverable error which probably happens when the prisma query engine has a panic ttps github com prisma prisma issues new body hi prisma team my prisma client just crashed this is the report versions name version node os darwin prisma client query engine query engine database undefined query executeraw query parameters logs tryloadenv environment variables loaded from com env tryloadenv environment variables loaded from com env prisma client clientversion engine cwd com engine search for query engine in com modules prisma engine search for query engine in com modules prisma plusx execution permissions of com modules prisma engine darwin are fine plusx execution permissions of com modules prisma engine darwin are fine engine engine flags engine enable raw queries engine unix path engine sock engine engine engine stdout starting a postgresql pool with connections engine stdout started http server on http sock engine search for query engine in com modules prisma prisma client prisma jwt claims user id prisma client prisma client call plusx execution permissions of com modules prisma engine darwin are fine prisma client prisma jwt claims user id prisma client prisma client call engine stdout panic in cargo rs out of bounds the len is but the index is engine typeerror this currentrequestpromise cancel is not a function engine at nodeengine handlepanic com modules js engine at nodeengine seterror com modules js engine at linestream com modules js engine at linestream emit js engine at addchunk stream readable js engine at readableaddchunk stream readable js engine at linestream readable push stream readable js engine at linestream transform push stream transform js engine at linestream pushbuffer com modules js engine at linestream transform com modules js out of bounds the len is but the index is cargo rs engine engine engine error socketerror closed engine at socket onsocketclose com modules js engine at socket emit js engine at pipe js engine name engine code err socket engine engine engine engine error error connect econnrefused sock engine at pipeconnectwrap afterconnect oncomplete js engine errno engine code engine syscall engine address sock engine engine engine client version engine engine version query engine engine active provider postgresql engine engine error error connect econnrefused sock engine at pipeconnectwrap afterconnect oncomplete js engine errno engine code engine syscall engine address sock engine engine client snippet please fill your code snippet here schema please add your schema here if possible title panic in cargo rs out of bounds the len is but the index is template bug report md if you want the prisma team to look into it please open the link above 🙏 to increase the chance of success please post your schema and a snippet of how you used prisma client in the issue at prismaclientfetcher request users m go src github com prisma issue node modules prisma client runtime index js at processticksandrejections internal process task queues js clientversion how to reproduce steps to reproduce the behavior go to change run see error js const prismaclient require prisma client const client new prismaclient async function main await client connect const id await client executeraw set jwt claims user id id const result await client executeraw select jwt claims user id console log result main catch console error finally client disconnect expected behavior should be able to click on the issue to create an issue prisma information your prisma schema prisma client queries do not include your database credentials when sharing your prisma schema environment setup os database node js version prisma version environment variables loaded from env prisma cli prisma client current platform darwin query engine query engine at node modules prisma engines query engine darwin migration engine migration engine cli at node modules prisma engines migration engine darwin introspection engine introspection core at node modules prisma engines introspection engine darwin format binary prisma fmt at node modules prisma engines prisma fmt darwin studio
| 1
|
4,013
| 6,949,379,920
|
IssuesEvent
|
2017-12-06 06:07:19
|
log2timeline/plaso
|
https://api.github.com/repos/log2timeline/plaso
|
closed
|
"preprocess" on a directory should not stop processing
|
bug preprocessing UX
|
**Plaso version:**
$ python tools/log2timeline.py --version
plaso - log2timeline version 20171120
**Operating system Plaso is running on:**
$ uname -a
Linux grimbergen 4.4.0-97-generic #120~14.04.1-Ubuntu SMP Wed Sep 20 15:53:13 UTC 2017 x86_64 x86_64 x86_64 GNU/Linu
**Installation method:**
Git clone sources
**Description of problem:**
When running log2timeline.py on a directory, preprocess is turned on, when the documentation says otherwise:
```
-p, --preprocess Turn on preprocessing. Preprocessing is turned on by
default when parsing image files, however if a mount
point is being parsed then this parameter needs to be
set manually.
```
This was detected when debugging issue #1509
**Debug output/tracebacks:**
**Source data:**
**To do:**
* [x] Update `--help` information
* removed the option
* https://codereview.appspot.com/337130043/
* [x] "preprocess" on a directory should not stop/block processing
* https://github.com/log2timeline/dfvfs/issues/247
|
1.0
|
"preprocess" on a directory should not stop processing - **Plaso version:**
$ python tools/log2timeline.py --version
plaso - log2timeline version 20171120
**Operating system Plaso is running on:**
$ uname -a
Linux grimbergen 4.4.0-97-generic #120~14.04.1-Ubuntu SMP Wed Sep 20 15:53:13 UTC 2017 x86_64 x86_64 x86_64 GNU/Linu
**Installation method:**
Git clone sources
**Description of problem:**
When running log2timeline.py on a directory, preprocess is turned on, when the documentation says otherwise:
```
-p, --preprocess Turn on preprocessing. Preprocessing is turned on by
default when parsing image files, however if a mount
point is being parsed then this parameter needs to be
set manually.
```
This was detected when debugging issue #1509
**Debug output/tracebacks:**
**Source data:**
**To do:**
* [x] Update `--help` information
* removed the option
* https://codereview.appspot.com/337130043/
* [x] "preprocess" on a directory should not stop/block processing
* https://github.com/log2timeline/dfvfs/issues/247
|
process
|
preprocess on a directory should not stop processing plaso version python tools py version plaso version operating system plaso is running on uname a linux grimbergen generic ubuntu smp wed sep utc gnu linu installation method git clone sources description of problem when running py on a directory preprocess is turned on when the documentation says otherwise p preprocess turn on preprocessing preprocessing is turned on by default when parsing image files however if a mount point is being parsed then this parameter needs to be set manually this was detected when debugging issue debug output tracebacks source data to do update help information removed the option preprocess on a directory should not stop block processing
| 1
|
18,357
| 24,487,523,492
|
IssuesEvent
|
2022-10-09 16:36:02
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
closed
|
Deprecate Process Child Watchers
|
expert-asyncio 3.12 expert-multiprocessing
|
BPO | [38591](https://bugs.python.org/issue38591)
--- | :---
Nosy | @vstinner, @benjaminp, @asvetlov, @1st1, @aeros
<sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup>
<details><summary>Show more details</summary><p>
GitHub fields:
```python
assignee = 'https://github.com/aeros'
closed_at = None
created_at = <Date 2019-10-25.19:23:29.645>
labels = ['3.9', 'expert-asyncio']
title = 'Deprecate Process Child Watchers'
updated_at = <Date 2019-11-15.13:21:36.776>
user = 'https://github.com/aeros'
```
bugs.python.org fields:
```python
activity = <Date 2019-11-15.13:21:36.776>
actor = 'aeros'
assignee = 'aeros'
closed = False
closed_date = None
closer = None
components = ['asyncio']
creation = <Date 2019-10-25.19:23:29.645>
creator = 'aeros'
dependencies = []
files = []
hgrepos = []
issue_num = 38591
keywords = []
message_count = 22.0
messages = ['355372', '355373', '355375', '355376', '355379', '355380', '355381', '355390', '355394', '355395', '355396', '355409', '355419', '355421', '356002', '356586', '356594', '356596', '356599', '356639', '356669', '356672']
nosy_count = 5.0
nosy_names = ['vstinner', 'benjamin.peterson', 'asvetlov', 'yselivanov', 'aeros']
pr_nums = []
priority = 'normal'
resolution = None
stage = None
status = 'open'
superseder = None
type = None
url = 'https://bugs.python.org/issue38591'
versions = ['Python 3.9']
```
</p></details>
|
1.0
|
Deprecate Process Child Watchers - BPO | [38591](https://bugs.python.org/issue38591)
--- | :---
Nosy | @vstinner, @benjaminp, @asvetlov, @1st1, @aeros
<sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup>
<details><summary>Show more details</summary><p>
GitHub fields:
```python
assignee = 'https://github.com/aeros'
closed_at = None
created_at = <Date 2019-10-25.19:23:29.645>
labels = ['3.9', 'expert-asyncio']
title = 'Deprecate Process Child Watchers'
updated_at = <Date 2019-11-15.13:21:36.776>
user = 'https://github.com/aeros'
```
bugs.python.org fields:
```python
activity = <Date 2019-11-15.13:21:36.776>
actor = 'aeros'
assignee = 'aeros'
closed = False
closed_date = None
closer = None
components = ['asyncio']
creation = <Date 2019-10-25.19:23:29.645>
creator = 'aeros'
dependencies = []
files = []
hgrepos = []
issue_num = 38591
keywords = []
message_count = 22.0
messages = ['355372', '355373', '355375', '355376', '355379', '355380', '355381', '355390', '355394', '355395', '355396', '355409', '355419', '355421', '356002', '356586', '356594', '356596', '356599', '356639', '356669', '356672']
nosy_count = 5.0
nosy_names = ['vstinner', 'benjamin.peterson', 'asvetlov', 'yselivanov', 'aeros']
pr_nums = []
priority = 'normal'
resolution = None
stage = None
status = 'open'
superseder = None
type = None
url = 'https://bugs.python.org/issue38591'
versions = ['Python 3.9']
```
</p></details>
|
process
|
deprecate process child watchers bpo nosy vstinner benjaminp asvetlov aeros note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee closed at none created at labels title deprecate process child watchers updated at user bugs python org fields python activity actor aeros assignee aeros closed false closed date none closer none components creation creator aeros dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type none url versions
| 1
|
15,723
| 19,897,786,319
|
IssuesEvent
|
2022-01-25 02:24:54
|
quark-engine/quark-engine
|
https://api.github.com/repos/quark-engine/quark-engine
|
closed
|
AttributeError occurs at the Rizin-based library.
|
work-in-progress issue-processing-state-04
|
**Describe the bug**
When running analysis on [this sample](https://apkpure.com/es/nodo-play/nodo.play) with the Rizin-based library, Quark crashes with an AttributeError. Here is the stack trace.
```
Traceback (most recent call last):
File "/usr/lib/python3.8/runpy.py", line 194, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/usr/lib/python3.8/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/home/haeter/.vscode/extensions/ms-python.python-2021.12.1559732655/pythonFiles/lib/python/debugpy/__main__.py", line 45, in <module>
cli.main()
File "/home/haeter/.vscode/extensions/ms-python.python-2021.12.1559732655/pythonFiles/lib/python/debugpy/../debugpy/server/cli.py", line 444, in main
run()
File "/home/haeter/.vscode/extensions/ms-python.python-2021.12.1559732655/pythonFiles/lib/python/debugpy/../debugpy/server/cli.py", line 285, in run_file
runpy.run_path(target_as_str, run_name=compat.force_str("__main__"))
File "/usr/lib/python3.8/runpy.py", line 265, in run_path
return _run_module_code(code, init_globals, run_name,
File "/usr/lib/python3.8/runpy.py", line 97, in _run_module_code
_run_code(code, mod_globals, init_globals,
File "/usr/lib/python3.8/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "quark/cli.py", line 410, in <module>
entry_point()
File "/home/haeter/.local/share/virtualenvs/quark-engine-5hUC8hAC/lib/python3.8/site-packages/click/core.py", line 1128, in __call__
return self.main(*args, **kwargs)
File "/home/haeter/.local/share/virtualenvs/quark-engine-5hUC8hAC/lib/python3.8/site-packages/click/core.py", line 1053, in main
rv = self.invoke(ctx)
File "/home/haeter/.local/share/virtualenvs/quark-engine-5hUC8hAC/lib/python3.8/site-packages/click/core.py", line 1395, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/home/haeter/.local/share/virtualenvs/quark-engine-5hUC8hAC/lib/python3.8/site-packages/click/core.py", line 754, in invoke
return __callback(*args, **kwargs)
File "quark/cli.py", line 315, in entry_point
data.run(rule_checker)
File "/home/haeter/Documents/Quark/quark-engine/quark/core/quark.py", line 426, in run
mutual_parent_function_list = self.find_intersection(
File "/home/haeter/Documents/Quark/quark-engine/quark/core/quark.py", line 109, in find_intersection
return self.method_recursive_search(
File "/home/haeter/Documents/Quark/quark-engine/quark/core/quark.py", line 132, in method_recursive_search
if self.apkinfo.upperfunc(method):
File "/home/haeter/Documents/Quark/quark-engine/quark/core/rzapkinfo.py", line 159, in upperfunc
cache = method_object.cache
AttributeError: 'NoneType' object has no attribute 'cache'
```
**To Reproduce**
1. Download [this sample](https://apkpure.com/es/nodo-play/nodo.play)
2. Run this command `quark --core-library rizin -a nodoPlay.apk`
3. See the error
**Environment**
+ Ubuntu 20.04
+ Quark-Engine v21.11.2
+ Rizin 0.2.1
|
1.0
|
AttributeError occurs at the Rizin-based library. - **Describe the bug**
When running analysis on [this sample](https://apkpure.com/es/nodo-play/nodo.play) with the Rizin-based library, Quark crashes with an AttributeError. Here is the stack trace.
```
Traceback (most recent call last):
File "/usr/lib/python3.8/runpy.py", line 194, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/usr/lib/python3.8/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "/home/haeter/.vscode/extensions/ms-python.python-2021.12.1559732655/pythonFiles/lib/python/debugpy/__main__.py", line 45, in <module>
cli.main()
File "/home/haeter/.vscode/extensions/ms-python.python-2021.12.1559732655/pythonFiles/lib/python/debugpy/../debugpy/server/cli.py", line 444, in main
run()
File "/home/haeter/.vscode/extensions/ms-python.python-2021.12.1559732655/pythonFiles/lib/python/debugpy/../debugpy/server/cli.py", line 285, in run_file
runpy.run_path(target_as_str, run_name=compat.force_str("__main__"))
File "/usr/lib/python3.8/runpy.py", line 265, in run_path
return _run_module_code(code, init_globals, run_name,
File "/usr/lib/python3.8/runpy.py", line 97, in _run_module_code
_run_code(code, mod_globals, init_globals,
File "/usr/lib/python3.8/runpy.py", line 87, in _run_code
exec(code, run_globals)
File "quark/cli.py", line 410, in <module>
entry_point()
File "/home/haeter/.local/share/virtualenvs/quark-engine-5hUC8hAC/lib/python3.8/site-packages/click/core.py", line 1128, in __call__
return self.main(*args, **kwargs)
File "/home/haeter/.local/share/virtualenvs/quark-engine-5hUC8hAC/lib/python3.8/site-packages/click/core.py", line 1053, in main
rv = self.invoke(ctx)
File "/home/haeter/.local/share/virtualenvs/quark-engine-5hUC8hAC/lib/python3.8/site-packages/click/core.py", line 1395, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/home/haeter/.local/share/virtualenvs/quark-engine-5hUC8hAC/lib/python3.8/site-packages/click/core.py", line 754, in invoke
return __callback(*args, **kwargs)
File "quark/cli.py", line 315, in entry_point
data.run(rule_checker)
File "/home/haeter/Documents/Quark/quark-engine/quark/core/quark.py", line 426, in run
mutual_parent_function_list = self.find_intersection(
File "/home/haeter/Documents/Quark/quark-engine/quark/core/quark.py", line 109, in find_intersection
return self.method_recursive_search(
File "/home/haeter/Documents/Quark/quark-engine/quark/core/quark.py", line 132, in method_recursive_search
if self.apkinfo.upperfunc(method):
File "/home/haeter/Documents/Quark/quark-engine/quark/core/rzapkinfo.py", line 159, in upperfunc
cache = method_object.cache
AttributeError: 'NoneType' object has no attribute 'cache'
```
**To Reproduce**
1. Download [this sample](https://apkpure.com/es/nodo-play/nodo.play)
2. Run this command `quark --core-library rizin -a nodoPlay.apk`
3. See the error
**Environment**
+ Ubuntu 20.04
+ Quark-Engine v21.11.2
+ Rizin 0.2.1
|
process
|
attributeerror occurs at the rizin based library describe the bug when running analysis on with the rizin based library quark crashes with an attributeerror here is the stack trace traceback most recent call last file usr lib runpy py line in run module as main return run code code main globals none file usr lib runpy py line in run code exec code run globals file home haeter vscode extensions ms python python pythonfiles lib python debugpy main py line in cli main file home haeter vscode extensions ms python python pythonfiles lib python debugpy debugpy server cli py line in main run file home haeter vscode extensions ms python python pythonfiles lib python debugpy debugpy server cli py line in run file runpy run path target as str run name compat force str main file usr lib runpy py line in run path return run module code code init globals run name file usr lib runpy py line in run module code run code code mod globals init globals file usr lib runpy py line in run code exec code run globals file quark cli py line in entry point file home haeter local share virtualenvs quark engine lib site packages click core py line in call return self main args kwargs file home haeter local share virtualenvs quark engine lib site packages click core py line in main rv self invoke ctx file home haeter local share virtualenvs quark engine lib site packages click core py line in invoke return ctx invoke self callback ctx params file home haeter local share virtualenvs quark engine lib site packages click core py line in invoke return callback args kwargs file quark cli py line in entry point data run rule checker file home haeter documents quark quark engine quark core quark py line in run mutual parent function list self find intersection file home haeter documents quark quark engine quark core quark py line in find intersection return self method recursive search file home haeter documents quark quark engine quark core quark py line in method recursive search if self apkinfo upperfunc method file home haeter documents quark quark engine quark core rzapkinfo py line in upperfunc cache method object cache attributeerror nonetype object has no attribute cache to reproduce download run this command quark core library rizin a nodoplay apk see the error environment ubuntu quark engine rizin
| 1
|
107,084
| 11,517,062,673
|
IssuesEvent
|
2020-02-14 07:19:05
|
argoproj-labs/argocd-notifications
|
https://api.github.com/repos/argoproj-labs/argocd-notifications
|
opened
|
Document slack bot usage and configuration
|
documentation
|
Document slack bot configuration steps and available commands
|
1.0
|
Document slack bot usage and configuration - Document slack bot configuration steps and available commands
|
non_process
|
document slack bot usage and configuration document slack bot configuration steps and available commands
| 0
|
60,894
| 12,134,768,002
|
IssuesEvent
|
2020-04-23 11:20:12
|
weso/shex-lite
|
https://api.github.com/repos/weso/shex-lite
|
opened
|
[SLI-00XX] Automatic generation of imports for the standard java library
|
affects/code-generation difficulty/medium priority/low status/awaiting-triage type/feature-request
|
@DaniFdezAlvarez proposes that the java code generator include a flag that allows to automatically generate the imports of the classes corresponding to the standard java library.
|
1.0
|
[SLI-00XX] Automatic generation of imports for the standard java library - @DaniFdezAlvarez proposes that the java code generator include a flag that allows to automatically generate the imports of the classes corresponding to the standard java library.
|
non_process
|
automatic generation of imports for the standard java library danifdezalvarez proposes that the java code generator include a flag that allows to automatically generate the imports of the classes corresponding to the standard java library
| 0
|
3,401
| 13,674,786,479
|
IssuesEvent
|
2020-09-29 11:47:11
|
elastic/beats
|
https://api.github.com/repos/elastic/beats
|
closed
|
Beats package got an error when building for darwin/amd64
|
[zube]: Ready automation bug
|
https://github.com/elastic/beats/issues/15953 was closed as the CI workers are now back to business.
[build for macos](https://beats-ci.elastic.co/job/elastic+beats+master+multijob-package-darwin/348/) has failed with:
```
18:14:33 Status: Downloaded newer image for docker.elastic.co/beats-dev/golang-crossbuild:1.13.7-darwin
18:14:34 >> Building using: cmd='build/mage-linux-amd64 golangCrossBuild', env=[CC=o64-clang, CXX=o64-clang++, GOARCH=amd64, GOARM=, GOOS=darwin, PLATFORM_ID=darwin-amd64]
18:14:34 sh: 1: build/mage-linux-amd64: not found
18:14:34 Error: failed building for darwin/amd64: exit status 127
18:14:34 failed building for darwin/amd64: exit status 127
18:14:34 package ran for 2m26.996460457s
18:14:34 Error: running "docker run --env EXEC_UID=596 --env EXEC_GID=503 --rm --env MAGEFILE_VERBOSE= --env MAGEFILE_TIMEOUT= -v /private/var/lib/jenkins/workspace/elastic+beats+master+multijob-package-darwin/src/github.com/elastic/beats:/go/src/github.com/elastic/beats -w /go/src/github.com/elastic/beats/auditbeat docker.elastic.co/beats-dev/golang-crossbuild:1.13.7-darwin --build-cmd build/mage-linux-amd64 golangCrossBuild -p darwin/amd64" failed with exit code 1
18:14:34 make[1]: *** [release] Error 1
```
## Actions
- Debug what's going on
- Reproduce locally if possible and document the steps to help with similar issues in the future.
- Propose a fix if possible or contact/transfer this ticket
|
1.0
|
Beats package got an error when building for darwin/amd64 - https://github.com/elastic/beats/issues/15953 was closed as the CI workers are now back to business.
[build for macos](https://beats-ci.elastic.co/job/elastic+beats+master+multijob-package-darwin/348/) has failed with:
```
18:14:33 Status: Downloaded newer image for docker.elastic.co/beats-dev/golang-crossbuild:1.13.7-darwin
18:14:34 >> Building using: cmd='build/mage-linux-amd64 golangCrossBuild', env=[CC=o64-clang, CXX=o64-clang++, GOARCH=amd64, GOARM=, GOOS=darwin, PLATFORM_ID=darwin-amd64]
18:14:34 sh: 1: build/mage-linux-amd64: not found
18:14:34 Error: failed building for darwin/amd64: exit status 127
18:14:34 failed building for darwin/amd64: exit status 127
18:14:34 package ran for 2m26.996460457s
18:14:34 Error: running "docker run --env EXEC_UID=596 --env EXEC_GID=503 --rm --env MAGEFILE_VERBOSE= --env MAGEFILE_TIMEOUT= -v /private/var/lib/jenkins/workspace/elastic+beats+master+multijob-package-darwin/src/github.com/elastic/beats:/go/src/github.com/elastic/beats -w /go/src/github.com/elastic/beats/auditbeat docker.elastic.co/beats-dev/golang-crossbuild:1.13.7-darwin --build-cmd build/mage-linux-amd64 golangCrossBuild -p darwin/amd64" failed with exit code 1
18:14:34 make[1]: *** [release] Error 1
```
## Actions
- Debug what's going on
- Reproduce locally if possible and document the steps to help with similar issues in the future.
- Propose a fix if possible or contact/transfer this ticket
|
non_process
|
beats package got an error when building for darwin was closed as the ci workers are now back to business has failed with status downloaded newer image for docker elastic co beats dev golang crossbuild darwin building using cmd build mage linux golangcrossbuild env sh build mage linux not found error failed building for darwin exit status failed building for darwin exit status package ran for error running docker run env exec uid env exec gid rm env magefile verbose env magefile timeout v private var lib jenkins workspace elastic beats master multijob package darwin src github com elastic beats go src github com elastic beats w go src github com elastic beats auditbeat docker elastic co beats dev golang crossbuild darwin build cmd build mage linux golangcrossbuild p darwin failed with exit code make error actions debug what s going on reproduce locally if possible and document the steps to help with similar issues in the future propose a fix if possible or contact transfer this ticket
| 0
|
165,042
| 13,964,718,806
|
IssuesEvent
|
2020-10-25 19:18:10
|
flameshot-org/flameshot
|
https://api.github.com/repos/flameshot-org/flameshot
|
closed
|
flameshot does not work with Ubuntu20.10
|
Documentation Linux
|
<!--
If you don't know how to get some of the following information from your computer, visit:
https://flameshot.js.org/#/data/issue-reporting
-->
**Flameshot version**
Flameshot v0.8.5
Compiled with Qt 5.12.8
**Describe the bug**
just can not be started
**To Reproduce**
click the flameshot icon
or in terminal type flameshot and click enter
**Expected behavior**
work well with Ubuntu20.10
**System Information**
Distributor ID: Ubuntu
Description: Ubuntu 20.10
Release: 20.10
Codename: groovy
|
1.0
|
flameshot does not work with Ubuntu20.10 - <!--
If you don't know how to get some of the following information from your computer, visit:
https://flameshot.js.org/#/data/issue-reporting
-->
**Flameshot version**
Flameshot v0.8.5
Compiled with Qt 5.12.8
**Describe the bug**
just can not be started
**To Reproduce**
click the flameshot icon
or in terminal type flameshot and click enter
**Expected behavior**
work well with Ubuntu20.10
**System Information**
Distributor ID: Ubuntu
Description: Ubuntu 20.10
Release: 20.10
Codename: groovy
|
non_process
|
flameshot does not work with if you don t know how to get some of the following information from your computer visit flameshot version flameshot compiled with qt describe the bug just can not be started to reproduce click the flameshot icon or in terminal type flameshot and click enter expected behavior work well with system information distributor id ubuntu description ubuntu release codename groovy
| 0
|
75,825
| 9,887,785,839
|
IssuesEvent
|
2019-06-25 09:57:48
|
numpy/numpy
|
https://api.github.com/repos/numpy/numpy
|
closed
|
DOC: string operations section doesn't specify module
|
04 - Documentation component: numpy.doc
|
The section https://docs.scipy.org/doc/numpy/reference/routines.html has a link to "string operations". The header of the routines page states that you import with "import numpy as np". Subsections that require imports list the module in the name of the link. For example "Linear algebra (numpy.linalg)". "String operations" has no such specification.
When you go to the page link (https://docs.scipy.org/doc/numpy/reference/routines.char.html) it starts with "This module provides..." The module name is not specified; what module is "this"? It is not in the main numpy namespace; for example I cannot write np.lower('HI').
I assume these are the routines from numpy.core.chararray, since "numpy.core.chararray.lower("HI") works, but who can say without running grep over the library?
My suggestion is to add "(numpy.core.chararray)", or whatever the correct module is, to the title of this page.
|
1.0
|
DOC: string operations section doesn't specify module - The section https://docs.scipy.org/doc/numpy/reference/routines.html has a link to "string operations". The header of the routines page states that you import with "import numpy as np". Subsections that require imports list the module in the name of the link. For example "Linear algebra (numpy.linalg)". "String operations" has no such specification.
When you go to the page link (https://docs.scipy.org/doc/numpy/reference/routines.char.html) it starts with "This module provides..." The module name is not specified; what module is "this"? It is not in the main numpy namespace; for example I cannot write np.lower('HI').
I assume these are the routines from numpy.core.chararray, since "numpy.core.chararray.lower("HI") works, but who can say without running grep over the library?
My suggestion is to add "(numpy.core.chararray)", or whatever the correct module is, to the title of this page.
|
non_process
|
doc string operations section doesn t specify module the section has a link to string operations the header of the routines page states that you import with import numpy as np subsections that require imports list the module in the name of the link for example linear algebra numpy linalg string operations has no such specification when you go to the page link it starts with this module provides the module name is not specified what module is this it is not in the main numpy namespace for example i cannot write np lower hi i assume these are the routines from numpy core chararray since numpy core chararray lower hi works but who can say without running grep over the library my suggestion is to add numpy core chararray or whatever the correct module is to the title of this page
| 0
|
120,862
| 25,885,396,739
|
IssuesEvent
|
2022-12-14 14:13:56
|
Clueless-Community/seamless-ui
|
https://api.github.com/repos/Clueless-Community/seamless-ui
|
opened
|
Somewhat Duplicate Navbar 02 & 26
|
codepeak22
|
## **_Nav 02_** and _**Nav 26**_ are also somewhat similar & nav 26 does not follows the navbar design system in terms of padding

So, What should I do ? Just improve the padding of nav 26 or remove it ?
|
1.0
|
Somewhat Duplicate Navbar 02 & 26 - ## **_Nav 02_** and _**Nav 26**_ are also somewhat similar & nav 26 does not follows the navbar design system in terms of padding

So, What should I do ? Just improve the padding of nav 26 or remove it ?
|
non_process
|
somewhat duplicate navbar nav and nav are also somewhat similar nav does not follows the navbar design system in terms of padding so what should i do just improve the padding of nav or remove it
| 0
|
182,110
| 6,667,217,326
|
IssuesEvent
|
2017-10-03 11:39:08
|
HelpyTeam/HelpyWeb
|
https://api.github.com/repos/HelpyTeam/HelpyWeb
|
opened
|
Update front-end
|
Front-end priority/3 To-do
|
# Overview
Update the front-end
# Target
After update, please note the components that you had updated and send pull request.
|
1.0
|
Update front-end - # Overview
Update the front-end
# Target
After update, please note the components that you had updated and send pull request.
|
non_process
|
update front end overview update the front end target after update please note the components that you had updated and send pull request
| 0
|
236,943
| 18,144,676,397
|
IssuesEvent
|
2021-09-25 07:48:34
|
kubernetes/autoscaler
|
https://api.github.com/repos/kubernetes/autoscaler
|
closed
|
VPA - Document the current recommendation algorithm
|
vertical-pod-autoscaler kind/documentation lifecycle/rotten
|
Document
* how recommendations are calculated out of raw samples for CPU and Memory.
* when it is reasonable to expect a stable recommendation for a new workload
Please note that VPA recommendation algorithm is not part of the API and is subject to change without notice
|
1.0
|
VPA - Document the current recommendation algorithm - Document
* how recommendations are calculated out of raw samples for CPU and Memory.
* when it is reasonable to expect a stable recommendation for a new workload
Please note that VPA recommendation algorithm is not part of the API and is subject to change without notice
|
non_process
|
vpa document the current recommendation algorithm document how recommendations are calculated out of raw samples for cpu and memory when it is reasonable to expect a stable recommendation for a new workload please note that vpa recommendation algorithm is not part of the api and is subject to change without notice
| 0
|
14,164
| 2,791,389,963
|
IssuesEvent
|
2015-05-10 03:12:31
|
jimradford/superputty
|
https://api.github.com/repos/jimradford/superputty
|
closed
|
home directory assumed rather than read
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. open scp session with non standard home (ex root -- home is /root not
/home/root)
2. connection opens but unable to navigate because /home/username does not exist
3.
What is the expected output? What do you see instead?
home is read from remote server and used for scp
What version of the product are you using? On what operating system?
1.4.0.4 windows server 2008 r2
Please provide any additional information below.
```
Original issue reported on code.google.com by `ada...@scheblein.net` on 28 May 2013 at 3:47
|
1.0
|
home directory assumed rather than read - ```
What steps will reproduce the problem?
1. open scp session with non standard home (ex root -- home is /root not
/home/root)
2. connection opens but unable to navigate because /home/username does not exist
3.
What is the expected output? What do you see instead?
home is read from remote server and used for scp
What version of the product are you using? On what operating system?
1.4.0.4 windows server 2008 r2
Please provide any additional information below.
```
Original issue reported on code.google.com by `ada...@scheblein.net` on 28 May 2013 at 3:47
|
non_process
|
home directory assumed rather than read what steps will reproduce the problem open scp session with non standard home ex root home is root not home root connection opens but unable to navigate because home username does not exist what is the expected output what do you see instead home is read from remote server and used for scp what version of the product are you using on what operating system windows server please provide any additional information below original issue reported on code google com by ada scheblein net on may at
| 0
|
206,777
| 15,774,176,404
|
IssuesEvent
|
2021-04-01 00:34:57
|
pathfinder-for-autonomous-navigation/FlightSoftware
|
https://api.github.com/repos/pathfinder-for-autonomous-navigation/FlightSoftware
|
closed
|
[PTest] Test RadioSession with AMC
|
functional testing gsw telemetry
|
Currently, the way RadioSession makes and stores statefields in a json for uplinks might not allow changes to states already written to or the addition of states to the packet after writing multiple states once. We need to adjust RadioSession as necessary so that the states written in each iteration of the AMC get written to an uplink, and test that this behavior works successfully.
|
1.0
|
[PTest] Test RadioSession with AMC - Currently, the way RadioSession makes and stores statefields in a json for uplinks might not allow changes to states already written to or the addition of states to the packet after writing multiple states once. We need to adjust RadioSession as necessary so that the states written in each iteration of the AMC get written to an uplink, and test that this behavior works successfully.
|
non_process
|
test radiosession with amc currently the way radiosession makes and stores statefields in a json for uplinks might not allow changes to states already written to or the addition of states to the packet after writing multiple states once we need to adjust radiosession as necessary so that the states written in each iteration of the amc get written to an uplink and test that this behavior works successfully
| 0
|
308,038
| 23,228,255,707
|
IssuesEvent
|
2022-08-03 04:07:17
|
cainmagi/FFmpeg-Encoder-Decoder-for-Python
|
https://api.github.com/repos/cainmagi/FFmpeg-Encoder-Decoder-for-Python
|
closed
|
Frame count and metadata access
|
to be solved documentation
|
### Get started
- [X] I have read [Contributing guidelines](https://github.com/cainmagi/FFmpeg-Encoder-Decoder-for-Python/blob/master/CONTRIBUTING.md).
- [X] I agree to follow the [Code of Conduct](https://github.com/cainmagi/FFmpeg-Encoder-Decoder-for-Python/blob/master/CODE_OF_CONDUCT.md).
- [X] I have confirmed that my issue is not duplicated with an existing issue.
### Problem
Hi, I'm currently testing this library and this looks promising, but I didn't found any way to access the frame count from python (although it seems it is available in the c++ private variable?)
I think it will be good if we can access the metadata of the video.
here is my sample script:
```
import mpegCoder
from PySide2.QtWidgets import QWidget, QLabel, QVBoxLayout, QApplication
from PySide2.QtGui import QImage, QPixmap
class Player(QWidget):
def __init__(self, parent=None):
super(Player, self).__init__(parent)
self._layout = QVBoxLayout(self)
self.image = QLabel()
self._layout.addWidget(self.image)
decoder = mpegCoder.MpegDecoder()
decoder.setParameter(nthread=4)
opened = decoder.FFmpegSetup('30.mp4')
if opened:
param = decoder.getParameter()
fps = float(param['frameRate'][0]) / param['frameRate'][1] # is this correct way to get fps?
framecount = 0 # How to get framecount of the video?
frame = 0
p = decoder.ExtractFrame(frame, 1)
decoder.clear() #do we need to clear the decoder?
data = p[0].data
height = p[0].shape[0]
width = p[0].shape[1]
channel = p[0].shape[2]
self.image.setPixmap(QPixmap.fromImage(QImage(data, width, height, channel*width, QImage.Format_RGB888)))
if __name__ == "__main__":
import sys
app = QApplication(sys.argv)
p = Player()
p.show()
app.exec_()
```
### Required feature
Frame count or metadata reading
### mpegCoder version
3.2.4
### Additional context
_No response_
Also will be waiting for audio support ;)
|
1.0
|
Frame count and metadata access - ### Get started
- [X] I have read [Contributing guidelines](https://github.com/cainmagi/FFmpeg-Encoder-Decoder-for-Python/blob/master/CONTRIBUTING.md).
- [X] I agree to follow the [Code of Conduct](https://github.com/cainmagi/FFmpeg-Encoder-Decoder-for-Python/blob/master/CODE_OF_CONDUCT.md).
- [X] I have confirmed that my issue is not duplicated with an existing issue.
### Problem
Hi, I'm currently testing this library and this looks promising, but I didn't found any way to access the frame count from python (although it seems it is available in the c++ private variable?)
I think it will be good if we can access the metadata of the video.
here is my sample script:
```
import mpegCoder
from PySide2.QtWidgets import QWidget, QLabel, QVBoxLayout, QApplication
from PySide2.QtGui import QImage, QPixmap
class Player(QWidget):
def __init__(self, parent=None):
super(Player, self).__init__(parent)
self._layout = QVBoxLayout(self)
self.image = QLabel()
self._layout.addWidget(self.image)
decoder = mpegCoder.MpegDecoder()
decoder.setParameter(nthread=4)
opened = decoder.FFmpegSetup('30.mp4')
if opened:
param = decoder.getParameter()
fps = float(param['frameRate'][0]) / param['frameRate'][1] # is this correct way to get fps?
framecount = 0 # How to get framecount of the video?
frame = 0
p = decoder.ExtractFrame(frame, 1)
decoder.clear() #do we need to clear the decoder?
data = p[0].data
height = p[0].shape[0]
width = p[0].shape[1]
channel = p[0].shape[2]
self.image.setPixmap(QPixmap.fromImage(QImage(data, width, height, channel*width, QImage.Format_RGB888)))
if __name__ == "__main__":
import sys
app = QApplication(sys.argv)
p = Player()
p.show()
app.exec_()
```
### Required feature
Frame count or metadata reading
### mpegCoder version
3.2.4
### Additional context
_No response_
Also will be waiting for audio support ;)
|
non_process
|
frame count and metadata access get started i have read i agree to follow the i have confirmed that my issue is not duplicated with an existing issue problem hi i m currently testing this library and this looks promising but i didn t found any way to access the frame count from python although it seems it is available in the c private variable i think it will be good if we can access the metadata of the video here is my sample script import mpegcoder from qtwidgets import qwidget qlabel qvboxlayout qapplication from qtgui import qimage qpixmap class player qwidget def init self parent none super player self init parent self layout qvboxlayout self self image qlabel self layout addwidget self image decoder mpegcoder mpegdecoder decoder setparameter nthread opened decoder ffmpegsetup if opened param decoder getparameter fps float param param is this correct way to get fps framecount how to get framecount of the video frame p decoder extractframe frame decoder clear do we need to clear the decoder data p data height p shape width p shape channel p shape self image setpixmap qpixmap fromimage qimage data width height channel width qimage format if name main import sys app qapplication sys argv p player p show app exec required feature frame count or metadata reading mpegcoder version additional context no response also will be waiting for audio support
| 0
|
12,057
| 14,739,483,748
|
IssuesEvent
|
2021-01-07 07:18:13
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
Mobile app > Password creation > Check password criteria after password is first entered
|
Android Bug P2 Process: Reopened Process: Track 3 iOS
|
Mobile app > Password creation > Sign up, reset password, change password flows
Check for the user-entered password meeting the criteria at the point the password is first entered rather than after it is confirmed in the 'Confirm password' field.
|
2.0
|
Mobile app > Password creation > Check password criteria after password is first entered - Mobile app > Password creation > Sign up, reset password, change password flows
Check for the user-entered password meeting the criteria at the point the password is first entered rather than after it is confirmed in the 'Confirm password' field.
|
process
|
mobile app password creation check password criteria after password is first entered mobile app password creation sign up reset password change password flows check for the user entered password meeting the criteria at the point the password is first entered rather than after it is confirmed in the confirm password field
| 1
|
15,352
| 19,522,523,109
|
IssuesEvent
|
2021-12-29 21:41:53
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
opened
|
Error shown for certain column types when getting results from cache
|
Type:Bug Priority:P2 Database/Redshift Querying/Processor .Backend
|
**Describe the bug**
When returning an `interval` column type via cached results, then an error is shown in the frontend instead of the results.
I'm unsure if this is specific to Redshift or generally for unknown column types. I've not been able to reproduce on Postgres.
**To Reproduce**
1. Admin > Settings > Caching: Duration=`0.0001` and TTL=`10000`
2. Native query > **Redshift** > `select interval '1 day'` - run query - see results - save question
3. Do browser refresh a couple of time, so it starts returning cached results (or is supposed to)
The returned "result" is:
`{"nippy/unthawable":{"type":"serializable","cause":"quarantined","class-name":"com.amazon.redshift.util.RedshiftInterval","content":"0xACED0005"}}`
And the log only contains the following related to this, no stacktraces or anything real errors:
`WARN metabase.driver.common Don't know how to map class 'class com.amazon.redshift.util.RedshiftInterval' to a Field base_type, falling back to :type/*.`

**Expected behavior**
1. Correct results from cache, of course
2. An error or stracktrace in the log.
**Information about your Metabase Installation:**
Tested 1.41.5 and master `b9bee5d`
|
1.0
|
Error shown for certain column types when getting results from cache - **Describe the bug**
When returning an `interval` column type via cached results, then an error is shown in the frontend instead of the results.
I'm unsure if this is specific to Redshift or generally for unknown column types. I've not been able to reproduce on Postgres.
**To Reproduce**
1. Admin > Settings > Caching: Duration=`0.0001` and TTL=`10000`
2. Native query > **Redshift** > `select interval '1 day'` - run query - see results - save question
3. Do browser refresh a couple of time, so it starts returning cached results (or is supposed to)
The returned "result" is:
`{"nippy/unthawable":{"type":"serializable","cause":"quarantined","class-name":"com.amazon.redshift.util.RedshiftInterval","content":"0xACED0005"}}`
And the log only contains the following related to this, no stacktraces or anything real errors:
`WARN metabase.driver.common Don't know how to map class 'class com.amazon.redshift.util.RedshiftInterval' to a Field base_type, falling back to :type/*.`

**Expected behavior**
1. Correct results from cache, of course
2. An error or stracktrace in the log.
**Information about your Metabase Installation:**
Tested 1.41.5 and master `b9bee5d`
|
process
|
error shown for certain column types when getting results from cache describe the bug when returning an interval column type via cached results then an error is shown in the frontend instead of the results i m unsure if this is specific to redshift or generally for unknown column types i ve not been able to reproduce on postgres to reproduce admin settings caching duration and ttl native query redshift select interval day run query see results save question do browser refresh a couple of time so it starts returning cached results or is supposed to the returned result is nippy unthawable type serializable cause quarantined class name com amazon redshift util redshiftinterval content and the log only contains the following related to this no stacktraces or anything real errors warn metabase driver common don t know how to map class class com amazon redshift util redshiftinterval to a field base type falling back to type expected behavior correct results from cache of course an error or stracktrace in the log information about your metabase installation tested and master
| 1
|
212,025
| 7,226,164,235
|
IssuesEvent
|
2018-02-10 05:33:14
|
brijeshshah13/OpenCI
|
https://api.github.com/repos/brijeshshah13/OpenCI
|
opened
|
Implement Search Button in Main Layout
|
Priority: High Type: Good first issue
|
**Actual Behavior**
Not yet implemented.
**Expected Behavior**
Clicking on the Search button should allow the user to search from the repositories in the current tab, i.e., either `Public` or `Private`.
**Steps to reproduce it**
Click on the search icon in the top right corner of the main layout.
|
1.0
|
Implement Search Button in Main Layout - **Actual Behavior**
Not yet implemented.
**Expected Behavior**
Clicking on the Search button should allow the user to search from the repositories in the current tab, i.e., either `Public` or `Private`.
**Steps to reproduce it**
Click on the search icon in the top right corner of the main layout.
|
non_process
|
implement search button in main layout actual behavior not yet implemented expected behavior clicking on the search button should allow the user to search from the repositories in the current tab i e either public or private steps to reproduce it click on the search icon in the top right corner of the main layout
| 0
|
2,017
| 4,837,505,477
|
IssuesEvent
|
2016-11-08 22:49:17
|
P0cL4bs/WiFi-Pumpkin
|
https://api.github.com/repos/P0cL4bs/WiFi-Pumpkin
|
closed
|
Internet interface set up bridge ppp0
|
enhancement in process solved
|
## What's the problem (or question)?
You plan to do support other interfaces, except eth0? Or, please tell me how to set up bridge ppp0 to eth0.
p.s.: sorry for my English
#### Please tell us details about your environment.
* Card wireless adapters name (please check if support AP/mode): wn722n
* Version used tool: 0.8.3
* Virtual Machine (yes or no and which): no
* Operating System and version: kali linux 2016.2 w43
|
1.0
|
Internet interface set up bridge ppp0 - ## What's the problem (or question)?
You plan to do support other interfaces, except eth0? Or, please tell me how to set up bridge ppp0 to eth0.
p.s.: sorry for my English
#### Please tell us details about your environment.
* Card wireless adapters name (please check if support AP/mode): wn722n
* Version used tool: 0.8.3
* Virtual Machine (yes or no and which): no
* Operating System and version: kali linux 2016.2 w43
|
process
|
internet interface set up bridge what s the problem or question you plan to do support other interfaces except or please tell me how to set up bridge to p s sorry for my english please tell us details about your environment card wireless adapters name please check if support ap mode version used tool virtual machine yes or no and which no operating system and version kali linux
| 1
|
9,650
| 12,620,814,570
|
IssuesEvent
|
2020-06-13 08:57:51
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
`source("....", echo=TRUE)` produces prompts wrongly in iESS `*R*`
|
bug process:eval
|
As the subject says: The current "prompt searching" and similar hoop jumping really kill the nice `echo = TRUE` behavior of `source()` by adding extra newlines the result of which is very ugly.
I have set `ess-eval-visibly` to (non default) `nil` .. but don't really expect it to make a difference
|
1.0
|
`source("....", echo=TRUE)` produces prompts wrongly in iESS `*R*` - As the subject says: The current "prompt searching" and similar hoop jumping really kill the nice `echo = TRUE` behavior of `source()` by adding extra newlines the result of which is very ugly.
I have set `ess-eval-visibly` to (non default) `nil` .. but don't really expect it to make a difference
|
process
|
source echo true produces prompts wrongly in iess r as the subject says the current prompt searching and similar hoop jumping really kill the nice echo true behavior of source by adding extra newlines the result of which is very ugly i have set ess eval visibly to non default nil but don t really expect it to make a difference
| 1
|
20,354
| 27,013,414,194
|
IssuesEvent
|
2023-02-10 17:09:35
|
mpi-forum/mpi-issues
|
https://api.github.com/repos/mpi-forum/mpi-issues
|
opened
|
'Pending operation' not defined, pending proper definition
|
chap-contexts chap-process chap-terms chap-collective
|
# Problem
The text uses the term "pending operation" in some places to refer to an "active operation" (as in: it's completion is pending). Nowhere is it defined in what state a pending operation is supposed to be so in some contexts this can be confusing. With semantic terms, the proper term to use is "active".
Examples:
§11.10.4
> `MPI_COMM_DISCONNECT` has the same action as `MPI_COMM_FREE`, except that it waits for *pending* communication to finish internally and enables the guarantee about the behavior of disconnected processes.
would be
> `MPI_COMM_DISCONNECT` has the same action as `MPI_COMM_FREE`, except that it waits for *active* communication to finish internally and enables the guarantee about the behavior of disconnected processes.
(resolved in https://github.com/mpi-forum/mpi-issues/issues/543, see also https://github.com/mpi-forum/mpi-issues/issues/536)
§6.12:
>That is, an MPI implementation should be able to support a large number of *pending* nonblocking operations.
would be
> That is, an MPI implementation should be able to support a large number of *active* nonblocking operations.
§7.1.2
> This allows the invocation of the library even if there are *pending* communications on “other” communicators, and avoids the need to synchronize entry or exit into library code. *Pending* point-to-point communications are also guaranteed not to interfere with collective communications within a single communicator.
would be
> This allows the invocation of the library even if there are *active* communications on “other” communicators, and avoids the need to synchronize entry or exit into library code. *Active* point-to-point communications are also guaranteed not to interfere with collective communications within a single communicator.
§7.4.3
> Any *pending* operations that use this communicator will complete normally
would be
> Any *active* operations that use this communicator will complete normally
§...
There are many more locations using 'pending'
# Proposal
See solution in issue https://github.com/mpi-forum/mpi-issues/issues/543
One must be careful, because *active* may be wrong (as in https://github.com/mpi-forum/mpi-issues/issues/543).
# Changes to the Text
See above. This should be part of the "cleanup" for MPI 4.1.
Alternatively, the term "pending" could be properly in the terms section.
(Risky, because its meaning may be different in different contexts)
# Impact on Implementations
None (unless there it renders some implementations non-compliant due to confusion stemming from the usage of the term "pending")
# Impact on Users
Clear use of terms helps users better understand the text without having to reason about what a "pending operation" is.
# References and Pull Requests
<!-- Insert any internal (other issues and pull requests) or external (websites, papers, etc.) references here. You may come back and edit the description later to add references. -->
|
1.0
|
'Pending operation' not defined, pending proper definition - # Problem
The text uses the term "pending operation" in some places to refer to an "active operation" (as in: it's completion is pending). Nowhere is it defined in what state a pending operation is supposed to be so in some contexts this can be confusing. With semantic terms, the proper term to use is "active".
Examples:
§11.10.4
> `MPI_COMM_DISCONNECT` has the same action as `MPI_COMM_FREE`, except that it waits for *pending* communication to finish internally and enables the guarantee about the behavior of disconnected processes.
would be
> `MPI_COMM_DISCONNECT` has the same action as `MPI_COMM_FREE`, except that it waits for *active* communication to finish internally and enables the guarantee about the behavior of disconnected processes.
(resolved in https://github.com/mpi-forum/mpi-issues/issues/543, see also https://github.com/mpi-forum/mpi-issues/issues/536)
§6.12:
>That is, an MPI implementation should be able to support a large number of *pending* nonblocking operations.
would be
> That is, an MPI implementation should be able to support a large number of *active* nonblocking operations.
§7.1.2
> This allows the invocation of the library even if there are *pending* communications on “other” communicators, and avoids the need to synchronize entry or exit into library code. *Pending* point-to-point communications are also guaranteed not to interfere with collective communications within a single communicator.
would be
> This allows the invocation of the library even if there are *active* communications on “other” communicators, and avoids the need to synchronize entry or exit into library code. *Active* point-to-point communications are also guaranteed not to interfere with collective communications within a single communicator.
§7.4.3
> Any *pending* operations that use this communicator will complete normally
would be
> Any *active* operations that use this communicator will complete normally
§...
There are many more locations using 'pending'
# Proposal
See solution in issue https://github.com/mpi-forum/mpi-issues/issues/543
One must be careful, because *active* may be wrong (as in https://github.com/mpi-forum/mpi-issues/issues/543).
# Changes to the Text
See above. This should be part of the "cleanup" for MPI 4.1.
Alternatively, the term "pending" could be properly in the terms section.
(Risky, because its meaning may be different in different contexts)
# Impact on Implementations
None (unless there it renders some implementations non-compliant due to confusion stemming from the usage of the term "pending")
# Impact on Users
Clear use of terms helps users better understand the text without having to reason about what a "pending operation" is.
# References and Pull Requests
<!-- Insert any internal (other issues and pull requests) or external (websites, papers, etc.) references here. You may come back and edit the description later to add references. -->
|
process
|
pending operation not defined pending proper definition problem the text uses the term pending operation in some places to refer to an active operation as in it s completion is pending nowhere is it defined in what state a pending operation is supposed to be so in some contexts this can be confusing with semantic terms the proper term to use is active examples § mpi comm disconnect has the same action as mpi comm free except that it waits for pending communication to finish internally and enables the guarantee about the behavior of disconnected processes would be mpi comm disconnect has the same action as mpi comm free except that it waits for active communication to finish internally and enables the guarantee about the behavior of disconnected processes resolved in see also § that is an mpi implementation should be able to support a large number of pending nonblocking operations would be that is an mpi implementation should be able to support a large number of active nonblocking operations § this allows the invocation of the library even if there are pending communications on “other” communicators and avoids the need to synchronize entry or exit into library code pending point to point communications are also guaranteed not to interfere with collective communications within a single communicator would be this allows the invocation of the library even if there are active communications on “other” communicators and avoids the need to synchronize entry or exit into library code active point to point communications are also guaranteed not to interfere with collective communications within a single communicator § any pending operations that use this communicator will complete normally would be any active operations that use this communicator will complete normally § there are many more locations using pending proposal see solution in issue one must be careful because active may be wrong as in changes to the text see above this should be part of the cleanup for mpi alternatively the term pending could be properly in the terms section risky because its meaning may be different in different contexts impact on implementations none unless there it renders some implementations non compliant due to confusion stemming from the usage of the term pending impact on users clear use of terms helps users better understand the text without having to reason about what a pending operation is references and pull requests
| 1
|
17,689
| 23,535,852,911
|
IssuesEvent
|
2022-08-19 20:39:07
|
nghi-huynh/HPA_HuBMAP_Kaggle_competition
|
https://api.github.com/repos/nghi-huynh/HPA_HuBMAP_Kaggle_competition
|
opened
|
Fine-tuning on augmented dataset
|
data preprocessing data preparation
|
- [ ] Prepare tiles (256x256) based on augmented dataset
- [ ] Update + organize pretrained models
|
1.0
|
Fine-tuning on augmented dataset - - [ ] Prepare tiles (256x256) based on augmented dataset
- [ ] Update + organize pretrained models
|
process
|
fine tuning on augmented dataset prepare tiles based on augmented dataset update organize pretrained models
| 1
|
608,029
| 18,796,433,624
|
IssuesEvent
|
2021-11-08 23:04:28
|
carbon-design-system/carbon-addons-iot-react
|
https://api.github.com/repos/carbon-design-system/carbon-addons-iot-react
|
opened
|
[PageTitleBar] Style audit
|
status: needs triage :mag: status: needs priority :inbox_tray: design reconciliation
|
<!--
Use this template if you want to request a new feature, or a change to an
existing feature.
If you'd like to request an entirely new component, please use the component request template instead.
If you are reporting a bug or problem, please use the bug template instead.
-->
### What package is this for?
- [x] React
- [ ] Angular
## 1. Padding adjustment
On the dynamic page header, there is an extra 2px in the collapsed header state, making it 42px high and not 40px.
### Current
Storybook: https://next.carbon-addons-iot-react.com/?path=/story/1-watson-iot-pagetitlebar--with-dynamic-scrolling
<img width="535" alt="Screen Shot 2021-11-08 at 4 10 47 PM" src="https://user-images.githubusercontent.com/43549567/140818822-38e90aaa-203b-4c24-91e2-ce3018521236.png">
### Proposed
<img width="778" alt="Screen Shot 2021-11-08 at 5 52 36 PM" src="https://user-images.githubusercontent.com/43549567/140830946-c35eb0a6-59c0-4c0d-93df-bb26a541151f.png">
## 2. Dividing line between buttons for accessibility
Similar to how a dividing line was adding between button groups in places like [modals](https://next.carbon-addons-iot-react.com/?path=/story/3-carbon-modal--default), we need a dividing line between any secondary/primary button pairings for accessibility
### Current
Storybook: https://next.carbon-addons-iot-react.com/?path=/story/1-watson-iot-pagetitlebar--with-dynamic-scrolling
<img width="440" alt="Screen Shot 2021-11-08 at 4 32 14 PM" src="https://user-images.githubusercontent.com/43549567/140821478-571ddfd6-742a-4926-a9dc-cc065b62c3b1.png">
Storybook: https://next.carbon-addons-iot-react.com/?path=/story/1-watson-iot-pagetitlebar--with-condensed-header
### Proposed
<img width="778" alt="Screen Shot 2021-11-08 at 5 52 36 PM" src="https://user-images.githubusercontent.com/43549567/140830946-c35eb0a6-59c0-4c0d-93df-bb26a541151f.png">
## 3. Line below tabs in collapsed state
Similar to the dynamic header with no tabs, there should be a dividing line to create separation from the tabs and content on scroll.
### Current
Storybook: https://next.carbon-addons-iot-react.com/?path=/story/1-watson-iot-pagetitlebar--with-dynamic-scrolling-and-tabs
<img width="1564" alt="Screen Shot 2021-11-08 at 4 34 19 PM" src="https://user-images.githubusercontent.com/43549567/140821898-bac5288e-e637-4fa5-8003-c209d2d26a85.png">
### Proposed
<img width="1070" alt="Screen Shot 2021-11-08 at 5 57 02 PM" src="https://user-images.githubusercontent.com/43549567/140831405-001cf275-a922-42bc-9be9-3e8090c8f324.png">
## 4. Dynamic header with tabs enhancement
Can we add a variation/prop to have both the collapsed header and tabs both sticky? This is an enhancement we're making on the design side as well based on feedback so you don't lose your page-level actions and breadcrumb/page title.
### Proposed
<img width="1627" alt="Screen Shot 2021-11-08 at 5 57 43 PM" src="https://user-images.githubusercontent.com/43549567/140831477-ccc7da92-6493-43df-affe-a1230dddba1c.png">
|
1.0
|
[PageTitleBar] Style audit - <!--
Use this template if you want to request a new feature, or a change to an
existing feature.
If you'd like to request an entirely new component, please use the component request template instead.
If you are reporting a bug or problem, please use the bug template instead.
-->
### What package is this for?
- [x] React
- [ ] Angular
## 1. Padding adjustment
On the dynamic page header, there is an extra 2px in the collapsed header state, making it 42px high and not 40px.
### Current
Storybook: https://next.carbon-addons-iot-react.com/?path=/story/1-watson-iot-pagetitlebar--with-dynamic-scrolling
<img width="535" alt="Screen Shot 2021-11-08 at 4 10 47 PM" src="https://user-images.githubusercontent.com/43549567/140818822-38e90aaa-203b-4c24-91e2-ce3018521236.png">
### Proposed
<img width="778" alt="Screen Shot 2021-11-08 at 5 52 36 PM" src="https://user-images.githubusercontent.com/43549567/140830946-c35eb0a6-59c0-4c0d-93df-bb26a541151f.png">
## 2. Dividing line between buttons for accessibility
Similar to how a dividing line was adding between button groups in places like [modals](https://next.carbon-addons-iot-react.com/?path=/story/3-carbon-modal--default), we need a dividing line between any secondary/primary button pairings for accessibility
### Current
Storybook: https://next.carbon-addons-iot-react.com/?path=/story/1-watson-iot-pagetitlebar--with-dynamic-scrolling
<img width="440" alt="Screen Shot 2021-11-08 at 4 32 14 PM" src="https://user-images.githubusercontent.com/43549567/140821478-571ddfd6-742a-4926-a9dc-cc065b62c3b1.png">
Storybook: https://next.carbon-addons-iot-react.com/?path=/story/1-watson-iot-pagetitlebar--with-condensed-header
### Proposed
<img width="778" alt="Screen Shot 2021-11-08 at 5 52 36 PM" src="https://user-images.githubusercontent.com/43549567/140830946-c35eb0a6-59c0-4c0d-93df-bb26a541151f.png">
## 3. Line below tabs in collapsed state
Similar to the dynamic header with no tabs, there should be a dividing line to create separation from the tabs and content on scroll.
### Current
Storybook: https://next.carbon-addons-iot-react.com/?path=/story/1-watson-iot-pagetitlebar--with-dynamic-scrolling-and-tabs
<img width="1564" alt="Screen Shot 2021-11-08 at 4 34 19 PM" src="https://user-images.githubusercontent.com/43549567/140821898-bac5288e-e637-4fa5-8003-c209d2d26a85.png">
### Proposed
<img width="1070" alt="Screen Shot 2021-11-08 at 5 57 02 PM" src="https://user-images.githubusercontent.com/43549567/140831405-001cf275-a922-42bc-9be9-3e8090c8f324.png">
## 4. Dynamic header with tabs enhancement
Can we add a variation/prop to have both the collapsed header and tabs both sticky? This is an enhancement we're making on the design side as well based on feedback so you don't lose your page-level actions and breadcrumb/page title.
### Proposed
<img width="1627" alt="Screen Shot 2021-11-08 at 5 57 43 PM" src="https://user-images.githubusercontent.com/43549567/140831477-ccc7da92-6493-43df-affe-a1230dddba1c.png">
|
non_process
|
style audit use this template if you want to request a new feature or a change to an existing feature if you d like to request an entirely new component please use the component request template instead if you are reporting a bug or problem please use the bug template instead what package is this for react angular padding adjustment on the dynamic page header there is an extra in the collapsed header state making it high and not current storybook img width alt screen shot at pm src proposed img width alt screen shot at pm src dividing line between buttons for accessibility similar to how a dividing line was adding between button groups in places like we need a dividing line between any secondary primary button pairings for accessibility current storybook img width alt screen shot at pm src storybook proposed img width alt screen shot at pm src line below tabs in collapsed state similar to the dynamic header with no tabs there should be a dividing line to create separation from the tabs and content on scroll current storybook img width alt screen shot at pm src proposed img width alt screen shot at pm src dynamic header with tabs enhancement can we add a variation prop to have both the collapsed header and tabs both sticky this is an enhancement we re making on the design side as well based on feedback so you don t lose your page level actions and breadcrumb page title proposed img width alt screen shot at pm src
| 0
|
10,159
| 13,044,162,642
|
IssuesEvent
|
2020-07-29 03:47:34
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `JsonSearchSig` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `JsonSearchSig` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @iosmanthus
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `JsonSearchSig` from TiDB -
## Description
Port the scalar function `JsonSearchSig` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @iosmanthus
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function jsonsearchsig from tidb description port the scalar function jsonsearchsig from tidb to coprocessor score mentor s iosmanthus recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
4,114
| 7,058,369,513
|
IssuesEvent
|
2018-01-04 20:05:26
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
opened
|
[Regression] Tests failed in System.ServiceProcess - FileNotFoundException
|
area-System.ServiceProcess bug os-windows
|
Failure is on all Windows OS versions (not Linux/Mac).
Likely regression from #26136 (@danmosemsft) - see @weshaggard's comment. I wonder why CI didn't catch it. I thought we use the same Helix infrastructure ...
## Types of failures
Affected tests:
* System.ServiceProcess.Tests.ServiceBaseTests /
* TestOnContinueBeforePause
* TestOnExecuteCustomCommand
* TestOnPauseAndContinueThenStop
* TestOnPauseThenStop
* TestOnStartThenStop
* TestOnStartWithArgsThenStop
* System.ServiceProcess.Tests.ServiceControllerTests / StartWithArguments
Failure:
```
System.IO.FileNotFoundException : Could not find file 'C:\\dotnetbuild\\work\\0488636f-ecff-467c-b6ef-343e05f4beae\\Work\\a8a48bf5-b649-4f63-999b-f9895293cdbd\\Unzip\\xunit.console.netcore.exe.9c0ebd24-e19c-43b3-97f7-3164daf3f557.log'.
at System.IO.FileStream.ValidateFileHandle(SafeFileHandle fileHandle) in E:\A\_work\1709\s\src\mscorlib\shared\System\IO\FileStream.Windows.cs:line 1661
at System.IO.FileStream.CreateFileOpenHandle(FileMode mode, FileShare share, FileOptions options) in E:\A\_work\1709\s\src\mscorlib\shared\System\IO\FileStream.Win32.cs:line 42
at System.IO.FileStream..ctor(String path, FileMode mode, FileAccess access, FileShare share, Int32 bufferSize, FileOptions options) in E:\A\_work\1709\s\src\mscorlib\shared\System\IO\FileStream.cs:line 229
at System.ServiceProcess.Tests.TestServiceProvider.GetServiceOutput() in E:\A\_work\1410\s\corefx\src\System.ServiceProcess.ServiceController\tests\TestServiceProvider.cs:line 126
at System.ServiceProcess.Tests.ServiceBaseTests.TestOnContinueBeforePause() in E:\A\_work\1410\s\corefx\src\System.ServiceProcess.ServiceController\tests\ServiceBaseTests.cs:line 174
```
## History of failures
Builds:
* [20180104.02](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180104.02/workItem/System.ServiceProcess.ServiceController.Tests) - 63 failures (7 tests, 9 platforms each)
* [20180104.04](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180104.04/workItem/System.ServiceProcess.ServiceController.Tests) - 63 failures (7 tests, 9 platforms each)
|
1.0
|
[Regression] Tests failed in System.ServiceProcess - FileNotFoundException - Failure is on all Windows OS versions (not Linux/Mac).
Likely regression from #26136 (@danmosemsft) - see @weshaggard's comment. I wonder why CI didn't catch it. I thought we use the same Helix infrastructure ...
## Types of failures
Affected tests:
* System.ServiceProcess.Tests.ServiceBaseTests /
* TestOnContinueBeforePause
* TestOnExecuteCustomCommand
* TestOnPauseAndContinueThenStop
* TestOnPauseThenStop
* TestOnStartThenStop
* TestOnStartWithArgsThenStop
* System.ServiceProcess.Tests.ServiceControllerTests / StartWithArguments
Failure:
```
System.IO.FileNotFoundException : Could not find file 'C:\\dotnetbuild\\work\\0488636f-ecff-467c-b6ef-343e05f4beae\\Work\\a8a48bf5-b649-4f63-999b-f9895293cdbd\\Unzip\\xunit.console.netcore.exe.9c0ebd24-e19c-43b3-97f7-3164daf3f557.log'.
at System.IO.FileStream.ValidateFileHandle(SafeFileHandle fileHandle) in E:\A\_work\1709\s\src\mscorlib\shared\System\IO\FileStream.Windows.cs:line 1661
at System.IO.FileStream.CreateFileOpenHandle(FileMode mode, FileShare share, FileOptions options) in E:\A\_work\1709\s\src\mscorlib\shared\System\IO\FileStream.Win32.cs:line 42
at System.IO.FileStream..ctor(String path, FileMode mode, FileAccess access, FileShare share, Int32 bufferSize, FileOptions options) in E:\A\_work\1709\s\src\mscorlib\shared\System\IO\FileStream.cs:line 229
at System.ServiceProcess.Tests.TestServiceProvider.GetServiceOutput() in E:\A\_work\1410\s\corefx\src\System.ServiceProcess.ServiceController\tests\TestServiceProvider.cs:line 126
at System.ServiceProcess.Tests.ServiceBaseTests.TestOnContinueBeforePause() in E:\A\_work\1410\s\corefx\src\System.ServiceProcess.ServiceController\tests\ServiceBaseTests.cs:line 174
```
## History of failures
Builds:
* [20180104.02](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180104.02/workItem/System.ServiceProcess.ServiceController.Tests) - 63 failures (7 tests, 9 platforms each)
* [20180104.04](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180104.04/workItem/System.ServiceProcess.ServiceController.Tests) - 63 failures (7 tests, 9 platforms each)
|
process
|
tests failed in system serviceprocess filenotfoundexception failure is on all windows os versions not linux mac likely regression from danmosemsft see weshaggard s comment i wonder why ci didn t catch it i thought we use the same helix infrastructure types of failures affected tests system serviceprocess tests servicebasetests testoncontinuebeforepause testonexecutecustomcommand testonpauseandcontinuethenstop testonpausethenstop testonstartthenstop testonstartwithargsthenstop system serviceprocess tests servicecontrollertests startwitharguments failure system io filenotfoundexception could not find file c dotnetbuild work ecff work unzip xunit console netcore exe log at system io filestream validatefilehandle safefilehandle filehandle in e a work s src mscorlib shared system io filestream windows cs line at system io filestream createfileopenhandle filemode mode fileshare share fileoptions options in e a work s src mscorlib shared system io filestream cs line at system io filestream ctor string path filemode mode fileaccess access fileshare share buffersize fileoptions options in e a work s src mscorlib shared system io filestream cs line at system serviceprocess tests testserviceprovider getserviceoutput in e a work s corefx src system serviceprocess servicecontroller tests testserviceprovider cs line at system serviceprocess tests servicebasetests testoncontinuebeforepause in e a work s corefx src system serviceprocess servicecontroller tests servicebasetests cs line history of failures builds failures tests platforms each failures tests platforms each
| 1
|
18,382
| 24,512,612,241
|
IssuesEvent
|
2022-10-10 23:49:27
|
pyanodon/pybugreports
|
https://api.github.com/repos/pyanodon/pybugreports
|
closed
|
Incompatible with Realistic Reactors - missing crafting category
|
mod:pypostprocessing postprocess-fail compatibility
|
### Mod source
PyAE Beta
### Which mod are you having an issue with?
- [ ] pyalienlife
- [ ] pyalternativeenergy
- [ ] pycoalprocessing
- [ ] pyfusionenergy
- [ ] pyhightech
- [ ] pyindustry
- [ ] pypetroleumhandling
- [X] pypostprocessing
- [ ] pyrawores
### Operating system
>=Windows 10
### What kind of issue is this?
- [ ] Compatibility
- [ ] Locale (names, descriptions, unknown keys)
- [ ] Graphical
- [ ] Crash
- [ ] Progression
- [ ] Balance
- [X] Pypostprocessing failure
- [ ] Other
### What is the problem?
31.942 Script @__pypostprocessing__/data-final-fixes.lua:142: AUTOTECH START
32.013 Error ModManager.cpp:1558: Failed to load mod "pypostprocessing":
ERROR: Missing crafting category: water-cooling (ingredients: 0, fluids in: 1, fluids out:0), for __START__ / water-cooling
stack traceback:
[C]: in function 'error'
__pypostprocessing__/prototypes/functions/data_parser.lua:239: in function 'parse_recipe'
__pypostprocessing__/prototypes/functions/data_parser.lua:116: in function 'run'
__pypostprocessing__/prototypes/functions/auto_tech.lua:34: in function 'run'
__pypostprocessing__/data-final-fixes.lua:144: in main chunk
32.014 Loading mod core 0.0.0 (data.lua)
32.217 Checksum for core: 1476961332
32.249 Error ModManager.cpp:1558: Error in assignID: recipe-category with name 'crafting' does not exist.
### Steps to reproduce
Install https://mods.factorio.com/mod/RealisticReactors
### Additional context
_No response_
### Log file
_No response_
|
2.0
|
Incompatible with Realistic Reactors - missing crafting category - ### Mod source
PyAE Beta
### Which mod are you having an issue with?
- [ ] pyalienlife
- [ ] pyalternativeenergy
- [ ] pycoalprocessing
- [ ] pyfusionenergy
- [ ] pyhightech
- [ ] pyindustry
- [ ] pypetroleumhandling
- [X] pypostprocessing
- [ ] pyrawores
### Operating system
>=Windows 10
### What kind of issue is this?
- [ ] Compatibility
- [ ] Locale (names, descriptions, unknown keys)
- [ ] Graphical
- [ ] Crash
- [ ] Progression
- [ ] Balance
- [X] Pypostprocessing failure
- [ ] Other
### What is the problem?
31.942 Script @__pypostprocessing__/data-final-fixes.lua:142: AUTOTECH START
32.013 Error ModManager.cpp:1558: Failed to load mod "pypostprocessing":
ERROR: Missing crafting category: water-cooling (ingredients: 0, fluids in: 1, fluids out:0), for __START__ / water-cooling
stack traceback:
[C]: in function 'error'
__pypostprocessing__/prototypes/functions/data_parser.lua:239: in function 'parse_recipe'
__pypostprocessing__/prototypes/functions/data_parser.lua:116: in function 'run'
__pypostprocessing__/prototypes/functions/auto_tech.lua:34: in function 'run'
__pypostprocessing__/data-final-fixes.lua:144: in main chunk
32.014 Loading mod core 0.0.0 (data.lua)
32.217 Checksum for core: 1476961332
32.249 Error ModManager.cpp:1558: Error in assignID: recipe-category with name 'crafting' does not exist.
### Steps to reproduce
Install https://mods.factorio.com/mod/RealisticReactors
### Additional context
_No response_
### Log file
_No response_
|
process
|
incompatible with realistic reactors missing crafting category mod source pyae beta which mod are you having an issue with pyalienlife pyalternativeenergy pycoalprocessing pyfusionenergy pyhightech pyindustry pypetroleumhandling pypostprocessing pyrawores operating system windows what kind of issue is this compatibility locale names descriptions unknown keys graphical crash progression balance pypostprocessing failure other what is the problem script pypostprocessing data final fixes lua autotech start error modmanager cpp failed to load mod pypostprocessing error missing crafting category water cooling ingredients fluids in fluids out for start water cooling stack traceback in function error pypostprocessing prototypes functions data parser lua in function parse recipe pypostprocessing prototypes functions data parser lua in function run pypostprocessing prototypes functions auto tech lua in function run pypostprocessing data final fixes lua in main chunk loading mod core data lua checksum for core error modmanager cpp error in assignid recipe category with name crafting does not exist steps to reproduce install additional context no response log file no response
| 1
|
45,113
| 18,412,550,861
|
IssuesEvent
|
2021-10-13 07:52:51
|
microsoft/BotFramework-Composer
|
https://api.github.com/repos/microsoft/BotFramework-Composer
|
closed
|
Problem while converting skill from Composer 1.4.x to Composer 2.x.x
|
Type: Bug customer-reported Bot Services customer-replied-to
|
<!-- Please search for your feature request before creating a new one. >
<!-- Complete the necessary portions of this template and delete the rest. -->
## Describe the bug
We have a Bot Composer 1.4.1 skill, we faced problems while converting it to the new Composer 2.x.x skill. The Composer skill we are using has a custom runtime with additional features supporting our requirements. After we got the "EISDIR: illegal operation on a directory" error. We looked up this issue and found a few that are related to this issue, no luck though (https://github.com/microsoft/BotFramework-Composer/issues/7444, https://github.com/microsoft/BotFramework-Composer/issues/7373).
After opening the new converted project which had errors on conversion, saw errors throughout the skill about missing schemas and unsupported dialogs. (Screenshots attached below)
We were also wondering, what the standard conversion from 1.4.x to 2.x.x actually does other than adding the new adaptive runtime. If that's something that can be done manually.
**Things we've tried:** using other versions of NPM, starting composer with different privileges, with different subversions of bot composer 2.1.
## Version
<!-- What version of the Composer are you using? Paste the build SHA found on the about page (`/about`). -->
Version: 2.1.1
Electron: 8.2.4
Chrome: 80.0.3987.165
NodeJS: 12.13.0
V8: 8.0.426.27-electron.0
## Browser
<!-- What browser are you using? -->
- [ ] Electron distribution
- [x] Chrome
- [ ] Safari
- [ ] Firefox
- [ ] Edge
## OS
<!-- What operating system are you using? -->
- [ ] macOS
- [x] Windows
- [ ] Ubuntu
## To Reproduce
Steps to reproduce the behavior:
1. Launch Bot Composer 2.1.1.
2. Open an existing 1.4.x based skill.
3. Click on "Convert".
4. Create a new bot project with the following settings. (unlike what's in issue https://github.com/microsoft/BotFramework-Composer/issues/7373)

## Expected behavior
<!-- Give a clear and concise description of what you expected to happen. -->
For it to at least convert the bare minimum of what's possible with automated conversion.
## Screenshots
<!-- If applicable, add screenshots/gif/video to help explain your problem. -->



## Additional context
<!-- Add any other context about the problem here. -->
|
1.0
|
Problem while converting skill from Composer 1.4.x to Composer 2.x.x -
<!-- Please search for your feature request before creating a new one. >
<!-- Complete the necessary portions of this template and delete the rest. -->
## Describe the bug
We have a Bot Composer 1.4.1 skill, we faced problems while converting it to the new Composer 2.x.x skill. The Composer skill we are using has a custom runtime with additional features supporting our requirements. After we got the "EISDIR: illegal operation on a directory" error. We looked up this issue and found a few that are related to this issue, no luck though (https://github.com/microsoft/BotFramework-Composer/issues/7444, https://github.com/microsoft/BotFramework-Composer/issues/7373).
After opening the new converted project which had errors on conversion, saw errors throughout the skill about missing schemas and unsupported dialogs. (Screenshots attached below)
We were also wondering, what the standard conversion from 1.4.x to 2.x.x actually does other than adding the new adaptive runtime. If that's something that can be done manually.
**Things we've tried:** using other versions of NPM, starting composer with different privileges, with different subversions of bot composer 2.1.
## Version
<!-- What version of the Composer are you using? Paste the build SHA found on the about page (`/about`). -->
Version: 2.1.1
Electron: 8.2.4
Chrome: 80.0.3987.165
NodeJS: 12.13.0
V8: 8.0.426.27-electron.0
## Browser
<!-- What browser are you using? -->
- [ ] Electron distribution
- [x] Chrome
- [ ] Safari
- [ ] Firefox
- [ ] Edge
## OS
<!-- What operating system are you using? -->
- [ ] macOS
- [x] Windows
- [ ] Ubuntu
## To Reproduce
Steps to reproduce the behavior:
1. Launch Bot Composer 2.1.1.
2. Open an existing 1.4.x based skill.
3. Click on "Convert".
4. Create a new bot project with the following settings. (unlike what's in issue https://github.com/microsoft/BotFramework-Composer/issues/7373)

## Expected behavior
<!-- Give a clear and concise description of what you expected to happen. -->
For it to at least convert the bare minimum of what's possible with automated conversion.
## Screenshots
<!-- If applicable, add screenshots/gif/video to help explain your problem. -->



## Additional context
<!-- Add any other context about the problem here. -->
|
non_process
|
problem while converting skill from composer x to composer x x describe the bug we have a bot composer skill we faced problems while converting it to the new composer x x skill the composer skill we are using has a custom runtime with additional features supporting our requirements after we got the eisdir illegal operation on a directory error we looked up this issue and found a few that are related to this issue no luck though after opening the new converted project which had errors on conversion saw errors throughout the skill about missing schemas and unsupported dialogs screenshots attached below we were also wondering what the standard conversion from x to x x actually does other than adding the new adaptive runtime if that s something that can be done manually things we ve tried using other versions of npm starting composer with different privileges with different subversions of bot composer version version electron chrome nodejs electron browser electron distribution chrome safari firefox edge os macos windows ubuntu to reproduce steps to reproduce the behavior launch bot composer open an existing x based skill click on convert create a new bot project with the following settings unlike what s in issue expected behavior for it to at least convert the bare minimum of what s possible with automated conversion screenshots additional context
| 0
|
8,585
| 4,281,997,713
|
IssuesEvent
|
2016-07-15 07:04:38
|
cake-build/cake
|
https://api.github.com/repos/cake-build/cake
|
closed
|
Correct code styling issues in Tests
|
Build
|
The spacing around the keyword 'typeof' is invalid, found in:
* src\Cake.Core.Tests\Fixtures\MethodAliasGeneratorFixture.cs:L20-20
* src\Cake.Core.Tests\Fixtures\MethodAliasGeneratorFixture.cs:L21-21
* src\Cake.Core.Tests\Fixtures\PropertyAliasGeneratorFixture.cs:L20-20
* src\Cake.Core.Tests\Fixtures\PropertyAliasGeneratorFixture.cs:L21-21
The spacing around the keyword 'switch' is invalid, found in:
src\Cake.Common.Tests\Fixtures\Tools\WiX\HeatFixture.cs:L36-36
Fields must be declared with private access. Use properties to expose fields found in:
* src\Cake.Common.Tests\Fixtures\Tools\GitReleaseNotesRunnerFixture.cs:L12-12
* src\Cake.Common.Tests\Fixtures\Tools\ReportUnit\ReportUnitDirectoryFixture.cs:L12-12
* src\Cake.Common.Tests\Fixtures\Tools\ReportUnit\ReportUnitDirectoryFixture.cs:L13-13
* src\Cake.Common.Tests\Fixtures\Tools\ReportUnit\ReportUnitFileFixture.cs:L12-12
* src\Cake.Common.Tests\Fixtures\Tools\ReportUnit\ReportUnitFileFixture.cs:L13-13
Invalid spacing around the closing curly bracket, found in:
* src\Cake.Common.Tests\Fixtures\Tools\CakeRunnerFixture.cs:L13-13
* src\Cake.Common.Tests\Fixtures\Tools\NuGet\SetApiKey\NuGetSetApiKeyFixture.cs:L21-21
* src\Cake.Common.Tests\Fixtures\Tools\NuGet\Sources\NuGetSourcesFixture.cs:L25-25
* src\Cake.Common.Tests\Fixtures\XmlDocExampleCodeParserFixture.cs:L34-34
An opening curly bracket must not be followed by a blank line, found in:
* src\Cake.Common.Tests\Fixtures\XmlPokeFixture.cs:L66-66
* src\Cake.Common.Tests\Fixtures\XmlPokeFixture.cs:L93-93
Invalid spacing around the semicolon, found in:
* src\Cake.Common.Tests\Fixtures\Tools\CakeRunnerFixture.cs:L13-13
Invalid spacing around the closing generic bracket, found in:
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L525-525
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L539-539
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L540-540
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L542-542
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L544-544
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L558-558
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L559-559
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L561-561
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L563-563
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L577-577
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L578-578
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L579-579
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L589-589
A closing curly bracket must not be preceded by a blank line, found in:
* src\Cake.Common.Tests\Unit\Build\Bamboo\Data\BambooPlanInfoTests.cs:L107-107
* src\Cake.Common.Tests\Unit\Build\Bamboo\Data\BambooPlanInfoTests.cs:L109-109
* src\Cake.Common.Tests\Unit\Build\BuildSystemTests.cs:L117-117
* src\Cake.Common.Tests\Unit\Build\BuildSystemTests.cs:L97-97
* src\Cake.Common.Tests\Unit\Tools\DotCover\Analyse\DotCoverAnalyserTests.cs:L45-45
* src\Cake.Common.Tests\Unit\Tools\DotCover\Cover\DotCoverCovererTests.cs:L45-45
* src\Cake.Common.Tests\Unit\Tools\SpecFlow\TestExecutionReport\SpecFlowTestExecutionReporterTests.cs:L46-46
* src\Cake.Core.Tests\Unit\Packaging\PackageReferenceTests.cs:L131-131
The code must not contain multiple blank lines in a row, found in:
* src\Cake.Common.Tests\Fixtures\Build\ContinuaCIInfoFixture.cs:L69-69
* src\Cake.Common.Tests\Fixtures\Tools\SignToolResolverFixture.cs:L19-19
* src\Cake.Common.Tests\Unit\Text\TextTransformationAliasesTests.cs:L329-329
* src\Cake.Common.Tests\Unit\Tools\Chocolatey\Push\ChocolateyPusherTests.cs:L145-145
* src\Cake.Common.Tests\Unit\Tools\NuGet\Restore\NuGetRestorerTests.cs:L252-252
* src\Cake.Common.Tests\Unit\Tools\NuGet\Restore\NuGetRestorerTests.cs:L253-253
* src\Cake.Core.Tests\Unit\CakeTaskTests.cs:L48-48
* src\Cake.Core.Tests\Unit\Configuration\Parser\ConfigurationTokenizerTests.cs:L96-96
* src\Cake.Core.Tests\Unit\IO\GlobberTests.cs:L431-431
* src\Cake.Core.Tests\Unit\IO\PathComparerTests.cs:L56-56
* src\Cake.NuGet.Tests\Properties\AssemblyInfo.cs:L20-20
* src\Cake.Testing.Xunit\WindowsTheory.cs:L19-19
An opening curly bracket must not be followed by a blank line, found in:
* src\Cake.Common.Tests\Fixtures\XmlPokeFixture.cs:L66-66
* src\Cake.Common.Tests\Fixtures\XmlPokeFixture.cs:L93-93
* src\Cake.Common.Tests\Unit\Build\Bamboo\Data\BambooCommitInfoTests.cs:L25-25
* src\Cake.Common.Tests\Unit\Build\Bamboo\Data\BambooRepositoryInfoTests.cs:L10-10
* src\Cake.Common.Tests\Unit\Build\Bitrise\Data\BitriseDirectoryInfoTests.cs:L43-43
* src\Cake.Common.Tests\Unit\Build\Jenkins\Data\JenkinsBuildInfoTests.cs:L91-91
* src\Cake.Common.Tests\Unit\Solution\Project\ProjectParserTests.cs:L50-50
* src\Cake.Common.Tests\Unit\Tools\NuGet\Pack\NuGetPackerTests.cs:L801-801
* src\Cake.Common.Tests\Unit\Tools\NuGet\Restore\NuGetRestorerTests.cs:L274-274
A single-line comment must not be followed by a blank line, found in:
* src\Cake.NuGet.Tests\Unit\NuGetAssemblyCompatibilityFilterTests.cs:L22-22
The spacing around the symbol '=' is invalide, found in:
* src\Cake.Common.Tests\Unit\Tools\SpecFlow\TestExecutionReport\SpecFlowTestExecutionReporterTests.cs:L234-234
|
1.0
|
Correct code styling issues in Tests - The spacing around the keyword 'typeof' is invalid, found in:
* src\Cake.Core.Tests\Fixtures\MethodAliasGeneratorFixture.cs:L20-20
* src\Cake.Core.Tests\Fixtures\MethodAliasGeneratorFixture.cs:L21-21
* src\Cake.Core.Tests\Fixtures\PropertyAliasGeneratorFixture.cs:L20-20
* src\Cake.Core.Tests\Fixtures\PropertyAliasGeneratorFixture.cs:L21-21
The spacing around the keyword 'switch' is invalid, found in:
src\Cake.Common.Tests\Fixtures\Tools\WiX\HeatFixture.cs:L36-36
Fields must be declared with private access. Use properties to expose fields found in:
* src\Cake.Common.Tests\Fixtures\Tools\GitReleaseNotesRunnerFixture.cs:L12-12
* src\Cake.Common.Tests\Fixtures\Tools\ReportUnit\ReportUnitDirectoryFixture.cs:L12-12
* src\Cake.Common.Tests\Fixtures\Tools\ReportUnit\ReportUnitDirectoryFixture.cs:L13-13
* src\Cake.Common.Tests\Fixtures\Tools\ReportUnit\ReportUnitFileFixture.cs:L12-12
* src\Cake.Common.Tests\Fixtures\Tools\ReportUnit\ReportUnitFileFixture.cs:L13-13
Invalid spacing around the closing curly bracket, found in:
* src\Cake.Common.Tests\Fixtures\Tools\CakeRunnerFixture.cs:L13-13
* src\Cake.Common.Tests\Fixtures\Tools\NuGet\SetApiKey\NuGetSetApiKeyFixture.cs:L21-21
* src\Cake.Common.Tests\Fixtures\Tools\NuGet\Sources\NuGetSourcesFixture.cs:L25-25
* src\Cake.Common.Tests\Fixtures\XmlDocExampleCodeParserFixture.cs:L34-34
An opening curly bracket must not be followed by a blank line, found in:
* src\Cake.Common.Tests\Fixtures\XmlPokeFixture.cs:L66-66
* src\Cake.Common.Tests\Fixtures\XmlPokeFixture.cs:L93-93
Invalid spacing around the semicolon, found in:
* src\Cake.Common.Tests\Fixtures\Tools\CakeRunnerFixture.cs:L13-13
Invalid spacing around the closing generic bracket, found in:
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L525-525
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L539-539
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L540-540
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L542-542
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L544-544
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L558-558
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L559-559
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L561-561
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L563-563
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L577-577
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L578-578
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L579-579
* src\Cake.Common.Tests\Unit\IO\DirectoryAliasesTests.cs:L589-589
A closing curly bracket must not be preceded by a blank line, found in:
* src\Cake.Common.Tests\Unit\Build\Bamboo\Data\BambooPlanInfoTests.cs:L107-107
* src\Cake.Common.Tests\Unit\Build\Bamboo\Data\BambooPlanInfoTests.cs:L109-109
* src\Cake.Common.Tests\Unit\Build\BuildSystemTests.cs:L117-117
* src\Cake.Common.Tests\Unit\Build\BuildSystemTests.cs:L97-97
* src\Cake.Common.Tests\Unit\Tools\DotCover\Analyse\DotCoverAnalyserTests.cs:L45-45
* src\Cake.Common.Tests\Unit\Tools\DotCover\Cover\DotCoverCovererTests.cs:L45-45
* src\Cake.Common.Tests\Unit\Tools\SpecFlow\TestExecutionReport\SpecFlowTestExecutionReporterTests.cs:L46-46
* src\Cake.Core.Tests\Unit\Packaging\PackageReferenceTests.cs:L131-131
The code must not contain multiple blank lines in a row, found in:
* src\Cake.Common.Tests\Fixtures\Build\ContinuaCIInfoFixture.cs:L69-69
* src\Cake.Common.Tests\Fixtures\Tools\SignToolResolverFixture.cs:L19-19
* src\Cake.Common.Tests\Unit\Text\TextTransformationAliasesTests.cs:L329-329
* src\Cake.Common.Tests\Unit\Tools\Chocolatey\Push\ChocolateyPusherTests.cs:L145-145
* src\Cake.Common.Tests\Unit\Tools\NuGet\Restore\NuGetRestorerTests.cs:L252-252
* src\Cake.Common.Tests\Unit\Tools\NuGet\Restore\NuGetRestorerTests.cs:L253-253
* src\Cake.Core.Tests\Unit\CakeTaskTests.cs:L48-48
* src\Cake.Core.Tests\Unit\Configuration\Parser\ConfigurationTokenizerTests.cs:L96-96
* src\Cake.Core.Tests\Unit\IO\GlobberTests.cs:L431-431
* src\Cake.Core.Tests\Unit\IO\PathComparerTests.cs:L56-56
* src\Cake.NuGet.Tests\Properties\AssemblyInfo.cs:L20-20
* src\Cake.Testing.Xunit\WindowsTheory.cs:L19-19
An opening curly bracket must not be followed by a blank line, found in:
* src\Cake.Common.Tests\Fixtures\XmlPokeFixture.cs:L66-66
* src\Cake.Common.Tests\Fixtures\XmlPokeFixture.cs:L93-93
* src\Cake.Common.Tests\Unit\Build\Bamboo\Data\BambooCommitInfoTests.cs:L25-25
* src\Cake.Common.Tests\Unit\Build\Bamboo\Data\BambooRepositoryInfoTests.cs:L10-10
* src\Cake.Common.Tests\Unit\Build\Bitrise\Data\BitriseDirectoryInfoTests.cs:L43-43
* src\Cake.Common.Tests\Unit\Build\Jenkins\Data\JenkinsBuildInfoTests.cs:L91-91
* src\Cake.Common.Tests\Unit\Solution\Project\ProjectParserTests.cs:L50-50
* src\Cake.Common.Tests\Unit\Tools\NuGet\Pack\NuGetPackerTests.cs:L801-801
* src\Cake.Common.Tests\Unit\Tools\NuGet\Restore\NuGetRestorerTests.cs:L274-274
A single-line comment must not be followed by a blank line, found in:
* src\Cake.NuGet.Tests\Unit\NuGetAssemblyCompatibilityFilterTests.cs:L22-22
The spacing around the symbol '=' is invalide, found in:
* src\Cake.Common.Tests\Unit\Tools\SpecFlow\TestExecutionReport\SpecFlowTestExecutionReporterTests.cs:L234-234
|
non_process
|
correct code styling issues in tests the spacing around the keyword typeof is invalid found in src cake core tests fixtures methodaliasgeneratorfixture cs src cake core tests fixtures methodaliasgeneratorfixture cs src cake core tests fixtures propertyaliasgeneratorfixture cs src cake core tests fixtures propertyaliasgeneratorfixture cs the spacing around the keyword switch is invalid found in src cake common tests fixtures tools wix heatfixture cs fields must be declared with private access use properties to expose fields found in src cake common tests fixtures tools gitreleasenotesrunnerfixture cs src cake common tests fixtures tools reportunit reportunitdirectoryfixture cs src cake common tests fixtures tools reportunit reportunitdirectoryfixture cs src cake common tests fixtures tools reportunit reportunitfilefixture cs src cake common tests fixtures tools reportunit reportunitfilefixture cs invalid spacing around the closing curly bracket found in src cake common tests fixtures tools cakerunnerfixture cs src cake common tests fixtures tools nuget setapikey nugetsetapikeyfixture cs src cake common tests fixtures tools nuget sources nugetsourcesfixture cs src cake common tests fixtures xmldocexamplecodeparserfixture cs an opening curly bracket must not be followed by a blank line found in src cake common tests fixtures xmlpokefixture cs src cake common tests fixtures xmlpokefixture cs invalid spacing around the semicolon found in src cake common tests fixtures tools cakerunnerfixture cs invalid spacing around the closing generic bracket found in src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs src cake common tests unit io directoryaliasestests cs a closing curly bracket must not be preceded by a blank line found in src cake common tests unit build bamboo data bambooplaninfotests cs src cake common tests unit build bamboo data bambooplaninfotests cs src cake common tests unit build buildsystemtests cs src cake common tests unit build buildsystemtests cs src cake common tests unit tools dotcover analyse dotcoveranalysertests cs src cake common tests unit tools dotcover cover dotcovercoverertests cs src cake common tests unit tools specflow testexecutionreport specflowtestexecutionreportertests cs src cake core tests unit packaging packagereferencetests cs the code must not contain multiple blank lines in a row found in src cake common tests fixtures build continuaciinfofixture cs src cake common tests fixtures tools signtoolresolverfixture cs src cake common tests unit text texttransformationaliasestests cs src cake common tests unit tools chocolatey push chocolateypushertests cs src cake common tests unit tools nuget restore nugetrestorertests cs src cake common tests unit tools nuget restore nugetrestorertests cs src cake core tests unit caketasktests cs src cake core tests unit configuration parser configurationtokenizertests cs src cake core tests unit io globbertests cs src cake core tests unit io pathcomparertests cs src cake nuget tests properties assemblyinfo cs src cake testing xunit windowstheory cs an opening curly bracket must not be followed by a blank line found in src cake common tests fixtures xmlpokefixture cs src cake common tests fixtures xmlpokefixture cs src cake common tests unit build bamboo data bamboocommitinfotests cs src cake common tests unit build bamboo data bamboorepositoryinfotests cs src cake common tests unit build bitrise data bitrisedirectoryinfotests cs src cake common tests unit build jenkins data jenkinsbuildinfotests cs src cake common tests unit solution project projectparsertests cs src cake common tests unit tools nuget pack nugetpackertests cs src cake common tests unit tools nuget restore nugetrestorertests cs a single line comment must not be followed by a blank line found in src cake nuget tests unit nugetassemblycompatibilityfiltertests cs the spacing around the symbol is invalide found in src cake common tests unit tools specflow testexecutionreport specflowtestexecutionreportertests cs
| 0
|
159,664
| 20,085,888,947
|
IssuesEvent
|
2022-02-05 01:07:19
|
AkshayMukkavilli/Tensorflow
|
https://api.github.com/repos/AkshayMukkavilli/Tensorflow
|
opened
|
CVE-2022-21730 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl
|
security vulnerability
|
## CVE-2022-21730 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /Tensorflow/src/requirements.txt</p>
<p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Tensorflow is an Open Source Machine Learning Framework. The implementation of `FractionalAvgPoolGrad` does not consider cases where the input tensors are invalid allowing an attacker to read from outside of bounds of heap. The fix will be included in TensorFlow 2.8.0. We will also cherrypick this commit on TensorFlow 2.7.1, TensorFlow 2.6.3, and TensorFlow 2.5.3, as these are also affected and still in supported range.
<p>Publish Date: 2022-02-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21730>CVE-2022-21730</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-vjg4-v33c-ggc4">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-vjg4-v33c-ggc4</a></p>
<p>Release Date: 2022-02-03</p>
<p>Fix Resolution: tensorflow - 2.5.3,2.6.3,2.7.1;tensorflow-cpu - 2.5.3,2.6.3,2.7.1;tensorflow-gpu - 2.5.3,2.6.3,2.7.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-21730 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2022-21730 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /Tensorflow/src/requirements.txt</p>
<p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Tensorflow is an Open Source Machine Learning Framework. The implementation of `FractionalAvgPoolGrad` does not consider cases where the input tensors are invalid allowing an attacker to read from outside of bounds of heap. The fix will be included in TensorFlow 2.8.0. We will also cherrypick this commit on TensorFlow 2.7.1, TensorFlow 2.6.3, and TensorFlow 2.5.3, as these are also affected and still in supported range.
<p>Publish Date: 2022-02-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21730>CVE-2022-21730</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-vjg4-v33c-ggc4">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-vjg4-v33c-ggc4</a></p>
<p>Release Date: 2022-02-03</p>
<p>Fix Resolution: tensorflow - 2.5.3,2.6.3,2.7.1;tensorflow-cpu - 2.5.3,2.6.3,2.7.1;tensorflow-gpu - 2.5.3,2.6.3,2.7.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file tensorflow src requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an open source machine learning framework the implementation of fractionalavgpoolgrad does not consider cases where the input tensors are invalid allowing an attacker to read from outside of bounds of heap the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
| 0
|
22,253
| 30,802,634,911
|
IssuesEvent
|
2023-08-01 03:34:47
|
emily-writes-poems/emily-writes-poems-processing
|
https://api.github.com/repos/emily-writes-poems/emily-writes-poems-processing
|
closed
|
redesign display of linked poems
|
processing refinement
|
would be great if this was a separate modal that looks nice and has:
- list of currently linked poems, with options to delete link
- form to select another poem to link
|
1.0
|
redesign display of linked poems - would be great if this was a separate modal that looks nice and has:
- list of currently linked poems, with options to delete link
- form to select another poem to link
|
process
|
redesign display of linked poems would be great if this was a separate modal that looks nice and has list of currently linked poems with options to delete link form to select another poem to link
| 1
|
9,457
| 12,438,476,473
|
IssuesEvent
|
2020-05-26 08:29:35
|
prisma/migrate
|
https://api.github.com/repos/prisma/migrate
|
closed
|
Truncate migration README SQL statements with a semicolon
|
kind/improvement process/candidate
|
## Problem
In the `Database Steps` section of a migration's README.md, there's a list of SQL statements that Migrate took to reach a specific state.
## Suggested solution
If each statement were to be truncated with a semi-colon, it would make it easy to copy & paste the generated SQL directly into a SQL editor and execute them in one go.
Tiny, tiny QoL thing really.
## Alternatives
I can go over and type in the semi-colons myself too, but who wants to do that
## Additional context
This is the SQL generated for a migration I just `save`ed:
```sql
CREATE TABLE "public"."User" (
"active" boolean NOT NULL DEFAULT true,"createdAt" timestamp(3) NOT NULL DEFAULT CURRENT_TIMESTAMP,"email" text NOT NULL ,"id" text NOT NULL ,"name" text ,"registryAccessToken" text NOT NULL ,"updatedAt" timestamp(3) NOT NULL ,"updatedBy" text ,
PRIMARY KEY ("id"))
CREATE UNIQUE INDEX "User.email" ON "public"."User"("email")
CREATE UNIQUE INDEX "User.registryAccessToken" ON "public"."User"("registryAccessToken")
ALTER TABLE "public"."User" ADD FOREIGN KEY ("updatedBy")REFERENCES "public"."User"("id") ON DELETE SET NULL ON UPDATE CASCADE
DROP TYPE "ScoutingFile"
```
|
1.0
|
Truncate migration README SQL statements with a semicolon - ## Problem
In the `Database Steps` section of a migration's README.md, there's a list of SQL statements that Migrate took to reach a specific state.
## Suggested solution
If each statement were to be truncated with a semi-colon, it would make it easy to copy & paste the generated SQL directly into a SQL editor and execute them in one go.
Tiny, tiny QoL thing really.
## Alternatives
I can go over and type in the semi-colons myself too, but who wants to do that
## Additional context
This is the SQL generated for a migration I just `save`ed:
```sql
CREATE TABLE "public"."User" (
"active" boolean NOT NULL DEFAULT true,"createdAt" timestamp(3) NOT NULL DEFAULT CURRENT_TIMESTAMP,"email" text NOT NULL ,"id" text NOT NULL ,"name" text ,"registryAccessToken" text NOT NULL ,"updatedAt" timestamp(3) NOT NULL ,"updatedBy" text ,
PRIMARY KEY ("id"))
CREATE UNIQUE INDEX "User.email" ON "public"."User"("email")
CREATE UNIQUE INDEX "User.registryAccessToken" ON "public"."User"("registryAccessToken")
ALTER TABLE "public"."User" ADD FOREIGN KEY ("updatedBy")REFERENCES "public"."User"("id") ON DELETE SET NULL ON UPDATE CASCADE
DROP TYPE "ScoutingFile"
```
|
process
|
truncate migration readme sql statements with a semicolon problem in the database steps section of a migration s readme md there s a list of sql statements that migrate took to reach a specific state suggested solution if each statement were to be truncated with a semi colon it would make it easy to copy paste the generated sql directly into a sql editor and execute them in one go tiny tiny qol thing really alternatives i can go over and type in the semi colons myself too but who wants to do that additional context this is the sql generated for a migration i just save ed sql create table public user active boolean not null default true createdat timestamp not null default current timestamp email text not null id text not null name text registryaccesstoken text not null updatedat timestamp not null updatedby text primary key id create unique index user email on public user email create unique index user registryaccesstoken on public user registryaccesstoken alter table public user add foreign key updatedby references public user id on delete set null on update cascade drop type scoutingfile
| 1
|
21,487
| 7,030,599,315
|
IssuesEvent
|
2017-12-26 11:13:16
|
Microsoft/WindowsTemplateStudio
|
https://api.github.com/repos/Microsoft/WindowsTemplateStudio
|
closed
|
Build dev.templates.tests.gen_20171226.1 failed
|
bug vsts-build
|
## Build dev.templates.tests.gen_20171226.1
- **Build result:** `failed`
- **Build queued:** 12/26/2017 6:03:58 AM
- **Build duration:** 90.34 minutes
### Details
Build [dev.templates.tests.gen_20171226.1](https://winappstudio.visualstudio.com/web/build.aspx?pcguid=a4ef43be-68ce-4195-a619-079b4d9834c2&builduri=vstfs%3a%2f%2f%2fBuild%2fBuild%2f24537) failed
+ The job running on agent wtsb6 has exceeded the maximum execution time of 01:30:00.
+ The operation was canceled.
Find detailed information in the [build log files](https://uwpctdiags.blob.core.windows.net/buildlogs/dev.templates.tests.gen_20171226.1_logs.zip)
|
1.0
|
Build dev.templates.tests.gen_20171226.1 failed - ## Build dev.templates.tests.gen_20171226.1
- **Build result:** `failed`
- **Build queued:** 12/26/2017 6:03:58 AM
- **Build duration:** 90.34 minutes
### Details
Build [dev.templates.tests.gen_20171226.1](https://winappstudio.visualstudio.com/web/build.aspx?pcguid=a4ef43be-68ce-4195-a619-079b4d9834c2&builduri=vstfs%3a%2f%2f%2fBuild%2fBuild%2f24537) failed
+ The job running on agent wtsb6 has exceeded the maximum execution time of 01:30:00.
+ The operation was canceled.
Find detailed information in the [build log files](https://uwpctdiags.blob.core.windows.net/buildlogs/dev.templates.tests.gen_20171226.1_logs.zip)
|
non_process
|
build dev templates tests gen failed build dev templates tests gen build result failed build queued am build duration minutes details build failed the job running on agent has exceeded the maximum execution time of the operation was canceled find detailed information in the
| 0
|
20,770
| 27,503,972,291
|
IssuesEvent
|
2023-03-06 00:26:12
|
jart/blink
|
https://api.github.com/repos/jart/blink
|
closed
|
Weird RNG behavior (running rust binary)
|
microprocessor
|
It seems that I have a weird random generator issue:
```
use rand::Rng;
use rand::SeedableRng;
use rand_chacha::ChaChaRng;
pub fn main() {
let mut rng = ChaChaRng::seed_from_u64(0x42);
for _ in 1..1000 {
let x: [u8; 8] = rng.gen();
println!("{:?}", x);
}
}
```
After ~100 iterations (approximately 1 kilobyte of random data) it starts to generate the same values, repeating each 1, 2 or 4:
```
[143, 209, 221, 216, 248, 42, 173, 182]
[196, 71, 126, 101, 236, 239, 50, 206]
[143, 209, 221, 216, 248, 42, 173, 182]
[196, 71, 126, 101, 236, 239, 50, 206]
[197, 244, 234, 254, 111, 244, 53, 113]
[74, 199, 130, 9, 246, 198, 169, 107]
[197, 244, 234, 254, 111, 244, 53, 113]
[74, 199, 130, 9, 246, 198, 169, 107]
[169, 224, 50, 100, 60, 228, 4, 26]
[60, 29, 188, 159, 8, 15, 198, 36]
[169, 224, 50, 100, 60, 228, 4, 26]
[60, 29, 188, 159, 8, 15, 198, 36]
[169, 224, 50, 100, 60, 228, 4, 26]
[60, 29, 188, 159, 8, 15, 198, 36]
[169, 224, 50, 100, 60, 228, 4, 26]
```
I'm not sure it's a problem with rust / rng implementation or something like that.
But `qemu` works as expected. Any hints for fixing that?
|
1.0
|
Weird RNG behavior (running rust binary) - It seems that I have a weird random generator issue:
```
use rand::Rng;
use rand::SeedableRng;
use rand_chacha::ChaChaRng;
pub fn main() {
let mut rng = ChaChaRng::seed_from_u64(0x42);
for _ in 1..1000 {
let x: [u8; 8] = rng.gen();
println!("{:?}", x);
}
}
```
After ~100 iterations (approximately 1 kilobyte of random data) it starts to generate the same values, repeating each 1, 2 or 4:
```
[143, 209, 221, 216, 248, 42, 173, 182]
[196, 71, 126, 101, 236, 239, 50, 206]
[143, 209, 221, 216, 248, 42, 173, 182]
[196, 71, 126, 101, 236, 239, 50, 206]
[197, 244, 234, 254, 111, 244, 53, 113]
[74, 199, 130, 9, 246, 198, 169, 107]
[197, 244, 234, 254, 111, 244, 53, 113]
[74, 199, 130, 9, 246, 198, 169, 107]
[169, 224, 50, 100, 60, 228, 4, 26]
[60, 29, 188, 159, 8, 15, 198, 36]
[169, 224, 50, 100, 60, 228, 4, 26]
[60, 29, 188, 159, 8, 15, 198, 36]
[169, 224, 50, 100, 60, 228, 4, 26]
[60, 29, 188, 159, 8, 15, 198, 36]
[169, 224, 50, 100, 60, 228, 4, 26]
```
I'm not sure it's a problem with rust / rng implementation or something like that.
But `qemu` works as expected. Any hints for fixing that?
|
process
|
weird rng behavior running rust binary it seems that i have a weird random generator issue use rand rng use rand seedablerng use rand chacha chacharng pub fn main let mut rng chacharng seed from for in let x rng gen println x after iterations approximately kilobyte of random data it starts to generate the same values repeating each or i m not sure it s a problem with rust rng implementation or something like that but qemu works as expected any hints for fixing that
| 1
|
96,913
| 3,975,288,010
|
IssuesEvent
|
2016-05-05 03:50:32
|
Valhalla-Gaming/Tracker
|
https://api.github.com/repos/Valhalla-Gaming/Tracker
|
closed
|
I shoot magic into the Darkness
|
Priority-Normal Type-Quest
|
Quest: http://www.wowhead.com/quest=9700/i-shoot-magic-into-the-darkness
Start/End: http://www.wowhead.com/npc=17684/vindicator-boros
You can't verify the location of the portal, explored all around the structure and nothing happened
|
1.0
|
I shoot magic into the Darkness - Quest: http://www.wowhead.com/quest=9700/i-shoot-magic-into-the-darkness
Start/End: http://www.wowhead.com/npc=17684/vindicator-boros
You can't verify the location of the portal, explored all around the structure and nothing happened
|
non_process
|
i shoot magic into the darkness quest start end you can t verify the location of the portal explored all around the structure and nothing happened
| 0
|
10,508
| 13,281,412,965
|
IssuesEvent
|
2020-08-23 17:13:41
|
candango/myfuses
|
https://api.github.com/repos/candango/myfuses
|
closed
|
Move MyFusesContext to process
|
process refactory
|
MyFusesContext was placed on myfuses.util.context. It is clear that context is part of process and should be moved to the right place.
|
1.0
|
Move MyFusesContext to process - MyFusesContext was placed on myfuses.util.context. It is clear that context is part of process and should be moved to the right place.
|
process
|
move myfusescontext to process myfusescontext was placed on myfuses util context it is clear that context is part of process and should be moved to the right place
| 1
|
830,965
| 32,032,911,206
|
IssuesEvent
|
2023-09-22 13:30:51
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
STM32U5 ADC: Selected channels are not reset between reads
|
bug priority: low area: ADC platform: STM32
|
**Describe the bug**
Performing an ADC read on a channel will leave that channel activated in the channel select register (e.g. ADC4->CHSEL). This will cause issues when attempting to read from a new channel, as the hardware will attempt to read from both.
**To Reproduce**
This can be seen using the adc driver demo. I modified the `nucleo_u575zi_q.overlay` to use ADC4 and include another channel:
```
/ {
zephyr,user {
/* adjust channel number according to pinmux in board.dts */
io-channels = <&adc4 2>, <&adc4 3>;
};
};
&adc4 {
#address-cells = <1>;
#size-cells = <0>;
channel@2 {
reg = <2>;
zephyr,gain = "ADC_GAIN_1";
zephyr,reference = "ADC_REF_INTERNAL";
// Note the acquisiton time is set long to help demonstrate ISR firing multiple times for single request.
zephyr,acquisition-time = <ADC_ACQ_TIME_MAX>;
zephyr,resolution = <12>;
};
channel@3 {
reg = <3>;
zephyr,gain = "ADC_GAIN_1";
zephyr,reference = "ADC_REF_INTERNAL";
zephyr,acquisition-time = <ADC_ACQ_TIME_MAX>;
zephyr,resolution = <12>;
};
};
```
Then build the project with:
`west build --pristine -b nucleo_u575zi_q zephyr/samples/drivers/adc`
Running the project youll see the following:
1. first call to `adc_read` reading Channel 2:
* ADC4->CHSEL register == 0x04 (correct)
* `adc_stm32_isr` is hit once for the single conversion.
2. second call to `adc_read` reading Channel 8:
* ADC4->CHSEL register == 0x0c (incorrect, should be 0x08)
* `adc_stm32_isr`is hit twice, once for each channel. The destination buffer will overflow here from second read.
**Expected behavior**
A new read should not be impacted by any previous reads. Channels should be cleared before setting up the new read.
**Workaround:**
Reset the selected channels at the start of the `start_read` func in `adc_stm32.c`:
```c
static int start_read(const struct device *dev,
const struct adc_sequence *sequence)
{
...
#if defined(CONFIG_SOC_SERIES_STM32U5X)
// Note I've only tested with STM32U5 ADC4
if (adc == ADC4) {
LL_ADC_REG_SetSequencerChannels(adc, 0);
}
#endif
...
}
```
**Environment:**
- OS: MacOS
- Toolchain: Zephyr SDK,
- Zephyr v3.4
**Additional info:**
This issue might be present on other STM32 families as well. I dont any other family of boards to test, however a quick look over the `start_read` func doesnt seem to have any calls to reset the channels, and only calls to add new channels.
|
1.0
|
STM32U5 ADC: Selected channels are not reset between reads - **Describe the bug**
Performing an ADC read on a channel will leave that channel activated in the channel select register (e.g. ADC4->CHSEL). This will cause issues when attempting to read from a new channel, as the hardware will attempt to read from both.
**To Reproduce**
This can be seen using the adc driver demo. I modified the `nucleo_u575zi_q.overlay` to use ADC4 and include another channel:
```
/ {
zephyr,user {
/* adjust channel number according to pinmux in board.dts */
io-channels = <&adc4 2>, <&adc4 3>;
};
};
&adc4 {
#address-cells = <1>;
#size-cells = <0>;
channel@2 {
reg = <2>;
zephyr,gain = "ADC_GAIN_1";
zephyr,reference = "ADC_REF_INTERNAL";
// Note the acquisiton time is set long to help demonstrate ISR firing multiple times for single request.
zephyr,acquisition-time = <ADC_ACQ_TIME_MAX>;
zephyr,resolution = <12>;
};
channel@3 {
reg = <3>;
zephyr,gain = "ADC_GAIN_1";
zephyr,reference = "ADC_REF_INTERNAL";
zephyr,acquisition-time = <ADC_ACQ_TIME_MAX>;
zephyr,resolution = <12>;
};
};
```
Then build the project with:
`west build --pristine -b nucleo_u575zi_q zephyr/samples/drivers/adc`
Running the project youll see the following:
1. first call to `adc_read` reading Channel 2:
* ADC4->CHSEL register == 0x04 (correct)
* `adc_stm32_isr` is hit once for the single conversion.
2. second call to `adc_read` reading Channel 8:
* ADC4->CHSEL register == 0x0c (incorrect, should be 0x08)
* `adc_stm32_isr`is hit twice, once for each channel. The destination buffer will overflow here from second read.
**Expected behavior**
A new read should not be impacted by any previous reads. Channels should be cleared before setting up the new read.
**Workaround:**
Reset the selected channels at the start of the `start_read` func in `adc_stm32.c`:
```c
static int start_read(const struct device *dev,
const struct adc_sequence *sequence)
{
...
#if defined(CONFIG_SOC_SERIES_STM32U5X)
// Note I've only tested with STM32U5 ADC4
if (adc == ADC4) {
LL_ADC_REG_SetSequencerChannels(adc, 0);
}
#endif
...
}
```
**Environment:**
- OS: MacOS
- Toolchain: Zephyr SDK,
- Zephyr v3.4
**Additional info:**
This issue might be present on other STM32 families as well. I dont any other family of boards to test, however a quick look over the `start_read` func doesnt seem to have any calls to reset the channels, and only calls to add new channels.
|
non_process
|
adc selected channels are not reset between reads describe the bug performing an adc read on a channel will leave that channel activated in the channel select register e g chsel this will cause issues when attempting to read from a new channel as the hardware will attempt to read from both to reproduce this can be seen using the adc driver demo i modified the nucleo q overlay to use and include another channel zephyr user adjust channel number according to pinmux in board dts io channels address cells size cells channel reg zephyr gain adc gain zephyr reference adc ref internal note the acquisiton time is set long to help demonstrate isr firing multiple times for single request zephyr acquisition time zephyr resolution channel reg zephyr gain adc gain zephyr reference adc ref internal zephyr acquisition time zephyr resolution then build the project with west build pristine b nucleo q zephyr samples drivers adc running the project youll see the following first call to adc read reading channel chsel register correct adc isr is hit once for the single conversion second call to adc read reading channel chsel register incorrect should be adc isr is hit twice once for each channel the destination buffer will overflow here from second read expected behavior a new read should not be impacted by any previous reads channels should be cleared before setting up the new read workaround reset the selected channels at the start of the start read func in adc c c static int start read const struct device dev const struct adc sequence sequence if defined config soc series note i ve only tested with if adc ll adc reg setsequencerchannels adc endif environment os macos toolchain zephyr sdk zephyr additional info this issue might be present on other families as well i dont any other family of boards to test however a quick look over the start read func doesnt seem to have any calls to reset the channels and only calls to add new channels
| 0
|
148
| 2,577,616,680
|
IssuesEvent
|
2015-02-12 18:05:07
|
cfpb/hmda-viz-prototype
|
https://api.github.com/repos/cfpb/hmda-viz-prototype
|
opened
|
Database to print out State and MSA name
|
Processing
|
Set up a database to write a query that would print out the friendly state name and friendly msa names.
|
1.0
|
Database to print out State and MSA name - Set up a database to write a query that would print out the friendly state name and friendly msa names.
|
process
|
database to print out state and msa name set up a database to write a query that would print out the friendly state name and friendly msa names
| 1
|
384,079
| 26,574,525,908
|
IssuesEvent
|
2023-01-21 16:34:02
|
dzully/google-places-autocomplete
|
https://api.github.com/repos/dzully/google-places-autocomplete
|
closed
|
Missing unit test for reducer and HOC
|
documentation todo
|
The current codebase for the project is missing unit tests for the reducer and higher-order component (HOC). This poses a risk for future development, as changes to these areas may break existing functionality without proper testing. It is important to have unit tests in place to ensure that these crucial parts of the application are functioning correctly.
The missing unit tests should cover all possible actions and state changes within the reducer, and should test the HOC's ability to properly pass down the map data to the wrapped component.
I propose that we implement these missing tests as soon as possible to ensure the stability and maintainability of the codebase. This will also make it easier for other developers to understand the codebase and make changes with confidence.
|
1.0
|
Missing unit test for reducer and HOC - The current codebase for the project is missing unit tests for the reducer and higher-order component (HOC). This poses a risk for future development, as changes to these areas may break existing functionality without proper testing. It is important to have unit tests in place to ensure that these crucial parts of the application are functioning correctly.
The missing unit tests should cover all possible actions and state changes within the reducer, and should test the HOC's ability to properly pass down the map data to the wrapped component.
I propose that we implement these missing tests as soon as possible to ensure the stability and maintainability of the codebase. This will also make it easier for other developers to understand the codebase and make changes with confidence.
|
non_process
|
missing unit test for reducer and hoc the current codebase for the project is missing unit tests for the reducer and higher order component hoc this poses a risk for future development as changes to these areas may break existing functionality without proper testing it is important to have unit tests in place to ensure that these crucial parts of the application are functioning correctly the missing unit tests should cover all possible actions and state changes within the reducer and should test the hoc s ability to properly pass down the map data to the wrapped component i propose that we implement these missing tests as soon as possible to ensure the stability and maintainability of the codebase this will also make it easier for other developers to understand the codebase and make changes with confidence
| 0
|
258,852
| 22,353,460,999
|
IssuesEvent
|
2022-06-15 13:54:27
|
MherZaqaryan/BedWars1058-Cosmetics-Issues
|
https://api.github.com/repos/MherZaqaryan/BedWars1058-Cosmetics-Issues
|
closed
|
TPS drop during Cold Snap Dance
|
bug test needed
|
**Describe the bug**
When you use Cold Snap dance during the dance server TPS drop fast
**How to reproduce**
Steps to reproduce the behavior:
1. Select Cold Snap dance
2. Win a game
3. during the dance the server TPS drop
**Screenshots**
Yt Video: https://youtu.be/53x1IFsK4Lo
**Server Information (please complete the following information):**
- OS: [Linux]
- Server Software: [Paper]
- Software Version: [1.8.8]
- BedWars1058 Version: [22.3.4]
- Cosmetics Addon: [1.0.0]
**Additional context**
Server's informations
- Ryzen 9 3900x (6 core)
- 16gb ram DDR4 (3200MHz) -> 14 dedicated to server
|
1.0
|
TPS drop during Cold Snap Dance - **Describe the bug**
When you use Cold Snap dance during the dance server TPS drop fast
**How to reproduce**
Steps to reproduce the behavior:
1. Select Cold Snap dance
2. Win a game
3. during the dance the server TPS drop
**Screenshots**
Yt Video: https://youtu.be/53x1IFsK4Lo
**Server Information (please complete the following information):**
- OS: [Linux]
- Server Software: [Paper]
- Software Version: [1.8.8]
- BedWars1058 Version: [22.3.4]
- Cosmetics Addon: [1.0.0]
**Additional context**
Server's informations
- Ryzen 9 3900x (6 core)
- 16gb ram DDR4 (3200MHz) -> 14 dedicated to server
|
non_process
|
tps drop during cold snap dance describe the bug when you use cold snap dance during the dance server tps drop fast how to reproduce steps to reproduce the behavior select cold snap dance win a game during the dance the server tps drop screenshots yt video server information please complete the following information os server software software version version cosmetics addon additional context server s informations ryzen core ram dedicated to server
| 0
|
343,758
| 30,687,454,750
|
IssuesEvent
|
2023-07-26 13:14:54
|
sarahsporck/Migration-Test-Repository
|
https://api.github.com/repos/sarahsporck/Migration-Test-Repository
|
closed
|
IGAPP-396: Determining gps location fails
|
bug ready Native testbirds
|
# Max Ammann - 23.11.2020, 11:56:38
Maybe we need to use the GPlay services if they are available. This table shows how many seconds the loading took:
690659 2s
690667 0s
690819 90s
690967 2s
690974 0s
691072 20s
693708 999s
693709 120s
Research whether this can be improved by using other geolocation libraries from google.
#### Expected Results:
Since GPS is activated, I expect to see a list of "Nearby places" ("Orte in der Nähe"). My phone knows my position, as can be seen on the Google Maps screenshot.
#### Actual Results:
My position cannot be set; after a while a message is displayed: "Standortbestimmung fehlgeschlagen".
#### Steps to Reproduce:
* I go through the step-by-step onboarding and end up on the "Location search".
https://nest.testbirds.com/client/bug/show/283545
**Environment**: -
**Linked issues:**
- Relates: IGAPP-110
|
1.0
|
IGAPP-396: Determining gps location fails -
# Max Ammann - 23.11.2020, 11:56:38
Maybe we need to use the GPlay services if they are available. This table shows how many seconds the loading took:
690659 2s
690667 0s
690819 90s
690967 2s
690974 0s
691072 20s
693708 999s
693709 120s
Research whether this can be improved by using other geolocation libraries from google.
#### Expected Results:
Since GPS is activated, I expect to see a list of "Nearby places" ("Orte in der Nähe"). My phone knows my position, as can be seen on the Google Maps screenshot.
#### Actual Results:
My position cannot be set; after a while a message is displayed: "Standortbestimmung fehlgeschlagen".
#### Steps to Reproduce:
* I go through the step-by-step onboarding and end up on the "Location search".
https://nest.testbirds.com/client/bug/show/283545
**Environment**: -
**Linked issues:**
- Relates: IGAPP-110
|
non_process
|
igapp determining gps location fails max ammann maybe we need to use the gplay services if they are available this table shows how many seconds the loading took research whether this can be improved by using other geolocation libraries from google expected results since gps is activated i expect to see a list of nearby places orte in der nähe my phone knows my position as can be seen on the google maps screenshot actual results my position cannot be set after a while a message is displayed standortbestimmung fehlgeschlagen steps to reproduce i go through the step by step onboarding and end up on the location search environment linked issues relates igapp
| 0
|
2,751
| 2,607,938,429
|
IssuesEvent
|
2015-02-26 00:29:50
|
chrsmithdemos/minify
|
https://api.github.com/repos/chrsmithdemos/minify
|
opened
|
Rewrite and version problem
|
auto-migrated Priority-Medium Type-Defect
|
```
modRewriteWorks: false
http://examplem.com/min/?g=keyName&1111
modRewriteWorks: true
http://examplem.com/min/g=keyName&1111 the second url starts with & and ? is
missing. For me the browsers doesn't take the version number it should be
http://examplem.com/min/g=keyName?1111 because it was rewriten
```
-----
Original issue reported on code.google.com by `ciprian.mihalache@originals.ro` on 23 Apr 2014 at 6:10
|
1.0
|
Rewrite and version problem - ```
modRewriteWorks: false
http://examplem.com/min/?g=keyName&1111
modRewriteWorks: true
http://examplem.com/min/g=keyName&1111 the second url starts with & and ? is
missing. For me the browsers doesn't take the version number it should be
http://examplem.com/min/g=keyName?1111 because it was rewriten
```
-----
Original issue reported on code.google.com by `ciprian.mihalache@originals.ro` on 23 Apr 2014 at 6:10
|
non_process
|
rewrite and version problem modrewriteworks false modrewriteworks true the second url starts with and is missing for me the browsers doesn t take the version number it should be because it was rewriten original issue reported on code google com by ciprian mihalache originals ro on apr at
| 0
|
107,066
| 9,201,475,878
|
IssuesEvent
|
2019-03-07 19:44:03
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: backupTPCC failed
|
C-test-failure O-roachtest O-robot
|
SHA: https://github.com/cockroachdb/cockroach/commits/b80d241d5693d6ba2bde01ae9167b49b04b47226
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=backupTPCC PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1133193&tab=buildLog
```
The test failed on master:
cluster.go:1585,backup.go:127,test.go:1206: pq: AS OF SYSTEM TIME: cannot specify timestamp in the future
```
|
2.0
|
roachtest: backupTPCC failed - SHA: https://github.com/cockroachdb/cockroach/commits/b80d241d5693d6ba2bde01ae9167b49b04b47226
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=backupTPCC PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1133193&tab=buildLog
```
The test failed on master:
cluster.go:1585,backup.go:127,test.go:1206: pq: AS OF SYSTEM TIME: cannot specify timestamp in the future
```
|
non_process
|
roachtest backuptpcc failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests backuptpcc pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on master cluster go backup go test go pq as of system time cannot specify timestamp in the future
| 0
|
105,158
| 11,433,998,485
|
IssuesEvent
|
2020-02-04 16:36:47
|
LLK/scratch-gui
|
https://api.github.com/repos/LLK/scratch-gui
|
opened
|
Project state reducer and state machine are confusing, need documentation
|
documentation ergonomics
|
### Expected Behavior
Programmers understand project state reducer and state machine
### Actual Behavior
Project state reducer and state machine cause developers to wish they went into the restaurant business, cousin Sheila seems to have a peaceful life as a baker
### Steps to Reproduce
Try to understand the project state reducer and state machine
|
1.0
|
Project state reducer and state machine are confusing, need documentation - ### Expected Behavior
Programmers understand project state reducer and state machine
### Actual Behavior
Project state reducer and state machine cause developers to wish they went into the restaurant business, cousin Sheila seems to have a peaceful life as a baker
### Steps to Reproduce
Try to understand the project state reducer and state machine
|
non_process
|
project state reducer and state machine are confusing need documentation expected behavior programmers understand project state reducer and state machine actual behavior project state reducer and state machine cause developers to wish they went into the restaurant business cousin sheila seems to have a peaceful life as a baker steps to reproduce try to understand the project state reducer and state machine
| 0
|
528,533
| 15,369,120,119
|
IssuesEvent
|
2021-03-02 06:51:37
|
Xmetalfanx/website
|
https://api.github.com/repos/Xmetalfanx/website
|
closed
|
Mobile Footer issue
|
Priority bug
|
things dont look right on smaller layouts when it comes to the footer
i think it would look ok if i just (on smaller screens only) removed the "site-made" div
|
1.0
|
Mobile Footer issue - things dont look right on smaller layouts when it comes to the footer
i think it would look ok if i just (on smaller screens only) removed the "site-made" div
|
non_process
|
mobile footer issue things dont look right on smaller layouts when it comes to the footer i think it would look ok if i just on smaller screens only removed the site made div
| 0
|
38,728
| 5,002,543,014
|
IssuesEvent
|
2016-12-11 13:08:32
|
taviandir/FoodPlannerApp
|
https://api.github.com/repos/taviandir/FoodPlannerApp
|
closed
|
Ingredients
|
design decision feature recipes
|
Should have persisted list of ingredients that can be added to. Not sure how to manage quantities just yet. Default quantifier? Forced quantifier?
Ingredient types? (Not really sure if this adds anything of value (for v1.0))
- Vegetable
- Fruit
- Meat
- Dairy
- ...etc
Should also feature a list of possible quantifiers, such as:
- grams
- (quantity/blank)
- dl
- ml
- teaspoons
- ...etc
The thought behind this is so that the ingredients can be bundled and added to each other from multiple recipes in order to create a shopping list.
For example,
10 - ml - worcester sauce
3 - (blank) - carrots
500 - g - minced meat
Pattern is obviously "AMOUNT QUANTIFIER NAME". Problem is it wont necessarily be the same kind of quantifier for one single ingredient.
MODIFIER: Freetext wanted at the end of the ingredient? For example "(for garnish)", "(optional)", "(mashed)".
|
1.0
|
Ingredients - Should have persisted list of ingredients that can be added to. Not sure how to manage quantities just yet. Default quantifier? Forced quantifier?
Ingredient types? (Not really sure if this adds anything of value (for v1.0))
- Vegetable
- Fruit
- Meat
- Dairy
- ...etc
Should also feature a list of possible quantifiers, such as:
- grams
- (quantity/blank)
- dl
- ml
- teaspoons
- ...etc
The thought behind this is so that the ingredients can be bundled and added to each other from multiple recipes in order to create a shopping list.
For example,
10 - ml - worcester sauce
3 - (blank) - carrots
500 - g - minced meat
Pattern is obviously "AMOUNT QUANTIFIER NAME". Problem is it wont necessarily be the same kind of quantifier for one single ingredient.
MODIFIER: Freetext wanted at the end of the ingredient? For example "(for garnish)", "(optional)", "(mashed)".
|
non_process
|
ingredients should have persisted list of ingredients that can be added to not sure how to manage quantities just yet default quantifier forced quantifier ingredient types not really sure if this adds anything of value for vegetable fruit meat dairy etc should also feature a list of possible quantifiers such as grams quantity blank dl ml teaspoons etc the thought behind this is so that the ingredients can be bundled and added to each other from multiple recipes in order to create a shopping list for example ml worcester sauce blank carrots g minced meat pattern is obviously amount quantifier name problem is it wont necessarily be the same kind of quantifier for one single ingredient modifier freetext wanted at the end of the ingredient for example for garnish optional mashed
| 0
|
153,191
| 24,086,176,943
|
IssuesEvent
|
2022-09-19 11:08:58
|
TheSuperHackers/GeneralsGamePatch
|
https://api.github.com/repos/TheSuperHackers/GeneralsGamePatch
|
closed
|
Demo GLA Bomb Truck deals same damage as regular Bomb Truck
|
Design Controversial Minor
|
Demo GLA Bomb Truck deals same damage as regular Bomb Truck. Is this an unexpected problem?
```
;------------------------------------------------------------------------------
;The basic explosive damage inflicted to nearby units when detonated.
;------------------------------------------------------------------------------
Weapon BombTruckDefaultBombDamage
PrimaryDamage = 1000.0
PrimaryDamageRadius = 40.0
SecondaryDamage = 100.0
SecondaryDamageRadius = 65.0
DamageType = EXPLOSION
DeathType = EXPLODED
WeaponSpeed = 99999.0
ProjectileObject = NONE
DamageDealtAtSelfPosition = Yes ; this is a suicide bomber... remember?
RadiusDamageAffects = SELF ALLIES ENEMIES NEUTRALS
DelayBetweenShots = 0
ClipSize = 1
ClipReloadTime = 0
AutoReloadsClip = No
;**** DON'T USE ANY EFFECTS!!! ****
End
;------------------------------------------------------------------------------
;A larger explosive damage inflicted immediately to nearby units if the
;bomb-truck has the high-explosive upgrade.
;------------------------------------------------------------------------------
Weapon BombTruckHighExplosionBombDamage
PrimaryDamage = 2000.0
PrimaryDamageRadius = 50.0
SecondaryDamage = 200.0
SecondaryDamageRadius = 85.0
DamageType = EXPLOSION
DeathType = EXPLODED
ProjectileObject = NONE
RadiusDamageAffects = SELF ALLIES ENEMIES NEUTRALS
DelayBetweenShots = 0 ; time between shots, msec
ClipSize = 1 ; how many shots in a Clip (0 == infinite)
ClipReloadTime = 0 ; how long to reload a Clip, msec
AutoReloadsClip = No
AntiGround = Yes
DamageDealtAtSelfPosition = Yes
WeaponBonus = PLAYER_UPGRADE DAMAGE 125%
;**** DON'T USE ANY EFFECTS!!! ****
End
;------------------------------------------------------------------------------
;DEMO GENERAL VARIATION
;The basic explosive damage inflicted to nearby units when detonated.
;------------------------------------------------------------------------------
Weapon Demo_BombTruckDefaultBombDamage
PrimaryDamage = 1000.0
PrimaryDamageRadius = 40.0
SecondaryDamage = 100.0
SecondaryDamageRadius = 65.0
DamageType = EXPLOSION
DeathType = EXPLODED
WeaponSpeed = 99999.0
ProjectileObject = NONE
DamageDealtAtSelfPosition = Yes ; this is a suicide bomber... remember?
RadiusDamageAffects = SELF ALLIES ENEMIES NEUTRALS
DelayBetweenShots = 0
ClipSize = 1
ClipReloadTime = 0
AutoReloadsClip = No
;**** DON'T USE ANY EFFECTS!!! ****
End
;------------------------------------------------------------------------------
;DEMO GENERAL VARIATION
;A larger explosive damage inflicted immediately to nearby units if the
;bomb-truck has the high-explosive upgrade.
;------------------------------------------------------------------------------
Weapon Demo_BombTruckHighExplosionBombDamage
PrimaryDamage = 2000.0
PrimaryDamageRadius = 50.0
SecondaryDamage = 200.0
SecondaryDamageRadius = 85.0
DamageType = EXPLOSION
DeathType = EXPLODED
ProjectileObject = NONE
RadiusDamageAffects = SELF ALLIES ENEMIES NEUTRALS
DelayBetweenShots = 0 ; time between shots, msec
ClipSize = 1 ; how many shots in a Clip (0 == infinite)
ClipReloadTime = 0 ; how long to reload a Clip, msec
AutoReloadsClip = No
AntiGround = Yes
DamageDealtAtSelfPosition = Yes
WeaponBonus = PLAYER_UPGRADE DAMAGE 125%
;**** DON'T USE ANY EFFECTS!!! ****
End
```
|
1.0
|
Demo GLA Bomb Truck deals same damage as regular Bomb Truck - Demo GLA Bomb Truck deals same damage as regular Bomb Truck. Is this an unexpected problem?
```
;------------------------------------------------------------------------------
;The basic explosive damage inflicted to nearby units when detonated.
;------------------------------------------------------------------------------
Weapon BombTruckDefaultBombDamage
PrimaryDamage = 1000.0
PrimaryDamageRadius = 40.0
SecondaryDamage = 100.0
SecondaryDamageRadius = 65.0
DamageType = EXPLOSION
DeathType = EXPLODED
WeaponSpeed = 99999.0
ProjectileObject = NONE
DamageDealtAtSelfPosition = Yes ; this is a suicide bomber... remember?
RadiusDamageAffects = SELF ALLIES ENEMIES NEUTRALS
DelayBetweenShots = 0
ClipSize = 1
ClipReloadTime = 0
AutoReloadsClip = No
;**** DON'T USE ANY EFFECTS!!! ****
End
;------------------------------------------------------------------------------
;A larger explosive damage inflicted immediately to nearby units if the
;bomb-truck has the high-explosive upgrade.
;------------------------------------------------------------------------------
Weapon BombTruckHighExplosionBombDamage
PrimaryDamage = 2000.0
PrimaryDamageRadius = 50.0
SecondaryDamage = 200.0
SecondaryDamageRadius = 85.0
DamageType = EXPLOSION
DeathType = EXPLODED
ProjectileObject = NONE
RadiusDamageAffects = SELF ALLIES ENEMIES NEUTRALS
DelayBetweenShots = 0 ; time between shots, msec
ClipSize = 1 ; how many shots in a Clip (0 == infinite)
ClipReloadTime = 0 ; how long to reload a Clip, msec
AutoReloadsClip = No
AntiGround = Yes
DamageDealtAtSelfPosition = Yes
WeaponBonus = PLAYER_UPGRADE DAMAGE 125%
;**** DON'T USE ANY EFFECTS!!! ****
End
;------------------------------------------------------------------------------
;DEMO GENERAL VARIATION
;The basic explosive damage inflicted to nearby units when detonated.
;------------------------------------------------------------------------------
Weapon Demo_BombTruckDefaultBombDamage
PrimaryDamage = 1000.0
PrimaryDamageRadius = 40.0
SecondaryDamage = 100.0
SecondaryDamageRadius = 65.0
DamageType = EXPLOSION
DeathType = EXPLODED
WeaponSpeed = 99999.0
ProjectileObject = NONE
DamageDealtAtSelfPosition = Yes ; this is a suicide bomber... remember?
RadiusDamageAffects = SELF ALLIES ENEMIES NEUTRALS
DelayBetweenShots = 0
ClipSize = 1
ClipReloadTime = 0
AutoReloadsClip = No
;**** DON'T USE ANY EFFECTS!!! ****
End
;------------------------------------------------------------------------------
;DEMO GENERAL VARIATION
;A larger explosive damage inflicted immediately to nearby units if the
;bomb-truck has the high-explosive upgrade.
;------------------------------------------------------------------------------
Weapon Demo_BombTruckHighExplosionBombDamage
PrimaryDamage = 2000.0
PrimaryDamageRadius = 50.0
SecondaryDamage = 200.0
SecondaryDamageRadius = 85.0
DamageType = EXPLOSION
DeathType = EXPLODED
ProjectileObject = NONE
RadiusDamageAffects = SELF ALLIES ENEMIES NEUTRALS
DelayBetweenShots = 0 ; time between shots, msec
ClipSize = 1 ; how many shots in a Clip (0 == infinite)
ClipReloadTime = 0 ; how long to reload a Clip, msec
AutoReloadsClip = No
AntiGround = Yes
DamageDealtAtSelfPosition = Yes
WeaponBonus = PLAYER_UPGRADE DAMAGE 125%
;**** DON'T USE ANY EFFECTS!!! ****
End
```
|
non_process
|
demo gla bomb truck deals same damage as regular bomb truck demo gla bomb truck deals same damage as regular bomb truck is this an unexpected problem the basic explosive damage inflicted to nearby units when detonated weapon bombtruckdefaultbombdamage primarydamage primarydamageradius secondarydamage secondarydamageradius damagetype explosion deathtype exploded weaponspeed projectileobject none damagedealtatselfposition yes this is a suicide bomber remember radiusdamageaffects self allies enemies neutrals delaybetweenshots clipsize clipreloadtime autoreloadsclip no don t use any effects end a larger explosive damage inflicted immediately to nearby units if the bomb truck has the high explosive upgrade weapon bombtruckhighexplosionbombdamage primarydamage primarydamageradius secondarydamage secondarydamageradius damagetype explosion deathtype exploded projectileobject none radiusdamageaffects self allies enemies neutrals delaybetweenshots time between shots msec clipsize how many shots in a clip infinite clipreloadtime how long to reload a clip msec autoreloadsclip no antiground yes damagedealtatselfposition yes weaponbonus player upgrade damage don t use any effects end demo general variation the basic explosive damage inflicted to nearby units when detonated weapon demo bombtruckdefaultbombdamage primarydamage primarydamageradius secondarydamage secondarydamageradius damagetype explosion deathtype exploded weaponspeed projectileobject none damagedealtatselfposition yes this is a suicide bomber remember radiusdamageaffects self allies enemies neutrals delaybetweenshots clipsize clipreloadtime autoreloadsclip no don t use any effects end demo general variation a larger explosive damage inflicted immediately to nearby units if the bomb truck has the high explosive upgrade weapon demo bombtruckhighexplosionbombdamage primarydamage primarydamageradius secondarydamage secondarydamageradius damagetype explosion deathtype exploded projectileobject none radiusdamageaffects self allies enemies neutrals delaybetweenshots time between shots msec clipsize how many shots in a clip infinite clipreloadtime how long to reload a clip msec autoreloadsclip no antiground yes damagedealtatselfposition yes weaponbonus player upgrade damage don t use any effects end
| 0
|
54,115
| 6,796,182,333
|
IssuesEvent
|
2017-11-01 18:07:43
|
opencollective/opencollective
|
https://api.github.com/repos/opencollective/opencollective
|
opened
|
revise how backers avatars are shown
|
design
|
Avatars from individuals making large contributions are lost in the ui. They are smaller than companies that gave much less. We need to find a way of showing this better. We are only incentivizing corporate sponsors like this.
cc / @cuiki @xdamman
Thank you for taking the time to report an issue 🙏
The easier it is for us to reproduce it, the faster we can solve it.
So please try to be as complete as possible when filing your issue.
***
URL:
Logged in as: (just provide your username or the url of your profile)
Error message if any:
Expected result:
Browser:
Bonus point if you can add a screenshot :-)
Thank you and have a wonderful day/evening!
(and sorry for inconvenience! We'll do our best to solve this ASAP)
|
1.0
|
revise how backers avatars are shown - Avatars from individuals making large contributions are lost in the ui. They are smaller than companies that gave much less. We need to find a way of showing this better. We are only incentivizing corporate sponsors like this.
cc / @cuiki @xdamman
Thank you for taking the time to report an issue 🙏
The easier it is for us to reproduce it, the faster we can solve it.
So please try to be as complete as possible when filing your issue.
***
URL:
Logged in as: (just provide your username or the url of your profile)
Error message if any:
Expected result:
Browser:
Bonus point if you can add a screenshot :-)
Thank you and have a wonderful day/evening!
(and sorry for inconvenience! We'll do our best to solve this ASAP)
|
non_process
|
revise how backers avatars are shown avatars from individuals making large contributions are lost in the ui they are smaller than companies that gave much less we need to find a way of showing this better we are only incentivizing corporate sponsors like this cc cuiki xdamman thank you for taking the time to report an issue 🙏 the easier it is for us to reproduce it the faster we can solve it so please try to be as complete as possible when filing your issue url logged in as just provide your username or the url of your profile error message if any expected result browser bonus point if you can add a screenshot thank you and have a wonderful day evening and sorry for inconvenience we ll do our best to solve this asap
| 0
|
11,962
| 14,728,064,850
|
IssuesEvent
|
2021-01-06 09:26:58
|
yuta252/startlens_web_backend
|
https://api.github.com/repos/yuta252/startlens_web_backend
|
closed
|
Exhibitモデルにおける複数画像のアップロード
|
dev process
|
## 概要
CarrierWaveを利用してフロントエンドからbase64で送信したAPIリクエスト(画像を複数含む)に対し、bas64のデコードを行いExhibitモデルに複数の画像を登録する処理を実装する。
## 変更点
---
- [x] Exhibitモデルの作成
- 特にimagesカラムを作成し複数の画像を投稿できるようにする
- [x] Exhibitコントローラーの作成
- [x] レスポンス時のシリアライザーの作成
- [x] 画像を保存するPictureモデルの作成
- [x] Rspecによる画像投稿テスト
## 課題
---
- "CarrierWaveで複数画像をアップロードする"の通りに、imagesをJson型のカラムとしuploaderをマウントすればcarrierwaveで複数画像を登録することが可能である。
- ただし、今回はフロントからbase64エンコードされた画像データからデコード処理を行うため単純に上記の方法を採用してもworkしない
- そこで、"railsで複数画像を一度にアップロードする方法"の記事を参考に、ExhibitにPictureオブジェクトを1 to manyで対応させ、Pictureオブジェクト自体に唯一の画像を持たせる方法を検討。
## 参照
---
- [CarrierWaveで複数画像をアップロードする](https://qiita.com/tanutanu/items/47f8a229ef52cae3c251)
- [railsで複数画像を一度にアップロードする方法](https://qiita.com/maru_u/items/238027a1d3580e04cb12)
## 備考
---
|
1.0
|
Exhibitモデルにおける複数画像のアップロード - ## 概要
CarrierWaveを利用してフロントエンドからbase64で送信したAPIリクエスト(画像を複数含む)に対し、bas64のデコードを行いExhibitモデルに複数の画像を登録する処理を実装する。
## 変更点
---
- [x] Exhibitモデルの作成
- 特にimagesカラムを作成し複数の画像を投稿できるようにする
- [x] Exhibitコントローラーの作成
- [x] レスポンス時のシリアライザーの作成
- [x] 画像を保存するPictureモデルの作成
- [x] Rspecによる画像投稿テスト
## 課題
---
- "CarrierWaveで複数画像をアップロードする"の通りに、imagesをJson型のカラムとしuploaderをマウントすればcarrierwaveで複数画像を登録することが可能である。
- ただし、今回はフロントからbase64エンコードされた画像データからデコード処理を行うため単純に上記の方法を採用してもworkしない
- そこで、"railsで複数画像を一度にアップロードする方法"の記事を参考に、ExhibitにPictureオブジェクトを1 to manyで対応させ、Pictureオブジェクト自体に唯一の画像を持たせる方法を検討。
## 参照
---
- [CarrierWaveで複数画像をアップロードする](https://qiita.com/tanutanu/items/47f8a229ef52cae3c251)
- [railsで複数画像を一度にアップロードする方法](https://qiita.com/maru_u/items/238027a1d3580e04cb12)
## 備考
---
|
process
|
exhibitモデルにおける複数画像のアップロード 概要 (画像を複数含む)に対し、 。 変更点 exhibitモデルの作成 特にimagesカラムを作成し複数の画像を投稿できるようにする exhibitコントローラーの作成 レスポンス時のシリアライザーの作成 画像を保存するpictureモデルの作成 rspecによる画像投稿テスト 課題 carrierwaveで複数画像をアップロードする の通りに、imagesをjson型のカラムとしuploaderをマウントすればcarrierwaveで複数画像を登録することが可能である。 ただし、 そこで、 railsで複数画像を一度にアップロードする方法 の記事を参考に、 to manyで対応させ、pictureオブジェクト自体に唯一の画像を持たせる方法を検討。 参照 備考
| 1
|
54,516
| 6,393,636,982
|
IssuesEvent
|
2017-08-04 08:05:35
|
LiskHQ/lisk-nano
|
https://api.github.com/repos/LiskHQ/lisk-nano
|
closed
|
Add unit tests for sign/verify message
|
easy test
|
### Expected behaviour
There should be unit tests for sign/verify message
### Actual behaviour
There are no unit tests for sign/verify message
|
1.0
|
Add unit tests for sign/verify message - ### Expected behaviour
There should be unit tests for sign/verify message
### Actual behaviour
There are no unit tests for sign/verify message
|
non_process
|
add unit tests for sign verify message expected behaviour there should be unit tests for sign verify message actual behaviour there are no unit tests for sign verify message
| 0
|
437,054
| 30,585,480,379
|
IssuesEvent
|
2023-07-21 13:03:33
|
mercedes-benz/sechub-plugin-intellij
|
https://api.github.com/repos/mercedes-benz/sechub-plugin-intellij
|
closed
|
CLA information is not up to date
|
documentation
|
The [CONTRIBUTING](https://github.com/mercedes-benz/sechub-plugin-intellij/blob/master/CONTRIBUTING.md) file currently has a reference to the CLA for Daimler and the email address there is also for Daimler.
|
1.0
|
CLA information is not up to date - The [CONTRIBUTING](https://github.com/mercedes-benz/sechub-plugin-intellij/blob/master/CONTRIBUTING.md) file currently has a reference to the CLA for Daimler and the email address there is also for Daimler.
|
non_process
|
cla information is not up to date the file currently has a reference to the cla for daimler and the email address there is also for daimler
| 0
|
16,009
| 20,188,223,573
|
IssuesEvent
|
2022-02-11 01:19:25
|
savitamittalmsft/WAS-SEC-TEST
|
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
|
opened
|
Build a security containment strategy
|
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Security & Compliance Network Security
|
<a href="https://docs.microsoft.com/azure/architecture/framework/security/design-network-segmentation">Build a security containment strategy</a>
<p><b>Why Consider This?</b></p>
'Assume compromise' is the recommended cybersecurity mindset, and the ability to contain an attacker is vital in protecting information systems. Model an attacker able to achieve a foothold at various points within the workload, and establish controls to mitigate further expansion.
<p><b>Context</b></p>
<p><span>Containment of attack vectors within an environment is critical. However, to be effective in cloud environments, traditional approaches may prove inadequate and security organizations need to evolve their methods. </span></p><p><span>An evolving emerging best practice recommendation is to adopt a Zero Trust strategy based on user, device, and application identities. In contrast to network access controls that are based on elements such as source and destination IP address, protocols, and port numbers, Zero Trust enforces and validates access control at "access time". This avoids the need to play a prediction game for an entire deployment, network, or subnet - only the destination resource needs to provide the necessary access controls.</span></p><ul style="list-style-type:disc"><li value="1" style="text-indent: 0px;"><span>Azure Network Security Groups can be used for basic layer 3 "amp; 4 access controls between Azure Virtual Networks, their subnets, and the Internet.</span></li><li value="2" style="margin-right: 0px;text-indent: 0px;"><span>Azure Web Application Firewall and the Azure Firewall can be used for more advanced network access controls that require application layer support.</span></li><li value="3" style="margin-right: 0px;text-indent: 0px;"><span>Local Admin Password Solution (LAPS) or a third-party Privileged Access Management can set strong local admin passwords and just in time access to them</span></li></ul>
<p><b>Suggested Actions</b></p>
<p><span>Create a risk containment strategy that blends proven approaches including:</span></p><ul style="list-style-type:disc"><li value="1" style="text-indent: 0px;"><span>Existing network security controls and practices</span></li><li value="2" style="margin-right: 0px;text-indent: 0px;"><span>Native security controls available in Azure</span></li><li value="3" style="margin-right: 0px;text-indent: 0px;"><span>Zero trust approaches</span></li></ul>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/azure/architecture/framework/Security/network-security-containment#build-a-security-containment-strategy" target="_blank"><span>Build a security containment strategy</span></a><span /></p>
|
1.0
|
Build a security containment strategy - <a href="https://docs.microsoft.com/azure/architecture/framework/security/design-network-segmentation">Build a security containment strategy</a>
<p><b>Why Consider This?</b></p>
'Assume compromise' is the recommended cybersecurity mindset, and the ability to contain an attacker is vital in protecting information systems. Model an attacker able to achieve a foothold at various points within the workload, and establish controls to mitigate further expansion.
<p><b>Context</b></p>
<p><span>Containment of attack vectors within an environment is critical. However, to be effective in cloud environments, traditional approaches may prove inadequate and security organizations need to evolve their methods. </span></p><p><span>An evolving emerging best practice recommendation is to adopt a Zero Trust strategy based on user, device, and application identities. In contrast to network access controls that are based on elements such as source and destination IP address, protocols, and port numbers, Zero Trust enforces and validates access control at "access time". This avoids the need to play a prediction game for an entire deployment, network, or subnet - only the destination resource needs to provide the necessary access controls.</span></p><ul style="list-style-type:disc"><li value="1" style="text-indent: 0px;"><span>Azure Network Security Groups can be used for basic layer 3 "amp; 4 access controls between Azure Virtual Networks, their subnets, and the Internet.</span></li><li value="2" style="margin-right: 0px;text-indent: 0px;"><span>Azure Web Application Firewall and the Azure Firewall can be used for more advanced network access controls that require application layer support.</span></li><li value="3" style="margin-right: 0px;text-indent: 0px;"><span>Local Admin Password Solution (LAPS) or a third-party Privileged Access Management can set strong local admin passwords and just in time access to them</span></li></ul>
<p><b>Suggested Actions</b></p>
<p><span>Create a risk containment strategy that blends proven approaches including:</span></p><ul style="list-style-type:disc"><li value="1" style="text-indent: 0px;"><span>Existing network security controls and practices</span></li><li value="2" style="margin-right: 0px;text-indent: 0px;"><span>Native security controls available in Azure</span></li><li value="3" style="margin-right: 0px;text-indent: 0px;"><span>Zero trust approaches</span></li></ul>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/azure/architecture/framework/Security/network-security-containment#build-a-security-containment-strategy" target="_blank"><span>Build a security containment strategy</span></a><span /></p>
|
process
|
build a security containment strategy why consider this assume compromise is the recommended cybersecurity mindset and the ability to contain an attacker is vital in protecting information systems model an attacker able to achieve a foothold at various points within the workload and establish controls to mitigate further expansion context containment of attack vectors within an environment is critical however to be effective in cloud environments traditional approaches may prove inadequate and security organizations need to evolve their methods an evolving emerging best practice recommendation is to adopt a zero trust strategy based on user device and application identities in contrast to network access controls that are based on elements such as source and destination ip address protocols and port numbers zero trust enforces and validates access control at access time this avoids the need to play a prediction game for an entire deployment network or subnet only the destination resource needs to provide the necessary access controls azure network security groups can be used for basic layer amp access controls between azure virtual networks their subnets and the internet azure web application firewall and the azure firewall can be used for more advanced network access controls that require application layer support local admin password solution laps or a third party privileged access management can set strong local admin passwords and just in time access to them suggested actions create a risk containment strategy that blends proven approaches including existing network security controls and practices native security controls available in azure zero trust approaches learn more build a security containment strategy
| 1
|
189,028
| 6,793,241,463
|
IssuesEvent
|
2017-11-01 06:17:13
|
cuappdev/podcast-ios
|
https://api.github.com/repos/cuappdev/podcast-ios
|
closed
|
Tag buttons show up as "and ..." / tags should probably be redesigned to look like hashtags
|
Priority: Low Type: Bug
|
<img width="374" alt="screen shot 2017-08-30 at 10 13 24 pm" src="https://user-images.githubusercontent.com/14966713/29903206-9db7b184-8dd0-11e7-8c86-58f1bd455de7.png">
|
1.0
|
Tag buttons show up as "and ..." / tags should probably be redesigned to look like hashtags - <img width="374" alt="screen shot 2017-08-30 at 10 13 24 pm" src="https://user-images.githubusercontent.com/14966713/29903206-9db7b184-8dd0-11e7-8c86-58f1bd455de7.png">
|
non_process
|
tag buttons show up as and tags should probably be redesigned to look like hashtags img width alt screen shot at pm src
| 0
|
41,674
| 10,563,627,485
|
IssuesEvent
|
2019-10-04 21:32:35
|
mpawlow/github-milestone-generator
|
https://api.github.com/repos/mpawlow/github-milestone-generator
|
closed
|
Exception encountered trying to close milestones: TypeError: expected string or bytes-like object
|
priority-high severity-major type-defect
|
## Exception
```
$ python index.py -c -r dap/dap-planning
[2019-10-03 18:50:59,222] [INFO ] [__main__] [-- GITHUB MILESTONE GENERATOR ------------------------------------------------
[2019-10-03 18:50:59,222] [INFO ] [__main__] Parsed Command-line Arguments:
GitHub API Domain: github.ibm.com.
GitHub Repository: dap/dap-planning.
New Milestone Name: None.
New Milestone Due Date: None.
Close Milestones: True.
Logging Level: INFO.
Traceback (most recent call last):
File "index.py", line 744, in <module>
_main()
File "index.py", line 693, in _main
date=args.milestone_due_date)
File "index.py", line 244, in _validate_milestone_due_date
status = _validate_iso_8601_date(date)
File "index.py", line 259, in _validate_iso_8601_date
match = re.fullmatch(ISO_8601_DATE_REGEX, date)
File "C:\Users\MikePawlowski\AppData\Local\Programs\Python\Python37\lib\re.py", line 178, in fullmatch
return _compile(pattern, flags).fullmatch(string)
TypeError: expected string or bytes-like object
(github-milestone-generator)
```
|
1.0
|
Exception encountered trying to close milestones: TypeError: expected string or bytes-like object - ## Exception
```
$ python index.py -c -r dap/dap-planning
[2019-10-03 18:50:59,222] [INFO ] [__main__] [-- GITHUB MILESTONE GENERATOR ------------------------------------------------
[2019-10-03 18:50:59,222] [INFO ] [__main__] Parsed Command-line Arguments:
GitHub API Domain: github.ibm.com.
GitHub Repository: dap/dap-planning.
New Milestone Name: None.
New Milestone Due Date: None.
Close Milestones: True.
Logging Level: INFO.
Traceback (most recent call last):
File "index.py", line 744, in <module>
_main()
File "index.py", line 693, in _main
date=args.milestone_due_date)
File "index.py", line 244, in _validate_milestone_due_date
status = _validate_iso_8601_date(date)
File "index.py", line 259, in _validate_iso_8601_date
match = re.fullmatch(ISO_8601_DATE_REGEX, date)
File "C:\Users\MikePawlowski\AppData\Local\Programs\Python\Python37\lib\re.py", line 178, in fullmatch
return _compile(pattern, flags).fullmatch(string)
TypeError: expected string or bytes-like object
(github-milestone-generator)
```
|
non_process
|
exception encountered trying to close milestones typeerror expected string or bytes like object exception python index py c r dap dap planning github milestone generator parsed command line arguments github api domain github ibm com github repository dap dap planning new milestone name none new milestone due date none close milestones true logging level info traceback most recent call last file index py line in main file index py line in main date args milestone due date file index py line in validate milestone due date status validate iso date date file index py line in validate iso date match re fullmatch iso date regex date file c users mikepawlowski appdata local programs python lib re py line in fullmatch return compile pattern flags fullmatch string typeerror expected string or bytes like object github milestone generator
| 0
|
16,953
| 22,307,732,915
|
IssuesEvent
|
2022-06-13 14:22:20
|
alphagov/govuk-design-system
|
https://api.github.com/repos/alphagov/govuk-design-system
|
closed
|
Investigate platform options for Design System Day 2022
|
🕔 weeks process shared ownership
|
## What
Agree which platforms we will use for video, note taking and collaboration activities.
## Why
So we can set expectations with attendees and facilitators, and secure budget approval if needed.
## Done when
- [x] Investigated options
- [x] Platform agreed for video
- [x] Platform agreed for note taking
- [x] Platform agreed for collaboration
- [x] Get access to a Padlet account - @kellylee-gds
- [x] Check Zoom setup/settings/features/how it works
|
1.0
|
Investigate platform options for Design System Day 2022 - ## What
Agree which platforms we will use for video, note taking and collaboration activities.
## Why
So we can set expectations with attendees and facilitators, and secure budget approval if needed.
## Done when
- [x] Investigated options
- [x] Platform agreed for video
- [x] Platform agreed for note taking
- [x] Platform agreed for collaboration
- [x] Get access to a Padlet account - @kellylee-gds
- [x] Check Zoom setup/settings/features/how it works
|
process
|
investigate platform options for design system day what agree which platforms we will use for video note taking and collaboration activities why so we can set expectations with attendees and facilitators and secure budget approval if needed done when investigated options platform agreed for video platform agreed for note taking platform agreed for collaboration get access to a padlet account kellylee gds check zoom setup settings features how it works
| 1
|
11,984
| 14,737,120,204
|
IssuesEvent
|
2021-01-07 00:55:44
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
portland-068 sab not working
|
anc-ops anc-process anp-1 ant-bug ant-child/secondary ant-support
|
In GitLab by @kdjstudios on Apr 17, 2018, 14:16
**Submitted by:** Lettice
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-17-58335/conversation
**Server:** Internal
**Client/Site:** Portland
**Account:** ALL
**Issue:**
I’m sending this email because after I login to SAB and try to enter an account number I get an error message that states We’re sorry, but something went wrong.
If someone can please assist with this issue.
|
1.0
|
portland-068 sab not working - In GitLab by @kdjstudios on Apr 17, 2018, 14:16
**Submitted by:** Lettice
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-17-58335/conversation
**Server:** Internal
**Client/Site:** Portland
**Account:** ALL
**Issue:**
I’m sending this email because after I login to SAB and try to enter an account number I get an error message that states We’re sorry, but something went wrong.
If someone can please assist with this issue.
|
process
|
portland sab not working in gitlab by kdjstudios on apr submitted by lettice helpdesk server internal client site portland account all issue i’m sending this email because after i login to sab and try to enter an account number i get an error message that states we’re sorry but something went wrong if someone can please assist with this issue
| 1
|
685,848
| 23,469,188,018
|
IssuesEvent
|
2022-08-16 19:54:17
|
Unity-Technologies/com.unity.netcode.gameobjects
|
https://api.github.com/repos/Unity-Technologies/com.unity.netcode.gameobjects
|
closed
|
NetworkAnimator.Cleanup() removes OnClientConnectedCallback only if IsServer is true, but ShutdownInternal() sets it to false beforehand.
|
type:bug priority:medium stat:imported
|
### Description
When shutting down a server/host and starting it again with a NetworkAnimator on e.g. the player, an error gets thrown:
```
NullReferenceException: Object reference not set to an instance of an object
Unity.Netcode.Components.NetworkAnimator.OnClientConnectedCallback (System.UInt64 playerId) (at Library/PackageCache/com.unity.netcode.gameobjects@1.0.0/Components/NetworkAnimator.cs:447)
Unity.Netcode.NetworkManager.InvokeOnClientConnectedCallback (System.UInt64 clientId) (at Library/PackageCache/com.unity.netcode.gameobjects@1.0.0/Runtime/Core/NetworkManager.cs:379)
Unity.Netcode.NetworkManager.HandleConnectionApproval (System.UInt64 ownerClientId, Unity.Netcode.NetworkManager+ConnectionApprovalResponse response) (at Library/PackageCache/com.unity.netcode.gameobjects@1.0.0/Runtime/Core/NetworkManager.cs:2077)
Unity.Netcode.NetworkManager.StartHost () (at Library/PackageCache/com.unity.netcode.gameobjects@1.0.0/Runtime/Core/NetworkManager.cs:1126)
````
It's pretty clear why that happens:
For the NetworkAnimator, OnNetworkSpawn:
```
if (IsServer)
{
NetworkManager.OnClientConnectedCallback += OnClientConnectedCallback;
}
```
Now, on Cleanup(), the callback is removed again:
```
private void Cleanup()
{
...
if (IsServer)
{
NetworkManager.OnClientConnectedCallback -= OnClientConnectedCallback;
}
...
```
However, cleanup gets called here:
```
internal void ShutdownInternal()
{
....
IsConnectedClient = false;
IsServer = false;
IsClient = false;
this.UnregisterAllNetworkUpdates();
....
if (SpawnManager != null)
{
SpawnManager.DespawnAndDestroyNetworkObjects(); <---- this is where Cleanup gets called
SpawnManager.ServerResetShudownStateForSceneObjects();
SpawnManager = null;
}
}
Then, the callback is not being cleaned up, but the object behind it is deleted. Connecting again then tries to execute that cleanup but fails, obviously.
### Reproduce Steps
1. Add NetworkAnimator to e.g. PlayerPrefab
2. StartHost()
3. StopHost()
4. StartHost() again.
Not 100% sure this is enough as we're doing a few initialization things inbetween, but the issue should be quite clear from the code.
### Actual Outcome
Callback doesn't get removed
### Expected Outcome
Callback gets removed
### Screenshots
If applicable, add screenshots to help explain your problem.
### Environment
- OS: Any
- Unity Version: 2021.3.5
- Netcode Version: 1.0.0
- Netcode Commit: https://github.com/Unity-Technologies/com.unity.netcode.gameobjects/commit/39a244854e323e5e0f444bc17bb7662f66ce28f2#diff-d4e79dceee7c1288ed9747284d2b61057e03cc974cfb21af8afb2704b1fada99R258
### Additional Context
|
1.0
|
NetworkAnimator.Cleanup() removes OnClientConnectedCallback only if IsServer is true, but ShutdownInternal() sets it to false beforehand. - ### Description
When shutting down a server/host and starting it again with a NetworkAnimator on e.g. the player, an error gets thrown:
```
NullReferenceException: Object reference not set to an instance of an object
Unity.Netcode.Components.NetworkAnimator.OnClientConnectedCallback (System.UInt64 playerId) (at Library/PackageCache/com.unity.netcode.gameobjects@1.0.0/Components/NetworkAnimator.cs:447)
Unity.Netcode.NetworkManager.InvokeOnClientConnectedCallback (System.UInt64 clientId) (at Library/PackageCache/com.unity.netcode.gameobjects@1.0.0/Runtime/Core/NetworkManager.cs:379)
Unity.Netcode.NetworkManager.HandleConnectionApproval (System.UInt64 ownerClientId, Unity.Netcode.NetworkManager+ConnectionApprovalResponse response) (at Library/PackageCache/com.unity.netcode.gameobjects@1.0.0/Runtime/Core/NetworkManager.cs:2077)
Unity.Netcode.NetworkManager.StartHost () (at Library/PackageCache/com.unity.netcode.gameobjects@1.0.0/Runtime/Core/NetworkManager.cs:1126)
````
It's pretty clear why that happens:
For the NetworkAnimator, OnNetworkSpawn:
```
if (IsServer)
{
NetworkManager.OnClientConnectedCallback += OnClientConnectedCallback;
}
```
Now, on Cleanup(), the callback is removed again:
```
private void Cleanup()
{
...
if (IsServer)
{
NetworkManager.OnClientConnectedCallback -= OnClientConnectedCallback;
}
...
```
However, cleanup gets called here:
```
internal void ShutdownInternal()
{
....
IsConnectedClient = false;
IsServer = false;
IsClient = false;
this.UnregisterAllNetworkUpdates();
....
if (SpawnManager != null)
{
SpawnManager.DespawnAndDestroyNetworkObjects(); <---- this is where Cleanup gets called
SpawnManager.ServerResetShudownStateForSceneObjects();
SpawnManager = null;
}
}
Then, the callback is not being cleaned up, but the object behind it is deleted. Connecting again then tries to execute that cleanup but fails, obviously.
### Reproduce Steps
1. Add NetworkAnimator to e.g. PlayerPrefab
2. StartHost()
3. StopHost()
4. StartHost() again.
Not 100% sure this is enough as we're doing a few initialization things inbetween, but the issue should be quite clear from the code.
### Actual Outcome
Callback doesn't get removed
### Expected Outcome
Callback gets removed
### Screenshots
If applicable, add screenshots to help explain your problem.
### Environment
- OS: Any
- Unity Version: 2021.3.5
- Netcode Version: 1.0.0
- Netcode Commit: https://github.com/Unity-Technologies/com.unity.netcode.gameobjects/commit/39a244854e323e5e0f444bc17bb7662f66ce28f2#diff-d4e79dceee7c1288ed9747284d2b61057e03cc974cfb21af8afb2704b1fada99R258
### Additional Context
|
non_process
|
networkanimator cleanup removes onclientconnectedcallback only if isserver is true but shutdowninternal sets it to false beforehand description when shutting down a server host and starting it again with a networkanimator on e g the player an error gets thrown nullreferenceexception object reference not set to an instance of an object unity netcode components networkanimator onclientconnectedcallback system playerid at library packagecache com unity netcode gameobjects components networkanimator cs unity netcode networkmanager invokeonclientconnectedcallback system clientid at library packagecache com unity netcode gameobjects runtime core networkmanager cs unity netcode networkmanager handleconnectionapproval system ownerclientid unity netcode networkmanager connectionapprovalresponse response at library packagecache com unity netcode gameobjects runtime core networkmanager cs unity netcode networkmanager starthost at library packagecache com unity netcode gameobjects runtime core networkmanager cs it s pretty clear why that happens for the networkanimator onnetworkspawn if isserver networkmanager onclientconnectedcallback onclientconnectedcallback now on cleanup the callback is removed again private void cleanup if isserver networkmanager onclientconnectedcallback onclientconnectedcallback however cleanup gets called here internal void shutdowninternal isconnectedclient false isserver false isclient false this unregisterallnetworkupdates if spawnmanager null spawnmanager despawnanddestroynetworkobjects this is where cleanup gets called spawnmanager serverresetshudownstateforsceneobjects spawnmanager null then the callback is not being cleaned up but the object behind it is deleted connecting again then tries to execute that cleanup but fails obviously reproduce steps add networkanimator to e g playerprefab starthost stophost starthost again not sure this is enough as we re doing a few initialization things inbetween but the issue should be quite clear from the code actual outcome callback doesn t get removed expected outcome callback gets removed screenshots if applicable add screenshots to help explain your problem environment os any unity version netcode version netcode commit additional context
| 0
|
3,503
| 6,557,662,230
|
IssuesEvent
|
2017-09-06 18:17:24
|
w3c/activitypub
|
https://api.github.com/repos/w3c/activitypub
|
closed
|
sharedInbox / siteInbox type endpoint (publicInbox, but not just for public posts)
|
Needs Process Help
|
Currently, the most general way to post an activity in ActivityPub is to post it to a user's `inbox` endpoint. However, since well known figures with many subscribers would result in many posts to many users at once, we've made an exception for public posts, which may be posted to the [publicInbox](https://www.w3.org/TR/activitypub/#public-inbox-delivery), which may be shared amongst users on a site.
On the call today, we found that this was not enough for Mastodon. On Mastodon, followers-only posts are common. Gargron gave an example that they have over 12k users, and should every followers-only post result in 12k HTTP requests given that many users are on shared servers?
Gargron suggested that Mastodon will probably reuse the publicInbox endpoint for this purpose. While I personally strongly prefer the delivery to inboxes approach, I think we need to address this. It's clear that Mastodon will do something to the effect in its implementation, so I think we need to get this right in ActivityPub itself, otherwise we could end up in the same space as what's happening in OStatus right now. One could easily see an implementation like Mastodon posting private content to the publicInbox endpoint and expecting servers to filter delivery based on content, and other servers not being aware and unintentionally delivering that information publicly to their users. That would be bad!
So, I think we should rename `publicInbox` to something like `sharedInbox` or `siteInbox` and change its behavior.
* For posts that are addressed to the special Public collection, the behavior is pretty much the same as currently for `publicInbox`.
* For non-public posts, we have two options...
- Servers can look at the addressing of the post and identify what recipients it should deliver to. We could maybe restrict this to being done for followers-style collections only, or it could also look at the individuals addressed on to, cc, etc. At any rate, a receiving server will have to do its "best knowledge" of determining who a post should be delivered to. This may be less precise; it could fail for users that *should have* been Block'ed. Errors in the network could also mean that the followers list is somewhat out of sync between servers.
- Alternately, a sending server could, as a separate part of the message, specify an exact list of recipients relevant to that server. For 12k or even 1M followers, this could be a large post (though it could be done in multiple posts) but it would be less large than 12k individual HTTP POSTs. But it would be much more precise, and more respectful of things like blocklists (which currently we specify *not* federating across servers to protect users.)
|
1.0
|
sharedInbox / siteInbox type endpoint (publicInbox, but not just for public posts) - Currently, the most general way to post an activity in ActivityPub is to post it to a user's `inbox` endpoint. However, since well known figures with many subscribers would result in many posts to many users at once, we've made an exception for public posts, which may be posted to the [publicInbox](https://www.w3.org/TR/activitypub/#public-inbox-delivery), which may be shared amongst users on a site.
On the call today, we found that this was not enough for Mastodon. On Mastodon, followers-only posts are common. Gargron gave an example that they have over 12k users, and should every followers-only post result in 12k HTTP requests given that many users are on shared servers?
Gargron suggested that Mastodon will probably reuse the publicInbox endpoint for this purpose. While I personally strongly prefer the delivery to inboxes approach, I think we need to address this. It's clear that Mastodon will do something to the effect in its implementation, so I think we need to get this right in ActivityPub itself, otherwise we could end up in the same space as what's happening in OStatus right now. One could easily see an implementation like Mastodon posting private content to the publicInbox endpoint and expecting servers to filter delivery based on content, and other servers not being aware and unintentionally delivering that information publicly to their users. That would be bad!
So, I think we should rename `publicInbox` to something like `sharedInbox` or `siteInbox` and change its behavior.
* For posts that are addressed to the special Public collection, the behavior is pretty much the same as currently for `publicInbox`.
* For non-public posts, we have two options...
- Servers can look at the addressing of the post and identify what recipients it should deliver to. We could maybe restrict this to being done for followers-style collections only, or it could also look at the individuals addressed on to, cc, etc. At any rate, a receiving server will have to do its "best knowledge" of determining who a post should be delivered to. This may be less precise; it could fail for users that *should have* been Block'ed. Errors in the network could also mean that the followers list is somewhat out of sync between servers.
- Alternately, a sending server could, as a separate part of the message, specify an exact list of recipients relevant to that server. For 12k or even 1M followers, this could be a large post (though it could be done in multiple posts) but it would be less large than 12k individual HTTP POSTs. But it would be much more precise, and more respectful of things like blocklists (which currently we specify *not* federating across servers to protect users.)
|
process
|
sharedinbox siteinbox type endpoint publicinbox but not just for public posts currently the most general way to post an activity in activitypub is to post it to a user s inbox endpoint however since well known figures with many subscribers would result in many posts to many users at once we ve made an exception for public posts which may be posted to the which may be shared amongst users on a site on the call today we found that this was not enough for mastodon on mastodon followers only posts are common gargron gave an example that they have over users and should every followers only post result in http requests given that many users are on shared servers gargron suggested that mastodon will probably reuse the publicinbox endpoint for this purpose while i personally strongly prefer the delivery to inboxes approach i think we need to address this it s clear that mastodon will do something to the effect in its implementation so i think we need to get this right in activitypub itself otherwise we could end up in the same space as what s happening in ostatus right now one could easily see an implementation like mastodon posting private content to the publicinbox endpoint and expecting servers to filter delivery based on content and other servers not being aware and unintentionally delivering that information publicly to their users that would be bad so i think we should rename publicinbox to something like sharedinbox or siteinbox and change its behavior for posts that are addressed to the special public collection the behavior is pretty much the same as currently for publicinbox for non public posts we have two options servers can look at the addressing of the post and identify what recipients it should deliver to we could maybe restrict this to being done for followers style collections only or it could also look at the individuals addressed on to cc etc at any rate a receiving server will have to do its best knowledge of determining who a post should be delivered to this may be less precise it could fail for users that should have been block ed errors in the network could also mean that the followers list is somewhat out of sync between servers alternately a sending server could as a separate part of the message specify an exact list of recipients relevant to that server for or even followers this could be a large post though it could be done in multiple posts but it would be less large than individual http posts but it would be much more precise and more respectful of things like blocklists which currently we specify not federating across servers to protect users
| 1
|
6,481
| 9,553,686,681
|
IssuesEvent
|
2019-05-02 19:56:42
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
No Such File or Directory
|
command-line options log-processing question
|
$ goaccess -p /usr/local/etc/goaccess/goaccess.conf mylog.log-20190502 ~/desktop/alpha.html
GoAccess - version 1.3 - Apr 29 2019 17:12:23
Config file: /usr/local/etc/goaccess/goaccess.conf
Fatal error has occurred
Error occurred at: src/parser.c - read_log - 2728
Unable to open the specified log file. No such file or directory
$ ll
-rwxr-xr-x 1 rich admin 21K Oct 15 2018 mylog.log-20181003
-rwxr-xr-x@ 1 rich admin 1.1M May 2 07:59 mylog.log-20190502
For some reason, the log file cannot be opened. It's there (last line) and it's copied and pasted so I'm not overlooking it.
Anything I should try?
|
1.0
|
No Such File or Directory - $ goaccess -p /usr/local/etc/goaccess/goaccess.conf mylog.log-20190502 ~/desktop/alpha.html
GoAccess - version 1.3 - Apr 29 2019 17:12:23
Config file: /usr/local/etc/goaccess/goaccess.conf
Fatal error has occurred
Error occurred at: src/parser.c - read_log - 2728
Unable to open the specified log file. No such file or directory
$ ll
-rwxr-xr-x 1 rich admin 21K Oct 15 2018 mylog.log-20181003
-rwxr-xr-x@ 1 rich admin 1.1M May 2 07:59 mylog.log-20190502
For some reason, the log file cannot be opened. It's there (last line) and it's copied and pasted so I'm not overlooking it.
Anything I should try?
|
process
|
no such file or directory goaccess p usr local etc goaccess goaccess conf mylog log desktop alpha html goaccess version apr config file usr local etc goaccess goaccess conf fatal error has occurred error occurred at src parser c read log unable to open the specified log file no such file or directory ll rwxr xr x rich admin oct mylog log rwxr xr x rich admin may mylog log for some reason the log file cannot be opened it s there last line and it s copied and pasted so i m not overlooking it anything i should try
| 1
|
76,284
| 21,328,216,394
|
IssuesEvent
|
2022-04-18 03:39:25
|
opensearch-project/OpenSearch-Dashboards
|
https://api.github.com/repos/opensearch-project/OpenSearch-Dashboards
|
reopened
|
[BUG] Re2 Requires a higher NODE_MODULE_VERSION
|
bug build v2.0.0
|
**Describe the bug**
node-re2 so that we can safely use the regular expressions specified by end-users in Timelion
see version conflict error when using v1.2 on arm instance.
v1.2:
<img width="1533" alt="Screen Shot 2022-04-07 at 22 52 08" src="https://user-images.githubusercontent.com/79961084/162476818-6a07f2cd-db15-4c44-8ed3-2d9eb5cf7788.png">
main (2.0):
<img width="1654" alt="Screen Shot 2022-04-08 at 10 13 33" src="https://user-images.githubusercontent.com/79961084/162490263-6f8244b0-abae-464a-8745-f8cdc1047b58.png">
**To Reproduce**
create v1.2 graviton2 domain
logon to OpenSearch and load sample web log data
go to visualize and select to create a timelion visualization
run expression:
```
.es(*).label('all', '.*')
```
**Expected behavior**
fix the error
**OpenSearch Version**
v1.2 +
**Dashboards Version**
v1.2 +
**Plugins**
no plugin
**Host/Environment (please complete the following information):**
- OS: [e.g. iOS]
- Browser and version [e.g. 22]
|
1.0
|
[BUG] Re2 Requires a higher NODE_MODULE_VERSION - **Describe the bug**
node-re2 so that we can safely use the regular expressions specified by end-users in Timelion
see version conflict error when using v1.2 on arm instance.
v1.2:
<img width="1533" alt="Screen Shot 2022-04-07 at 22 52 08" src="https://user-images.githubusercontent.com/79961084/162476818-6a07f2cd-db15-4c44-8ed3-2d9eb5cf7788.png">
main (2.0):
<img width="1654" alt="Screen Shot 2022-04-08 at 10 13 33" src="https://user-images.githubusercontent.com/79961084/162490263-6f8244b0-abae-464a-8745-f8cdc1047b58.png">
**To Reproduce**
create v1.2 graviton2 domain
logon to OpenSearch and load sample web log data
go to visualize and select to create a timelion visualization
run expression:
```
.es(*).label('all', '.*')
```
**Expected behavior**
fix the error
**OpenSearch Version**
v1.2 +
**Dashboards Version**
v1.2 +
**Plugins**
no plugin
**Host/Environment (please complete the following information):**
- OS: [e.g. iOS]
- Browser and version [e.g. 22]
|
non_process
|
requires a higher node module version describe the bug node so that we can safely use the regular expressions specified by end users in timelion see version conflict error when using on arm instance img width alt screen shot at src main img width alt screen shot at src to reproduce create domain logon to opensearch and load sample web log data go to visualize and select to create a timelion visualization run expression es label all expected behavior fix the error opensearch version dashboards version plugins no plugin host environment please complete the following information os browser and version
| 0
|
174,507
| 14,484,668,657
|
IssuesEvent
|
2020-12-10 16:37:45
|
Blank1611/jsonmergeutils
|
https://api.github.com/repos/Blank1611/jsonmergeutils
|
closed
|
[feature] Update readme
|
documentation good first issue
|
Update the readme to reflect the input method changes and usage
based on PR https://github.com/Blank1611/jsonmergeutils/pull/14
|
1.0
|
[feature] Update readme - Update the readme to reflect the input method changes and usage
based on PR https://github.com/Blank1611/jsonmergeutils/pull/14
|
non_process
|
update readme update the readme to reflect the input method changes and usage based on pr
| 0
|
19,841
| 26,244,311,541
|
IssuesEvent
|
2023-01-05 14:07:14
|
RIOT-OS/RIOT
|
https://api.github.com/repos/RIOT-OS/RIOT
|
closed
|
sys/ztimer: Potential footgun when mixing timestamps of ztimer clocks
|
Process: API change Area: timers Area: sys State: don't stale
|
#### Description
Each ztimer clock can potentially be backed by different hardware. E.g. on a given system `ZTIMER_USEC` could use `periph_timer` as backend, while `ZTIMER_MSEC` could use `periph_rtt` as backend. A call to `ztimer_now()` will give the current counter value of the underlying hardware (after first applying frequency conversion and bit extension as needed). When comparing timestamps of different clocks this can be problematic (that is, after converting them to the same time unit), because:
1. Each hardware clock's epoch might be different. E.g. after a warm boot the RTT might resume counting from the value it had, while the high frequency timer will be reinitialized with zero.
2. Clocks will drift in regard to each other
#### Steps to reproduce the issue
```C
static ztimer_now_t point_in_time_a_us;
static ztimer_now_t point_in_time_b_ms;
static ztimer_now_t point_in_time_c_ms;
static void event_a_callback(void) {
point_in_time_a_us = ztimer_now(ZTIMER_USEC);
}
static void event_b_callback(void) {
point_in_time_b_ms = ztimer_now(ZTIMER_MSEC);
}
static void event_c_callback(void) {
point_in_time_c_ms = ztimer_now(ZTIMER_MSEC);
ztimer_now_t duration_a_b_us = point_in_time_b_ms * USEC_PER_MSEC - point_in_time_a_us;
ztimer_now_t duration_b_c_ms = point_in_time_c_ms - point_in_time_b_ms;
}
```
#### Expected results
Both `duration_a_b_us` and `duration_b_c_ms` are correct.
#### Actual results
If `ZTIMER_MSEC` and `ZTIMER_USEC` are not provided by the same hardware clock, `duration_a_b_us` will likely be incorrect.
#### Versions
Current `master`.
|
1.0
|
sys/ztimer: Potential footgun when mixing timestamps of ztimer clocks - #### Description
Each ztimer clock can potentially be backed by different hardware. E.g. on a given system `ZTIMER_USEC` could use `periph_timer` as backend, while `ZTIMER_MSEC` could use `periph_rtt` as backend. A call to `ztimer_now()` will give the current counter value of the underlying hardware (after first applying frequency conversion and bit extension as needed). When comparing timestamps of different clocks this can be problematic (that is, after converting them to the same time unit), because:
1. Each hardware clock's epoch might be different. E.g. after a warm boot the RTT might resume counting from the value it had, while the high frequency timer will be reinitialized with zero.
2. Clocks will drift in regard to each other
#### Steps to reproduce the issue
```C
static ztimer_now_t point_in_time_a_us;
static ztimer_now_t point_in_time_b_ms;
static ztimer_now_t point_in_time_c_ms;
static void event_a_callback(void) {
point_in_time_a_us = ztimer_now(ZTIMER_USEC);
}
static void event_b_callback(void) {
point_in_time_b_ms = ztimer_now(ZTIMER_MSEC);
}
static void event_c_callback(void) {
point_in_time_c_ms = ztimer_now(ZTIMER_MSEC);
ztimer_now_t duration_a_b_us = point_in_time_b_ms * USEC_PER_MSEC - point_in_time_a_us;
ztimer_now_t duration_b_c_ms = point_in_time_c_ms - point_in_time_b_ms;
}
```
#### Expected results
Both `duration_a_b_us` and `duration_b_c_ms` are correct.
#### Actual results
If `ZTIMER_MSEC` and `ZTIMER_USEC` are not provided by the same hardware clock, `duration_a_b_us` will likely be incorrect.
#### Versions
Current `master`.
|
process
|
sys ztimer potential footgun when mixing timestamps of ztimer clocks description each ztimer clock can potentially be backed by different hardware e g on a given system ztimer usec could use periph timer as backend while ztimer msec could use periph rtt as backend a call to ztimer now will give the current counter value of the underlying hardware after first applying frequency conversion and bit extension as needed when comparing timestamps of different clocks this can be problematic that is after converting them to the same time unit because each hardware clock s epoch might be different e g after a warm boot the rtt might resume counting from the value it had while the high frequency timer will be reinitialized with zero clocks will drift in regard to each other steps to reproduce the issue c static ztimer now t point in time a us static ztimer now t point in time b ms static ztimer now t point in time c ms static void event a callback void point in time a us ztimer now ztimer usec static void event b callback void point in time b ms ztimer now ztimer msec static void event c callback void point in time c ms ztimer now ztimer msec ztimer now t duration a b us point in time b ms usec per msec point in time a us ztimer now t duration b c ms point in time c ms point in time b ms expected results both duration a b us and duration b c ms are correct actual results if ztimer msec and ztimer usec are not provided by the same hardware clock duration a b us will likely be incorrect versions current master
| 1
|
8,218
| 11,406,062,318
|
IssuesEvent
|
2020-01-31 13:33:54
|
prisma/prisma2
|
https://api.github.com/repos/prisma/prisma2
|
opened
|
Native Enums only
|
kind/improvement process/candidate
|
Right now Prisma has its own notion of Enums, we want to remove that and replace it with fully native Enums that the databases support themselves.
|
1.0
|
Native Enums only - Right now Prisma has its own notion of Enums, we want to remove that and replace it with fully native Enums that the databases support themselves.
|
process
|
native enums only right now prisma has its own notion of enums we want to remove that and replace it with fully native enums that the databases support themselves
| 1
|
173,700
| 21,177,049,199
|
IssuesEvent
|
2022-04-08 01:51:06
|
ibm-cio-vulnerability-scanning/insomnia
|
https://api.github.com/repos/ibm-cio-vulnerability-scanning/insomnia
|
reopened
|
CVE-2021-37712 (High) detected in tar-4.4.13.tgz
|
security vulnerability
|
## CVE-2021-37712 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.13.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p>
<p>
Dependency Hierarchy:
- webpack-4.44.1.tgz (Root Library)
- watchpack-1.7.4.tgz
- watchpack-chokidar2-2.0.0.tgz
- chokidar-2.1.8.tgz
- fsevents-1.2.11.tgz
- node-pre-gyp-0.14.0.tgz
- :x: **tar-4.4.13.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ibm-cio-vulnerability-scanning/insomnia/commit/c08890295e602a2b6dc87c5303b52c137d02f0c9">c08890295e602a2b6dc87c5303b52c137d02f0c9</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value. Additionally, on Windows systems, long path portions would resolve to the same file system entities as their 8.3 "short path" counterparts. A specially crafted tar archive could thus include a directory with one form of the path, followed by a symbolic link with a different string that resolves to the same file system entity, followed by a file using the first form. By first creating a directory, and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-qq89-hq3f-393p.
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37712>CVE-2021-37712</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p">https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution (tar): 4.4.18</p>
<p>Direct dependency fix Resolution (webpack): 4.44.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-37712 (High) detected in tar-4.4.13.tgz - ## CVE-2021-37712 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.13.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p>
<p>
Dependency Hierarchy:
- webpack-4.44.1.tgz (Root Library)
- watchpack-1.7.4.tgz
- watchpack-chokidar2-2.0.0.tgz
- chokidar-2.1.8.tgz
- fsevents-1.2.11.tgz
- node-pre-gyp-0.14.0.tgz
- :x: **tar-4.4.13.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ibm-cio-vulnerability-scanning/insomnia/commit/c08890295e602a2b6dc87c5303b52c137d02f0c9">c08890295e602a2b6dc87c5303b52c137d02f0c9</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value. Additionally, on Windows systems, long path portions would resolve to the same file system entities as their 8.3 "short path" counterparts. A specially crafted tar archive could thus include a directory with one form of the path, followed by a symbolic link with a different string that resolves to the same file system entity, followed by a file using the first form. By first creating a directory, and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-qq89-hq3f-393p.
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37712>CVE-2021-37712</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p">https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution (tar): 4.4.18</p>
<p>Direct dependency fix Resolution (webpack): 4.44.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href dependency hierarchy webpack tgz root library watchpack tgz watchpack tgz chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in head commit a href found in base branch develop vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value additionally on windows systems long path portions would resolve to the same file system entities as their short path counterparts a specially crafted tar archive could thus include a directory with one form of the path followed by a symbolic link with a different string that resolves to the same file system entity followed by a file using the first form by first creating a directory and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar direct dependency fix resolution webpack step up your open source security game with whitesource
| 0
|
1,715
| 3,422,362,398
|
IssuesEvent
|
2015-12-08 22:41:13
|
ikreymer/netcapsule
|
https://api.github.com/repos/ikreymer/netcapsule
|
opened
|
Add apparmour profiles to limit exec capabilities in containers
|
security
|
Create profiles for different browsers, probably using
https://github.com/jfrazelle/bane to restrict execution with browser containers
This should address some (but not all) of the more serious security issues.
|
True
|
Add apparmour profiles to limit exec capabilities in containers - Create profiles for different browsers, probably using
https://github.com/jfrazelle/bane to restrict execution with browser containers
This should address some (but not all) of the more serious security issues.
|
non_process
|
add apparmour profiles to limit exec capabilities in containers create profiles for different browsers probably using to restrict execution with browser containers this should address some but not all of the more serious security issues
| 0
|
7,506
| 10,587,644,228
|
IssuesEvent
|
2019-10-08 22:49:04
|
googleapis/google-api-java-client-services
|
https://api.github.com/repos/googleapis/google-api-java-client-services
|
closed
|
Add clirr test to ensure no breaking changes
|
type: process
|
We can use the `maven-clirr-plugin` to test against the previously released version.
|
1.0
|
Add clirr test to ensure no breaking changes - We can use the `maven-clirr-plugin` to test against the previously released version.
|
process
|
add clirr test to ensure no breaking changes we can use the maven clirr plugin to test against the previously released version
| 1
|
13,596
| 16,168,307,066
|
IssuesEvent
|
2021-05-01 23:58:06
|
hasura/ask-me-anything
|
https://api.github.com/repos/hasura/ask-me-anything
|
opened
|
How does one build an ODBC connecting string to connect to SQL Server?
|
good first issue next-up-for-ama processing-for-shortvid question
|
ODBC Format Docs: https://docs.microsoft.com/en-us/previous-versions/windows/desktop/ms722656(v=vs.85)
Important notes:
* Connection strings over a standard connection to SQL Server.
* Connection strings in relation to "trusted" connections.
* Connection strings and the specifics of local development with Docker containers.
|
1.0
|
How does one build an ODBC connecting string to connect to SQL Server? - ODBC Format Docs: https://docs.microsoft.com/en-us/previous-versions/windows/desktop/ms722656(v=vs.85)
Important notes:
* Connection strings over a standard connection to SQL Server.
* Connection strings in relation to "trusted" connections.
* Connection strings and the specifics of local development with Docker containers.
|
process
|
how does one build an odbc connecting string to connect to sql server odbc format docs important notes connection strings over a standard connection to sql server connection strings in relation to trusted connections connection strings and the specifics of local development with docker containers
| 1
|
25,482
| 3,933,049,766
|
IssuesEvent
|
2016-04-25 17:48:10
|
OSTraining/OSCampus
|
https://api.github.com/repos/OSTraining/OSCampus
|
closed
|
Add published date to class display
|
Design
|
@billtomczak It seems to be based on when the class was first created rather than the published date:
https://www.ostraining.com/new-classes/
@htmgarcia Could we add the published date to this view?
|
1.0
|
Add published date to class display - @billtomczak It seems to be based on when the class was first created rather than the published date:
https://www.ostraining.com/new-classes/
@htmgarcia Could we add the published date to this view?
|
non_process
|
add published date to class display billtomczak it seems to be based on when the class was first created rather than the published date htmgarcia could we add the published date to this view
| 0
|
21,066
| 11,089,988,039
|
IssuesEvent
|
2019-12-14 22:43:37
|
reactioncommerce/reaction
|
https://api.github.com/repos/reactioncommerce/reaction
|
closed
|
Use npm for package management.
|
performance
|
Specifically we need to accomplish a package.json that has all the import/dependencies/aliases configured + allows for custom addition/delete of dependencies on the actual client projects. Suggest that this is done using a package.json templates, that the developer can then commit/edit.
|
True
|
Use npm for package management. - Specifically we need to accomplish a package.json that has all the import/dependencies/aliases configured + allows for custom addition/delete of dependencies on the actual client projects. Suggest that this is done using a package.json templates, that the developer can then commit/edit.
|
non_process
|
use npm for package management specifically we need to accomplish a package json that has all the import dependencies aliases configured allows for custom addition delete of dependencies on the actual client projects suggest that this is done using a package json templates that the developer can then commit edit
| 0
|
82,540
| 15,646,592,557
|
IssuesEvent
|
2021-03-23 01:16:59
|
veshitala/datadiscussforum
|
https://api.github.com/repos/veshitala/datadiscussforum
|
opened
|
CVE-2020-11023 (Medium) detected in jquery-2.1.4.min.js, jquery-1.12.2.min.js
|
security vulnerability
|
## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.12.2.min.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: /datadiscussforum/django_simple_forum/templates/dashboard/dashboard_login.html</p>
<p>Path to vulnerable library: datadiscussforum/django_simple_forum/templates/dashboard/dashboard_login.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.12.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.2/jquery.min.js</a></p>
<p>Path to dependency file: /datadiscussforum/django_simple_forum/templates/dashboard/dashboard_login.html</p>
<p>Path to vulnerable library: datadiscussforum/django_simple_forum/templates/dashboard/dashboard_login.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.12.2.min.js** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-11023 (Medium) detected in jquery-2.1.4.min.js, jquery-1.12.2.min.js - ## CVE-2020-11023 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.12.2.min.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: /datadiscussforum/django_simple_forum/templates/dashboard/dashboard_login.html</p>
<p>Path to vulnerable library: datadiscussforum/django_simple_forum/templates/dashboard/dashboard_login.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.12.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.12.2/jquery.min.js</a></p>
<p>Path to dependency file: /datadiscussforum/django_simple_forum/templates/dashboard/dashboard_login.html</p>
<p>Path to vulnerable library: datadiscussforum/django_simple_forum/templates/dashboard/dashboard_login.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.12.2.min.js** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in jquery min js jquery min js cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file datadiscussforum django simple forum templates dashboard dashboard login html path to vulnerable library datadiscussforum django simple forum templates dashboard dashboard login html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file datadiscussforum django simple forum templates dashboard dashboard login html path to vulnerable library datadiscussforum django simple forum templates dashboard dashboard login html dependency hierarchy x jquery min js vulnerable library vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
21,347
| 29,172,549,313
|
IssuesEvent
|
2023-05-19 04:09:51
|
phuocduong-agilityio/internship-huy-dao
|
https://api.github.com/repos/phuocduong-agilityio/internship-huy-dao
|
opened
|
Fix the README.md file
|
In-process
|
- [ ] What is the purpose of the `Introduce` section?
- [ ] Unclear targets.
- [ ] For the Technical section, we should have a short introduction for each technical item we use. I think it would be better to add an official site for reference.
- [ ] Guideline does not work.
|
1.0
|
Fix the README.md file - - [ ] What is the purpose of the `Introduce` section?
- [ ] Unclear targets.
- [ ] For the Technical section, we should have a short introduction for each technical item we use. I think it would be better to add an official site for reference.
- [ ] Guideline does not work.
|
process
|
fix the readme md file what is the purpose of the introduce section unclear targets for the technical section we should have a short introduction for each technical item we use i think it would be better to add an official site for reference guideline does not work
| 1
|
20,463
| 27,128,797,549
|
IssuesEvent
|
2023-02-16 08:12:52
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Remove `--python_path` flag
|
P3 type: process team-Rules-Python stale
|
This had to be broken out of #7375 because it turns out `python_path` is [used on windows](https://github.com/bazelbuild/bazel/blob/49107ad79ef08811db22636928dfd113a9acf902/src/main/cpp/blaze_util_windows.cc#L1483-L1491) to [communicate](https://github.com/bazelbuild/bazel/blob/f59fad7c97ed9b4d5ae577e6ef50cefd3927e4bd/src/main/cpp/blaze_util_platform.h#L265-L269) the location of the Python interpreter even when it's not otherwise available on the `PATH` at execution time.
We can possibly switch this bit of client-server communication to use a hidden flag in place of `--python_path`. Alternatively (and probably preferably), we can replace this logic with a repo rule that detects the host platform's Python interpreter, and feed that into the Python toolchain. Another option is to deprecate this bit of niceness and instead require that the interpreter either be available on `PATH` at execution time or else have its path embedded into an appropriate `py_runtime` of a toolchain definition.
|
1.0
|
Remove `--python_path` flag - This had to be broken out of #7375 because it turns out `python_path` is [used on windows](https://github.com/bazelbuild/bazel/blob/49107ad79ef08811db22636928dfd113a9acf902/src/main/cpp/blaze_util_windows.cc#L1483-L1491) to [communicate](https://github.com/bazelbuild/bazel/blob/f59fad7c97ed9b4d5ae577e6ef50cefd3927e4bd/src/main/cpp/blaze_util_platform.h#L265-L269) the location of the Python interpreter even when it's not otherwise available on the `PATH` at execution time.
We can possibly switch this bit of client-server communication to use a hidden flag in place of `--python_path`. Alternatively (and probably preferably), we can replace this logic with a repo rule that detects the host platform's Python interpreter, and feed that into the Python toolchain. Another option is to deprecate this bit of niceness and instead require that the interpreter either be available on `PATH` at execution time or else have its path embedded into an appropriate `py_runtime` of a toolchain definition.
|
process
|
remove python path flag this had to be broken out of because it turns out python path is to the location of the python interpreter even when it s not otherwise available on the path at execution time we can possibly switch this bit of client server communication to use a hidden flag in place of python path alternatively and probably preferably we can replace this logic with a repo rule that detects the host platform s python interpreter and feed that into the python toolchain another option is to deprecate this bit of niceness and instead require that the interpreter either be available on path at execution time or else have its path embedded into an appropriate py runtime of a toolchain definition
| 1
|
21,353
| 29,184,799,615
|
IssuesEvent
|
2023-05-19 14:37:02
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
[API proposal]: add ProcessStartInfo constructor that accepts IEnumerable<string> arguments.
|
api-approved area-System.Diagnostics.Process good first issue help wanted in-pr
|
## Background and motivation
Starting a process should be easy when you have an `IEnumerable<string>` with arguments.
This is true if you can use the static `Process.Start` overload that accepts an `IEnumerable`.
If you need to use `ProcessStartInfo` to make additional configuration, like redirecting the streams, there is no way to pass in the `IEnumerable<string>`.
`Process.ArgumentList` is a `Collection<string>`. Because it doesn't have an `AddRange`, the user needs to add a loop to add the arguments one-by-one.
## API Proposal
A constructor can be added to `ProcessStartInfo`.
```cs
ProcessStartInfo(string fileName, IEnumerable<string> arguments)
```
This is consistent with rest of the API of `Process` where `IEnumerable<string>` overloads were added to existing APIs that accepted `string arguments`.
## API Usage
```cs
ProcessStartInfo psi = new("filename", GetArguments());
```
## Alternative designs
## Risks
|
1.0
|
[API proposal]: add ProcessStartInfo constructor that accepts IEnumerable<string> arguments. - ## Background and motivation
Starting a process should be easy when you have an `IEnumerable<string>` with arguments.
This is true if you can use the static `Process.Start` overload that accepts an `IEnumerable`.
If you need to use `ProcessStartInfo` to make additional configuration, like redirecting the streams, there is no way to pass in the `IEnumerable<string>`.
`Process.ArgumentList` is a `Collection<string>`. Because it doesn't have an `AddRange`, the user needs to add a loop to add the arguments one-by-one.
## API Proposal
A constructor can be added to `ProcessStartInfo`.
```cs
ProcessStartInfo(string fileName, IEnumerable<string> arguments)
```
This is consistent with rest of the API of `Process` where `IEnumerable<string>` overloads were added to existing APIs that accepted `string arguments`.
## API Usage
```cs
ProcessStartInfo psi = new("filename", GetArguments());
```
## Alternative designs
## Risks
|
process
|
add processstartinfo constructor that accepts ienumerable arguments background and motivation starting a process should be easy when you have an ienumerable with arguments this is true if you can use the static process start overload that accepts an ienumerable if you need to use processstartinfo to make additional configuration like redirecting the streams there is no way to pass in the ienumerable process argumentlist is a collection because it doesn t have an addrange the user needs to add a loop to add the arguments one by one api proposal a constructor can be added to processstartinfo cs processstartinfo string filename ienumerable arguments this is consistent with rest of the api of process where ienumerable overloads were added to existing apis that accepted string arguments api usage cs processstartinfo psi new filename getarguments alternative designs risks
| 1
|
217,750
| 16,887,173,251
|
IssuesEvent
|
2021-06-23 02:52:11
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
[failing test] kubeadm-kinder-latest: task-03-init
|
area/kubeadm kind/failing-test sig/cluster-lifecycle triage/accepted
|
<!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI -->
#### Which jobs are failing:
https://k8s-testgrid.appspot.com/sig-cluster-lifecycle-kubeadm#kubeadm-kinder-latest
#### Which test(s) are failing:
kubeadm-init
#### Since when has it been failing:
06-22
#### Testgrid link:
https://k8s-testgrid.appspot.com/sig-cluster-lifecycle-kubeadm#kubeadm-kinder-latest
#### Reason for failure:
```
[preflight] Some fatal errors occurred:
[ERROR ImagePull]: failed to pull image k8s.gcr.io/kube-apiserver:v1.22.0-alpha.3.377_6dd9deea3daa98: output: time="2021-06-22T17:39:50Z" level=fatal msg="pulling image failed: rpc error: code = NotFound desc = failed to pull and unpack image \"k8s.gcr.io/kube-apiserver:v1.22.0-alpha.3.377_6dd9deea3daa98\": failed to resolve reference \"k8s.gcr.io/kube-apiserver:v1.22.0-alpha.3.377_6dd9deea3daa98\": k8s.gcr.io/kube-apiserver:v1.22.0-alpha.3.377_6dd9deea3daa98: not found"
, error: exit status 1
[ERROR ImagePull]: failed to pull image k8s.gcr.io/kube-controller-manager:v1.22.0-alpha.3.377_6dd9deea3daa98: output: time="2021-06-22T17:39:50Z" level=fatal msg="pulling image failed: rpc error: code = NotFound desc = failed to pull and unpack image \"k8s.gcr.io/kube-controller-manager:v1.22.0-alpha.3.377_6dd9deea3daa98\": failed to resolve reference \"k8s.gcr.io/kube-controller-manager:v1.22.0-alpha.3.377_6dd9deea3daa98\": k8s.gcr.io/kube-controller-manager:v1.22.0-alpha.3.377_6dd9deea3daa98: not found"
, error: exit status 1
[ERROR ImagePull]: failed to pull image k8s.gcr.io/kube-scheduler:v1.22.0-alpha.3.377_6dd9deea3daa98: output: time="2021-06-22T17:39:50Z" level=fatal msg="pulling image failed: rpc error: code = NotFound desc = failed to pull and unpack image \"k8s.gcr.io/kube-scheduler:v1.22.0-alpha.3.377_6dd9deea3daa98\": failed to resolve reference \"k8s.gcr.io/kube-scheduler:v1.22.0-alpha.3.377_6dd9deea3daa98\": k8s.gcr.io/kube-scheduler:v1.22.0-alpha.3.377_6dd9deea3daa98: not found"
, error: exit status 1
[ERROR ImagePull]: failed to pull image k8s.gcr.io/kube-proxy:v1.22.0-alpha.3.377_6dd9deea3daa98: output: time="2021-06-22T17:39:51Z" level=fatal msg="pulling image failed: rpc error: code = NotFound desc = failed to pull and unpack image \"k8s.gcr.io/kube-proxy:v1.22.0-alpha.3.377_6dd9deea3daa98\": failed to resolve reference \"k8s.gcr.io/kube-proxy:v1.22.0-alpha.3.377_6dd9deea3daa98\": k8s.gcr.io/kube-proxy:v1.22.0-alpha.3.377_6dd9deea3daa98: not found"
, error: exit status 1
```
#### Anything else we need to know:
all kubeadm jobs are failing against k/k master.
is this failing after we merged?
https://github.com/kubernetes/kubernetes/pull/102901
|
1.0
|
[failing test] kubeadm-kinder-latest: task-03-init - <!-- Please only use this template for submitting reports about continuously failing tests or jobs in Kubernetes CI -->
#### Which jobs are failing:
https://k8s-testgrid.appspot.com/sig-cluster-lifecycle-kubeadm#kubeadm-kinder-latest
#### Which test(s) are failing:
kubeadm-init
#### Since when has it been failing:
06-22
#### Testgrid link:
https://k8s-testgrid.appspot.com/sig-cluster-lifecycle-kubeadm#kubeadm-kinder-latest
#### Reason for failure:
```
[preflight] Some fatal errors occurred:
[ERROR ImagePull]: failed to pull image k8s.gcr.io/kube-apiserver:v1.22.0-alpha.3.377_6dd9deea3daa98: output: time="2021-06-22T17:39:50Z" level=fatal msg="pulling image failed: rpc error: code = NotFound desc = failed to pull and unpack image \"k8s.gcr.io/kube-apiserver:v1.22.0-alpha.3.377_6dd9deea3daa98\": failed to resolve reference \"k8s.gcr.io/kube-apiserver:v1.22.0-alpha.3.377_6dd9deea3daa98\": k8s.gcr.io/kube-apiserver:v1.22.0-alpha.3.377_6dd9deea3daa98: not found"
, error: exit status 1
[ERROR ImagePull]: failed to pull image k8s.gcr.io/kube-controller-manager:v1.22.0-alpha.3.377_6dd9deea3daa98: output: time="2021-06-22T17:39:50Z" level=fatal msg="pulling image failed: rpc error: code = NotFound desc = failed to pull and unpack image \"k8s.gcr.io/kube-controller-manager:v1.22.0-alpha.3.377_6dd9deea3daa98\": failed to resolve reference \"k8s.gcr.io/kube-controller-manager:v1.22.0-alpha.3.377_6dd9deea3daa98\": k8s.gcr.io/kube-controller-manager:v1.22.0-alpha.3.377_6dd9deea3daa98: not found"
, error: exit status 1
[ERROR ImagePull]: failed to pull image k8s.gcr.io/kube-scheduler:v1.22.0-alpha.3.377_6dd9deea3daa98: output: time="2021-06-22T17:39:50Z" level=fatal msg="pulling image failed: rpc error: code = NotFound desc = failed to pull and unpack image \"k8s.gcr.io/kube-scheduler:v1.22.0-alpha.3.377_6dd9deea3daa98\": failed to resolve reference \"k8s.gcr.io/kube-scheduler:v1.22.0-alpha.3.377_6dd9deea3daa98\": k8s.gcr.io/kube-scheduler:v1.22.0-alpha.3.377_6dd9deea3daa98: not found"
, error: exit status 1
[ERROR ImagePull]: failed to pull image k8s.gcr.io/kube-proxy:v1.22.0-alpha.3.377_6dd9deea3daa98: output: time="2021-06-22T17:39:51Z" level=fatal msg="pulling image failed: rpc error: code = NotFound desc = failed to pull and unpack image \"k8s.gcr.io/kube-proxy:v1.22.0-alpha.3.377_6dd9deea3daa98\": failed to resolve reference \"k8s.gcr.io/kube-proxy:v1.22.0-alpha.3.377_6dd9deea3daa98\": k8s.gcr.io/kube-proxy:v1.22.0-alpha.3.377_6dd9deea3daa98: not found"
, error: exit status 1
```
#### Anything else we need to know:
all kubeadm jobs are failing against k/k master.
is this failing after we merged?
https://github.com/kubernetes/kubernetes/pull/102901
|
non_process
|
kubeadm kinder latest task init which jobs are failing which test s are failing kubeadm init since when has it been failing testgrid link reason for failure some fatal errors occurred failed to pull image gcr io kube apiserver alpha output time level fatal msg pulling image failed rpc error code notfound desc failed to pull and unpack image gcr io kube apiserver alpha failed to resolve reference gcr io kube apiserver alpha gcr io kube apiserver alpha not found error exit status failed to pull image gcr io kube controller manager alpha output time level fatal msg pulling image failed rpc error code notfound desc failed to pull and unpack image gcr io kube controller manager alpha failed to resolve reference gcr io kube controller manager alpha gcr io kube controller manager alpha not found error exit status failed to pull image gcr io kube scheduler alpha output time level fatal msg pulling image failed rpc error code notfound desc failed to pull and unpack image gcr io kube scheduler alpha failed to resolve reference gcr io kube scheduler alpha gcr io kube scheduler alpha not found error exit status failed to pull image gcr io kube proxy alpha output time level fatal msg pulling image failed rpc error code notfound desc failed to pull and unpack image gcr io kube proxy alpha failed to resolve reference gcr io kube proxy alpha gcr io kube proxy alpha not found error exit status anything else we need to know all kubeadm jobs are failing against k k master is this failing after we merged
| 0
|
366,195
| 10,818,220,444
|
IssuesEvent
|
2019-11-08 11:30:44
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
mega.nz - see bug description
|
browser-fenix engine-gecko priority-important
|
<!-- @browser: Firefox Mobile 70.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://mega.nz/fm/r5pX0QbR
**Browser / Version**: Firefox Mobile 70.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: Streaming via Mega lags device a lot
**Steps to Reproduce**:
I played the video fullscreen and the phone will start to stutter and lag until it just crash.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
mega.nz - see bug description - <!-- @browser: Firefox Mobile 70.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://mega.nz/fm/r5pX0QbR
**Browser / Version**: Firefox Mobile 70.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: Streaming via Mega lags device a lot
**Steps to Reproduce**:
I played the video fullscreen and the phone will start to stutter and lag until it just crash.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
mega nz see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description streaming via mega lags device a lot steps to reproduce i played the video fullscreen and the phone will start to stutter and lag until it just crash browser configuration none from with ❤️
| 0
|
13,602
| 16,189,808,062
|
IssuesEvent
|
2021-05-04 06:38:05
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Improve diagnostics for timeouts in Migrate
|
process/candidate team/migrations
|
## Problem
People are encountering confusing database timeout messages
## Suggested solution
Include more information about the origin of the timeouts in the error messages, as well as how to remedy them.
## Alternatives
Increase the default timeout for migrate. We may want to do this regardless.
## Additional context
Related issues:
- https://github.com/prisma/prisma/issues/6768
- https://github.com/prisma/prisma/issues/6224
- ...
|
1.0
|
Improve diagnostics for timeouts in Migrate - ## Problem
People are encountering confusing database timeout messages
## Suggested solution
Include more information about the origin of the timeouts in the error messages, as well as how to remedy them.
## Alternatives
Increase the default timeout for migrate. We may want to do this regardless.
## Additional context
Related issues:
- https://github.com/prisma/prisma/issues/6768
- https://github.com/prisma/prisma/issues/6224
- ...
|
process
|
improve diagnostics for timeouts in migrate problem people are encountering confusing database timeout messages suggested solution include more information about the origin of the timeouts in the error messages as well as how to remedy them alternatives increase the default timeout for migrate we may want to do this regardless additional context related issues
| 1
|
4,545
| 7,375,045,348
|
IssuesEvent
|
2018-03-13 22:29:27
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
App won't work after deployment to Azure
|
cosmos-db cxp in-process product-question resolved-by-customer triaged
|
I added requirements (flask==0.12.1, pymongo==3.5.1) and my cosmos db connection string and deployment to Azure went successfull. But anyway, something crashes out there as app website is not accessible. Any suggestions what trouble might be?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: e178cd44-55d5-23ab-d965-94c47ada47dc
* Version Independent ID: c0f2ba6b-9c90-8975-6796-009417be4ffa
* Content: [Azure Cosmos DB: Build a Flask web app with Python and the Azure Cosmos DB MongoDB API | Microsoft Docs](https://docs.microsoft.com/en-us/azure/cosmos-db/create-mongodb-flask)
* Content Source: [articles/cosmos-db/create-mongodb-flask.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cosmos-db/create-mongodb-flask.md)
* Service: **cosmos-db**
* GitHub Login: @hshapiro
* Microsoft Alias: **hshapiro**
|
1.0
|
App won't work after deployment to Azure - I added requirements (flask==0.12.1, pymongo==3.5.1) and my cosmos db connection string and deployment to Azure went successfull. But anyway, something crashes out there as app website is not accessible. Any suggestions what trouble might be?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: e178cd44-55d5-23ab-d965-94c47ada47dc
* Version Independent ID: c0f2ba6b-9c90-8975-6796-009417be4ffa
* Content: [Azure Cosmos DB: Build a Flask web app with Python and the Azure Cosmos DB MongoDB API | Microsoft Docs](https://docs.microsoft.com/en-us/azure/cosmos-db/create-mongodb-flask)
* Content Source: [articles/cosmos-db/create-mongodb-flask.md](https://github.com/Microsoft/azure-docs/blob/master/articles/cosmos-db/create-mongodb-flask.md)
* Service: **cosmos-db**
* GitHub Login: @hshapiro
* Microsoft Alias: **hshapiro**
|
process
|
app won t work after deployment to azure i added requirements flask pymongo and my cosmos db connection string and deployment to azure went successfull but anyway something crashes out there as app website is not accessible any suggestions what trouble might be document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service cosmos db github login hshapiro microsoft alias hshapiro
| 1
|
17,244
| 23,023,894,994
|
IssuesEvent
|
2022-07-22 07:40:35
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
[processor/transform] The `set` function cannot set a field to `nil`
|
bug good first issue priority:p2 processor/transform
|
**Describe the bug**
The Telemetry Query Language has the concept of `nil`, but the set function doesn't understand it. You cannot do `set(attribute["test"], nil)`. This is because of the [nil check in the set function](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/4d5ee4a1b9ec41b182cd4bd6990e3e26a6ebb535/processor/transformprocessor/internal/common/func_set.go#L20) and also because of the setAttr function in each signal's context ([traces for example](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/4d5ee4a1b9ec41b182cd4bd6990e3e26a6ebb535/processor/transformprocessor/internal/traces/traces.go#L548)) doesn't know how to handle `nil`.
The TQL/processor need updated to better support setting nil.
|
1.0
|
[processor/transform] The `set` function cannot set a field to `nil` - **Describe the bug**
The Telemetry Query Language has the concept of `nil`, but the set function doesn't understand it. You cannot do `set(attribute["test"], nil)`. This is because of the [nil check in the set function](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/4d5ee4a1b9ec41b182cd4bd6990e3e26a6ebb535/processor/transformprocessor/internal/common/func_set.go#L20) and also because of the setAttr function in each signal's context ([traces for example](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/4d5ee4a1b9ec41b182cd4bd6990e3e26a6ebb535/processor/transformprocessor/internal/traces/traces.go#L548)) doesn't know how to handle `nil`.
The TQL/processor need updated to better support setting nil.
|
process
|
the set function cannot set a field to nil describe the bug the telemetry query language has the concept of nil but the set function doesn t understand it you cannot do set attribute nil this is because of the and also because of the setattr function in each signal s context doesn t know how to handle nil the tql processor need updated to better support setting nil
| 1
|
92,096
| 15,856,860,205
|
IssuesEvent
|
2021-04-08 03:21:31
|
tlubitz/hv_lubitz
|
https://api.github.com/repos/tlubitz/hv_lubitz
|
opened
|
CVE-2020-13596 (Medium) detected in Django-2.2.1.tar.gz
|
security vulnerability
|
## CVE-2020-13596 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-2.2.1.tar.gz</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/fd/70/36c08f4c3b22523173b3a5e21fbdaa137bdb1722b76f356e0e2d5d8aa645/Django-2.2.1.tar.gz">https://files.pythonhosted.org/packages/fd/70/36c08f4c3b22523173b3a5e21fbdaa137bdb1722b76f356e0e2d5d8aa645/Django-2.2.1.tar.gz</a></p>
<p>Path to dependency file: /hv_lubitz/requirements.txt</p>
<p>Path to vulnerable library: teSource-ArchiveExtractor_14b6631b-bd5c-4a6c-9e80-2d2264f0c033/20190507134225_14542/20190507134130_depth_0/6/Django-2.2.1.tar/Django-2.2.1</p>
<p>
Dependency Hierarchy:
- :x: **Django-2.2.1.tar.gz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in Django 2.2 before 2.2.13 and 3.0 before 3.0.7. Query parameters generated by the Django admin ForeignKeyRawIdWidget were not properly URL encoded, leading to a possibility of an XSS attack.
<p>Publish Date: 2020-06-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13596>CVE-2020-13596</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.djangoproject.com/weblog/2020/jun/03/security-releases/">https://www.djangoproject.com/weblog/2020/jun/03/security-releases/</a></p>
<p>Release Date: 2020-06-03</p>
<p>Fix Resolution: 3.0.7,2.2.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-13596 (Medium) detected in Django-2.2.1.tar.gz - ## CVE-2020-13596 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-2.2.1.tar.gz</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/fd/70/36c08f4c3b22523173b3a5e21fbdaa137bdb1722b76f356e0e2d5d8aa645/Django-2.2.1.tar.gz">https://files.pythonhosted.org/packages/fd/70/36c08f4c3b22523173b3a5e21fbdaa137bdb1722b76f356e0e2d5d8aa645/Django-2.2.1.tar.gz</a></p>
<p>Path to dependency file: /hv_lubitz/requirements.txt</p>
<p>Path to vulnerable library: teSource-ArchiveExtractor_14b6631b-bd5c-4a6c-9e80-2d2264f0c033/20190507134225_14542/20190507134130_depth_0/6/Django-2.2.1.tar/Django-2.2.1</p>
<p>
Dependency Hierarchy:
- :x: **Django-2.2.1.tar.gz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in Django 2.2 before 2.2.13 and 3.0 before 3.0.7. Query parameters generated by the Django admin ForeignKeyRawIdWidget were not properly URL encoded, leading to a possibility of an XSS attack.
<p>Publish Date: 2020-06-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13596>CVE-2020-13596</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.djangoproject.com/weblog/2020/jun/03/security-releases/">https://www.djangoproject.com/weblog/2020/jun/03/security-releases/</a></p>
<p>Release Date: 2020-06-03</p>
<p>Fix Resolution: 3.0.7,2.2.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in django tar gz cve medium severity vulnerability vulnerable library django tar gz a high level python web framework that encourages rapid development and clean pragmatic design library home page a href path to dependency file hv lubitz requirements txt path to vulnerable library tesource archiveextractor depth django tar django dependency hierarchy x django tar gz vulnerable library vulnerability details an issue was discovered in django before and before query parameters generated by the django admin foreignkeyrawidwidget were not properly url encoded leading to a possibility of an xss attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
17,151
| 22,702,452,125
|
IssuesEvent
|
2022-07-05 11:58:39
|
camunda/zeebe
|
https://api.github.com/repos/camunda/zeebe
|
closed
|
Test start instance anywhere for supported element types
|
team/process-automation area/test
|
The start process instance anywhere feature (#9366) can be used to start the instance at specific elements. Not all elements are supported. Those that are supported should be covered by a parameterized test.
|
1.0
|
Test start instance anywhere for supported element types - The start process instance anywhere feature (#9366) can be used to start the instance at specific elements. Not all elements are supported. Those that are supported should be covered by a parameterized test.
|
process
|
test start instance anywhere for supported element types the start process instance anywhere feature can be used to start the instance at specific elements not all elements are supported those that are supported should be covered by a parameterized test
| 1
|
169,306
| 20,834,549,163
|
IssuesEvent
|
2022-03-20 01:07:15
|
benlazarine/clank
|
https://api.github.com/repos/benlazarine/clank
|
opened
|
CVE-2020-14332 (Medium) detected in ansible-2.4.2.0.tar.gz
|
security vulnerability
|
## CVE-2020-14332 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.4.2.0.tar.gz</b></p></summary>
<p>Radically simple IT automation</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/4f/65/ae3ad8589c38f9e04ebc8a824c2880eb4f9e603a1f62b5f5a3f938e524b0/ansible-2.4.2.0.tar.gz">https://files.pythonhosted.org/packages/4f/65/ae3ad8589c38f9e04ebc8a824c2880eb4f9e603a1f62b5f5a3f938e524b0/ansible-2.4.2.0.tar.gz</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **ansible-2.4.2.0.tar.gz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Ansible Engine when using module_args. Tasks executed with check mode (--check-mode) do not properly neutralize sensitive data exposed in the event data. This flaw allows unauthorized users to read this data. The highest threat from this vulnerability is to confidentiality.
<p>Publish Date: 2020-09-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14332>CVE-2020-14332</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-14332">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-14332</a></p>
<p>Release Date: 2020-09-11</p>
<p>Fix Resolution: 2.8.14,2.9.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-14332 (Medium) detected in ansible-2.4.2.0.tar.gz - ## CVE-2020-14332 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.4.2.0.tar.gz</b></p></summary>
<p>Radically simple IT automation</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/4f/65/ae3ad8589c38f9e04ebc8a824c2880eb4f9e603a1f62b5f5a3f938e524b0/ansible-2.4.2.0.tar.gz">https://files.pythonhosted.org/packages/4f/65/ae3ad8589c38f9e04ebc8a824c2880eb4f9e603a1f62b5f5a3f938e524b0/ansible-2.4.2.0.tar.gz</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **ansible-2.4.2.0.tar.gz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Ansible Engine when using module_args. Tasks executed with check mode (--check-mode) do not properly neutralize sensitive data exposed in the event data. This flaw allows unauthorized users to read this data. The highest threat from this vulnerability is to confidentiality.
<p>Publish Date: 2020-09-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14332>CVE-2020-14332</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-14332">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-14332</a></p>
<p>Release Date: 2020-09-11</p>
<p>Fix Resolution: 2.8.14,2.9.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in ansible tar gz cve medium severity vulnerability vulnerable library ansible tar gz radically simple it automation library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy x ansible tar gz vulnerable library vulnerability details a flaw was found in the ansible engine when using module args tasks executed with check mode check mode do not properly neutralize sensitive data exposed in the event data this flaw allows unauthorized users to read this data the highest threat from this vulnerability is to confidentiality publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
111,029
| 4,448,782,674
|
IssuesEvent
|
2016-08-22 02:27:01
|
muicss/mui
|
https://api.github.com/repos/muicss/mui
|
closed
|
Muicss Input with floating label does not respond to programmatic changes (React)
|
bug help wanted priority: high
|
-Our project is written in React-Redux and uses the mui-css library.
We have an mui-css input that receives a value in one of two ways:
It can either receive a value like a regular input (when the user types into the field), or the value is passed in programmatically from a backend calculation. This input has a floating label. When the value is changed programmatically, the floating label does not respond. If the value is cleared out programmatically- the label remains suspended on top of the input, and does not return to it's original place (even when the user clicks outside the field). When a value is entered programmatically, the label does not lift, it remains behind the text.
How can we cause the floating label to respond to our programmatic changes?
Shouldn't the label be responding on it's own?
Here is a link to a sample project where we reproduced the bug: http://www.webpackbin.com/NJtdjOVFZ
Thank you in advance for your help.
|
1.0
|
Muicss Input with floating label does not respond to programmatic changes (React) - -Our project is written in React-Redux and uses the mui-css library.
We have an mui-css input that receives a value in one of two ways:
It can either receive a value like a regular input (when the user types into the field), or the value is passed in programmatically from a backend calculation. This input has a floating label. When the value is changed programmatically, the floating label does not respond. If the value is cleared out programmatically- the label remains suspended on top of the input, and does not return to it's original place (even when the user clicks outside the field). When a value is entered programmatically, the label does not lift, it remains behind the text.
How can we cause the floating label to respond to our programmatic changes?
Shouldn't the label be responding on it's own?
Here is a link to a sample project where we reproduced the bug: http://www.webpackbin.com/NJtdjOVFZ
Thank you in advance for your help.
|
non_process
|
muicss input with floating label does not respond to programmatic changes react our project is written in react redux and uses the mui css library we have an mui css input that receives a value in one of two ways it can either receive a value like a regular input when the user types into the field or the value is passed in programmatically from a backend calculation this input has a floating label when the value is changed programmatically the floating label does not respond if the value is cleared out programmatically the label remains suspended on top of the input and does not return to it s original place even when the user clicks outside the field when a value is entered programmatically the label does not lift it remains behind the text how can we cause the floating label to respond to our programmatic changes shouldn t the label be responding on it s own here is a link to a sample project where we reproduced the bug thank you in advance for your help
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.