Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
33,376 | 7,106,978,963 | IssuesEvent | 2018-01-16 18:21:13 | netty/netty | https://api.github.com/repos/netty/netty | closed | Wrapped direct ByteBufs do not respect the position of the underlying buffer | defect | ### Expected behavior
If I create a MappedByteBuffer then change it's position and limit then wrap that buffer, the ByteBuf should respect that position/offset.
### Actual behavior
`ReadOnlyUnsafeDirectByteBuf` does not respect the position of the underlying MappedByteBuffer and will start at the base address rather than the position.
### Steps to reproduce
```scala
val f = java.nio.channels.FileChannel.open(java.nio.file.FileSystems.getDefault.getPath("somefile"))
val m = f.map(java.nio.channels.FileChannel.MapMode.READ_ONLY, 0, f.size)
val b1 = io.netty.buffer.Unpooled.wrappedBuffer(m)
val dup = m.duplicate
dup.position(2)
dup.limit(4)
val b2 = io.netty.buffer.Unpooled.wrappedBuffer(dup)
b1.slice(2, 2) == b2 // this should be true, but is not because b2 will start at the wrapped buffer's 0 index
```
### Minimal yet complete reproducer code (or URL to code)
See above
### Netty version
4.1.19.Final
### JVM version (e.g. `java -version`)
java version "1.8.0_121"
Java(TM) SE Runtime Environment (build 1.8.0_121-b13)
Java HotSpot(TM) 64-Bit Server VM (build 25.121-b13, mixed mode)
### OS version (e.g. `uname -a`)
Linux sprsquish 4.10.0-42-generic #46-Ubuntu SMP Mon Dec 4 14:38:01 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux
| 1.0 | Wrapped direct ByteBufs do not respect the position of the underlying buffer - ### Expected behavior
If I create a MappedByteBuffer then change it's position and limit then wrap that buffer, the ByteBuf should respect that position/offset.
### Actual behavior
`ReadOnlyUnsafeDirectByteBuf` does not respect the position of the underlying MappedByteBuffer and will start at the base address rather than the position.
### Steps to reproduce
```scala
val f = java.nio.channels.FileChannel.open(java.nio.file.FileSystems.getDefault.getPath("somefile"))
val m = f.map(java.nio.channels.FileChannel.MapMode.READ_ONLY, 0, f.size)
val b1 = io.netty.buffer.Unpooled.wrappedBuffer(m)
val dup = m.duplicate
dup.position(2)
dup.limit(4)
val b2 = io.netty.buffer.Unpooled.wrappedBuffer(dup)
b1.slice(2, 2) == b2 // this should be true, but is not because b2 will start at the wrapped buffer's 0 index
```
### Minimal yet complete reproducer code (or URL to code)
See above
### Netty version
4.1.19.Final
### JVM version (e.g. `java -version`)
java version "1.8.0_121"
Java(TM) SE Runtime Environment (build 1.8.0_121-b13)
Java HotSpot(TM) 64-Bit Server VM (build 25.121-b13, mixed mode)
### OS version (e.g. `uname -a`)
Linux sprsquish 4.10.0-42-generic #46-Ubuntu SMP Mon Dec 4 14:38:01 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux
| defect | wrapped direct bytebufs do not respect the position of the underlying buffer expected behavior if i create a mappedbytebuffer then change it s position and limit then wrap that buffer the bytebuf should respect that position offset actual behavior readonlyunsafedirectbytebuf does not respect the position of the underlying mappedbytebuffer and will start at the base address rather than the position steps to reproduce scala val f java nio channels filechannel open java nio file filesystems getdefault getpath somefile val m f map java nio channels filechannel mapmode read only f size val io netty buffer unpooled wrappedbuffer m val dup m duplicate dup position dup limit val io netty buffer unpooled wrappedbuffer dup slice this should be true but is not because will start at the wrapped buffer s index minimal yet complete reproducer code or url to code see above netty version final jvm version e g java version java version java tm se runtime environment build java hotspot tm bit server vm build mixed mode os version e g uname a linux sprsquish generic ubuntu smp mon dec utc gnu linux | 1 |
43,855 | 11,865,628,145 | IssuesEvent | 2020-03-26 01:02:20 | richgel999/jpeg-compressor | https://api.github.com/repos/richgel999/jpeg-compressor | closed | Shortened RGBA->YCC code | Priority-Medium Type-Defect auto-migrated | ```
Attached patch
```
Original issue reported on code.google.com by `pornel...@gmail.com` on 29 Sep 2013 at 5:38
Attachments:
- [0001-Use-standard-header.patch](https://storage.googleapis.com/google-code-attachments/jpeg-compressor/issue-3/comment-0/0001-Use-standard-header.patch)
- [0002-Removed-duplicate-RGB-RGBA-conversions.patch](https://storage.googleapis.com/google-code-attachments/jpeg-compressor/issue-3/comment-0/0002-Removed-duplicate-RGB-RGBA-conversions.patch)
| 1.0 | Shortened RGBA->YCC code - ```
Attached patch
```
Original issue reported on code.google.com by `pornel...@gmail.com` on 29 Sep 2013 at 5:38
Attachments:
- [0001-Use-standard-header.patch](https://storage.googleapis.com/google-code-attachments/jpeg-compressor/issue-3/comment-0/0001-Use-standard-header.patch)
- [0002-Removed-duplicate-RGB-RGBA-conversions.patch](https://storage.googleapis.com/google-code-attachments/jpeg-compressor/issue-3/comment-0/0002-Removed-duplicate-RGB-RGBA-conversions.patch)
| defect | shortened rgba ycc code attached patch original issue reported on code google com by pornel gmail com on sep at attachments | 1 |
517,700 | 15,018,657,154 | IssuesEvent | 2021-02-01 12:32:00 | eventespresso/barista | https://api.github.com/repos/eventespresso/barista | closed | Move New Default Tickets UI Behind Feature Flag | C: UI/UX 🚽 D: EDTR ✏️ P2: HIGH priority 😮 S:1 new 👶🏻 T: task 🧹 | Title says it all. We need to move ALL aspects of the new Default Tickets UI behind a Feature Flag (FF) so that it is not accessible to users unless they possess the appropriate permission (let's call it "use_default_ticket_manager") | 1.0 | Move New Default Tickets UI Behind Feature Flag - Title says it all. We need to move ALL aspects of the new Default Tickets UI behind a Feature Flag (FF) so that it is not accessible to users unless they possess the appropriate permission (let's call it "use_default_ticket_manager") | non_defect | move new default tickets ui behind feature flag title says it all we need to move all aspects of the new default tickets ui behind a feature flag ff so that it is not accessible to users unless they possess the appropriate permission let s call it use default ticket manager | 0 |
47,465 | 13,056,197,234 | IssuesEvent | 2020-07-30 03:57:40 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | qt ports should check for opengl (Trac #596) | Migrated from Trac cmake defect | need to verify that qt isn't getting silently built w/o opengl support. People are getting the ports installed and then failing on OPENGL_FOUND when configuring the build of the glshovel... this should be impossible.
Migrated from https://code.icecube.wisc.edu/ticket/596
```json
{
"status": "closed",
"changetime": "2010-02-15T21:42:56",
"description": "need to verify that qt isn't getting silently built w/o opengl support. People are getting the ports installed and then failing on OPENGL_FOUND when configuring the build of the glshovel... this should be impossible.",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1266270176000000",
"component": "cmake",
"summary": "qt ports should check for opengl",
"priority": "normal",
"keywords": "",
"time": "2010-02-09T15:34:42",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| 1.0 | qt ports should check for opengl (Trac #596) - need to verify that qt isn't getting silently built w/o opengl support. People are getting the ports installed and then failing on OPENGL_FOUND when configuring the build of the glshovel... this should be impossible.
Migrated from https://code.icecube.wisc.edu/ticket/596
```json
{
"status": "closed",
"changetime": "2010-02-15T21:42:56",
"description": "need to verify that qt isn't getting silently built w/o opengl support. People are getting the ports installed and then failing on OPENGL_FOUND when configuring the build of the glshovel... this should be impossible.",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1266270176000000",
"component": "cmake",
"summary": "qt ports should check for opengl",
"priority": "normal",
"keywords": "",
"time": "2010-02-09T15:34:42",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| defect | qt ports should check for opengl trac need to verify that qt isn t getting silently built w o opengl support people are getting the ports installed and then failing on opengl found when configuring the build of the glshovel this should be impossible migrated from json status closed changetime description need to verify that qt isn t getting silently built w o opengl support people are getting the ports installed and then failing on opengl found when configuring the build of the glshovel this should be impossible reporter troy cc resolution fixed ts component cmake summary qt ports should check for opengl priority normal keywords time milestone owner troy type defect | 1 |
98,786 | 20,798,233,514 | IssuesEvent | 2022-03-17 11:24:36 | Onelinerhub/onelinerhub | https://api.github.com/repos/Onelinerhub/onelinerhub | opened | Short solution needed: "Custom log format" (nginx) | help wanted good first issue code nginx | Please help us write most modern and shortest code solution for this issue:
**Custom log format** (technology: [nginx](https://onelinerhub.com/nginx))
### Fast way
Just write the code solution in the comments.
### Prefered way
1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox).
2. Don't forget to use comments to make solution explained.
3. Link to this issue in comments of pull request. | 1.0 | Short solution needed: "Custom log format" (nginx) - Please help us write most modern and shortest code solution for this issue:
**Custom log format** (technology: [nginx](https://onelinerhub.com/nginx))
### Fast way
Just write the code solution in the comments.
### Prefered way
1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox).
2. Don't forget to use comments to make solution explained.
3. Link to this issue in comments of pull request. | non_defect | short solution needed custom log format nginx please help us write most modern and shortest code solution for this issue custom log format technology fast way just write the code solution in the comments prefered way create pull request with a new code file inside don t forget to use comments to make solution explained link to this issue in comments of pull request | 0 |
12,037 | 3,250,687,281 | IssuesEvent | 2015-10-19 03:19:47 | kumulsoft/Fixed-Assets | https://api.github.com/repos/kumulsoft/Fixed-Assets | closed | SETUP >> Manage Staff. Small Adjustments to the entry screen | bug enhancement Fixed Ready for testing UI | 1. Rename section 'Contact Information' to 'Staff Information'
2. Rename label 'Employee Name' to 'Staff Name'
3. Contact Type must be defaulted to 'Staff' and Read Only
4. Arrange Centre and Location to be Side by Side (like Division and Section)
5. Remove/Hide the Address section
6. Position Field too is missing, put it back

| 1.0 | SETUP >> Manage Staff. Small Adjustments to the entry screen - 1. Rename section 'Contact Information' to 'Staff Information'
2. Rename label 'Employee Name' to 'Staff Name'
3. Contact Type must be defaulted to 'Staff' and Read Only
4. Arrange Centre and Location to be Side by Side (like Division and Section)
5. Remove/Hide the Address section
6. Position Field too is missing, put it back

| non_defect | setup manage staff small adjustments to the entry screen rename section contact information to staff information rename label employee name to staff name contact type must be defaulted to staff and read only arrange centre and location to be side by side like division and section remove hide the address section position field too is missing put it back | 0 |
93,971 | 27,084,849,079 | IssuesEvent | 2023-02-14 16:18:20 | camunda/zeebe | https://api.github.com/repos/camunda/zeebe | closed | Maven downloads non-camunda artifacts from camunda-nexus instead of maven central | area/build component/build-pipeline | Since https://github.com/camunda/zeebe/pull/11498 for merged we observed maven to not just download camunda artifacts from the camunda nexus repo but also non camunda artifacts, e.g. https://github.com/camunda/zeebe/actions/runs/4173174037/jobs/7229130526
```
[INFO] --- maven-install-plugin:2.4:install (default-install) @ zeebe-journal ---
[INFO] Installing /runner/_work/zeebe/zeebe/journal/target/zeebe-journal-8.2.0-SNAPSHOT.jar to /home/runner/.m2/repository/io/camunda/zeebe-journal/8.2.0-SNAPSHOT/zeebe-journal-8.2.0-SNAPSHOT.jar
[INFO] Installing /runner/_work/zeebe/zeebe/journal/.flattened-pom.xml to /home/runner/.m2/repository/io/camunda/zeebe-journal/8.2.0-SNAPSHOT/zeebe-journal-8.2.0-SNAPSHOT.pom
[INFO] Downloaded from camunda-nexus: https://repository.nexus.camunda.cloud/content/groups/internal/io/netty/netty-handler/4.1.86.Final/netty-handler-4.1.86.Final.pom (4.5 kB at 14 kB/s)
[INFO] Downloading from camunda-nexus: https://repository.nexus.camunda.cloud/content/groups/internal/io/netty/netty-transport-native-unix-common/4.1.86.Final/netty-transport-native-unix-common-4.1.86.Final.pom
```
On Github hosted runners we observed the download from this maven repo to be slow sometimes, ideally only camunda artifacts should get downloaded from this mirror. | 2.0 | Maven downloads non-camunda artifacts from camunda-nexus instead of maven central - Since https://github.com/camunda/zeebe/pull/11498 for merged we observed maven to not just download camunda artifacts from the camunda nexus repo but also non camunda artifacts, e.g. https://github.com/camunda/zeebe/actions/runs/4173174037/jobs/7229130526
```
[INFO] --- maven-install-plugin:2.4:install (default-install) @ zeebe-journal ---
[INFO] Installing /runner/_work/zeebe/zeebe/journal/target/zeebe-journal-8.2.0-SNAPSHOT.jar to /home/runner/.m2/repository/io/camunda/zeebe-journal/8.2.0-SNAPSHOT/zeebe-journal-8.2.0-SNAPSHOT.jar
[INFO] Installing /runner/_work/zeebe/zeebe/journal/.flattened-pom.xml to /home/runner/.m2/repository/io/camunda/zeebe-journal/8.2.0-SNAPSHOT/zeebe-journal-8.2.0-SNAPSHOT.pom
[INFO] Downloaded from camunda-nexus: https://repository.nexus.camunda.cloud/content/groups/internal/io/netty/netty-handler/4.1.86.Final/netty-handler-4.1.86.Final.pom (4.5 kB at 14 kB/s)
[INFO] Downloading from camunda-nexus: https://repository.nexus.camunda.cloud/content/groups/internal/io/netty/netty-transport-native-unix-common/4.1.86.Final/netty-transport-native-unix-common-4.1.86.Final.pom
```
On Github hosted runners we observed the download from this maven repo to be slow sometimes, ideally only camunda artifacts should get downloaded from this mirror. | non_defect | maven downloads non camunda artifacts from camunda nexus instead of maven central since for merged we observed maven to not just download camunda artifacts from the camunda nexus repo but also non camunda artifacts e g maven install plugin install default install zeebe journal installing runner work zeebe zeebe journal target zeebe journal snapshot jar to home runner repository io camunda zeebe journal snapshot zeebe journal snapshot jar installing runner work zeebe zeebe journal flattened pom xml to home runner repository io camunda zeebe journal snapshot zeebe journal snapshot pom downloaded from camunda nexus kb at kb s downloading from camunda nexus on github hosted runners we observed the download from this maven repo to be slow sometimes ideally only camunda artifacts should get downloaded from this mirror | 0 |
76,678 | 26,553,671,760 | IssuesEvent | 2023-01-20 10:07:17 | decentraland/unity-renderer | https://api.github.com/repos/decentraland/unity-renderer | opened | [QA] Cannot enter worlds on desktop when using a deeplink with new decoupled loading screen on | medium defect | Description:
With the new loading screen decoupled from kernel it is not possible to enter any of the Worlds directly on desktop by using a deep link - loading gets stuck at 0%. With the decoupled loading screen feature flag off the loading proceeds normally without issues. Issue occurs only on desktop clients and not on web.
Reproduction rate:
5/5
Steps to reproduce:
1. Launch the desktop client on the release/release-20230119 branch with `ENABLE_DECOUPLED_LOADING_SCREEN&realm=menduz.dcl.eth` parameters.
2. Log in either via a guest or a wallet account.
3. Observe the loading screen get stuck at 0%.
Results:
With the decoupled loading screen feature flag on loading gets stuck at 0% and refuses to proceed on worlds.
Expected results:
With the decoupled loading screen feature flag on loading proceeds normally on worlds.
Notes:
- Platforms: Windows and MacOS desktop, wallet and guest login.
- Issue occurs for the [release/release-20230119 branch](https://github.com/decentraland/explorer-desktop/pull/464).
- Please see the attached screenshot and video for reference.

https://images.zenhubusercontent.com/337227404/15ace7b0-103b-4ea1-9d95-e54668102dc5/2023_01_20_10_57_44.mp4
| 1.0 | [QA] Cannot enter worlds on desktop when using a deeplink with new decoupled loading screen on - Description:
With the new loading screen decoupled from kernel it is not possible to enter any of the Worlds directly on desktop by using a deep link - loading gets stuck at 0%. With the decoupled loading screen feature flag off the loading proceeds normally without issues. Issue occurs only on desktop clients and not on web.
Reproduction rate:
5/5
Steps to reproduce:
1. Launch the desktop client on the release/release-20230119 branch with `ENABLE_DECOUPLED_LOADING_SCREEN&realm=menduz.dcl.eth` parameters.
2. Log in either via a guest or a wallet account.
3. Observe the loading screen get stuck at 0%.
Results:
With the decoupled loading screen feature flag on loading gets stuck at 0% and refuses to proceed on worlds.
Expected results:
With the decoupled loading screen feature flag on loading proceeds normally on worlds.
Notes:
- Platforms: Windows and MacOS desktop, wallet and guest login.
- Issue occurs for the [release/release-20230119 branch](https://github.com/decentraland/explorer-desktop/pull/464).
- Please see the attached screenshot and video for reference.

https://images.zenhubusercontent.com/337227404/15ace7b0-103b-4ea1-9d95-e54668102dc5/2023_01_20_10_57_44.mp4
| defect | cannot enter worlds on desktop when using a deeplink with new decoupled loading screen on description with the new loading screen decoupled from kernel it is not possible to enter any of the worlds directly on desktop by using a deep link loading gets stuck at with the decoupled loading screen feature flag off the loading proceeds normally without issues issue occurs only on desktop clients and not on web reproduction rate steps to reproduce launch the desktop client on the release release branch with enable decoupled loading screen realm menduz dcl eth parameters log in either via a guest or a wallet account observe the loading screen get stuck at results with the decoupled loading screen feature flag on loading gets stuck at and refuses to proceed on worlds expected results with the decoupled loading screen feature flag on loading proceeds normally on worlds notes platforms windows and macos desktop wallet and guest login issue occurs for the please see the attached screenshot and video for reference | 1 |
40,919 | 10,225,606,945 | IssuesEvent | 2019-08-16 15:33:11 | google/auto | https://api.github.com/repos/google/auto | closed | Autofactory compilation fails on Java 8 | Component: factory P3 type=defect | [ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile (default-testCompile) on project auto-factory: Compilation failure: Compilation failure:
[ERROR] /home/zubzub/git/auto/factory/src/test/java/com/google/auto/factory/processor/AutoFactoryDeclarationTest.java:[18,71] package com.google.auto.factory.processor.AutoFactoryDeclaration does not exist
[ERROR] /home/zubzub/git/auto/factory/src/test/java/com/google/auto/factory/processor/AutoFactoryDeclarationTest.java:[18,1] static import only from classes and interfaces
| 1.0 | Autofactory compilation fails on Java 8 - [ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile (default-testCompile) on project auto-factory: Compilation failure: Compilation failure:
[ERROR] /home/zubzub/git/auto/factory/src/test/java/com/google/auto/factory/processor/AutoFactoryDeclarationTest.java:[18,71] package com.google.auto.factory.processor.AutoFactoryDeclaration does not exist
[ERROR] /home/zubzub/git/auto/factory/src/test/java/com/google/auto/factory/processor/AutoFactoryDeclarationTest.java:[18,1] static import only from classes and interfaces
| defect | autofactory compilation fails on java failed to execute goal org apache maven plugins maven compiler plugin testcompile default testcompile on project auto factory compilation failure compilation failure home zubzub git auto factory src test java com google auto factory processor autofactorydeclarationtest java package com google auto factory processor autofactorydeclaration does not exist home zubzub git auto factory src test java com google auto factory processor autofactorydeclarationtest java static import only from classes and interfaces | 1 |
68,064 | 21,468,818,139 | IssuesEvent | 2022-04-26 07:38:01 | klubcoin/lcn-mobile | https://api.github.com/repos/klubcoin/lcn-mobile | opened | Klubcoin app should navigate to Dashboard while press back handler button om Purchase Methods screen | Defect Should Have Task Major | ### **Description:**
Klubcoin app should navigate to Dashboard while press back handler button om Purchase Methods screen.
**Build Environment:** Staging Candidate Environment
**Affects Version:**
**Device Platform:** Android
**Device OS:** Android 12
**Test Device:** Samsung S21
### **Pre-condition:**
1. User successfully installed Klubcoin App
2. User has an existing Klubcoin Wallet Account
4. User is currently at Klubcoin Dashboard
### **Steps to Reproduce:**
1. Tap Hamburger Button
2. Tap Buy
3. Tap Back Handler button
### **Expected Result:**
Navigate to Dashboard.
### **Actual Result:**
Navigate to another Purchase Methods screen.
| 1.0 | Klubcoin app should navigate to Dashboard while press back handler button om Purchase Methods screen - ### **Description:**
Klubcoin app should navigate to Dashboard while press back handler button om Purchase Methods screen.
**Build Environment:** Staging Candidate Environment
**Affects Version:**
**Device Platform:** Android
**Device OS:** Android 12
**Test Device:** Samsung S21
### **Pre-condition:**
1. User successfully installed Klubcoin App
2. User has an existing Klubcoin Wallet Account
4. User is currently at Klubcoin Dashboard
### **Steps to Reproduce:**
1. Tap Hamburger Button
2. Tap Buy
3. Tap Back Handler button
### **Expected Result:**
Navigate to Dashboard.
### **Actual Result:**
Navigate to another Purchase Methods screen.
| defect | klubcoin app should navigate to dashboard while press back handler button om purchase methods screen description klubcoin app should navigate to dashboard while press back handler button om purchase methods screen build environment staging candidate environment affects version device platform android device os android test device samsung pre condition user successfully installed klubcoin app user has an existing klubcoin wallet account user is currently at klubcoin dashboard steps to reproduce tap hamburger button tap buy tap back handler button expected result navigate to dashboard actual result navigate to another purchase methods screen | 1 |
12,972 | 2,732,346,269 | IssuesEvent | 2015-04-17 04:48:04 | rasmus/fast-member | https://api.github.com/repos/rasmus/fast-member | closed | TypeAccessor.WriteSetter throws exception, if property setter is not public | auto-migrated Priority-Medium Type-Defect | ```
The method TypeAccessor.WriteSetter throws an ArgumentNullException, if the
target type has a property with a non-public setter. (Line 141:
http://code.google.com/p/fast-member/source/browse/FastMember/TypeAccessor.cs#14
1)
Example:
class TestTarget
{
public int Id { get; private set; }
}
I am not sure why `prop.GetSetMethod()` returns `null`, although
`prop.CanWrite` obviously returned `true` in line 129.
```
Original issue reported on code.google.com by `schuster...@gmail.com` on 27 Jan 2012 at 8:22 | 1.0 | TypeAccessor.WriteSetter throws exception, if property setter is not public - ```
The method TypeAccessor.WriteSetter throws an ArgumentNullException, if the
target type has a property with a non-public setter. (Line 141:
http://code.google.com/p/fast-member/source/browse/FastMember/TypeAccessor.cs#14
1)
Example:
class TestTarget
{
public int Id { get; private set; }
}
I am not sure why `prop.GetSetMethod()` returns `null`, although
`prop.CanWrite` obviously returned `true` in line 129.
```
Original issue reported on code.google.com by `schuster...@gmail.com` on 27 Jan 2012 at 8:22 | defect | typeaccessor writesetter throws exception if property setter is not public the method typeaccessor writesetter throws an argumentnullexception if the target type has a property with a non public setter line example class testtarget public int id get private set i am not sure why prop getsetmethod returns null although prop canwrite obviously returned true in line original issue reported on code google com by schuster gmail com on jan at | 1 |
21,497 | 3,734,176,998 | IssuesEvent | 2016-03-08 04:54:40 | department-of-veterans-affairs/roadrunner | https://api.github.com/repos/department-of-veterans-affairs/roadrunner | opened | best steez for links on landing page? | design | Wondering how best to style these links - buttons or just add link styling?

If you don't have time I can take another look but didn't want to delay getting the rest of this pr out :) | 1.0 | best steez for links on landing page? - Wondering how best to style these links - buttons or just add link styling?

If you don't have time I can take another look but didn't want to delay getting the rest of this pr out :) | non_defect | best steez for links on landing page wondering how best to style these links buttons or just add link styling if you don t have time i can take another look but didn t want to delay getting the rest of this pr out | 0 |
11,207 | 3,193,179,500 | IssuesEvent | 2015-09-30 02:29:05 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | TestProcWithExceededActionQueueDepth is flaky | area/platform/mesos kind/flake priority/P0 team/test-infra | @jdef @karlkfi @davidopp
Can we get this fixed ASAP?
Thanks!
```
proc_test.go:288: starting test case nested at 2015-09-25 00:21:13.875176228 +0000 UTC
proc_test.go:304: delegate chain invoked for nested at 2015-09-25 00:21:13.930546473 +0000 UTC
proc_test.go:323: executing deferred action: nested at 2015-09-25 00:21:13.933936289 +0000 UTC
proc_test.go:335: runDelegationTest received executed signal at 2015-09-25 00:21:13.949041708 +0000 UTC
proc_test.go:290: runDelegationTest finished at 2015-09-25 00:21:14.035916116 +0000 UTC
proc_test.go:390: unexpected error: cannot execute action because process has terminated
```
https://app.shippable.com/builds/56048e527291610b002dea79 | 1.0 | TestProcWithExceededActionQueueDepth is flaky - @jdef @karlkfi @davidopp
Can we get this fixed ASAP?
Thanks!
```
proc_test.go:288: starting test case nested at 2015-09-25 00:21:13.875176228 +0000 UTC
proc_test.go:304: delegate chain invoked for nested at 2015-09-25 00:21:13.930546473 +0000 UTC
proc_test.go:323: executing deferred action: nested at 2015-09-25 00:21:13.933936289 +0000 UTC
proc_test.go:335: runDelegationTest received executed signal at 2015-09-25 00:21:13.949041708 +0000 UTC
proc_test.go:290: runDelegationTest finished at 2015-09-25 00:21:14.035916116 +0000 UTC
proc_test.go:390: unexpected error: cannot execute action because process has terminated
```
https://app.shippable.com/builds/56048e527291610b002dea79 | non_defect | testprocwithexceededactionqueuedepth is flaky jdef karlkfi davidopp can we get this fixed asap thanks proc test go starting test case nested at utc proc test go delegate chain invoked for nested at utc proc test go executing deferred action nested at utc proc test go rundelegationtest received executed signal at utc proc test go rundelegationtest finished at utc proc test go unexpected error cannot execute action because process has terminated | 0 |
282,572 | 24,485,657,290 | IssuesEvent | 2022-10-09 11:56:41 | Tencent/bk-ci | https://api.github.com/repos/Tencent/bk-ci | closed | bug:质量红线导致的流水线失败缺少错误信息 | kind/bug for gray for test done area/ci/backend tested streams/tested streams/for test streams/for gray | 问题:质量红线导致的流水线失败缺少错误信息,会导致统计等功能异常。
措施:质量红线导致的流水线失败应记录错误信息,记录的错误信息要注意区分errorType的类型 | 4.0 | bug:质量红线导致的流水线失败缺少错误信息 - 问题:质量红线导致的流水线失败缺少错误信息,会导致统计等功能异常。
措施:质量红线导致的流水线失败应记录错误信息,记录的错误信息要注意区分errorType的类型 | non_defect | bug 质量红线导致的流水线失败缺少错误信息 问题:质量红线导致的流水线失败缺少错误信息,会导致统计等功能异常。 措施:质量红线导致的流水线失败应记录错误信息,记录的错误信息要注意区分errortype的类型 | 0 |
53,085 | 13,260,877,196 | IssuesEvent | 2020-08-20 18:54:58 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | SLALIB/C needs a real makefile (Trac #677) | Migrated from Trac defect tools/ports | makefile is currently a PoS. needs to be re-done with proper make constructs, and variables. (ie: CC, CFLAGS)
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/677">https://code.icecube.wisc.edu/projects/icecube/ticket/677</a>, reported by negaand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-05-29T21:39:28",
"_ts": "1338327568000000",
"description": "makefile is currently a PoS. needs to be re-done with proper make constructs, and variables. (ie: CC, CFLAGS)",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2012-05-29T19:00:40",
"component": "tools/ports",
"summary": "SLALIB/C needs a real makefile",
"priority": "normal",
"keywords": "slalib",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | SLALIB/C needs a real makefile (Trac #677) - makefile is currently a PoS. needs to be re-done with proper make constructs, and variables. (ie: CC, CFLAGS)
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/677">https://code.icecube.wisc.edu/projects/icecube/ticket/677</a>, reported by negaand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-05-29T21:39:28",
"_ts": "1338327568000000",
"description": "makefile is currently a PoS. needs to be re-done with proper make constructs, and variables. (ie: CC, CFLAGS)",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2012-05-29T19:00:40",
"component": "tools/ports",
"summary": "SLALIB/C needs a real makefile",
"priority": "normal",
"keywords": "slalib",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | slalib c needs a real makefile trac makefile is currently a pos needs to be re done with proper make constructs and variables ie cc cflags migrated from json status closed changetime ts description makefile is currently a pos needs to be re done with proper make constructs and variables ie cc cflags reporter nega cc resolution fixed time component tools ports summary slalib c needs a real makefile priority normal keywords slalib milestone owner nega type defect | 1 |
8,368 | 2,982,464,693 | IssuesEvent | 2015-07-17 11:26:06 | gheber/kenzo | https://api.github.com/repos/gheber/kenzo | opened | ECL (15.2.21) test failure | bug test | --------------------------------
CHECK-HAT []:
Unexpected Error: #<a SIMPLE-TYPE-ERROR>
The assertion (EQ :EQUAL
(CAT:A-CMPR3 CAT:CMPR CAT::DEL-I-DEL-J CAT::DEL-J-DEL-I+1)) failed..
-------------------------------- | 1.0 | ECL (15.2.21) test failure - --------------------------------
CHECK-HAT []:
Unexpected Error: #<a SIMPLE-TYPE-ERROR>
The assertion (EQ :EQUAL
(CAT:A-CMPR3 CAT:CMPR CAT::DEL-I-DEL-J CAT::DEL-J-DEL-I+1)) failed..
-------------------------------- | non_defect | ecl test failure check hat unexpected error the assertion eq equal cat a cat cmpr cat del i del j cat del j del i failed | 0 |
159,287 | 13,759,755,441 | IssuesEvent | 2020-10-07 03:59:25 | dankamongmen/notcurses | https://api.github.com/repos/dankamongmen/notcurses | opened | reels ought support gapless mode | documentation enhancement | Currently, `ncreel` assumes (and enforces) exactly one line of gap between tablets. Users might want to run gapless, or even with multiline gaps. We ought support the former (0 lines of gap), and if we're gonna do that, it's really no trouble to support N lines of gap (except that 0 lines can be specified with a flag, while supporting arbitrarily many rows requires a new field in `ncreel_options`). Yeah, I think simply supporting gapless mode is sufficient. Extra gap lines can, after all, be effected by simply leaving empty lines in a tablet draw call. Yep, do the flag. | 1.0 | reels ought support gapless mode - Currently, `ncreel` assumes (and enforces) exactly one line of gap between tablets. Users might want to run gapless, or even with multiline gaps. We ought support the former (0 lines of gap), and if we're gonna do that, it's really no trouble to support N lines of gap (except that 0 lines can be specified with a flag, while supporting arbitrarily many rows requires a new field in `ncreel_options`). Yeah, I think simply supporting gapless mode is sufficient. Extra gap lines can, after all, be effected by simply leaving empty lines in a tablet draw call. Yep, do the flag. | non_defect | reels ought support gapless mode currently ncreel assumes and enforces exactly one line of gap between tablets users might want to run gapless or even with multiline gaps we ought support the former lines of gap and if we re gonna do that it s really no trouble to support n lines of gap except that lines can be specified with a flag while supporting arbitrarily many rows requires a new field in ncreel options yeah i think simply supporting gapless mode is sufficient extra gap lines can after all be effected by simply leaving empty lines in a tablet draw call yep do the flag | 0 |
87,611 | 17,346,590,394 | IssuesEvent | 2021-07-29 00:17:11 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | opened | superpmi: parallel mode doesn't handle crash of superpmi | area-CodeGen-coreclr | When investigating #56156, I found that if superpmi crashes in one of multiple superpmi parallel replay processes, that crash isn't caught and the superpmi replay "succeeds".
This was on Linux/arm32. Perhaps we don't catch SIGBUS "Bus error" exceptions, compared to others? We do have try/catch in place.
The crash is:
```
Thread 1 "superpmi" received signal SIGBUS, Bus error.
Compiler::impImportStaticReadOnlyField (this=0xd17844, fldAddr=0xd665d9, lclTyp=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:7569
7569 lval = *((__int64*)fldAddr);
(gdb) bt
#0 Compiler::impImportStaticReadOnlyField (this=0xd17844, fldAddr=0xd665d9, lclTyp=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:7569
#1 0xf5f6e1ae in Compiler::impImportBlockCode (this=<optimized out>, block=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:14882
#2 0xf5f7266e in Compiler::impImportBlock(BasicBlock*)::$_0::operator()(Compiler::impImportBlock(BasicBlock*)::FilterVerificationExceptionsParam*) const (this=<optimized out>, pParam=0xfffed640)
at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:17695
#3 Compiler::impImportBlock (this=0xd17844, block=0xd25178) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:17705
#4 0xf5f746fc in Compiler::impImport (this=0xd17844) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:18778
#5 0xf5f131be in Compiler::fgImport (this=0xd17844) at /home/bruce/gh/runtime2/src/coreclr/jit/flowgraph.cpp:625
#6 0xf6018f36 in Phase::Run (this=0xfffed7b4) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.cpp:61
#7 0xf5ed3b20 in DoPhase (_compiler=0xd17844, _phase=PHASE_IMPORTATION, _action=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.h:136
#8 Compiler::compCompile (this=0xd17844, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:4511
#9 0xf5ed6aa2 in Compiler::compCompileHelper (this=0xd17844, classPtr=<optimized out>, compHnd=<optimized out>, methodInfo=<optimized out>, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0)
at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:6407
#10 0xf5ed58d4 in Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::$_11::operator()(Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::__JITParam*) const (
this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5684
#11 Compiler::compCompile (this=0xd17844, classPtr=0xefe9f000, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5703
#12 0xf5ed745e in jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::{lambda(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*)#1}::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7053
#13 jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7078
#14 jitNativeCode (methodHnd=0xe7767664, classPtr=0xefe9f000, compHnd=0x5a9c58, methodInfo=0xd239f0, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0, inlineInfoPtr=0xfffede80)
at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7080
#15 0xf5f030cc in Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::$_0::operator()(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) const (this=<optimized out>, pParam=0xfffede6c)
at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:973
#16 Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::$_0::__invoke(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) (pParam=0xfffede6c)
at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:921
#17 0x004197c0 in RunWithErrorTrap(void (*)(void*), void*)::$_0::operator()(RunWithErrorTrap(void (*)(void*), void*)::TrapParam*) const (this=<optimized out>, pTrapParam=<optimized out>)
at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi-shared/errorhandling.cpp:138
#18 RunWithErrorTrap (function=<optimized out>, param=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi-shared/errorhandling.cpp:140
#19 0xf5f0174a in Compiler::eeRunWithErrorTrap<Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param>(void (*)(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*), Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) (this=0xd147ac, function=0xd665d9, param=0xfffede6c) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.h:7957
#20 Compiler::fgInvokeInlineeCompiler (this=0xd147ac, call=0xd23860, inlineResult=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:920
#21 0xf5fee9f0 in Compiler::fgMorphCallInlineHelper (this=0xd147ac, call=0xd23860, result=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/morph.cpp:6721
#22 0xf5fee816 in Compiler::fgMorphCallInline (this=0xd147ac, call=0xd23860, inlineResult=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/morph.cpp:6603
#23 0xf5f0020a in Compiler::fgInline (this=0xd147ac) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:153
#24 0xf6018f36 in Phase::Run (this=0xfffee63c) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.cpp:61
#25 0xf5ed3cb6 in DoPhase (_compiler=0xd147ac, _phase=PHASE_MORPH_INLINE, _action=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.h:136
#26 Compiler::compCompile (this=0xd147ac, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:4684
#27 0xf5ed6aa2 in Compiler::compCompileHelper (this=0xd147ac, classPtr=<optimized out>, compHnd=<optimized out>, methodInfo=<optimized out>, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0)
at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:6407
#28 0xf5ed58d4 in Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::$_11::operator()(Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::__JITParam*) const (
this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5684
#29 Compiler::compCompile (this=0xd147ac, classPtr=0xe77f360c, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5703
#30 0xf5ed745e in jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::{lambda(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*)#1}::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7053
#31 jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7078
#32 jitNativeCode (methodHnd=0xe77f9c54, classPtr=0xe77f360c, compHnd=0x5a9c58, methodInfo=0xfffeeda4, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0, inlineInfoPtr=0x0)
at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7080
#33 0xf5edf25e in CILJit::compileMethod (this=<optimized out>, compHnd=0x5a9c58, methodInfo=0xfffeeda4, flags=<optimized out>, entryAddress=0xfffeed90, nativeSizeOfCode=0xfffeed8c)
at /home/bruce/gh/runtime2/src/coreclr/jit/ee_il_dll.cpp:276
#34 0x0040d474 in JitInstance::CompileMethod(MethodContext*, int, bool)::$_0::operator()(JitInstance::CompileMethod(MethodContext*, int, bool)::Param*) const (this=<optimized out>, pParam=0xfffeed94)
at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/jitinstance.cpp:314
#35 JitInstance::CompileMethod (this=0x59d608, MethodToCompile=<optimized out>, mcIndex=0x2a03d, collectThroughput=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/jitinstance.cpp:373
#36 0x00410d90 in main (argc=<optimized out>, argv=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/superpmi.cpp:361
(gdb) ? fldAddr
Undefined command: "". Try "help".
(gdb) print fldAddr
$1 = (void *) 0xd665d9
```
The "successful" run looks like:
```
$ /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/superpmi -p /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/libclrjit.so /home/brucefo/bugs/spmicollect/small.mch
Using child (/home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/superpmi) with args ( /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/libclrjit.so /home/brucefo/bugs/spmicollect/small.mch)
workerCount=4, skipCleanup=0.
Loaded 3 Jitted 3 FailedCompile 0 Excluded 0 Missing 0
```
But there are 5 method contexts in this example: the first process that was supposed to handle 2 of them crashed.
| 1.0 | superpmi: parallel mode doesn't handle crash of superpmi - When investigating #56156, I found that if superpmi crashes in one of multiple superpmi parallel replay processes, that crash isn't caught and the superpmi replay "succeeds".
This was on Linux/arm32. Perhaps we don't catch SIGBUS "Bus error" exceptions, compared to others? We do have try/catch in place.
The crash is:
```
Thread 1 "superpmi" received signal SIGBUS, Bus error.
Compiler::impImportStaticReadOnlyField (this=0xd17844, fldAddr=0xd665d9, lclTyp=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:7569
7569 lval = *((__int64*)fldAddr);
(gdb) bt
#0 Compiler::impImportStaticReadOnlyField (this=0xd17844, fldAddr=0xd665d9, lclTyp=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:7569
#1 0xf5f6e1ae in Compiler::impImportBlockCode (this=<optimized out>, block=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:14882
#2 0xf5f7266e in Compiler::impImportBlock(BasicBlock*)::$_0::operator()(Compiler::impImportBlock(BasicBlock*)::FilterVerificationExceptionsParam*) const (this=<optimized out>, pParam=0xfffed640)
at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:17695
#3 Compiler::impImportBlock (this=0xd17844, block=0xd25178) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:17705
#4 0xf5f746fc in Compiler::impImport (this=0xd17844) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:18778
#5 0xf5f131be in Compiler::fgImport (this=0xd17844) at /home/bruce/gh/runtime2/src/coreclr/jit/flowgraph.cpp:625
#6 0xf6018f36 in Phase::Run (this=0xfffed7b4) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.cpp:61
#7 0xf5ed3b20 in DoPhase (_compiler=0xd17844, _phase=PHASE_IMPORTATION, _action=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.h:136
#8 Compiler::compCompile (this=0xd17844, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:4511
#9 0xf5ed6aa2 in Compiler::compCompileHelper (this=0xd17844, classPtr=<optimized out>, compHnd=<optimized out>, methodInfo=<optimized out>, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0)
at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:6407
#10 0xf5ed58d4 in Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::$_11::operator()(Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::__JITParam*) const (
this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5684
#11 Compiler::compCompile (this=0xd17844, classPtr=0xefe9f000, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5703
#12 0xf5ed745e in jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::{lambda(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*)#1}::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7053
#13 jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7078
#14 jitNativeCode (methodHnd=0xe7767664, classPtr=0xefe9f000, compHnd=0x5a9c58, methodInfo=0xd239f0, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0, inlineInfoPtr=0xfffede80)
at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7080
#15 0xf5f030cc in Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::$_0::operator()(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) const (this=<optimized out>, pParam=0xfffede6c)
at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:973
#16 Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::$_0::__invoke(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) (pParam=0xfffede6c)
at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:921
#17 0x004197c0 in RunWithErrorTrap(void (*)(void*), void*)::$_0::operator()(RunWithErrorTrap(void (*)(void*), void*)::TrapParam*) const (this=<optimized out>, pTrapParam=<optimized out>)
at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi-shared/errorhandling.cpp:138
#18 RunWithErrorTrap (function=<optimized out>, param=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi-shared/errorhandling.cpp:140
#19 0xf5f0174a in Compiler::eeRunWithErrorTrap<Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param>(void (*)(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*), Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) (this=0xd147ac, function=0xd665d9, param=0xfffede6c) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.h:7957
#20 Compiler::fgInvokeInlineeCompiler (this=0xd147ac, call=0xd23860, inlineResult=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:920
#21 0xf5fee9f0 in Compiler::fgMorphCallInlineHelper (this=0xd147ac, call=0xd23860, result=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/morph.cpp:6721
#22 0xf5fee816 in Compiler::fgMorphCallInline (this=0xd147ac, call=0xd23860, inlineResult=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/morph.cpp:6603
#23 0xf5f0020a in Compiler::fgInline (this=0xd147ac) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:153
#24 0xf6018f36 in Phase::Run (this=0xfffee63c) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.cpp:61
#25 0xf5ed3cb6 in DoPhase (_compiler=0xd147ac, _phase=PHASE_MORPH_INLINE, _action=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.h:136
#26 Compiler::compCompile (this=0xd147ac, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:4684
#27 0xf5ed6aa2 in Compiler::compCompileHelper (this=0xd147ac, classPtr=<optimized out>, compHnd=<optimized out>, methodInfo=<optimized out>, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0)
at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:6407
#28 0xf5ed58d4 in Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::$_11::operator()(Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::__JITParam*) const (
this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5684
#29 Compiler::compCompile (this=0xd147ac, classPtr=0xe77f360c, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5703
#30 0xf5ed745e in jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::{lambda(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*)#1}::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7053
#31 jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7078
#32 jitNativeCode (methodHnd=0xe77f9c54, classPtr=0xe77f360c, compHnd=0x5a9c58, methodInfo=0xfffeeda4, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0, inlineInfoPtr=0x0)
at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7080
#33 0xf5edf25e in CILJit::compileMethod (this=<optimized out>, compHnd=0x5a9c58, methodInfo=0xfffeeda4, flags=<optimized out>, entryAddress=0xfffeed90, nativeSizeOfCode=0xfffeed8c)
at /home/bruce/gh/runtime2/src/coreclr/jit/ee_il_dll.cpp:276
#34 0x0040d474 in JitInstance::CompileMethod(MethodContext*, int, bool)::$_0::operator()(JitInstance::CompileMethod(MethodContext*, int, bool)::Param*) const (this=<optimized out>, pParam=0xfffeed94)
at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/jitinstance.cpp:314
#35 JitInstance::CompileMethod (this=0x59d608, MethodToCompile=<optimized out>, mcIndex=0x2a03d, collectThroughput=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/jitinstance.cpp:373
#36 0x00410d90 in main (argc=<optimized out>, argv=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/superpmi.cpp:361
(gdb) ? fldAddr
Undefined command: "". Try "help".
(gdb) print fldAddr
$1 = (void *) 0xd665d9
```
The "successful" run looks like:
```
$ /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/superpmi -p /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/libclrjit.so /home/brucefo/bugs/spmicollect/small.mch
Using child (/home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/superpmi) with args ( /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/libclrjit.so /home/brucefo/bugs/spmicollect/small.mch)
workerCount=4, skipCleanup=0.
Loaded 3 Jitted 3 FailedCompile 0 Excluded 0 Missing 0
```
But there are 5 method contexts in this example: the first process that was supposed to handle 2 of them crashed.
| non_defect | superpmi parallel mode doesn t handle crash of superpmi when investigating i found that if superpmi crashes in one of multiple superpmi parallel replay processes that crash isn t caught and the superpmi replay succeeds this was on linux perhaps we don t catch sigbus bus error exceptions compared to others we do have try catch in place the crash is thread superpmi received signal sigbus bus error compiler impimportstaticreadonlyfield this fldaddr lcltyp at home bruce gh src coreclr jit importer cpp lval fldaddr gdb bt compiler impimportstaticreadonlyfield this fldaddr lcltyp at home bruce gh src coreclr jit importer cpp in compiler impimportblockcode this block at home bruce gh src coreclr jit importer cpp in compiler impimportblock basicblock operator compiler impimportblock basicblock filterverificationexceptionsparam const this pparam at home bruce gh src coreclr jit importer cpp compiler impimportblock this block at home bruce gh src coreclr jit importer cpp in compiler impimport this at home bruce gh src coreclr jit importer cpp in compiler fgimport this at home bruce gh src coreclr jit flowgraph cpp in phase run this at home bruce gh src coreclr jit phase cpp in dophase compiler phase phase importation action at home bruce gh src coreclr jit phase h compiler compcompile this methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in compiler compcompilehelper this classptr comphnd methodinfo methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in compiler compcompile corinfo module struct void unsigned int jitflags operator compiler compcompile corinfo module struct void unsigned int jitflags jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp compiler compcompile this classptr methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const lambda jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const jitparam operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp jitnativecode methodhnd classptr comphnd methodinfo methodcodeptr methodcodesize compileflags inlineinfoptr at home bruce gh src coreclr jit compiler cpp in compiler fginvokeinlineecompiler gentreecall inlineresult operator compiler fginvokeinlineecompiler gentreecall inlineresult param const this pparam at home bruce gh src coreclr jit fginline cpp compiler fginvokeinlineecompiler gentreecall inlineresult invoke compiler fginvokeinlineecompiler gentreecall inlineresult param pparam at home bruce gh src coreclr jit fginline cpp in runwitherrortrap void void void operator runwitherrortrap void void void trapparam const this ptrapparam at home bruce gh src coreclr toolbox superpmi superpmi shared errorhandling cpp runwitherrortrap function param at home bruce gh src coreclr toolbox superpmi superpmi shared errorhandling cpp in compiler eerunwitherrortrap void compiler fginvokeinlineecompiler gentreecall inlineresult param compiler fginvokeinlineecompiler gentreecall inlineresult param this function param at home bruce gh src coreclr jit compiler h compiler fginvokeinlineecompiler this call inlineresult at home bruce gh src coreclr jit fginline cpp in compiler fgmorphcallinlinehelper this call result at home bruce gh src coreclr jit morph cpp in compiler fgmorphcallinline this call inlineresult at home bruce gh src coreclr jit morph cpp in compiler fginline this at home bruce gh src coreclr jit fginline cpp in phase run this at home bruce gh src coreclr jit phase cpp in dophase compiler phase phase morph inline action at home bruce gh src coreclr jit phase h compiler compcompile this methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in compiler compcompilehelper this classptr comphnd methodinfo methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in compiler compcompile corinfo module struct void unsigned int jitflags operator compiler compcompile corinfo module struct void unsigned int jitflags jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp compiler compcompile this classptr methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const lambda jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const jitparam operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp jitnativecode methodhnd classptr comphnd methodinfo methodcodeptr methodcodesize compileflags inlineinfoptr at home bruce gh src coreclr jit compiler cpp in ciljit compilemethod this comphnd methodinfo flags entryaddress nativesizeofcode at home bruce gh src coreclr jit ee il dll cpp in jitinstance compilemethod methodcontext int bool operator jitinstance compilemethod methodcontext int bool param const this pparam at home bruce gh src coreclr toolbox superpmi superpmi jitinstance cpp jitinstance compilemethod this methodtocompile mcindex collectthroughput at home bruce gh src coreclr toolbox superpmi superpmi jitinstance cpp in main argc argv at home bruce gh src coreclr toolbox superpmi superpmi superpmi cpp gdb fldaddr undefined command try help gdb print fldaddr void the successful run looks like home brucefo build runtime artifacts tests coreclr linux arm checked tests core root superpmi p home brucefo build runtime artifacts tests coreclr linux arm checked tests core root libclrjit so home brucefo bugs spmicollect small mch using child home brucefo build runtime artifacts tests coreclr linux arm checked tests core root superpmi with args home brucefo build runtime artifacts tests coreclr linux arm checked tests core root libclrjit so home brucefo bugs spmicollect small mch workercount skipcleanup loaded jitted failedcompile excluded missing but there are method contexts in this example the first process that was supposed to handle of them crashed | 0 |
35,520 | 7,759,210,765 | IssuesEvent | 2018-05-31 22:21:58 | NREL/EnergyPlus | https://api.github.com/repos/NREL/EnergyPlus | closed | Multiple GenericContam controllers are not set correctly | Defect | Issue overview
--------------
Issue #6565 fixed a bug to set multiple CO2 controllers correctly. The same issue occurs to handle multiple generic contam controllers.
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version)
- Version of EnergyPlus (if using an intermediate build, include SHA)
- Unmethours link or helpdesk ticket number
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| 1.0 | Multiple GenericContam controllers are not set correctly - Issue overview
--------------
Issue #6565 fixed a bug to set multiple CO2 controllers correctly. The same issue occurs to handle multiple generic contam controllers.
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version)
- Version of EnergyPlus (if using an intermediate build, include SHA)
- Unmethours link or helpdesk ticket number
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| defect | multiple genericcontam controllers are not set correctly issue overview issue fixed a bug to set multiple controllers correctly the same issue occurs to handle multiple generic contam controllers details some additional details for this issue if relevant platform operating system version version of energyplus if using an intermediate build include sha unmethours link or helpdesk ticket number checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect | 1 |
96,997 | 3,980,280,968 | IssuesEvent | 2016-05-06 06:29:53 | xcat2/xcat-core | https://api.github.com/repos/xcat2/xcat-core | closed | [FVT] [21 Apr 2016] Need verify Ubuntu 16.04 final release support against the latest xCAT master build | component:os_provision priority:high type:feature | Ubuntu 16.04 will be released on April 21st, 2016. Please refer the Ubuntu 16.04 release schedule [1].
[1] https://wiki.ubuntu.com/XenialXerus/ReleaseSchedule | 1.0 | [FVT] [21 Apr 2016] Need verify Ubuntu 16.04 final release support against the latest xCAT master build - Ubuntu 16.04 will be released on April 21st, 2016. Please refer the Ubuntu 16.04 release schedule [1].
[1] https://wiki.ubuntu.com/XenialXerus/ReleaseSchedule | non_defect | need verify ubuntu final release support against the latest xcat master build ubuntu will be released on april please refer the ubuntu release schedule | 0 |
73,164 | 24,480,751,445 | IssuesEvent | 2022-10-08 20:00:54 | FreeRADIUS/freeradius-server | https://api.github.com/repos/FreeRADIUS/freeradius-server | opened | sqlcounter[103]: Counter attribute Session-Timeout MUST be integer64 | defect | ### What type of defect/bug is this?
Unexpected behaviour (obvious or verified by project member)
### How can the issue be reproduced?
@alandekok
how to fix this?
### Log output from the FreeRADIUS daemon
```shell
BusyBox v1.35.0 (2022-10-05 19:50:01 UTC) built-in shell (ash) _______ ________ __ | |.-----.-----.-----.| | | |.----.| |_ | - || _ | -__| || | | || _|| _| |_______|| __|_____|__|__||________||__| |____| |__| W I R E L E S S F R E E D O M ----------------------------------------------------- OpenWrt 22.03.0, r19685-512e76967f ----------------------------------------------------- === WARNING! ===================================== There is no root password defined on this device! Use the "passwd" command to set up a new password in order to prevent unauthorized SSH logins. -------------------------------------------------- root@OpenWrt:~# radiusd -X FreeRADIUS Version 3.0.21 Copyright (C) 1999-2019 The FreeRADIUS server project and contributors There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE You may redistribute copies of FreeRADIUS under the terms of the GNU General Public License For more information about these matters, see the file named COPYRIGHT Starting - reading configuration files ... including dictionary file /usr/share/freeradius3/dictionary including dictionary file /etc/freeradius3/dictionary including configuration file /etc/freeradius3/radiusd.conf including configuration file /etc/freeradius3/clients.conf including files in directory /etc/freeradius3/mods-enabled/ including configuration file /etc/freeradius3/mods-enabled/expiration including configuration file /etc/freeradius3/mods-enabled/expr including configuration file /etc/freeradius3/mods-available/sql including configuration file /etc/freeradius3/mods-config/sql/main/mysql/queries.conf including configuration file /etc/freeradius3/mods-available/sqlcounter including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/dailycounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/monthlycounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/expire_on_login.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/noresetcounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/chillispot_max_bytes.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/max_all_mb.conf including files in directory /etc/freeradius3/sites-enabled/ including configuration file /etc/freeradius3/sites-enabled/default main { security { allow_core_dumps = no } name = "radiusd" prefix = "/usr" localstatedir = "/var" logdir = "/var/log" run_dir = "/var/run/radiusd" } main { name = "radiusd" prefix = "/usr" localstatedir = "/var" sbindir = "/usr/sbin" logdir = "/var/log" run_dir = "/var/run/radiusd" libdir = "/usr/lib/freeradius3" radacctdir = "/var/db/radacct" hostname_lookups = no max_request_time = 30 cleanup_delay = 5 max_requests = 16384 pidfile = "/var/run/radiusd/radiusd.pid" checkrad = "/usr/sbin/checkrad" debug_level = 0 proxy_requests = yes log { stripped_names = no auth = no auth_badpass = no auth_goodpass = no colourise = yes msg_denied = "You are already logged in - access denied" } resources { } security { max_attributes = 200 reject_delay = 1.000000 status_server = yes } } radiusd: #### Loading Realms and Home Servers #### radiusd: #### Loading Clients #### client 192.168.0.1 { ipaddr = 192.168.0.1 require_message_authenticator = no secret = <<< secret >>> nas_type = "other" proto = "*" limit { max_connections = 16 lifetime = 0 idle_timeout = 30 } } Debugger not attached # Creating Auth-Type = mschap # Creating Auth-Type = digest radiusd: #### Instantiating modules #### modules { # Loaded module rlm_expiration # Loading module "expiration" from file /etc/freeradius3/mods-enabled/expiration # Loaded module rlm_expr # Loading module "expr" from file /etc/freeradius3/mods-enabled/expr expr { safe_characters = "@abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789.-_: /äéöüàâæçèéêëîïôœùûüaÿÄÉÖÜßÀÂÆÇÈÉÊËÎÏÔŒÙÛÜŸ" } # Loaded module rlm_sql # Loading module "sql" from file /etc/freeradius3/mods-available/sql sql { driver = "rlm_sql_mysql" server = "localhost" port = 0 login = "easyhotspot" password = <<< secret >>> radius_db = "hotspot" read_groups = yes read_profiles = yes read_clients = no delete_stale_sessions = yes sql_user_name = "%{User-Name}" default_user_profile = "" client_query = "SELECT id, nasname, shortname, type, secret, server FROM nas" authorize_check_query = "SELECT id, username, attribute, value, op FROM radcheck WHERE username = '%{SQL-User-Name}' ORDER BY id" authorize_reply_query = "SELECT id, username, attribute, value, op FROM radreply WHERE username = '%{SQL-User-Name}' ORDER BY id" authorize_group_check_query = "SELECT id, groupname, attribute, Value, op FROM radgroupcheck WHERE groupname = '%{SQL-Group}' ORDER BY id" authorize_group_reply_query = "SELECT id, groupname, attribute, value, op FROM radgroupreply WHERE groupname = '%{SQL-Group}' ORDER BY id" group_membership_query = "SELECT groupname FROM radusergroup WHERE username = '%{SQL-User-Name}' ORDER BY priority" simul_count_query = "SELECT COUNT(*) FROM radacct WHERE username = '%{SQL-User-Name}' AND acctstoptime IS NULL" simul_verify_query = "SELECT radacctid, acctsessionid, username, nasipaddress, nasportid, framedipaddress, callingstationid, framedprotocol FROM radacct WHERE username = '%{SQL-User-Name}' AND acctstoptime IS NULL" safe_characters = "@abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789.-_: /" auto_escape = no accounting { reference = "%{tolower:type.%{%{Acct-Status-Type}:-%{Request-Processing-Stage}}.query}" type { accounting-on { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = '%{%{integer:Event-Timestamp}:-%l}' - UNIX_TIMESTAMP(acctstarttime), acctterminatecause = '%{%{Acct-Terminate-Cause}:-NAS-Reboot}' WHERE acctstoptime IS NULL AND nasipaddress = '%{NAS-IP-Address}' AND acctstarttime <= FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l})" } accounting-off { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = '%{%{integer:Event-Timestamp}:-%l}' - UNIX_TIMESTAMP(acctstarttime), acctterminatecause = '%{%{Acct-Terminate-Cause}:-NAS-Reboot}' WHERE acctstoptime IS NULL AND nasipaddress = '%{NAS-IP-Address}' AND acctstarttime <= FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l})" } start { query = "INSERT INTO radacct (acctsessionid, acctuniqueid, username, realm, nasipaddress, nasportid, nasporttype, acctstarttime, acctupdatetime, acctstoptime, acctsessiontime, acctauthentic, connectinfo_start, connectinfo_stop, acctinputoctets, acctoutputoctets, calledstationid, callingstationid, acctterminatecause, servicetype, framedprotocol, framedipaddress, framedipv6address, framedipv6prefix, framedinterfaceid, delegatedipv6prefix) VALUES ('%{Acct-Session-Id}', '%{Acct-Unique-Session-Id}', '%{SQL-User-Name}', '%{Realm}', '%{NAS-IP-Address}', '%{%{NAS-Port-ID}:-%{NAS-Port}}', '%{NAS-Port-Type}', FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), NULL, '0', '%{Acct-Authentic}', '%{Connect-Info}', '', '0', '0', '%{Called-Station-Id}', '%{Calling-Station-Id}', '', '%{Service-Type}', '%{Framed-Protocol}', '%{Framed-IP-Address}', '%{Framed-IPv6-Address}', '%{Framed-IPv6-Prefix}', '%{Framed-Interface-Id}', '%{Delegated-IPv6-Prefix}')" } interim-update { query = "UPDATE radacct SET acctupdatetime = (@acctupdatetime_old:=acctupdatetime), acctupdatetime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctinterval = %{%{integer:Event-Timestamp}:-%l} - UNIX_TIMESTAMP(@acctupdatetime_old), framedipaddress = '%{Framed-IP-Address}', framedipv6address = '%{Framed-IPv6-Address}', framedipv6prefix = '%{Framed-IPv6-Prefix}', framedinterfaceid = '%{Framed-Interface-Id}', delegatedipv6prefix = '%{Delegated-IPv6-Prefix}', acctsessiontime = %{%{Acct-Session-Time}:-NULL}, acctinputoctets = '%{%{Acct-Input-Gigawords}:-0}' << 32 | '%{%{Acct-Input-Octets}:-0}', acctoutputoctets = '%{%{Acct-Output-Gigawords}:-0}' << 32 | '%{%{Acct-Output-Octets}:-0}' WHERE AcctUniqueId = '%{Acct-Unique-Session-Id}'" } stop { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = %{%{Acct-Session-Time}:-NULL}, acctinputoctets = '%{%{Acct-Input-Gigawords}:-0}' << 32 | '%{%{Acct-Input-Octets}:-0}', acctoutputoctets = '%{%{Acct-Output-Gigawords}:-0}' << 32 | '%{%{Acct-Output-Octets}:-0}', acctterminatecause = '%{Acct-Terminate-Cause}', connectinfo_stop = '%{Connect-Info}' WHERE AcctUniqueId = '%{Acct-Unique-Session-Id}'" } } } post-auth { reference = ".query" query = "INSERT INTO radpostauth (username, pass, reply, authdate) VALUES ( '%{SQL-User-Name}', '%{%{User-Password}:-%{Chap-Password}}', '%{reply:Packet-Type}', '%S.%M')" } } rlm_sql (sql): Driver rlm_sql_mysql (module rlm_sql_mysql) loaded and linked Creating attribute SQL-Group # Loaded module rlm_sqlcounter # Loading module "dailycounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter dailycounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(acctsessiontime - GREATEST((%%b - UNIX_TIMESTAMP(acctstarttime)), 0)) FROM radacct WHERE username = '%{User-Name}' AND UNIX_TIMESTAMP(acctstarttime) + acctsessiontime > '%%b'" reset = "daily" counter_name = "Daily-Session-Time" check_name = "Max-Daily-Session" reply_name = "Session-Timeout" } # Loading module "monthlycounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter monthlycounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(acctsessiontime - GREATEST((%%b - UNIX_TIMESTAMP(acctstarttime)), 0)) FROM radacct WHERE username='%{User-Name}' AND UNIX_TIMESTAMP(acctstarttime) + acctsessiontime > '%%b'" reset = "monthly" counter_name = "Monthly-Session-Time" check_name = "Max-Monthly-Session" reply_name = "Session-Timeout" } # Loading module "expire_on_login" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter expire_on_login { sql_module_instance = "sql" key = "User-Name" query = "SELECT IFNULL( MAX(TIME_TO_SEC(TIMEDIFF(NOW(), acctstarttime))),0) FROM radacct WHERE UserName='%{User-Name}' ORDER BY acctstarttime LIMIT 1;" reset = "never" counter_name = "Expire-After-Initial-Login" check_name = "Expire-After" reply_name = "Session-Timeout" } # Loading module "noresetcounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter noresetcounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(Acctsessiontime) FROM radacct WHERE UserName='%{User-Name}'" reset = "never" counter_name = "Session-Timeout" check_name = "Session-Timeout" reply_name = "Session-Timeout" } /etc/freeradius3/mods-available/sqlcounter[103]: Counter attribute Session-Timeout MUST be integer64 /etc/freeradius3/mods-available/sqlcounter[103]: Instantiation failed for module "noresetcounter" root@OpenWrt:~#```
```
### Relevant log output from client utilities
_No response_
### Backtrace from LLDB or GDB
_No response_ | 1.0 | sqlcounter[103]: Counter attribute Session-Timeout MUST be integer64 - ### What type of defect/bug is this?
Unexpected behaviour (obvious or verified by project member)
### How can the issue be reproduced?
@alandekok
how to fix this?
### Log output from the FreeRADIUS daemon
```shell
BusyBox v1.35.0 (2022-10-05 19:50:01 UTC) built-in shell (ash) _______ ________ __ | |.-----.-----.-----.| | | |.----.| |_ | - || _ | -__| || | | || _|| _| |_______|| __|_____|__|__||________||__| |____| |__| W I R E L E S S F R E E D O M ----------------------------------------------------- OpenWrt 22.03.0, r19685-512e76967f ----------------------------------------------------- === WARNING! ===================================== There is no root password defined on this device! Use the "passwd" command to set up a new password in order to prevent unauthorized SSH logins. -------------------------------------------------- root@OpenWrt:~# radiusd -X FreeRADIUS Version 3.0.21 Copyright (C) 1999-2019 The FreeRADIUS server project and contributors There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE You may redistribute copies of FreeRADIUS under the terms of the GNU General Public License For more information about these matters, see the file named COPYRIGHT Starting - reading configuration files ... including dictionary file /usr/share/freeradius3/dictionary including dictionary file /etc/freeradius3/dictionary including configuration file /etc/freeradius3/radiusd.conf including configuration file /etc/freeradius3/clients.conf including files in directory /etc/freeradius3/mods-enabled/ including configuration file /etc/freeradius3/mods-enabled/expiration including configuration file /etc/freeradius3/mods-enabled/expr including configuration file /etc/freeradius3/mods-available/sql including configuration file /etc/freeradius3/mods-config/sql/main/mysql/queries.conf including configuration file /etc/freeradius3/mods-available/sqlcounter including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/dailycounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/monthlycounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/expire_on_login.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/noresetcounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/chillispot_max_bytes.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/max_all_mb.conf including files in directory /etc/freeradius3/sites-enabled/ including configuration file /etc/freeradius3/sites-enabled/default main { security { allow_core_dumps = no } name = "radiusd" prefix = "/usr" localstatedir = "/var" logdir = "/var/log" run_dir = "/var/run/radiusd" } main { name = "radiusd" prefix = "/usr" localstatedir = "/var" sbindir = "/usr/sbin" logdir = "/var/log" run_dir = "/var/run/radiusd" libdir = "/usr/lib/freeradius3" radacctdir = "/var/db/radacct" hostname_lookups = no max_request_time = 30 cleanup_delay = 5 max_requests = 16384 pidfile = "/var/run/radiusd/radiusd.pid" checkrad = "/usr/sbin/checkrad" debug_level = 0 proxy_requests = yes log { stripped_names = no auth = no auth_badpass = no auth_goodpass = no colourise = yes msg_denied = "You are already logged in - access denied" } resources { } security { max_attributes = 200 reject_delay = 1.000000 status_server = yes } } radiusd: #### Loading Realms and Home Servers #### radiusd: #### Loading Clients #### client 192.168.0.1 { ipaddr = 192.168.0.1 require_message_authenticator = no secret = <<< secret >>> nas_type = "other" proto = "*" limit { max_connections = 16 lifetime = 0 idle_timeout = 30 } } Debugger not attached # Creating Auth-Type = mschap # Creating Auth-Type = digest radiusd: #### Instantiating modules #### modules { # Loaded module rlm_expiration # Loading module "expiration" from file /etc/freeradius3/mods-enabled/expiration # Loaded module rlm_expr # Loading module "expr" from file /etc/freeradius3/mods-enabled/expr expr { safe_characters = "@abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789.-_: /äéöüàâæçèéêëîïôœùûüaÿÄÉÖÜßÀÂÆÇÈÉÊËÎÏÔŒÙÛÜŸ" } # Loaded module rlm_sql # Loading module "sql" from file /etc/freeradius3/mods-available/sql sql { driver = "rlm_sql_mysql" server = "localhost" port = 0 login = "easyhotspot" password = <<< secret >>> radius_db = "hotspot" read_groups = yes read_profiles = yes read_clients = no delete_stale_sessions = yes sql_user_name = "%{User-Name}" default_user_profile = "" client_query = "SELECT id, nasname, shortname, type, secret, server FROM nas" authorize_check_query = "SELECT id, username, attribute, value, op FROM radcheck WHERE username = '%{SQL-User-Name}' ORDER BY id" authorize_reply_query = "SELECT id, username, attribute, value, op FROM radreply WHERE username = '%{SQL-User-Name}' ORDER BY id" authorize_group_check_query = "SELECT id, groupname, attribute, Value, op FROM radgroupcheck WHERE groupname = '%{SQL-Group}' ORDER BY id" authorize_group_reply_query = "SELECT id, groupname, attribute, value, op FROM radgroupreply WHERE groupname = '%{SQL-Group}' ORDER BY id" group_membership_query = "SELECT groupname FROM radusergroup WHERE username = '%{SQL-User-Name}' ORDER BY priority" simul_count_query = "SELECT COUNT(*) FROM radacct WHERE username = '%{SQL-User-Name}' AND acctstoptime IS NULL" simul_verify_query = "SELECT radacctid, acctsessionid, username, nasipaddress, nasportid, framedipaddress, callingstationid, framedprotocol FROM radacct WHERE username = '%{SQL-User-Name}' AND acctstoptime IS NULL" safe_characters = "@abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789.-_: /" auto_escape = no accounting { reference = "%{tolower:type.%{%{Acct-Status-Type}:-%{Request-Processing-Stage}}.query}" type { accounting-on { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = '%{%{integer:Event-Timestamp}:-%l}' - UNIX_TIMESTAMP(acctstarttime), acctterminatecause = '%{%{Acct-Terminate-Cause}:-NAS-Reboot}' WHERE acctstoptime IS NULL AND nasipaddress = '%{NAS-IP-Address}' AND acctstarttime <= FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l})" } accounting-off { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = '%{%{integer:Event-Timestamp}:-%l}' - UNIX_TIMESTAMP(acctstarttime), acctterminatecause = '%{%{Acct-Terminate-Cause}:-NAS-Reboot}' WHERE acctstoptime IS NULL AND nasipaddress = '%{NAS-IP-Address}' AND acctstarttime <= FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l})" } start { query = "INSERT INTO radacct (acctsessionid, acctuniqueid, username, realm, nasipaddress, nasportid, nasporttype, acctstarttime, acctupdatetime, acctstoptime, acctsessiontime, acctauthentic, connectinfo_start, connectinfo_stop, acctinputoctets, acctoutputoctets, calledstationid, callingstationid, acctterminatecause, servicetype, framedprotocol, framedipaddress, framedipv6address, framedipv6prefix, framedinterfaceid, delegatedipv6prefix) VALUES ('%{Acct-Session-Id}', '%{Acct-Unique-Session-Id}', '%{SQL-User-Name}', '%{Realm}', '%{NAS-IP-Address}', '%{%{NAS-Port-ID}:-%{NAS-Port}}', '%{NAS-Port-Type}', FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), NULL, '0', '%{Acct-Authentic}', '%{Connect-Info}', '', '0', '0', '%{Called-Station-Id}', '%{Calling-Station-Id}', '', '%{Service-Type}', '%{Framed-Protocol}', '%{Framed-IP-Address}', '%{Framed-IPv6-Address}', '%{Framed-IPv6-Prefix}', '%{Framed-Interface-Id}', '%{Delegated-IPv6-Prefix}')" } interim-update { query = "UPDATE radacct SET acctupdatetime = (@acctupdatetime_old:=acctupdatetime), acctupdatetime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctinterval = %{%{integer:Event-Timestamp}:-%l} - UNIX_TIMESTAMP(@acctupdatetime_old), framedipaddress = '%{Framed-IP-Address}', framedipv6address = '%{Framed-IPv6-Address}', framedipv6prefix = '%{Framed-IPv6-Prefix}', framedinterfaceid = '%{Framed-Interface-Id}', delegatedipv6prefix = '%{Delegated-IPv6-Prefix}', acctsessiontime = %{%{Acct-Session-Time}:-NULL}, acctinputoctets = '%{%{Acct-Input-Gigawords}:-0}' << 32 | '%{%{Acct-Input-Octets}:-0}', acctoutputoctets = '%{%{Acct-Output-Gigawords}:-0}' << 32 | '%{%{Acct-Output-Octets}:-0}' WHERE AcctUniqueId = '%{Acct-Unique-Session-Id}'" } stop { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = %{%{Acct-Session-Time}:-NULL}, acctinputoctets = '%{%{Acct-Input-Gigawords}:-0}' << 32 | '%{%{Acct-Input-Octets}:-0}', acctoutputoctets = '%{%{Acct-Output-Gigawords}:-0}' << 32 | '%{%{Acct-Output-Octets}:-0}', acctterminatecause = '%{Acct-Terminate-Cause}', connectinfo_stop = '%{Connect-Info}' WHERE AcctUniqueId = '%{Acct-Unique-Session-Id}'" } } } post-auth { reference = ".query" query = "INSERT INTO radpostauth (username, pass, reply, authdate) VALUES ( '%{SQL-User-Name}', '%{%{User-Password}:-%{Chap-Password}}', '%{reply:Packet-Type}', '%S.%M')" } } rlm_sql (sql): Driver rlm_sql_mysql (module rlm_sql_mysql) loaded and linked Creating attribute SQL-Group # Loaded module rlm_sqlcounter # Loading module "dailycounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter dailycounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(acctsessiontime - GREATEST((%%b - UNIX_TIMESTAMP(acctstarttime)), 0)) FROM radacct WHERE username = '%{User-Name}' AND UNIX_TIMESTAMP(acctstarttime) + acctsessiontime > '%%b'" reset = "daily" counter_name = "Daily-Session-Time" check_name = "Max-Daily-Session" reply_name = "Session-Timeout" } # Loading module "monthlycounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter monthlycounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(acctsessiontime - GREATEST((%%b - UNIX_TIMESTAMP(acctstarttime)), 0)) FROM radacct WHERE username='%{User-Name}' AND UNIX_TIMESTAMP(acctstarttime) + acctsessiontime > '%%b'" reset = "monthly" counter_name = "Monthly-Session-Time" check_name = "Max-Monthly-Session" reply_name = "Session-Timeout" } # Loading module "expire_on_login" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter expire_on_login { sql_module_instance = "sql" key = "User-Name" query = "SELECT IFNULL( MAX(TIME_TO_SEC(TIMEDIFF(NOW(), acctstarttime))),0) FROM radacct WHERE UserName='%{User-Name}' ORDER BY acctstarttime LIMIT 1;" reset = "never" counter_name = "Expire-After-Initial-Login" check_name = "Expire-After" reply_name = "Session-Timeout" } # Loading module "noresetcounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter noresetcounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(Acctsessiontime) FROM radacct WHERE UserName='%{User-Name}'" reset = "never" counter_name = "Session-Timeout" check_name = "Session-Timeout" reply_name = "Session-Timeout" } /etc/freeradius3/mods-available/sqlcounter[103]: Counter attribute Session-Timeout MUST be integer64 /etc/freeradius3/mods-available/sqlcounter[103]: Instantiation failed for module "noresetcounter" root@OpenWrt:~#```
```
### Relevant log output from client utilities
_No response_
### Backtrace from LLDB or GDB
_No response_ | defect | sqlcounter counter attribute session timeout must be what type of defect bug is this unexpected behaviour obvious or verified by project member how can the issue be reproduced alandekok how to fix this log output from the freeradius daemon shell busybox utc built in shell ash w i r e l e s s f r e e d o m openwrt warning there is no root password defined on this device use the passwd command to set up a new password in order to prevent unauthorized ssh logins root openwrt radiusd x freeradius version copyright c the freeradius server project and contributors there is no warranty not even for merchantability or fitness for a particular purpose you may redistribute copies of freeradius under the terms of the gnu general public license for more information about these matters see the file named copyright starting reading configuration files including dictionary file usr share dictionary including dictionary file etc dictionary including configuration file etc radiusd conf including configuration file etc clients conf including files in directory etc mods enabled including configuration file etc mods enabled expiration including configuration file etc mods enabled expr including configuration file etc mods available sql including configuration file etc mods config sql main mysql queries conf including configuration file etc mods available sqlcounter including configuration file etc mods config sql counter mysql dailycounter conf including configuration file etc mods config sql counter mysql monthlycounter conf including configuration file etc mods config sql counter mysql expire on login conf including configuration file etc mods config sql counter mysql noresetcounter conf including configuration file etc mods config sql counter mysql chillispot max bytes conf including configuration file etc mods config sql counter mysql max all mb conf including files in directory etc sites enabled including configuration file etc sites enabled default main security allow core dumps no name radiusd prefix usr localstatedir var logdir var log run dir var run radiusd main name radiusd prefix usr localstatedir var sbindir usr sbin logdir var log run dir var run radiusd libdir usr lib radacctdir var db radacct hostname lookups no max request time cleanup delay max requests pidfile var run radiusd radiusd pid checkrad usr sbin checkrad debug level proxy requests yes log stripped names no auth no auth badpass no auth goodpass no colourise yes msg denied you are already logged in access denied resources security max attributes reject delay status server yes radiusd loading realms and home servers radiusd loading clients client ipaddr require message authenticator no secret nas type other proto limit max connections lifetime idle timeout debugger not attached creating auth type mschap creating auth type digest radiusd instantiating modules modules loaded module rlm expiration loading module expiration from file etc mods enabled expiration loaded module rlm expr loading module expr from file etc mods enabled expr expr safe characters äéöüàâæçèéêëîïôœùûüaÿäéöüßàâæçèéêëîïôœùûüÿ loaded module rlm sql loading module sql from file etc mods available sql sql driver rlm sql mysql server localhost port login easyhotspot password radius db hotspot read groups yes read profiles yes read clients no delete stale sessions yes sql user name user name default user profile client query select id nasname shortname type secret server from nas authorize check query select id username attribute value op from radcheck where username sql user name order by id authorize reply query select id username attribute value op from radreply where username sql user name order by id authorize group check query select id groupname attribute value op from radgroupcheck where groupname sql group order by id authorize group reply query select id groupname attribute value op from radgroupreply where groupname sql group order by id group membership query select groupname from radusergroup where username sql user name order by priority simul count query select count from radacct where username sql user name and acctstoptime is null simul verify query select radacctid acctsessionid username nasipaddress nasportid framedipaddress callingstationid framedprotocol from radacct where username sql user name and acctstoptime is null safe characters auto escape no accounting reference tolower type acct status type request processing stage query type accounting on query update radacct set acctstoptime from unixtime integer event timestamp l acctsessiontime integer event timestamp l unix timestamp acctstarttime acctterminatecause acct terminate cause nas reboot where acctstoptime is null and nasipaddress nas ip address and acctstarttime b reset daily counter name daily session time check name max daily session reply name session timeout loading module monthlycounter from file etc mods available sqlcounter sqlcounter monthlycounter sql module instance sql key user name query select sum acctsessiontime greatest b unix timestamp acctstarttime from radacct where username user name and unix timestamp acctstarttime acctsessiontime b reset monthly counter name monthly session time check name max monthly session reply name session timeout loading module expire on login from file etc mods available sqlcounter sqlcounter expire on login sql module instance sql key user name query select ifnull max time to sec timediff now acctstarttime from radacct where username user name order by acctstarttime limit reset never counter name expire after initial login check name expire after reply name session timeout loading module noresetcounter from file etc mods available sqlcounter sqlcounter noresetcounter sql module instance sql key user name query select sum acctsessiontime from radacct where username user name reset never counter name session timeout check name session timeout reply name session timeout etc mods available sqlcounter counter attribute session timeout must be etc mods available sqlcounter instantiation failed for module noresetcounter root openwrt relevant log output from client utilities no response backtrace from lldb or gdb no response | 1 |
75,931 | 3,478,697,915 | IssuesEvent | 2015-12-28 14:46:23 | nim-lang/nimble | https://api.github.com/repos/nim-lang/nimble | opened | Removing a newer version of a bin pkg does not switch bin symlinks | Bug Low Priority | * Install two versions (e.g. 0.6.2 and 0.7.0) of Nimble
* The ``~/.nimble/bin/nimble`` will point to ``~/.nimble/pkgs/nimble-0.7.0/nimble``
* ``nimble remove nimble@0.7``
* The symlink in ``~/.nimble/bin/nimble`` still points to ``~/.nimble/pkgs/nimble-0.7.0/nimble`` | 1.0 | Removing a newer version of a bin pkg does not switch bin symlinks - * Install two versions (e.g. 0.6.2 and 0.7.0) of Nimble
* The ``~/.nimble/bin/nimble`` will point to ``~/.nimble/pkgs/nimble-0.7.0/nimble``
* ``nimble remove nimble@0.7``
* The symlink in ``~/.nimble/bin/nimble`` still points to ``~/.nimble/pkgs/nimble-0.7.0/nimble`` | non_defect | removing a newer version of a bin pkg does not switch bin symlinks install two versions e g and of nimble the nimble bin nimble will point to nimble pkgs nimble nimble nimble remove nimble the symlink in nimble bin nimble still points to nimble pkgs nimble nimble | 0 |
174,984 | 21,300,626,348 | IssuesEvent | 2022-04-15 02:16:56 | farooqmir/React-Redux-Demonstration-with-api | https://api.github.com/repos/farooqmir/React-Redux-Demonstration-with-api | opened | CVE-2021-43138 (High) detected in multiple libraries | security vulnerability | ## CVE-2021-43138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>async-1.0.0.tgz</b>, <b>async-0.9.2.tgz</b>, <b>async-1.5.2.tgz</b></p></summary>
<p>
<details><summary><b>async-1.0.0.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-1.0.0.tgz">https://registry.npmjs.org/async/-/async-1.0.0.tgz</a></p>
<p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p>
<p>Path to vulnerable library: /node_modules/winston/node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- prompt-1.0.0.tgz (Root Library)
- winston-2.1.1.tgz
- :x: **async-1.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>async-0.9.2.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-0.9.2.tgz">https://registry.npmjs.org/async/-/async-0.9.2.tgz</a></p>
<p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p>
<p>Path to vulnerable library: /node_modules/utile/node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- prompt-1.0.0.tgz (Root Library)
- utile-0.3.0.tgz
- :x: **async-0.9.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>async-1.5.2.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-1.5.2.tgz">https://registry.npmjs.org/async/-/async-1.5.2.tgz</a></p>
<p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p>
<p>Path to vulnerable library: /node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.5.tgz (Root Library)
- portscanner-2.1.1.tgz
- :x: **async-1.5.2.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability exists in Async through 3.2.1 (fixed in 3.2.2) , which could let a malicious user obtain privileges via the mapValues() method.
<p>Publish Date: 2022-04-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43138>CVE-2021-43138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p>
<p>Release Date: 2022-04-06</p>
<p>Fix Resolution: async - v3.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-43138 (High) detected in multiple libraries - ## CVE-2021-43138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>async-1.0.0.tgz</b>, <b>async-0.9.2.tgz</b>, <b>async-1.5.2.tgz</b></p></summary>
<p>
<details><summary><b>async-1.0.0.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-1.0.0.tgz">https://registry.npmjs.org/async/-/async-1.0.0.tgz</a></p>
<p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p>
<p>Path to vulnerable library: /node_modules/winston/node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- prompt-1.0.0.tgz (Root Library)
- winston-2.1.1.tgz
- :x: **async-1.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>async-0.9.2.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-0.9.2.tgz">https://registry.npmjs.org/async/-/async-0.9.2.tgz</a></p>
<p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p>
<p>Path to vulnerable library: /node_modules/utile/node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- prompt-1.0.0.tgz (Root Library)
- utile-0.3.0.tgz
- :x: **async-0.9.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>async-1.5.2.tgz</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://registry.npmjs.org/async/-/async-1.5.2.tgz">https://registry.npmjs.org/async/-/async-1.5.2.tgz</a></p>
<p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p>
<p>Path to vulnerable library: /node_modules/async/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.5.tgz (Root Library)
- portscanner-2.1.1.tgz
- :x: **async-1.5.2.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability exists in Async through 3.2.1 (fixed in 3.2.2) , which could let a malicious user obtain privileges via the mapValues() method.
<p>Publish Date: 2022-04-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43138>CVE-2021-43138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p>
<p>Release Date: 2022-04-06</p>
<p>Fix Resolution: async - v3.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries async tgz async tgz async tgz async tgz higher order functions and common patterns for asynchronous code library home page a href path to dependency file react redux demonstration with api package json path to vulnerable library node modules winston node modules async package json dependency hierarchy prompt tgz root library winston tgz x async tgz vulnerable library async tgz higher order functions and common patterns for asynchronous code library home page a href path to dependency file react redux demonstration with api package json path to vulnerable library node modules utile node modules async package json dependency hierarchy prompt tgz root library utile tgz x async tgz vulnerable library async tgz higher order functions and common patterns for asynchronous code library home page a href path to dependency file react redux demonstration with api package json path to vulnerable library node modules async package json dependency hierarchy browser sync tgz root library portscanner tgz x async tgz vulnerable library vulnerability details a vulnerability exists in async through fixed in which could let a malicious user obtain privileges via the mapvalues method publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution async step up your open source security game with whitesource | 0 |
6,421 | 9,333,326,449 | IssuesEvent | 2019-03-28 14:12:23 | gnosis/dex-contracts | https://api.github.com/repos/gnosis/dex-contracts | closed | Order Collection | requirement | One should be able to submit a *limit order* of the following form
- **Buy** at most `X` of token `i` for token `j` if price `p_{i, j} < P`
- **Sell** at most `Y` of token `i` for token `j` if price `p_{i, j} >= P`
@twalth3r could you please confirm the order types that should be collected by the contract?
Expressing a `Sell` Order in terms of integers should be as follows:
- `buyToken: uint8`
- `sellToken: uint8`
- `buyAmount: uint`
- `sellAmount: uint`
- `priceBuyToken: uint`
- `priceSellToken: uint` | 1.0 | Order Collection - One should be able to submit a *limit order* of the following form
- **Buy** at most `X` of token `i` for token `j` if price `p_{i, j} < P`
- **Sell** at most `Y` of token `i` for token `j` if price `p_{i, j} >= P`
@twalth3r could you please confirm the order types that should be collected by the contract?
Expressing a `Sell` Order in terms of integers should be as follows:
- `buyToken: uint8`
- `sellToken: uint8`
- `buyAmount: uint`
- `sellAmount: uint`
- `priceBuyToken: uint`
- `priceSellToken: uint` | non_defect | order collection one should be able to submit a limit order of the following form buy at most x of token i for token j if price p i j p sell at most y of token i for token j if price p i j p could you please confirm the order types that should be collected by the contract expressing a sell order in terms of integers should be as follows buytoken selltoken buyamount uint sellamount uint pricebuytoken uint priceselltoken uint | 0 |
24,088 | 2,666,017,262 | IssuesEvent | 2015-03-21 03:23:36 | scanmem/scanmem | https://api.github.com/repos/scanmem/scanmem | closed | Non-blocking attaching | auto-migrated Priority-Low Type-Enhancement | ```
To attach the target process without blocking it
```
Original issue reported on code.google.com by `coolwan...@gmail.com` on 25 Apr 2013 at 2:00 | 1.0 | Non-blocking attaching - ```
To attach the target process without blocking it
```
Original issue reported on code.google.com by `coolwan...@gmail.com` on 25 Apr 2013 at 2:00 | non_defect | non blocking attaching to attach the target process without blocking it original issue reported on code google com by coolwan gmail com on apr at | 0 |
14,705 | 2,831,388,625 | IssuesEvent | 2015-05-24 15:53:31 | nobodyguy/dslrdashboard | https://api.github.com/repos/nobodyguy/dslrdashboard | closed | LRTimelapse Holy Grail Mode - Shutter Exceeding Set Limit | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. LRtimelapse - Holy Grail Mode (Sunset)
2. Starting exposure 1/60s, ISO 50
3. set Limit longest shutter to 2s, highest ISO: 800, 1/2 stop increment
for both shutter and ISO.
What is the expected output? What do you see instead?
Expected DD to limit the shutter speed at 2s and proceed to increase ISO.
Instead the DD keep on increasing the shutter speed past 2s (until 8s, at that
point I stopped the timelapse recording) and the ISO was still kept at 50.
What version of the product are you using? On what operating system?
latest version (updated 21 Feb 2014) on Galaxy Note 2; Nikon d600 connected
using wifi
Please provide any additional information below.
Previous version's Auto Holy Grail works flawlessly.
Will try again tomorrow for sunrise to see if similar problem will recur
Many thanks for the awesome apps!
regards
Kamrul
```
Original issue reported on code.google.com by `kamrul.a...@gmail.com` on 22 Feb 2014 at 1:33 | 1.0 | LRTimelapse Holy Grail Mode - Shutter Exceeding Set Limit - ```
What steps will reproduce the problem?
1. LRtimelapse - Holy Grail Mode (Sunset)
2. Starting exposure 1/60s, ISO 50
3. set Limit longest shutter to 2s, highest ISO: 800, 1/2 stop increment
for both shutter and ISO.
What is the expected output? What do you see instead?
Expected DD to limit the shutter speed at 2s and proceed to increase ISO.
Instead the DD keep on increasing the shutter speed past 2s (until 8s, at that
point I stopped the timelapse recording) and the ISO was still kept at 50.
What version of the product are you using? On what operating system?
latest version (updated 21 Feb 2014) on Galaxy Note 2; Nikon d600 connected
using wifi
Please provide any additional information below.
Previous version's Auto Holy Grail works flawlessly.
Will try again tomorrow for sunrise to see if similar problem will recur
Many thanks for the awesome apps!
regards
Kamrul
```
Original issue reported on code.google.com by `kamrul.a...@gmail.com` on 22 Feb 2014 at 1:33 | defect | lrtimelapse holy grail mode shutter exceeding set limit what steps will reproduce the problem lrtimelapse holy grail mode sunset starting exposure iso set limit longest shutter to highest iso stop increment for both shutter and iso what is the expected output what do you see instead expected dd to limit the shutter speed at and proceed to increase iso instead the dd keep on increasing the shutter speed past until at that point i stopped the timelapse recording and the iso was still kept at what version of the product are you using on what operating system latest version updated feb on galaxy note nikon connected using wifi please provide any additional information below previous version s auto holy grail works flawlessly will try again tomorrow for sunrise to see if similar problem will recur many thanks for the awesome apps regards kamrul original issue reported on code google com by kamrul a gmail com on feb at | 1 |
127,531 | 5,031,946,942 | IssuesEvent | 2016-12-16 09:28:23 | itsyouonline/identityserver | https://api.github.com/repos/itsyouonline/identityserver | closed | Possibility not to send invitation emails | priority_major state_verification type_feature | Some services would like to send the invitation email themselves so the possibility for not sending out invitation emails should be added when an organization invites someone.
Proposed solution: add an `invitemail` parameter to the invitation creation api with the possible values of `default` and `none`, the default value is `default`. When `none` is supplied, no invitation email should be send. | 1.0 | Possibility not to send invitation emails - Some services would like to send the invitation email themselves so the possibility for not sending out invitation emails should be added when an organization invites someone.
Proposed solution: add an `invitemail` parameter to the invitation creation api with the possible values of `default` and `none`, the default value is `default`. When `none` is supplied, no invitation email should be send. | non_defect | possibility not to send invitation emails some services would like to send the invitation email themselves so the possibility for not sending out invitation emails should be added when an organization invites someone proposed solution add an invitemail parameter to the invitation creation api with the possible values of default and none the default value is default when none is supplied no invitation email should be send | 0 |
141,379 | 11,414,337,583 | IssuesEvent | 2020-02-02 02:01:27 | istio/istio | https://api.github.com/repos/istio/istio | closed | New Test, TestAuthorizationPolicies_ListNamespacesOfServiceRoles, flaky due to non deterministic sorting | area/security kind/test failure lifecycle/needs-triage | https://prow.istio.io/view/gcs/istio-prow/pr-logs/pull/istio_istio/17755/unit-tests_istio/1981 | 1.0 | New Test, TestAuthorizationPolicies_ListNamespacesOfServiceRoles, flaky due to non deterministic sorting - https://prow.istio.io/view/gcs/istio-prow/pr-logs/pull/istio_istio/17755/unit-tests_istio/1981 | non_defect | new test testauthorizationpolicies listnamespacesofserviceroles flaky due to non deterministic sorting | 0 |
83,942 | 15,720,761,497 | IssuesEvent | 2021-03-29 01:07:21 | andygonzalez2010/store | https://api.github.com/repos/andygonzalez2010/store | opened | WS-2017-3767 (Medium) detected in spring-security-web-5.1.5.RELEASE.jar | security vulnerability | ## WS-2017-3767 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-5.1.5.RELEASE.jar</b></p></summary>
<p>spring-security-web</p>
<p>Library home page: <a href="https://spring.io/spring-security">https://spring.io/spring-security</a></p>
<p>Path to dependency file: store/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/5.1.5.RELEASE/spring-security-web-5.1.5.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-security-2.1.4.RELEASE.jar (Root Library)
- :x: **spring-security-web-5.1.5.RELEASE.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-Site Request Forgery (CSRF) vulnerability was found in spring-security before 4.2.15, 5.0.15, 5.1.9, 5.2.3, and 5.3.1. SwitchUserFilter responds to all HTTP methods, making it vulnerable to CSRF attacks.
<p>Publish Date: 2017-01-03
<p>URL: <a href=https://github.com/spring-projects/spring-security/commit/eed71243cb86833e7edf230e5e43ad89b01142f9>WS-2017-3767</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/spring-projects/spring-security/releases/tag/5.3.1.RELEASE">https://github.com/spring-projects/spring-security/releases/tag/5.3.1.RELEASE</a></p>
<p>Release Date: 2017-01-03</p>
<p>Fix Resolution: org.springframework.security:spring-security-web:4.2.15,5.0.15,5.1.9,5.2.3,5.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2017-3767 (Medium) detected in spring-security-web-5.1.5.RELEASE.jar - ## WS-2017-3767 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-5.1.5.RELEASE.jar</b></p></summary>
<p>spring-security-web</p>
<p>Library home page: <a href="https://spring.io/spring-security">https://spring.io/spring-security</a></p>
<p>Path to dependency file: store/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/5.1.5.RELEASE/spring-security-web-5.1.5.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-security-2.1.4.RELEASE.jar (Root Library)
- :x: **spring-security-web-5.1.5.RELEASE.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-Site Request Forgery (CSRF) vulnerability was found in spring-security before 4.2.15, 5.0.15, 5.1.9, 5.2.3, and 5.3.1. SwitchUserFilter responds to all HTTP methods, making it vulnerable to CSRF attacks.
<p>Publish Date: 2017-01-03
<p>URL: <a href=https://github.com/spring-projects/spring-security/commit/eed71243cb86833e7edf230e5e43ad89b01142f9>WS-2017-3767</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/spring-projects/spring-security/releases/tag/5.3.1.RELEASE">https://github.com/spring-projects/spring-security/releases/tag/5.3.1.RELEASE</a></p>
<p>Release Date: 2017-01-03</p>
<p>Fix Resolution: org.springframework.security:spring-security-web:4.2.15,5.0.15,5.1.9,5.2.3,5.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | ws medium detected in spring security web release jar ws medium severity vulnerability vulnerable library spring security web release jar spring security web library home page a href path to dependency file store pom xml path to vulnerable library home wss scanner repository org springframework security spring security web release spring security web release jar dependency hierarchy spring boot starter security release jar root library x spring security web release jar vulnerable library found in base branch master vulnerability details cross site request forgery csrf vulnerability was found in spring security before and switchuserfilter responds to all http methods making it vulnerable to csrf attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework security spring security web step up your open source security game with whitesource | 0 |
53,871 | 13,262,408,600 | IssuesEvent | 2020-08-20 21:43:56 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | better error reporting from boost::python (Trac #2221) | Migrated from Trac combo core defect | I keep getting crashes from steamshovel wit the error message
```text
>>> libc++abi.dylib: terminating with uncaught exception of type boost::python::error_already_set
Abort trap: 6
```
Which is as far as I can tell boost::python's super helpful response to the python interpreter encountering an error. We need to find a way to print out the actual error in the python interpreter.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2221">https://code.icecube.wisc.edu/projects/icecube/ticket/2221</a>, reported by kjmeagher</summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-07-01T13:45:37",
"_ts": "1561988737587870",
"description": "I keep getting crashes from steamshovel wit the error message\n{{{\n>>> libc++abi.dylib: terminating with uncaught exception of type boost::python::error_already_set\nAbort trap: 6\n}}}\nWhich is as far as I can tell boost::python's super helpful response to the python interpreter encountering an error. We need to find a way to print out the actual error in the python interpreter.",
"reporter": "kjmeagher",
"cc": "",
"resolution": "wontfix",
"time": "2018-12-06T16:36:20",
"component": "combo core",
"summary": "better error reporting from boost::python",
"priority": "normal",
"keywords": "",
"milestone": "Autumnal Equinox 2019",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| 1.0 | better error reporting from boost::python (Trac #2221) - I keep getting crashes from steamshovel wit the error message
```text
>>> libc++abi.dylib: terminating with uncaught exception of type boost::python::error_already_set
Abort trap: 6
```
Which is as far as I can tell boost::python's super helpful response to the python interpreter encountering an error. We need to find a way to print out the actual error in the python interpreter.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2221">https://code.icecube.wisc.edu/projects/icecube/ticket/2221</a>, reported by kjmeagher</summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-07-01T13:45:37",
"_ts": "1561988737587870",
"description": "I keep getting crashes from steamshovel wit the error message\n{{{\n>>> libc++abi.dylib: terminating with uncaught exception of type boost::python::error_already_set\nAbort trap: 6\n}}}\nWhich is as far as I can tell boost::python's super helpful response to the python interpreter encountering an error. We need to find a way to print out the actual error in the python interpreter.",
"reporter": "kjmeagher",
"cc": "",
"resolution": "wontfix",
"time": "2018-12-06T16:36:20",
"component": "combo core",
"summary": "better error reporting from boost::python",
"priority": "normal",
"keywords": "",
"milestone": "Autumnal Equinox 2019",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| defect | better error reporting from boost python trac i keep getting crashes from steamshovel wit the error message text libc abi dylib terminating with uncaught exception of type boost python error already set abort trap which is as far as i can tell boost python s super helpful response to the python interpreter encountering an error we need to find a way to print out the actual error in the python interpreter migrated from json status closed changetime ts description i keep getting crashes from steamshovel wit the error message n n libc abi dylib terminating with uncaught exception of type boost python error already set nabort trap n nwhich is as far as i can tell boost python s super helpful response to the python interpreter encountering an error we need to find a way to print out the actual error in the python interpreter reporter kjmeagher cc resolution wontfix time component combo core summary better error reporting from boost python priority normal keywords milestone autumnal equinox owner type defect | 1 |
774,905 | 27,214,999,896 | IssuesEvent | 2023-02-20 20:36:43 | ascheid/itsg33-pbmm-issue-gen | https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen | closed | SA-18 TAMPER RESISTANCE AND DETECTION | Priority: None defined | (A) The organization implements a tamper protection program for the information system, system component, or information system service. | 1.0 | SA-18 TAMPER RESISTANCE AND DETECTION - (A) The organization implements a tamper protection program for the information system, system component, or information system service. | non_defect | sa tamper resistance and detection a the organization implements a tamper protection program for the information system system component or information system service | 0 |
116,084 | 17,358,340,407 | IssuesEvent | 2021-07-29 16:58:11 | Tim-sandbox/webgoat-trng | https://api.github.com/repos/Tim-sandbox/webgoat-trng | opened | CVE-2021-22112 (High) detected in spring-security-web-5.2.1.RELEASE.jar | security vulnerability | ## CVE-2021-22112 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-5.2.1.RELEASE.jar</b></p></summary>
<p>spring-security-web</p>
<p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p>
<p>Path to dependency file: webgoat-trng/webgoat-integration-tests/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/5.2.1.RELEASE/spring-security-web-5.2.1.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- webwolf-v8.1.0.jar (Root Library)
- spring-boot-starter-security-2.2.2.RELEASE.jar
- :x: **spring-security-web-5.2.1.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/webgoat-trng/commit/35c20cdfd5acad60536ca6b0402f619e18097690">35c20cdfd5acad60536ca6b0402f619e18097690</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Spring Security 5.4.x prior to 5.4.4, 5.3.x prior to 5.3.8.RELEASE, 5.2.x prior to 5.2.9.RELEASE, and older unsupported versions can fail to save the SecurityContext if it is changed more than once in a single request.A malicious user cannot cause the bug to happen (it must be programmed in). However, if the application's intent is to only allow the user to run with elevated privileges in a small portion of the application, the bug can be leveraged to extend those privileges to the rest of the application.
<p>Publish Date: 2021-02-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22112>CVE-2021-22112</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22112">https://tanzu.vmware.com/security/cve-2021-22112</a></p>
<p>Release Date: 2021-02-23</p>
<p>Fix Resolution: org.springframework.security:spring-security-web:5.2.9,5.3.8,5.4.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework.security","packageName":"spring-security-web","packageVersion":"5.2.1.RELEASE","packageFilePaths":["/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.owasp.webgoat:webwolf:v8.1.0;org.springframework.boot:spring-boot-starter-security:2.2.2.RELEASE;org.springframework.security:spring-security-web:5.2.1.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework.security:spring-security-web:5.2.9,5.3.8,5.4.4"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-22112","vulnerabilityDetails":"Spring Security 5.4.x prior to 5.4.4, 5.3.x prior to 5.3.8.RELEASE, 5.2.x prior to 5.2.9.RELEASE, and older unsupported versions can fail to save the SecurityContext if it is changed more than once in a single request.A malicious user cannot cause the bug to happen (it must be programmed in). However, if the application\u0027s intent is to only allow the user to run with elevated privileges in a small portion of the application, the bug can be leveraged to extend those privileges to the rest of the application.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22112","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-22112 (High) detected in spring-security-web-5.2.1.RELEASE.jar - ## CVE-2021-22112 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-5.2.1.RELEASE.jar</b></p></summary>
<p>spring-security-web</p>
<p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p>
<p>Path to dependency file: webgoat-trng/webgoat-integration-tests/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/5.2.1.RELEASE/spring-security-web-5.2.1.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- webwolf-v8.1.0.jar (Root Library)
- spring-boot-starter-security-2.2.2.RELEASE.jar
- :x: **spring-security-web-5.2.1.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/webgoat-trng/commit/35c20cdfd5acad60536ca6b0402f619e18097690">35c20cdfd5acad60536ca6b0402f619e18097690</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Spring Security 5.4.x prior to 5.4.4, 5.3.x prior to 5.3.8.RELEASE, 5.2.x prior to 5.2.9.RELEASE, and older unsupported versions can fail to save the SecurityContext if it is changed more than once in a single request.A malicious user cannot cause the bug to happen (it must be programmed in). However, if the application's intent is to only allow the user to run with elevated privileges in a small portion of the application, the bug can be leveraged to extend those privileges to the rest of the application.
<p>Publish Date: 2021-02-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22112>CVE-2021-22112</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22112">https://tanzu.vmware.com/security/cve-2021-22112</a></p>
<p>Release Date: 2021-02-23</p>
<p>Fix Resolution: org.springframework.security:spring-security-web:5.2.9,5.3.8,5.4.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework.security","packageName":"spring-security-web","packageVersion":"5.2.1.RELEASE","packageFilePaths":["/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.owasp.webgoat:webwolf:v8.1.0;org.springframework.boot:spring-boot-starter-security:2.2.2.RELEASE;org.springframework.security:spring-security-web:5.2.1.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework.security:spring-security-web:5.2.9,5.3.8,5.4.4"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-22112","vulnerabilityDetails":"Spring Security 5.4.x prior to 5.4.4, 5.3.x prior to 5.3.8.RELEASE, 5.2.x prior to 5.2.9.RELEASE, and older unsupported versions can fail to save the SecurityContext if it is changed more than once in a single request.A malicious user cannot cause the bug to happen (it must be programmed in). However, if the application\u0027s intent is to only allow the user to run with elevated privileges in a small portion of the application, the bug can be leveraged to extend those privileges to the rest of the application.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22112","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_defect | cve high detected in spring security web release jar cve high severity vulnerability vulnerable library spring security web release jar spring security web library home page a href path to dependency file webgoat trng webgoat integration tests pom xml path to vulnerable library home wss scanner repository org springframework security spring security web release spring security web release jar dependency hierarchy webwolf jar root library spring boot starter security release jar x spring security web release jar vulnerable library found in head commit a href found in base branch main vulnerability details spring security x prior to x prior to release x prior to release and older unsupported versions can fail to save the securitycontext if it is changed more than once in a single request a malicious user cannot cause the bug to happen it must be programmed in however if the application s intent is to only allow the user to run with elevated privileges in a small portion of the application the bug can be leveraged to extend those privileges to the rest of the application publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework security spring security web isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org owasp webgoat webwolf org springframework boot spring boot starter security release org springframework security spring security web release isminimumfixversionavailable true minimumfixversion org springframework security spring security web basebranches vulnerabilityidentifier cve vulnerabilitydetails spring security x prior to x prior to release x prior to release and older unsupported versions can fail to save the securitycontext if it is changed more than once in a single request a malicious user cannot cause the bug to happen it must be programmed in however if the application intent is to only allow the user to run with elevated privileges in a small portion of the application the bug can be leveraged to extend those privileges to the rest of the application vulnerabilityurl | 0 |
47,142 | 13,206,083,391 | IssuesEvent | 2020-08-14 19:22:47 | kenferrara/layer5 | https://api.github.com/repos/kenferrara/layer5 | opened | CVE-2020-8161 (High) detected in rack-1.6.13.gem | security vulnerability | ## CVE-2020-8161 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rack-1.6.13.gem</b></p></summary>
<p>Rack provides a minimal, modular and adaptable interface for developing
web applications in Ruby. By wrapping HTTP requests and responses in
the simplest way possible, it unifies and distills the API for web
servers, web frameworks, and software in between (the so-called
middleware) into a single method call.
Also see http://rack.github.io/.
</p>
<p>Library home page: <a href="https://rubygems.org/gems/rack-1.6.13.gem">https://rubygems.org/gems/rack-1.6.13.gem</a></p>
<p>
Dependency Hierarchy:
- rack-jekyll-0.5.0.gem (Root Library)
- :x: **rack-1.6.13.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/layer5/commit/eda0b835176f91fd68458c4f840b249240947718">eda0b835176f91fd68458c4f840b249240947718</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A directory traversal vulnerability exists in rack < 2.2.0 that allows an attacker perform directory traversal vulnerability in the Rack::Directory app that is bundled with Rack which could result in information disclosure.
<p>Publish Date: 2020-07-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8161>CVE-2020-8161</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/rack/rack/tree/2.2.0">https://github.com/rack/rack/tree/2.2.0</a></p>
<p>Release Date: 2020-06-01</p>
<p>Fix Resolution: 2.2.0,2.1.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Ruby","packageName":"rack","packageVersion":"1.6.13","isTransitiveDependency":true,"dependencyTree":"rack-jekyll:0.5.0;rack:1.6.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.2.0,2.1.3"}],"vulnerabilityIdentifier":"CVE-2020-8161","vulnerabilityDetails":"A directory traversal vulnerability exists in rack \u003c 2.2.0 that allows an attacker perform directory traversal vulnerability in the Rack::Directory app that is bundled with Rack which could result in information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8161","cvss3Severity":"high","cvss3Score":"8.6","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-8161 (High) detected in rack-1.6.13.gem - ## CVE-2020-8161 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rack-1.6.13.gem</b></p></summary>
<p>Rack provides a minimal, modular and adaptable interface for developing
web applications in Ruby. By wrapping HTTP requests and responses in
the simplest way possible, it unifies and distills the API for web
servers, web frameworks, and software in between (the so-called
middleware) into a single method call.
Also see http://rack.github.io/.
</p>
<p>Library home page: <a href="https://rubygems.org/gems/rack-1.6.13.gem">https://rubygems.org/gems/rack-1.6.13.gem</a></p>
<p>
Dependency Hierarchy:
- rack-jekyll-0.5.0.gem (Root Library)
- :x: **rack-1.6.13.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/layer5/commit/eda0b835176f91fd68458c4f840b249240947718">eda0b835176f91fd68458c4f840b249240947718</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A directory traversal vulnerability exists in rack < 2.2.0 that allows an attacker perform directory traversal vulnerability in the Rack::Directory app that is bundled with Rack which could result in information disclosure.
<p>Publish Date: 2020-07-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8161>CVE-2020-8161</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/rack/rack/tree/2.2.0">https://github.com/rack/rack/tree/2.2.0</a></p>
<p>Release Date: 2020-06-01</p>
<p>Fix Resolution: 2.2.0,2.1.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Ruby","packageName":"rack","packageVersion":"1.6.13","isTransitiveDependency":true,"dependencyTree":"rack-jekyll:0.5.0;rack:1.6.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.2.0,2.1.3"}],"vulnerabilityIdentifier":"CVE-2020-8161","vulnerabilityDetails":"A directory traversal vulnerability exists in rack \u003c 2.2.0 that allows an attacker perform directory traversal vulnerability in the Rack::Directory app that is bundled with Rack which could result in information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8161","cvss3Severity":"high","cvss3Score":"8.6","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_defect | cve high detected in rack gem cve high severity vulnerability vulnerable library rack gem rack provides a minimal modular and adaptable interface for developing web applications in ruby by wrapping http requests and responses in the simplest way possible it unifies and distills the api for web servers web frameworks and software in between the so called middleware into a single method call also see library home page a href dependency hierarchy rack jekyll gem root library x rack gem vulnerable library found in head commit a href vulnerability details a directory traversal vulnerability exists in rack that allows an attacker perform directory traversal vulnerability in the rack directory app that is bundled with rack which could result in information disclosure publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a directory traversal vulnerability exists in rack that allows an attacker perform directory traversal vulnerability in the rack directory app that is bundled with rack which could result in information disclosure vulnerabilityurl | 0 |
42,013 | 5,412,272,179 | IssuesEvent | 2017-03-01 14:09:32 | wellcometrust/wellcomecollection.org | https://api.github.com/repos/wellcometrust/wellcomecollection.org | closed | Explore - Article promo | design | The format text (example: Article) should be black (hex: 1d1d1d) and is clickable.
The user can click image or title to go into article, but they can also click on format and go to the index page for that format.
Hover for this text is the similar as regular body link, with teal underline appearing on hover

| 1.0 | Explore - Article promo - The format text (example: Article) should be black (hex: 1d1d1d) and is clickable.
The user can click image or title to go into article, but they can also click on format and go to the index page for that format.
Hover for this text is the similar as regular body link, with teal underline appearing on hover

| non_defect | explore article promo the format text example article should be black hex and is clickable the user can click image or title to go into article but they can also click on format and go to the index page for that format hover for this text is the similar as regular body link with teal underline appearing on hover | 0 |
66,349 | 20,157,930,774 | IssuesEvent | 2022-02-09 18:15:05 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | ZFS 2.1.2 + NFSv4 ARC stuck pruning/evicting, flatlines throughput | Type: Defect | <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | `Ubuntu`
Distribution Version | `20.04.3 LTS`
Kernel Version | `5.4.0-96-generic`
Architecture | `x86_64`
OpenZFS Version | `zfs-2.1.2-1`
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
ZFS is flatlined on throughput with an `arc_evict` and `arc_prune` process spinning at 100%. The workload is kernel NFS server (all NFSv4 clients) with ZFS 2.1.2 built from source. Characterized by high CPU `iowait` and throughput flatlining.
#### RAM Graph
<img width="1350" alt="Screen Shot 2022-02-09 at 9 06 55 AM" src="https://user-images.githubusercontent.com/16908622/153252531-9ee24c8e-3ecb-4242-b862-b860353f5625.png">
It's holding on to RAM pretty hard. This is where I begin to lose debugging expertise, having discovered slabs yesterday. :-)
#### top
```
top - 16:55:29 up 21 days, 14:56, 1 user, load average: 195.23, 124.60, 134.24
Tasks: 993 total, 2 running, 991 sleeping, 0 stopped, 0 zombie
%Cpu(s): 0.0 us, 7.3 sy, 0.0 ni, 5.0 id, 87.3 wa, 0.0 hi, 0.4 si, 0.0 st
MiB Mem : 128622.1 total, 2501.0 free, 121637.5 used, 4483.5 buff/cache
MiB Swap: 1907.0 total, 1042.7 free, 864.2 used. 6078.4 avail Mem
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
245944 root 20 0 0 0 0 R 100.0 0.0 549:36.86 arc_evict
245943 root 20 0 0 0 0 S 93.1 0.0 459:13.30 arc_prune
246555 root 20 0 953064 20876 2732 S 3.6 0.0 29:32.85 zed
3174406 root 20 0 12784 4872 3248 R 1.3 0.0 0:00.17 top
14154 root rt 0 356604 31088 8300 S 0.7 0.0 133:03.60 multipathd
3159749 root 20 0 0 0 0 D 0.7 0.0 0:00.11 nfsd
```
#### slabtop
```
# slabtop
Active / Total Objects (% used) : 215263329 / 222780310 (96.6%)
Active / Total Slabs (% used) : 6020764 / 6020764 (100.0%)
Active / Total Caches (% used) : 141 / 199 (70.9%)
Active / Total Size (% used) : 98093294.62K / 99549181.36K (98.5%)
Minimum / Average / Maximum Object : 0.01K / 0.45K / 16.75K
OBJS ACTIVE USE OBJ SIZE SLABS OBJ/SLAB CACHE SIZE NAME
27437256 27437256 100% 0.97K 831432 33 26605824K dnode_t
25866414 25866282 99% 0.38K 615867 42 9853872K dmu_buf_impl_t
24953184 24952001 99% 0.50K 779787 32 12476592K kmalloc-512
20654336 18708128 90% 0.06K 322724 64 1290896K kmalloc-64
14428134 14148450 98% 0.09K 343527 42 1374108K arc_buf_hdr_t_l2only
10290924 10290229 99% 0.09K 245022 42 980088K kmalloc-96
10017152 9881010 98% 0.03K 78259 128 313036K kmalloc-32
7602798 7575100 99% 0.19K 181019 42 1448152K dentry
7255278 7139757 98% 1.09K 250182 29 8005824K zfs_znode_cache
7225482 7140416 98% 0.24K 218954 33 1751632K sa_cache
5172144 5172144 100% 0.16K 107753 48 862024K nfsd4_stateids
5079424 5077737 99% 0.25K 158732 32 1269856K filp
4988088 4985288 99% 0.19K 118764 42 950112K cred_jar
4952610 4952610 100% 0.02K 29133 170 116532K lsm_file_cache
4951520 4951520 100% 0.28K 176840 28 1414720K nfsd4_files
4949376 4949376 100% 0.03K 38667 128 154668K fsnotify_mark_connector
4947816 4947816 100% 0.08K 97016 51 388064K Acpi-State
4947372 4947372 100% 0.11K 137427 36 549708K khugepaged_mm_slot
4862528 4858723 99% 0.06K 75977 64 303908K kmalloc-rcl-64
3813933 1054784 27% 0.31K 74783 51 1196528K arc_buf_hdr_t_full
3601017 3601017 100% 0.05K 49329 73 197316K nsproxy
3022149 1059649 35% 0.10K 77491 39 309964K abd_t
2598288 2591411 99% 0.57K 92796 28 1484736K radix_tree_node
1523068 1523068 100% 0.42K 41164 37 658624K nfsd4_openowners
1248128 1247029 99% 1.00K 39004 32 1248128K kmalloc-1k
1128681 1052509 93% 0.08K 22131 51 88524K arc_buf_t
960386 960375 99% 16.00K 480193 2 15366176K zio_buf_comb_16384
945488 942004 99% 8.00K 236372 4 7563904K kmalloc-8k
290745 268913 92% 0.10K 7455 39 29820K buffer_head
288000 287664 99% 0.13K 9600 30 38400K kernfs_node_cache
228160 226423 99% 0.06K 3565 64 14260K anon_vma_chain
211894 204295 96% 0.59K 3998 53 127936K inode_cache
209664 209664 100% 0.02K 819 256 3276K kmalloc-16
187432 187432 100% 0.07K 3347 56 13388K Acpi-Operand
169611 168672 99% 0.20K 4349 39 34792K vm_area_struct
121856 121856 100% 0.01K 238 512 952K kmalloc-8
116978 116702 99% 0.09K 2543 46 10172K anon_vma
107856 99165 91% 0.09K 2568 42 10272K kmalloc-rcl-96
```
#### zpool
```
# zpool status pod-10
pool: pod-10
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
pod-10 ONLINE 0 0 0
raidz3-0 ONLINE 0 0 0
35000c500ae29a4bb ONLINE 0 0 0
35000c500ae95def3 ONLINE 0 0 0
35000c500ae96d1d7 ONLINE 0 0 0
35000c500ae9729af ONLINE 0 0 0
35000c500ae97296f ONLINE 0 0 0
35000c500ae96daeb ONLINE 0 0 0
35000c500ae968243 ONLINE 0 0 0
35000c500ae97269b ONLINE 0 0 0
35000c500ae970cd7 ONLINE 0 0 0
35000c500ae975a7f ONLINE 0 0 0
35000c500ae957c3b ONLINE 0 0 0
raidz3-1 ONLINE 0 0 0
35000c500ae96870f ONLINE 0 0 0
35000c500ae2bc057 ONLINE 0 0 0
35000c500ae9733b7 ONLINE 0 0 0
35000c500ae2bc4db ONLINE 0 0 0
35000c500ae96b4bf ONLINE 0 0 0
35000c500ae970e3b ONLINE 0 0 0
35000c500ae957bab ONLINE 0 0 0
35000c500ae96aa6f ONLINE 0 0 0
35000c500ae96833f ONLINE 0 0 0
35000c500ae96a4a3 ONLINE 0 0 0
35000c500ae2a6e9b ONLINE 0 0 0
raidz3-2 ONLINE 0 0 0
35000c500ae96c3f7 ONLINE 0 0 0
35000c500ae972ddf ONLINE 0 0 0
35000c500ae96bb4f ONLINE 0 0 0
35000c500ae95d66f ONLINE 0 0 0
35000c500ae96777f ONLINE 0 0 0
35000c500ae60fec3 ONLINE 0 0 0
35000c500ae96fcef ONLINE 0 0 0
35000c500ae966b0b ONLINE 0 0 0
35000c500ae96c823 ONLINE 0 0 0
35000c500ae95e363 ONLINE 0 0 0
35000c500ae96fab3 ONLINE 0 0 0
raidz3-3 ONLINE 0 0 0
35000c500ae34ceeb ONLINE 0 0 0
35000c500ae34d4c7 ONLINE 0 0 0
35000c500ae970af7 ONLINE 0 0 0
35000c500ae9597cb ONLINE 0 0 0
35000c500ae62c433 ONLINE 0 0 0
35000c500ae968857 ONLINE 0 0 0
35000c500ae970167 ONLINE 0 0 0
35000c500ae63517f ONLINE 0 0 0
35000c500ae961313 ONLINE 0 0 0
35000c500ae95d53b ONLINE 0 0 0
35000c500ae95cc5b ONLINE 0 0 0
raidz3-4 ONLINE 0 0 0
35000c500ae9737c3 ONLINE 0 0 0
35000c500ae970feb ONLINE 0 0 0
35000c500ae9686f3 ONLINE 0 0 0
35000c500ae97387b ONLINE 0 0 0
35000c500ae97403f ONLINE 0 0 0
35000c500ae95711f ONLINE 0 0 0
35000c500ae96cb23 ONLINE 0 0 0
35000c500ae2a6db7 ONLINE 0 0 0
35000c500ae9681a3 ONLINE 0 0 0
35000c500ae9688b7 ONLINE 0 0 0
35000c500ae97404b ONLINE 0 0 0
raidz3-5 ONLINE 0 0 0
35000c500ae956b87 ONLINE 0 0 0
35000c500ae974bf3 ONLINE 0 0 0
35000c500ae9744fb ONLINE 0 0 0
35000c500ae29e993 ONLINE 0 0 0
35000c500ae96ef57 ONLINE 0 0 0
35000c500ae974d2b ONLINE 0 0 0
35000c500ae970a0f ONLINE 0 0 0
35000c500ae39455f ONLINE 0 0 0
35000c500ae29d97f ONLINE 0 0 0
35000c500ae95712b ONLINE 0 0 0
35000c500ae9742c3 ONLINE 0 0 0
raidz3-6 ONLINE 0 0 0
35000c500ae96fc4b ONLINE 0 0 0
35000c500ae955eef ONLINE 0 0 0
35000c500ae95c243 ONLINE 0 0 0
35000c500ae974057 ONLINE 0 0 0
35000c500ae95e4cb ONLINE 0 0 0
35000c500ae96eab3 ONLINE 0 0 0
35000c500ae96c9bb ONLINE 0 0 0
35000c500ae959fd7 ONLINE 0 0 0
35000c500ae2a709f ONLINE 0 0 0
35000c500ae96d793 ONLINE 0 0 0
35000c500ae9728ab ONLINE 0 0 0
raidz3-7 ONLINE 0 0 0
35000c500ae962117 ONLINE 0 0 0
35000c500ae4044bf ONLINE 0 0 0
35000c500ae393b9b ONLINE 0 0 0
35000c500ae974223 ONLINE 0 0 0
35000c500ae95f193 ONLINE 0 0 0
35000c500ae957abb ONLINE 0 0 0
35000c500ae9563a3 ONLINE 0 0 0
35000c500ae9683ab ONLINE 0 0 0
35000c500ae96df8f ONLINE 0 0 0
35000c500ae96d543 ONLINE 0 0 0
35000c500ae972b2f ONLINE 0 0 0
raidz3-8 ONLINE 0 0 0
35000c500ae960517 ONLINE 0 0 0
35000c500ae95555b ONLINE 0 0 0
35000c500ae974637 ONLINE 0 0 0
35000c500ae972dd7 ONLINE 0 0 0
35000c500ae4323c7 ONLINE 0 0 0
35000c500ae96d617 ONLINE 0 0 0
35000c500ae9638b7 ONLINE 0 0 0
35000c500ae96ea0f ONLINE 0 0 0
35000c500ae96e3eb ONLINE 0 0 0
35000c500ae29a417 ONLINE 0 0 0
35000c500ae96fda3 ONLINE 0 0 0
logs
mirror-9 ONLINE 0 0 0
wwn-0x55cd2e4152220c93-part1 ONLINE 0 0 0
wwn-0x55cd2e41519b023b-part1 ONLINE 0 0 0
cache
wwn-0x55cd2e4152220c93-part2 ONLINE 0 0 0
wwn-0x55cd2e41519b023b-part2 ONLINE 0 0 0
wwn-0x55cd2e415221c8c7 ONLINE 0 0 0
spares
35000c500ae97206f AVAIL
35000c500ae965dbb AVAIL
35000c500ae2b926f AVAIL
35000c500ae976187 AVAIL
35000c500ae96dcf7 AVAIL
35000c500ae958777 AVAIL
35000c500ae41875f AVAIL
errors: No known data errors
```
#### zfs fs
```
# zfs list
NAME USED AVAIL REFER MOUNTPOINT
pod-10 120T 926T 279K /pod-10
pod-10/pod-10 120T 926T 120T /srv/pod-10
```
```
# zfs get primarycache,secondarycache pod-10/pod-10
NAME PROPERTY VALUE SOURCE
pod-10/pod-10 primarycache metadata local
pod-10/pod-10 secondarycache metadata local
```
Intention is for ARC/L2ARC to be entirely metadata.
#### initial arcstats
arcstats reported metadata usage above the limit.
```
ARC size (current): 102.8 % 64.6 GiB
Target size (adaptive): 100.0 % 62.8 GiB
Min size (hard limit): 6.2 % 3.9 GiB
Max size (high water): 16:1 62.8 GiB
Most Frequently Used (MFU) cache size: 13.0 % 2.6 GiB
Most Recently Used (MRU) cache size: 87.0 % 17.7 GiB
Metadata cache size (hard limit): 75.0 % 47.1 GiB
Metadata cache size (current): 135.8 % 64.0 GiB
Dnode cache size (hard limit): 10.0 % 4.7 GiB
Dnode cache size (current): 537.3 % 25.3 GiB
```
#### initial adjustment
Attempted to remediate by increasing ARC size by half of remaining RAM and increasing the metadata allocation in the ARC.
```
echo 86762369024 > /sys/module/zfs/parameters/zfs_arc_max
echo 90 > /sys/module/zfs/parameters/zfs_arc_meta_limit_percent
echo 50 > /sys/module/zfs/parameters/zfs_arc_dnode_limit_percent
```
#### current issue
This remediated the issue temporarily, and the prune processes stopped, but we are back! :)
Stopped NFS server and it seems to be free-ing memory albeit very slowly, many `dp_sync_taskq` processes... But removing the NFS server is very un-ideal. Given they are both living in the kernel, it's difficult for me personally to determine who is eating up the memory.
```
top - 18:05:52 up 21 days, 16:07, 2 users, load average: 2.52, 7.31, 35.56
Tasks: 807 total, 1 running, 806 sleeping, 0 stopped, 0 zombie
%Cpu(s): 0.0 us, 1.5 sy, 0.0 ni, 95.9 id, 2.5 wa, 0.0 hi, 0.1 si, 0.0 st
MiB Mem : 128622.1 total, 17620.3 free, 108471.5 used, 2530.3 buff/cache
MiB Swap: 1907.0 total, 1764.7 free, 142.2 used. 19277.1 avail Mem
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
248929 root 20 0 0 0 0 S 2.3 0.0 349:34.92 txg_sync
248864 root 39 19 0 0 0 S 1.7 0.0 33:52.18 dp_sync_taskq
1767630 root 20 0 0 0 0 D 1.7 0.0 3:35.82 kworker/23:2+events
248854 root 39 19 0 0 0 S 1.3 0.0 33:48.50 dp_sync_taskq
248855 root 39 19 0 0 0 S 1.3 0.0 33:51.96 dp_sync_taskq
248856 root 39 19 0 0 0 S 1.3 0.0 33:49.72 dp_sync_taskq
248857 root 39 19 0 0 0 S 1.3 0.0 33:49.15 dp_sync_taskq
248859 root 39 19 0 0 0 S 1.3 0.0 33:54.62 dp_sync_taskq
248860 root 39 19 0 0 0 S 1.3 0.0 33:50.09 dp_sync_taskq
248862 root 39 19 0 0 0 S 1.3 0.0 33:49.23 dp_sync_taskq
248865 root 39 19 0 0 0 S 1.3 0.0 33:50.26 dp_sync_taskq
248867 root 39 19 0 0 0 S 1.3 0.0 33:48.57 dp_sync_taskq
248870 root 39 19 0 0 0 S 1.3 0.0 33:51.20 dp_sync_taskq
248871 root 39 19 0 0 0 S 1.3 0.0 33:48.09 dp_sync_taskq
248872 root 39 19 0 0 0 S 1.3 0.0 33:51.06 dp_sync_taskq
248873 root 39 19 0 0 0 S 1.3 0.0 33:46.16 dp_sync_taskq
```
```
# arc_summary
------------------------------------------------------------------------
ZFS Subsystem Report Wed Feb 09 16:23:14 2022
Linux 5.4.0-96-generic 2.1.2-1
Machine: r8-n9 (x86_64) 2.1.2-1
ARC status: HEALTHY
Memory throttle count: 0
ARC size (current): 90.1 % 72.8 GiB
Target size (adaptive): 4.9 % 3.9 GiB
Min size (hard limit): 4.9 % 3.9 GiB
Max size (high water): 20:1 80.8 GiB
Most Frequently Used (MFU) cache size: 13.8 % 3.0 GiB
Most Recently Used (MRU) cache size: 86.2 % 18.7 GiB
Metadata cache size (hard limit): 90.0 % 72.7 GiB
Metadata cache size (current): 100.2 % 72.8 GiB
Dnode cache size (hard limit): 50.0 % 36.4 GiB
Dnode cache size (current): 80.9 % 29.4 GiB
ARC hash breakdown:
Elements max: 15.7M
Elements current: 97.2 % 15.2M
Collisions: 1.1G
Chain max: 10
Chains: 3.9M
ARC misc:
Deleted: 28.5M
Mutex misses: 6.6G
Eviction skips: 93.2G
Eviction skips due to L2 writes: 10.3k
L2 cached evictions: 1.4 TiB
L2 eligible evictions: 73.1 GiB
L2 eligible MFU evictions: 4.8 % 3.5 GiB
L2 eligible MRU evictions: 95.2 % 69.6 GiB
L2 ineligible evictions: 707.3 GiB
ARC total accesses (hits + misses): 11.3G
Cache hit ratio: 86.4 % 9.8G
Cache miss ratio: 13.6 % 1.5G
Actual hit ratio (MFU + MRU hits): 86.3 % 9.8G
Data demand efficiency: 1.6 % 1.6G
Data prefetch efficiency: n/a 0
Cache hits by cache type:
Most frequently used (MFU): 96.3 % 9.4G
Most recently used (MRU): 3.6 % 356.4M
Most frequently used (MFU) ghost: < 0.1 % 1.3M
Most recently used (MRU) ghost: < 0.1 % 1.0M
Cache hits by data type:
Demand data: 0.3 % 25.4M
Demand prefetch data: 0.0 % 0
Demand metadata: 99.7 % 9.8G
Demand prefetch metadata: < 0.1 % 2.4M
Cache misses by data type:
Demand data: 99.3 % 1.5G
Demand prefetch data: 0.0 % 0
Demand metadata: 0.4 % 5.9M
Demand prefetch metadata: 0.3 % 4.8M
DMU prefetch efficiency: 2.7G
Hit ratio: 5.7 % 152.9M
Miss ratio: 94.3 % 2.5G
L2ARC status: HEALTHY
Low memory aborts: 24.5k
Free on write: 281
R/W clashes: 0
Bad checksums: 0
I/O errors: 0
L2ARC size (adaptive): 1.3 TiB
Compressed: 5.6 % 74.6 GiB
Header size: 0.1 % 1.3 GiB
MFU allocated size: 19.4 % 14.5 GiB
MRU allocated size: 85.0 % 63.4 GiB
Prefetch allocated size: 0.1 % 67.6 MiB
Data (buffer content) allocated size: 0.0 % 0 Bytes
Metadata (buffer content) allocated size: 104.5 % 77.9 GiB
L2ARC breakdown: 1.5G
Hit ratio: 0.2 % 3.8M
Miss ratio: 99.8 % 1.5G
Feeds: 1.8M
L2ARC writes:
Writes sent: 100 % 1.4M
L2ARC evicts:
Lock retries: 0
Upon reading: 0
Solaris Porting Layer (SPL):
spl_hostid 0
spl_hostid_path /etc/hostid
spl_kmem_alloc_max 1048576
spl_kmem_alloc_warn 65536
spl_kmem_cache_kmem_threads 4
spl_kmem_cache_magazine_size 0
spl_kmem_cache_max_size 32
spl_kmem_cache_obj_per_slab 8
spl_kmem_cache_reclaim 0
spl_kmem_cache_slab_limit 16384
spl_max_show_tasks 512
spl_panic_halt 0
spl_schedule_hrtimeout_slack_us 0
spl_taskq_kick 0
spl_taskq_thread_bind 0
spl_taskq_thread_dynamic 1
spl_taskq_thread_priority 1
spl_taskq_thread_sequential 4
Tunables:
dbuf_cache_hiwater_pct 10
dbuf_cache_lowater_pct 10
dbuf_cache_max_bytes 18446744073709551615
dbuf_cache_shift 5
dbuf_metadata_cache_max_bytes 18446744073709551615
dbuf_metadata_cache_shift 6
dmu_object_alloc_chunk_shift 7
dmu_prefetch_max 134217728
ignore_hole_birth 1
l2arc_feed_again 1
l2arc_feed_min_ms 200
l2arc_feed_secs 1
l2arc_headroom 2
l2arc_headroom_boost 200
l2arc_meta_percent 33
l2arc_mfuonly 0
l2arc_noprefetch 1
l2arc_norw 0
l2arc_rebuild_blocks_min_l2size 1073741824
l2arc_rebuild_enabled 1
l2arc_trim_ahead 0
l2arc_write_boost 8388608
l2arc_write_max 8388608
metaslab_aliquot 524288
metaslab_bias_enabled 1
metaslab_debug_load 0
metaslab_debug_unload 0
metaslab_df_max_search 16777216
metaslab_df_use_largest_segment 0
metaslab_force_ganging 16777217
metaslab_fragmentation_factor_enabled 1
metaslab_lba_weighting_enabled 1
metaslab_preload_enabled 1
metaslab_unload_delay 32
metaslab_unload_delay_ms 600000
send_holes_without_birth_time 1
spa_asize_inflation 24
spa_config_path /etc/zfs/zpool.cache
spa_load_print_vdev_tree 0
spa_load_verify_data 1
spa_load_verify_metadata 1
spa_load_verify_shift 4
spa_slop_shift 5
vdev_file_logical_ashift 9
vdev_file_physical_ashift 9
vdev_removal_max_span 32768
vdev_validate_skip 0
zap_iterate_prefetch 1
zfetch_array_rd_sz 1048576
zfetch_max_distance 8388608
zfetch_max_idistance 67108864
zfetch_max_streams 8
zfetch_min_sec_reap 2
zfs_abd_scatter_enabled 1
zfs_abd_scatter_max_order 10
zfs_abd_scatter_min_size 1536
zfs_admin_snapshot 0
zfs_allow_redacted_dataset_mount 0
zfs_arc_average_blocksize 8192
zfs_arc_dnode_limit 0
zfs_arc_dnode_limit_percent 50
zfs_arc_dnode_reduce_percent 10
zfs_arc_evict_batch_limit 10
zfs_arc_eviction_pct 200
zfs_arc_grow_retry 0
zfs_arc_lotsfree_percent 10
zfs_arc_max 86762369024
zfs_arc_meta_adjust_restarts 4096
zfs_arc_meta_limit 0
zfs_arc_meta_limit_percent 90
zfs_arc_meta_min 0
zfs_arc_meta_prune 10000
zfs_arc_meta_strategy 1
zfs_arc_min 0
zfs_arc_min_prefetch_ms 0
zfs_arc_min_prescient_prefetch_ms 0
zfs_arc_p_dampener_disable 1
zfs_arc_p_min_shift 0
zfs_arc_pc_percent 0
zfs_arc_shrink_shift 0
zfs_arc_shrinker_limit 10000
zfs_arc_sys_free 0
zfs_async_block_max_blocks 18446744073709551615
zfs_autoimport_disable 1
zfs_checksum_events_per_second 20
zfs_commit_timeout_pct 5
zfs_compressed_arc_enabled 1
zfs_condense_indirect_commit_entry_delay_ms 0
zfs_condense_indirect_obsolete_pct 25
zfs_condense_indirect_vdevs_enable 1
zfs_condense_max_obsolete_bytes 1073741824
zfs_condense_min_mapping_bytes 131072
zfs_dbgmsg_enable 1
zfs_dbgmsg_maxsize 4194304
zfs_dbuf_state_index 0
zfs_ddt_data_is_special 1
zfs_deadman_checktime_ms 60000
zfs_deadman_enabled 1
zfs_deadman_failmode wait
zfs_deadman_synctime_ms 600000
zfs_deadman_ziotime_ms 300000
zfs_dedup_prefetch 0
zfs_delay_min_dirty_percent 60
zfs_delay_scale 500000
zfs_delete_blocks 20480
zfs_dirty_data_max 4294967296
zfs_dirty_data_max_max 4294967296
zfs_dirty_data_max_max_percent 25
zfs_dirty_data_max_percent 10
zfs_dirty_data_sync_percent 20
zfs_disable_ivset_guid_check 0
zfs_dmu_offset_next_sync 0
zfs_embedded_slog_min_ms 64
zfs_expire_snapshot 300
zfs_fallocate_reserve_percent 110
zfs_flags 0
zfs_free_bpobj_enabled 1
zfs_free_leak_on_eio 0
zfs_free_min_time_ms 1000
zfs_history_output_max 1048576
zfs_immediate_write_sz 32768
zfs_initialize_chunk_size 1048576
zfs_initialize_value 16045690984833335022
zfs_keep_log_spacemaps_at_export 0
zfs_key_max_salt_uses 400000000
zfs_livelist_condense_new_alloc 0
zfs_livelist_condense_sync_cancel 0
zfs_livelist_condense_sync_pause 0
zfs_livelist_condense_zthr_cancel 0
zfs_livelist_condense_zthr_pause 0
zfs_livelist_max_entries 500000
zfs_livelist_min_percent_shared 75
zfs_lua_max_instrlimit 100000000
zfs_lua_max_memlimit 104857600
zfs_max_async_dedup_frees 100000
zfs_max_log_walking 5
zfs_max_logsm_summary_length 10
zfs_max_missing_tvds 0
zfs_max_nvlist_src_size 0
zfs_max_recordsize 1048576
zfs_metaslab_find_max_tries 100
zfs_metaslab_fragmentation_threshold 70
zfs_metaslab_max_size_cache_sec 3600
zfs_metaslab_mem_limit 25
zfs_metaslab_segment_weight_enabled 1
zfs_metaslab_switch_threshold 2
zfs_metaslab_try_hard_before_gang 0
zfs_mg_fragmentation_threshold 95
zfs_mg_noalloc_threshold 0
zfs_min_metaslabs_to_flush 1
zfs_multihost_fail_intervals 10
zfs_multihost_history 0
zfs_multihost_import_intervals 20
zfs_multihost_interval 1000
zfs_multilist_num_sublists 0
zfs_no_scrub_io 0
zfs_no_scrub_prefetch 0
zfs_nocacheflush 0
zfs_nopwrite_enabled 1
zfs_object_mutex_size 64
zfs_obsolete_min_time_ms 500
zfs_override_estimate_recordsize 0
zfs_pd_bytes_max 52428800
zfs_per_txg_dirty_frees_percent 5
zfs_prefetch_disable 0
zfs_read_history 0
zfs_read_history_hits 0
zfs_rebuild_max_segment 1048576
zfs_rebuild_scrub_enabled 1
zfs_rebuild_vdev_limit 33554432
zfs_reconstruct_indirect_combinations_max 4096
zfs_recover 0
zfs_recv_queue_ff 20
zfs_recv_queue_length 16777216
zfs_recv_write_batch_size 1048576
zfs_removal_ignore_errors 0
zfs_removal_suspend_progress 0
zfs_remove_max_segment 16777216
zfs_resilver_disable_defer 0
zfs_resilver_min_time_ms 3000
zfs_scan_checkpoint_intval 7200
zfs_scan_fill_weight 3
zfs_scan_ignore_errors 0
zfs_scan_issue_strategy 0
zfs_scan_legacy 0
zfs_scan_max_ext_gap 2097152
zfs_scan_mem_lim_fact 20
zfs_scan_mem_lim_soft_fact 20
zfs_scan_strict_mem_lim 0
zfs_scan_suspend_progress 0
zfs_scan_vdev_limit 4194304
zfs_scrub_min_time_ms 1000
zfs_send_corrupt_data 0
zfs_send_no_prefetch_queue_ff 20
zfs_send_no_prefetch_queue_length 1048576
zfs_send_queue_ff 20
zfs_send_queue_length 16777216
zfs_send_unmodified_spill_blocks 1
zfs_slow_io_events_per_second 20
zfs_spa_discard_memory_limit 16777216
zfs_special_class_metadata_reserve_pct 25
zfs_sync_pass_deferred_free 2
zfs_sync_pass_dont_compress 8
zfs_sync_pass_rewrite 2
zfs_sync_taskq_batch_pct 75
zfs_traverse_indirect_prefetch_limit 32
zfs_trim_extent_bytes_max 134217728
zfs_trim_extent_bytes_min 32768
zfs_trim_metaslab_skip 0
zfs_trim_queue_limit 10
zfs_trim_txg_batch 32
zfs_txg_history 100
zfs_txg_timeout 5
zfs_unflushed_log_block_max 262144
zfs_unflushed_log_block_min 1000
zfs_unflushed_log_block_pct 400
zfs_unflushed_max_mem_amt 1073741824
zfs_unflushed_max_mem_ppm 1000
zfs_unlink_suspend_progress 0
zfs_user_indirect_is_special 1
zfs_vdev_aggregate_trim 0
zfs_vdev_aggregation_limit 1048576
zfs_vdev_aggregation_limit_non_rotating 131072
zfs_vdev_async_read_max_active 3
zfs_vdev_async_read_min_active 1
zfs_vdev_async_write_active_max_dirty_percent 60
zfs_vdev_async_write_active_min_dirty_percent 30
zfs_vdev_async_write_max_active 10
zfs_vdev_async_write_min_active 2
zfs_vdev_cache_bshift 16
zfs_vdev_cache_max 16384
zfs_vdev_cache_size 0
zfs_vdev_default_ms_count 200
zfs_vdev_default_ms_shift 29
zfs_vdev_initializing_max_active 1
zfs_vdev_initializing_min_active 1
zfs_vdev_max_active 1000
zfs_vdev_max_auto_ashift 16
zfs_vdev_min_auto_ashift 9
zfs_vdev_min_ms_count 16
zfs_vdev_mirror_non_rotating_inc 0
zfs_vdev_mirror_non_rotating_seek_inc 1
zfs_vdev_mirror_rotating_inc 0
zfs_vdev_mirror_rotating_seek_inc 5
zfs_vdev_mirror_rotating_seek_offset 1048576
zfs_vdev_ms_count_limit 131072
zfs_vdev_nia_credit 5
zfs_vdev_nia_delay 5
zfs_vdev_queue_depth_pct 1000
zfs_vdev_raidz_impl cycle [fastest] original scalar sse2 ssse3 avx2 avx512f avx512bw
zfs_vdev_read_gap_limit 32768
zfs_vdev_rebuild_max_active 3
zfs_vdev_rebuild_min_active 1
zfs_vdev_removal_max_active 2
zfs_vdev_removal_min_active 1
zfs_vdev_scheduler unused
zfs_vdev_scrub_max_active 3
zfs_vdev_scrub_min_active 1
zfs_vdev_sync_read_max_active 10
zfs_vdev_sync_read_min_active 10
zfs_vdev_sync_write_max_active 10
zfs_vdev_sync_write_min_active 10
zfs_vdev_trim_max_active 2
zfs_vdev_trim_min_active 1
zfs_vdev_write_gap_limit 4096
zfs_vnops_read_chunk_size 1048576
zfs_zevent_len_max 512
zfs_zevent_retain_expire_secs 900
zfs_zevent_retain_max 2000
zfs_zil_clean_taskq_maxalloc 1048576
zfs_zil_clean_taskq_minalloc 1024
zfs_zil_clean_taskq_nthr_pct 100
zil_maxblocksize 131072
zil_nocacheflush 0
zil_replay_disable 0
zil_slog_bulk 786432
zio_deadman_log_all 0
zio_dva_throttle_enabled 1
zio_requeue_io_start_cut_in_line 1
zio_slow_io_ms 30000
zio_taskq_batch_pct 80
zio_taskq_batch_tpq 0
zvol_inhibit_dev 0
zvol_major 230
zvol_max_discard_blocks 16384
zvol_prefetch_bytes 131072
zvol_request_sync 0
zvol_threads 32
zvol_volmode 1
VDEV cache disabled, skipping section
ZIL committed transactions: 3.2G
Commit requests: 404.3M
Flushes to stable storage: 369.8M
Transactions to SLOG storage pool: 167.6 TiB 1.6G
Transactions to non-SLOG storage pool: 0 Bytes 0
```
#### zed logs
```
-- Logs begin at Wed 2022-01-12 17:47:33 UTC, end at Wed 2022-02-09 17:58:06 UTC. --
Jan 19 02:15:35 r8-n9 systemd[1]: Started ZFS Event Daemon (zed).
Jan 19 02:15:35 r8-n9 zed[246555]: ZFS Event Daemon 2.1.2-1 (PID 246555)
Jan 19 02:15:35 r8-n9 zed[246555]: Processing events since eid=0
Jan 19 02:16:16 r8-n9 zed[249150]: eid=38 class=config_sync pool='pod-10'
Jan 19 02:25:41 r8-n9 zed[256421]: eid=45 class=vdev_add pool='pod-10'
Feb 09 14:46:20 r8-n9 zed[1838695]: eid=58 class=delay pool='pod-10' vdev=35000c500ae95cc5b size=4096 offset=4740717096960 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1036122
Feb 09 14:46:20 r8-n9 zed[1838693]: eid=57 class=delay pool='pod-10' vdev=35000c500ae970e3b size=20480 offset=4785921613824 priority=0 err=0 flags=0x40080c80 delay=30174ms
Feb 09 14:46:20 r8-n9 zed[1838700]: eid=60 class=delay pool='pod-10' vdev=35000c500ae34d4c7 size=4096 offset=4731204771840 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1059557
Feb 09 14:46:20 r8-n9 zed[1838703]: eid=62 class=delay pool='pod-10' vdev=35000c500ae96a4a3 size=4096 offset=4785771388928 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1033656
Feb 09 14:46:21 r8-n9 zed[1838803]: eid=70 class=delay pool='pod-10' vdev=35000c500ae95d66f size=4096 offset=4774821298176 priority=0 err=0 flags=0x180980 delay=30547ms bookmark=269:0:0:1041464
Feb 09 14:46:21 r8-n9 zed[1838814]: eid=76 class=delay pool='pod-10' vdev=35000c500ae62c433 size=4096 offset=4721981607936 priority=0 err=0 flags=0x180980 delay=30159ms bookmark=269:0:0:1044753
Feb 09 14:46:22 r8-n9 zed[1838836]: eid=77 class=delay pool='pod-10' vdev=35000c500ae29d97f size=4096 offset=4781092179968 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1053826
Feb 09 14:46:22 r8-n9 zed[1838842]: eid=78 class=delay pool='pod-10' vdev=35000c500ae29a4bb size=4096 offset=4776777699328 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1031840
Feb 09 14:46:22 r8-n9 zed[1838846]: eid=82 class=delay pool='pod-10' vdev=35000c500ae96aa6f size=4096 offset=4786110545920 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1033105
Feb 09 14:46:22 r8-n9 zed[1838875]: eid=84 class=delay pool='pod-10' vdev=35000c500ae95555b size=53248 offset=4774958784512 priority=0 err=0 flags=0x40080c80 delay=30460ms
Feb 09 14:46:22 r8-n9 zed[1838878]: eid=83 class=delay pool='pod-10' vdev=35000c500ae957c3b size=4096 offset=4743807332352 priority=0 err=0 flags=0x180980 delay=30668ms bookmark=269:0:0:1054047
Feb 09 14:46:22 r8-n9 zed[1838884]: eid=87 class=delay pool='pod-10' vdev=35000c500ae96d793 size=4096 offset=4768750641152 priority=0 err=0 flags=0x180980 delay=30444ms bookmark=269:0:0:1042517
Feb 09 14:46:22 r8-n9 zed[1838935]: eid=89 class=delay pool='pod-10' vdev=35000c500ae972b2f size=4096 offset=4769971810304 priority=0 err=0 flags=0x180980 delay=30623ms bookmark=269:0:0:1033391
Feb 09 14:46:22 r8-n9 zed[1838942]: eid=91 class=delay pool='pod-10' vdev=35000c500ae9737c3 size=4096 offset=4728066199552 priority=0 err=0 flags=0x180980 delay=30623ms bookmark=269:0:0:1057873
Feb 09 14:46:22 r8-n9 zed[1838963]: eid=94 class=delay pool='pod-10' vdev=35000c500ae95d53b size=4096 offset=4712411389952 priority=0 err=0 flags=0x180980 delay=30831ms bookmark=269:0:0:1049636
Feb 09 14:46:22 r8-n9 zed[1838967]: eid=95 class=delay pool='pod-10' vdev=35000c500ae970cd7 size=4096 offset=4776225902592 priority=0 err=0 flags=0x180980 delay=30626ms bookmark=269:0:0:1046985
Feb 09 14:46:22 r8-n9 zed[1838972]: eid=97 class=delay pool='pod-10' vdev=35000c500ae2a6e9b size=4096 offset=4782662496256 priority=0 err=0 flags=0x180980 delay=30831ms bookmark=269:0:0:1048773
Feb 09 14:46:22 r8-n9 zed[1838969]: eid=96 class=delay pool='pod-10' vdev=35000c500ae955eef size=20480 offset=4774534471680 priority=0 err=0 flags=0x40080c80 delay=30626ms
```
#### vmstat
```
# cat /proc/vmstat
nr_free_pages 497162
nr_zone_inactive_anon 77677
nr_zone_active_anon 81958
nr_zone_inactive_file 16392
nr_zone_active_file 10007
nr_zone_unevictable 7960
nr_zone_write_pending 844
nr_mlock 7960
nr_page_table_pages 1963
nr_kernel_stack 19296
nr_bounce 0
nr_zspages 0
nr_free_cma 0
numa_hit 151121274140
numa_miss 652564522
numa_foreign 652564522
numa_interleave 96982
numa_local 133392770941
numa_other 18381067721
nr_inactive_anon 77677
nr_active_anon 81958
nr_inactive_file 16392
nr_active_file 10007
nr_unevictable 7960
nr_slab_reclaimable 1104575
nr_slab_unreclaimable 28741700
nr_isolated_anon 0
nr_isolated_file 0
workingset_nodes 1973
workingset_refault 78664
workingset_activate 26788
workingset_restore 14173
workingset_nodereclaim 816
nr_anon_pages 162434
nr_mapped 20118
nr_file_pages 32246
nr_dirty 844
nr_writeback 0
nr_writeback_temp 0
nr_shmem 369
nr_shmem_hugepages 0
nr_shmem_pmdmapped 0
nr_file_hugepages 0
nr_file_pmdmapped 0
nr_anon_transparent_hugepages 0
nr_unstable 0
nr_vmscan_write 68263
nr_vmscan_immediate_reclaim 641
nr_dirtied 3379059
nr_written 3113482
nr_kernel_misc_reclaimable 0
nr_dirty_threshold 73743
nr_dirty_background_threshold 36826
pgpgin 194613865828
pgpgout 594833359918
pswpin 4218
pswpout 68258
pgalloc_dma 0
pgalloc_dma32 324399540
pgalloc_normal 245349772461
pgalloc_movable 0
allocstall_dma 0
allocstall_dma32 0
allocstall_normal 1
allocstall_movable 2
pgskip_dma 0
pgskip_dma32 0
pgskip_normal 0
pgskip_movable 0
pgfree 245674776191
pgactivate 730915
pgdeactivate 358591
pglazyfree 6389
pgfault 243384077
pgmajfault 15967
pglazyfreed 0
pgrefill 382033
pgsteal_kswapd 292238
pgsteal_direct 1375
pgscan_kswapd 477828
pgscan_direct 1399
pgscan_direct_throttle 0
zone_reclaim_failed 0
pginodesteal 74
slabs_scanned 459729476
kswapd_inodesteal 26293
kswapd_low_wmark_hit_quickly 1
kswapd_high_wmark_hit_quickly 14
pageoutrun 277
pgrotated 69682
drop_pagecache 1
drop_slab 1
oom_kill 0
numa_pte_updates 4332995
numa_huge_pte_updates 12
numa_hint_faults 3863526
numa_hint_faults_local 3281256
numa_pages_migrated 327088
pgmigrate_success 411548
pgmigrate_fail 3419
compact_migrate_scanned 10896483
compact_free_scanned 460891
compact_isolated 170144
compact_stall 4
compact_fail 4
compact_success 0
compact_daemon_wake 255
compact_daemon_migrate_scanned 1507268
compact_daemon_free_scanned 317747
htlb_buddy_alloc_success 0
htlb_buddy_alloc_fail 0
unevictable_pgs_culled 81639
unevictable_pgs_scanned 0
unevictable_pgs_rescued 14620
unevictable_pgs_mlocked 26080
unevictable_pgs_munlocked 16730
unevictable_pgs_cleared 1390
unevictable_pgs_stranded 1390
thp_fault_alloc 10
thp_fault_fallback 0
thp_collapse_alloc 8
thp_collapse_alloc_failed 2
thp_file_alloc 0
thp_file_mapped 0
thp_split_page 0
thp_split_page_failed 0
thp_deferred_split_page 18
thp_split_pmd 9
thp_split_pud 0
thp_zero_page_alloc 0
thp_zero_page_alloc_failed 0
thp_swpout 0
thp_swpout_fallback 0
balloon_inflate 0
balloon_deflate 0
balloon_migrate 0
swap_ra 1519
swap_ra_hit 833
```
#### buddyinfo
```
Node: 0
Zone: DMA
Free KiB in zone: 15876.00
Fragment size Free fragments Total available KiB
4096 1 4.0
8192 0 0.0
16384 0 0.0
32768 0 0.0
65536 2 128.0
131072 1 128.0
262144 1 256.0
524288 0 0.0
1048576 1 1024.0
2097152 1 2048.0
4194304 3 12288.0
Zone: DMA32
Free KiB in zone: 252500.00
Fragment size Free fragments Total available KiB
4096 1923 7692.0
8192 5143 41144.0
16384 2581 41296.0
32768 142 4544.0
65536 26 1664.0
131072 16 2048.0
262144 78 19968.0
524288 70 35840.0
1048576 36 36864.0
2097152 22 45056.0
4194304 4 16384.0
Zone: Normal
Free KiB in zone: 107160.00
Fragment size Free fragments Total available KiB
4096 1586 6344.0
8192 8328 66624.0
16384 1987 31792.0
32768 27 864.0
65536 24 1536.0
131072 0 0.0
262144 0 0.0
524288 0 0.0
1048576 0 0.0
2097152 0 0.0
4194304 0 0.0
Node: 1
Zone: Normal
Free KiB in zone: 2152516.00
Fragment size Free fragments Total available KiB
4096 149157 596628.0
8192 107244 857952.0
16384 3545 56720.0
32768 4110 131520.0
65536 3156 201984.0
131072 1610 206080.0
262144 297 76032.0
524288 48 24576.0
1048576 1 1024.0
2097152 0 0.0
4194304 0 0.0
```
### Describe how to reproduce the problem
Uncertain but it has occurred on 2 separate servers so it is likely to happen again.
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
| 1.0 | ZFS 2.1.2 + NFSv4 ARC stuck pruning/evicting, flatlines throughput - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | `Ubuntu`
Distribution Version | `20.04.3 LTS`
Kernel Version | `5.4.0-96-generic`
Architecture | `x86_64`
OpenZFS Version | `zfs-2.1.2-1`
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
ZFS is flatlined on throughput with an `arc_evict` and `arc_prune` process spinning at 100%. The workload is kernel NFS server (all NFSv4 clients) with ZFS 2.1.2 built from source. Characterized by high CPU `iowait` and throughput flatlining.
#### RAM Graph
<img width="1350" alt="Screen Shot 2022-02-09 at 9 06 55 AM" src="https://user-images.githubusercontent.com/16908622/153252531-9ee24c8e-3ecb-4242-b862-b860353f5625.png">
It's holding on to RAM pretty hard. This is where I begin to lose debugging expertise, having discovered slabs yesterday. :-)
#### top
```
top - 16:55:29 up 21 days, 14:56, 1 user, load average: 195.23, 124.60, 134.24
Tasks: 993 total, 2 running, 991 sleeping, 0 stopped, 0 zombie
%Cpu(s): 0.0 us, 7.3 sy, 0.0 ni, 5.0 id, 87.3 wa, 0.0 hi, 0.4 si, 0.0 st
MiB Mem : 128622.1 total, 2501.0 free, 121637.5 used, 4483.5 buff/cache
MiB Swap: 1907.0 total, 1042.7 free, 864.2 used. 6078.4 avail Mem
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
245944 root 20 0 0 0 0 R 100.0 0.0 549:36.86 arc_evict
245943 root 20 0 0 0 0 S 93.1 0.0 459:13.30 arc_prune
246555 root 20 0 953064 20876 2732 S 3.6 0.0 29:32.85 zed
3174406 root 20 0 12784 4872 3248 R 1.3 0.0 0:00.17 top
14154 root rt 0 356604 31088 8300 S 0.7 0.0 133:03.60 multipathd
3159749 root 20 0 0 0 0 D 0.7 0.0 0:00.11 nfsd
```
#### slabtop
```
# slabtop
Active / Total Objects (% used) : 215263329 / 222780310 (96.6%)
Active / Total Slabs (% used) : 6020764 / 6020764 (100.0%)
Active / Total Caches (% used) : 141 / 199 (70.9%)
Active / Total Size (% used) : 98093294.62K / 99549181.36K (98.5%)
Minimum / Average / Maximum Object : 0.01K / 0.45K / 16.75K
OBJS ACTIVE USE OBJ SIZE SLABS OBJ/SLAB CACHE SIZE NAME
27437256 27437256 100% 0.97K 831432 33 26605824K dnode_t
25866414 25866282 99% 0.38K 615867 42 9853872K dmu_buf_impl_t
24953184 24952001 99% 0.50K 779787 32 12476592K kmalloc-512
20654336 18708128 90% 0.06K 322724 64 1290896K kmalloc-64
14428134 14148450 98% 0.09K 343527 42 1374108K arc_buf_hdr_t_l2only
10290924 10290229 99% 0.09K 245022 42 980088K kmalloc-96
10017152 9881010 98% 0.03K 78259 128 313036K kmalloc-32
7602798 7575100 99% 0.19K 181019 42 1448152K dentry
7255278 7139757 98% 1.09K 250182 29 8005824K zfs_znode_cache
7225482 7140416 98% 0.24K 218954 33 1751632K sa_cache
5172144 5172144 100% 0.16K 107753 48 862024K nfsd4_stateids
5079424 5077737 99% 0.25K 158732 32 1269856K filp
4988088 4985288 99% 0.19K 118764 42 950112K cred_jar
4952610 4952610 100% 0.02K 29133 170 116532K lsm_file_cache
4951520 4951520 100% 0.28K 176840 28 1414720K nfsd4_files
4949376 4949376 100% 0.03K 38667 128 154668K fsnotify_mark_connector
4947816 4947816 100% 0.08K 97016 51 388064K Acpi-State
4947372 4947372 100% 0.11K 137427 36 549708K khugepaged_mm_slot
4862528 4858723 99% 0.06K 75977 64 303908K kmalloc-rcl-64
3813933 1054784 27% 0.31K 74783 51 1196528K arc_buf_hdr_t_full
3601017 3601017 100% 0.05K 49329 73 197316K nsproxy
3022149 1059649 35% 0.10K 77491 39 309964K abd_t
2598288 2591411 99% 0.57K 92796 28 1484736K radix_tree_node
1523068 1523068 100% 0.42K 41164 37 658624K nfsd4_openowners
1248128 1247029 99% 1.00K 39004 32 1248128K kmalloc-1k
1128681 1052509 93% 0.08K 22131 51 88524K arc_buf_t
960386 960375 99% 16.00K 480193 2 15366176K zio_buf_comb_16384
945488 942004 99% 8.00K 236372 4 7563904K kmalloc-8k
290745 268913 92% 0.10K 7455 39 29820K buffer_head
288000 287664 99% 0.13K 9600 30 38400K kernfs_node_cache
228160 226423 99% 0.06K 3565 64 14260K anon_vma_chain
211894 204295 96% 0.59K 3998 53 127936K inode_cache
209664 209664 100% 0.02K 819 256 3276K kmalloc-16
187432 187432 100% 0.07K 3347 56 13388K Acpi-Operand
169611 168672 99% 0.20K 4349 39 34792K vm_area_struct
121856 121856 100% 0.01K 238 512 952K kmalloc-8
116978 116702 99% 0.09K 2543 46 10172K anon_vma
107856 99165 91% 0.09K 2568 42 10272K kmalloc-rcl-96
```
#### zpool
```
# zpool status pod-10
pool: pod-10
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
pod-10 ONLINE 0 0 0
raidz3-0 ONLINE 0 0 0
35000c500ae29a4bb ONLINE 0 0 0
35000c500ae95def3 ONLINE 0 0 0
35000c500ae96d1d7 ONLINE 0 0 0
35000c500ae9729af ONLINE 0 0 0
35000c500ae97296f ONLINE 0 0 0
35000c500ae96daeb ONLINE 0 0 0
35000c500ae968243 ONLINE 0 0 0
35000c500ae97269b ONLINE 0 0 0
35000c500ae970cd7 ONLINE 0 0 0
35000c500ae975a7f ONLINE 0 0 0
35000c500ae957c3b ONLINE 0 0 0
raidz3-1 ONLINE 0 0 0
35000c500ae96870f ONLINE 0 0 0
35000c500ae2bc057 ONLINE 0 0 0
35000c500ae9733b7 ONLINE 0 0 0
35000c500ae2bc4db ONLINE 0 0 0
35000c500ae96b4bf ONLINE 0 0 0
35000c500ae970e3b ONLINE 0 0 0
35000c500ae957bab ONLINE 0 0 0
35000c500ae96aa6f ONLINE 0 0 0
35000c500ae96833f ONLINE 0 0 0
35000c500ae96a4a3 ONLINE 0 0 0
35000c500ae2a6e9b ONLINE 0 0 0
raidz3-2 ONLINE 0 0 0
35000c500ae96c3f7 ONLINE 0 0 0
35000c500ae972ddf ONLINE 0 0 0
35000c500ae96bb4f ONLINE 0 0 0
35000c500ae95d66f ONLINE 0 0 0
35000c500ae96777f ONLINE 0 0 0
35000c500ae60fec3 ONLINE 0 0 0
35000c500ae96fcef ONLINE 0 0 0
35000c500ae966b0b ONLINE 0 0 0
35000c500ae96c823 ONLINE 0 0 0
35000c500ae95e363 ONLINE 0 0 0
35000c500ae96fab3 ONLINE 0 0 0
raidz3-3 ONLINE 0 0 0
35000c500ae34ceeb ONLINE 0 0 0
35000c500ae34d4c7 ONLINE 0 0 0
35000c500ae970af7 ONLINE 0 0 0
35000c500ae9597cb ONLINE 0 0 0
35000c500ae62c433 ONLINE 0 0 0
35000c500ae968857 ONLINE 0 0 0
35000c500ae970167 ONLINE 0 0 0
35000c500ae63517f ONLINE 0 0 0
35000c500ae961313 ONLINE 0 0 0
35000c500ae95d53b ONLINE 0 0 0
35000c500ae95cc5b ONLINE 0 0 0
raidz3-4 ONLINE 0 0 0
35000c500ae9737c3 ONLINE 0 0 0
35000c500ae970feb ONLINE 0 0 0
35000c500ae9686f3 ONLINE 0 0 0
35000c500ae97387b ONLINE 0 0 0
35000c500ae97403f ONLINE 0 0 0
35000c500ae95711f ONLINE 0 0 0
35000c500ae96cb23 ONLINE 0 0 0
35000c500ae2a6db7 ONLINE 0 0 0
35000c500ae9681a3 ONLINE 0 0 0
35000c500ae9688b7 ONLINE 0 0 0
35000c500ae97404b ONLINE 0 0 0
raidz3-5 ONLINE 0 0 0
35000c500ae956b87 ONLINE 0 0 0
35000c500ae974bf3 ONLINE 0 0 0
35000c500ae9744fb ONLINE 0 0 0
35000c500ae29e993 ONLINE 0 0 0
35000c500ae96ef57 ONLINE 0 0 0
35000c500ae974d2b ONLINE 0 0 0
35000c500ae970a0f ONLINE 0 0 0
35000c500ae39455f ONLINE 0 0 0
35000c500ae29d97f ONLINE 0 0 0
35000c500ae95712b ONLINE 0 0 0
35000c500ae9742c3 ONLINE 0 0 0
raidz3-6 ONLINE 0 0 0
35000c500ae96fc4b ONLINE 0 0 0
35000c500ae955eef ONLINE 0 0 0
35000c500ae95c243 ONLINE 0 0 0
35000c500ae974057 ONLINE 0 0 0
35000c500ae95e4cb ONLINE 0 0 0
35000c500ae96eab3 ONLINE 0 0 0
35000c500ae96c9bb ONLINE 0 0 0
35000c500ae959fd7 ONLINE 0 0 0
35000c500ae2a709f ONLINE 0 0 0
35000c500ae96d793 ONLINE 0 0 0
35000c500ae9728ab ONLINE 0 0 0
raidz3-7 ONLINE 0 0 0
35000c500ae962117 ONLINE 0 0 0
35000c500ae4044bf ONLINE 0 0 0
35000c500ae393b9b ONLINE 0 0 0
35000c500ae974223 ONLINE 0 0 0
35000c500ae95f193 ONLINE 0 0 0
35000c500ae957abb ONLINE 0 0 0
35000c500ae9563a3 ONLINE 0 0 0
35000c500ae9683ab ONLINE 0 0 0
35000c500ae96df8f ONLINE 0 0 0
35000c500ae96d543 ONLINE 0 0 0
35000c500ae972b2f ONLINE 0 0 0
raidz3-8 ONLINE 0 0 0
35000c500ae960517 ONLINE 0 0 0
35000c500ae95555b ONLINE 0 0 0
35000c500ae974637 ONLINE 0 0 0
35000c500ae972dd7 ONLINE 0 0 0
35000c500ae4323c7 ONLINE 0 0 0
35000c500ae96d617 ONLINE 0 0 0
35000c500ae9638b7 ONLINE 0 0 0
35000c500ae96ea0f ONLINE 0 0 0
35000c500ae96e3eb ONLINE 0 0 0
35000c500ae29a417 ONLINE 0 0 0
35000c500ae96fda3 ONLINE 0 0 0
logs
mirror-9 ONLINE 0 0 0
wwn-0x55cd2e4152220c93-part1 ONLINE 0 0 0
wwn-0x55cd2e41519b023b-part1 ONLINE 0 0 0
cache
wwn-0x55cd2e4152220c93-part2 ONLINE 0 0 0
wwn-0x55cd2e41519b023b-part2 ONLINE 0 0 0
wwn-0x55cd2e415221c8c7 ONLINE 0 0 0
spares
35000c500ae97206f AVAIL
35000c500ae965dbb AVAIL
35000c500ae2b926f AVAIL
35000c500ae976187 AVAIL
35000c500ae96dcf7 AVAIL
35000c500ae958777 AVAIL
35000c500ae41875f AVAIL
errors: No known data errors
```
#### zfs fs
```
# zfs list
NAME USED AVAIL REFER MOUNTPOINT
pod-10 120T 926T 279K /pod-10
pod-10/pod-10 120T 926T 120T /srv/pod-10
```
```
# zfs get primarycache,secondarycache pod-10/pod-10
NAME PROPERTY VALUE SOURCE
pod-10/pod-10 primarycache metadata local
pod-10/pod-10 secondarycache metadata local
```
Intention is for ARC/L2ARC to be entirely metadata.
#### initial arcstats
arcstats reported metadata usage above the limit.
```
ARC size (current): 102.8 % 64.6 GiB
Target size (adaptive): 100.0 % 62.8 GiB
Min size (hard limit): 6.2 % 3.9 GiB
Max size (high water): 16:1 62.8 GiB
Most Frequently Used (MFU) cache size: 13.0 % 2.6 GiB
Most Recently Used (MRU) cache size: 87.0 % 17.7 GiB
Metadata cache size (hard limit): 75.0 % 47.1 GiB
Metadata cache size (current): 135.8 % 64.0 GiB
Dnode cache size (hard limit): 10.0 % 4.7 GiB
Dnode cache size (current): 537.3 % 25.3 GiB
```
#### initial adjustment
Attempted to remediate by increasing ARC size by half of remaining RAM and increasing the metadata allocation in the ARC.
```
echo 86762369024 > /sys/module/zfs/parameters/zfs_arc_max
echo 90 > /sys/module/zfs/parameters/zfs_arc_meta_limit_percent
echo 50 > /sys/module/zfs/parameters/zfs_arc_dnode_limit_percent
```
#### current issue
This remediated the issue temporarily, and the prune processes stopped, but we are back! :)
Stopped NFS server and it seems to be free-ing memory albeit very slowly, many `dp_sync_taskq` processes... But removing the NFS server is very un-ideal. Given they are both living in the kernel, it's difficult for me personally to determine who is eating up the memory.
```
top - 18:05:52 up 21 days, 16:07, 2 users, load average: 2.52, 7.31, 35.56
Tasks: 807 total, 1 running, 806 sleeping, 0 stopped, 0 zombie
%Cpu(s): 0.0 us, 1.5 sy, 0.0 ni, 95.9 id, 2.5 wa, 0.0 hi, 0.1 si, 0.0 st
MiB Mem : 128622.1 total, 17620.3 free, 108471.5 used, 2530.3 buff/cache
MiB Swap: 1907.0 total, 1764.7 free, 142.2 used. 19277.1 avail Mem
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
248929 root 20 0 0 0 0 S 2.3 0.0 349:34.92 txg_sync
248864 root 39 19 0 0 0 S 1.7 0.0 33:52.18 dp_sync_taskq
1767630 root 20 0 0 0 0 D 1.7 0.0 3:35.82 kworker/23:2+events
248854 root 39 19 0 0 0 S 1.3 0.0 33:48.50 dp_sync_taskq
248855 root 39 19 0 0 0 S 1.3 0.0 33:51.96 dp_sync_taskq
248856 root 39 19 0 0 0 S 1.3 0.0 33:49.72 dp_sync_taskq
248857 root 39 19 0 0 0 S 1.3 0.0 33:49.15 dp_sync_taskq
248859 root 39 19 0 0 0 S 1.3 0.0 33:54.62 dp_sync_taskq
248860 root 39 19 0 0 0 S 1.3 0.0 33:50.09 dp_sync_taskq
248862 root 39 19 0 0 0 S 1.3 0.0 33:49.23 dp_sync_taskq
248865 root 39 19 0 0 0 S 1.3 0.0 33:50.26 dp_sync_taskq
248867 root 39 19 0 0 0 S 1.3 0.0 33:48.57 dp_sync_taskq
248870 root 39 19 0 0 0 S 1.3 0.0 33:51.20 dp_sync_taskq
248871 root 39 19 0 0 0 S 1.3 0.0 33:48.09 dp_sync_taskq
248872 root 39 19 0 0 0 S 1.3 0.0 33:51.06 dp_sync_taskq
248873 root 39 19 0 0 0 S 1.3 0.0 33:46.16 dp_sync_taskq
```
```
# arc_summary
------------------------------------------------------------------------
ZFS Subsystem Report Wed Feb 09 16:23:14 2022
Linux 5.4.0-96-generic 2.1.2-1
Machine: r8-n9 (x86_64) 2.1.2-1
ARC status: HEALTHY
Memory throttle count: 0
ARC size (current): 90.1 % 72.8 GiB
Target size (adaptive): 4.9 % 3.9 GiB
Min size (hard limit): 4.9 % 3.9 GiB
Max size (high water): 20:1 80.8 GiB
Most Frequently Used (MFU) cache size: 13.8 % 3.0 GiB
Most Recently Used (MRU) cache size: 86.2 % 18.7 GiB
Metadata cache size (hard limit): 90.0 % 72.7 GiB
Metadata cache size (current): 100.2 % 72.8 GiB
Dnode cache size (hard limit): 50.0 % 36.4 GiB
Dnode cache size (current): 80.9 % 29.4 GiB
ARC hash breakdown:
Elements max: 15.7M
Elements current: 97.2 % 15.2M
Collisions: 1.1G
Chain max: 10
Chains: 3.9M
ARC misc:
Deleted: 28.5M
Mutex misses: 6.6G
Eviction skips: 93.2G
Eviction skips due to L2 writes: 10.3k
L2 cached evictions: 1.4 TiB
L2 eligible evictions: 73.1 GiB
L2 eligible MFU evictions: 4.8 % 3.5 GiB
L2 eligible MRU evictions: 95.2 % 69.6 GiB
L2 ineligible evictions: 707.3 GiB
ARC total accesses (hits + misses): 11.3G
Cache hit ratio: 86.4 % 9.8G
Cache miss ratio: 13.6 % 1.5G
Actual hit ratio (MFU + MRU hits): 86.3 % 9.8G
Data demand efficiency: 1.6 % 1.6G
Data prefetch efficiency: n/a 0
Cache hits by cache type:
Most frequently used (MFU): 96.3 % 9.4G
Most recently used (MRU): 3.6 % 356.4M
Most frequently used (MFU) ghost: < 0.1 % 1.3M
Most recently used (MRU) ghost: < 0.1 % 1.0M
Cache hits by data type:
Demand data: 0.3 % 25.4M
Demand prefetch data: 0.0 % 0
Demand metadata: 99.7 % 9.8G
Demand prefetch metadata: < 0.1 % 2.4M
Cache misses by data type:
Demand data: 99.3 % 1.5G
Demand prefetch data: 0.0 % 0
Demand metadata: 0.4 % 5.9M
Demand prefetch metadata: 0.3 % 4.8M
DMU prefetch efficiency: 2.7G
Hit ratio: 5.7 % 152.9M
Miss ratio: 94.3 % 2.5G
L2ARC status: HEALTHY
Low memory aborts: 24.5k
Free on write: 281
R/W clashes: 0
Bad checksums: 0
I/O errors: 0
L2ARC size (adaptive): 1.3 TiB
Compressed: 5.6 % 74.6 GiB
Header size: 0.1 % 1.3 GiB
MFU allocated size: 19.4 % 14.5 GiB
MRU allocated size: 85.0 % 63.4 GiB
Prefetch allocated size: 0.1 % 67.6 MiB
Data (buffer content) allocated size: 0.0 % 0 Bytes
Metadata (buffer content) allocated size: 104.5 % 77.9 GiB
L2ARC breakdown: 1.5G
Hit ratio: 0.2 % 3.8M
Miss ratio: 99.8 % 1.5G
Feeds: 1.8M
L2ARC writes:
Writes sent: 100 % 1.4M
L2ARC evicts:
Lock retries: 0
Upon reading: 0
Solaris Porting Layer (SPL):
spl_hostid 0
spl_hostid_path /etc/hostid
spl_kmem_alloc_max 1048576
spl_kmem_alloc_warn 65536
spl_kmem_cache_kmem_threads 4
spl_kmem_cache_magazine_size 0
spl_kmem_cache_max_size 32
spl_kmem_cache_obj_per_slab 8
spl_kmem_cache_reclaim 0
spl_kmem_cache_slab_limit 16384
spl_max_show_tasks 512
spl_panic_halt 0
spl_schedule_hrtimeout_slack_us 0
spl_taskq_kick 0
spl_taskq_thread_bind 0
spl_taskq_thread_dynamic 1
spl_taskq_thread_priority 1
spl_taskq_thread_sequential 4
Tunables:
dbuf_cache_hiwater_pct 10
dbuf_cache_lowater_pct 10
dbuf_cache_max_bytes 18446744073709551615
dbuf_cache_shift 5
dbuf_metadata_cache_max_bytes 18446744073709551615
dbuf_metadata_cache_shift 6
dmu_object_alloc_chunk_shift 7
dmu_prefetch_max 134217728
ignore_hole_birth 1
l2arc_feed_again 1
l2arc_feed_min_ms 200
l2arc_feed_secs 1
l2arc_headroom 2
l2arc_headroom_boost 200
l2arc_meta_percent 33
l2arc_mfuonly 0
l2arc_noprefetch 1
l2arc_norw 0
l2arc_rebuild_blocks_min_l2size 1073741824
l2arc_rebuild_enabled 1
l2arc_trim_ahead 0
l2arc_write_boost 8388608
l2arc_write_max 8388608
metaslab_aliquot 524288
metaslab_bias_enabled 1
metaslab_debug_load 0
metaslab_debug_unload 0
metaslab_df_max_search 16777216
metaslab_df_use_largest_segment 0
metaslab_force_ganging 16777217
metaslab_fragmentation_factor_enabled 1
metaslab_lba_weighting_enabled 1
metaslab_preload_enabled 1
metaslab_unload_delay 32
metaslab_unload_delay_ms 600000
send_holes_without_birth_time 1
spa_asize_inflation 24
spa_config_path /etc/zfs/zpool.cache
spa_load_print_vdev_tree 0
spa_load_verify_data 1
spa_load_verify_metadata 1
spa_load_verify_shift 4
spa_slop_shift 5
vdev_file_logical_ashift 9
vdev_file_physical_ashift 9
vdev_removal_max_span 32768
vdev_validate_skip 0
zap_iterate_prefetch 1
zfetch_array_rd_sz 1048576
zfetch_max_distance 8388608
zfetch_max_idistance 67108864
zfetch_max_streams 8
zfetch_min_sec_reap 2
zfs_abd_scatter_enabled 1
zfs_abd_scatter_max_order 10
zfs_abd_scatter_min_size 1536
zfs_admin_snapshot 0
zfs_allow_redacted_dataset_mount 0
zfs_arc_average_blocksize 8192
zfs_arc_dnode_limit 0
zfs_arc_dnode_limit_percent 50
zfs_arc_dnode_reduce_percent 10
zfs_arc_evict_batch_limit 10
zfs_arc_eviction_pct 200
zfs_arc_grow_retry 0
zfs_arc_lotsfree_percent 10
zfs_arc_max 86762369024
zfs_arc_meta_adjust_restarts 4096
zfs_arc_meta_limit 0
zfs_arc_meta_limit_percent 90
zfs_arc_meta_min 0
zfs_arc_meta_prune 10000
zfs_arc_meta_strategy 1
zfs_arc_min 0
zfs_arc_min_prefetch_ms 0
zfs_arc_min_prescient_prefetch_ms 0
zfs_arc_p_dampener_disable 1
zfs_arc_p_min_shift 0
zfs_arc_pc_percent 0
zfs_arc_shrink_shift 0
zfs_arc_shrinker_limit 10000
zfs_arc_sys_free 0
zfs_async_block_max_blocks 18446744073709551615
zfs_autoimport_disable 1
zfs_checksum_events_per_second 20
zfs_commit_timeout_pct 5
zfs_compressed_arc_enabled 1
zfs_condense_indirect_commit_entry_delay_ms 0
zfs_condense_indirect_obsolete_pct 25
zfs_condense_indirect_vdevs_enable 1
zfs_condense_max_obsolete_bytes 1073741824
zfs_condense_min_mapping_bytes 131072
zfs_dbgmsg_enable 1
zfs_dbgmsg_maxsize 4194304
zfs_dbuf_state_index 0
zfs_ddt_data_is_special 1
zfs_deadman_checktime_ms 60000
zfs_deadman_enabled 1
zfs_deadman_failmode wait
zfs_deadman_synctime_ms 600000
zfs_deadman_ziotime_ms 300000
zfs_dedup_prefetch 0
zfs_delay_min_dirty_percent 60
zfs_delay_scale 500000
zfs_delete_blocks 20480
zfs_dirty_data_max 4294967296
zfs_dirty_data_max_max 4294967296
zfs_dirty_data_max_max_percent 25
zfs_dirty_data_max_percent 10
zfs_dirty_data_sync_percent 20
zfs_disable_ivset_guid_check 0
zfs_dmu_offset_next_sync 0
zfs_embedded_slog_min_ms 64
zfs_expire_snapshot 300
zfs_fallocate_reserve_percent 110
zfs_flags 0
zfs_free_bpobj_enabled 1
zfs_free_leak_on_eio 0
zfs_free_min_time_ms 1000
zfs_history_output_max 1048576
zfs_immediate_write_sz 32768
zfs_initialize_chunk_size 1048576
zfs_initialize_value 16045690984833335022
zfs_keep_log_spacemaps_at_export 0
zfs_key_max_salt_uses 400000000
zfs_livelist_condense_new_alloc 0
zfs_livelist_condense_sync_cancel 0
zfs_livelist_condense_sync_pause 0
zfs_livelist_condense_zthr_cancel 0
zfs_livelist_condense_zthr_pause 0
zfs_livelist_max_entries 500000
zfs_livelist_min_percent_shared 75
zfs_lua_max_instrlimit 100000000
zfs_lua_max_memlimit 104857600
zfs_max_async_dedup_frees 100000
zfs_max_log_walking 5
zfs_max_logsm_summary_length 10
zfs_max_missing_tvds 0
zfs_max_nvlist_src_size 0
zfs_max_recordsize 1048576
zfs_metaslab_find_max_tries 100
zfs_metaslab_fragmentation_threshold 70
zfs_metaslab_max_size_cache_sec 3600
zfs_metaslab_mem_limit 25
zfs_metaslab_segment_weight_enabled 1
zfs_metaslab_switch_threshold 2
zfs_metaslab_try_hard_before_gang 0
zfs_mg_fragmentation_threshold 95
zfs_mg_noalloc_threshold 0
zfs_min_metaslabs_to_flush 1
zfs_multihost_fail_intervals 10
zfs_multihost_history 0
zfs_multihost_import_intervals 20
zfs_multihost_interval 1000
zfs_multilist_num_sublists 0
zfs_no_scrub_io 0
zfs_no_scrub_prefetch 0
zfs_nocacheflush 0
zfs_nopwrite_enabled 1
zfs_object_mutex_size 64
zfs_obsolete_min_time_ms 500
zfs_override_estimate_recordsize 0
zfs_pd_bytes_max 52428800
zfs_per_txg_dirty_frees_percent 5
zfs_prefetch_disable 0
zfs_read_history 0
zfs_read_history_hits 0
zfs_rebuild_max_segment 1048576
zfs_rebuild_scrub_enabled 1
zfs_rebuild_vdev_limit 33554432
zfs_reconstruct_indirect_combinations_max 4096
zfs_recover 0
zfs_recv_queue_ff 20
zfs_recv_queue_length 16777216
zfs_recv_write_batch_size 1048576
zfs_removal_ignore_errors 0
zfs_removal_suspend_progress 0
zfs_remove_max_segment 16777216
zfs_resilver_disable_defer 0
zfs_resilver_min_time_ms 3000
zfs_scan_checkpoint_intval 7200
zfs_scan_fill_weight 3
zfs_scan_ignore_errors 0
zfs_scan_issue_strategy 0
zfs_scan_legacy 0
zfs_scan_max_ext_gap 2097152
zfs_scan_mem_lim_fact 20
zfs_scan_mem_lim_soft_fact 20
zfs_scan_strict_mem_lim 0
zfs_scan_suspend_progress 0
zfs_scan_vdev_limit 4194304
zfs_scrub_min_time_ms 1000
zfs_send_corrupt_data 0
zfs_send_no_prefetch_queue_ff 20
zfs_send_no_prefetch_queue_length 1048576
zfs_send_queue_ff 20
zfs_send_queue_length 16777216
zfs_send_unmodified_spill_blocks 1
zfs_slow_io_events_per_second 20
zfs_spa_discard_memory_limit 16777216
zfs_special_class_metadata_reserve_pct 25
zfs_sync_pass_deferred_free 2
zfs_sync_pass_dont_compress 8
zfs_sync_pass_rewrite 2
zfs_sync_taskq_batch_pct 75
zfs_traverse_indirect_prefetch_limit 32
zfs_trim_extent_bytes_max 134217728
zfs_trim_extent_bytes_min 32768
zfs_trim_metaslab_skip 0
zfs_trim_queue_limit 10
zfs_trim_txg_batch 32
zfs_txg_history 100
zfs_txg_timeout 5
zfs_unflushed_log_block_max 262144
zfs_unflushed_log_block_min 1000
zfs_unflushed_log_block_pct 400
zfs_unflushed_max_mem_amt 1073741824
zfs_unflushed_max_mem_ppm 1000
zfs_unlink_suspend_progress 0
zfs_user_indirect_is_special 1
zfs_vdev_aggregate_trim 0
zfs_vdev_aggregation_limit 1048576
zfs_vdev_aggregation_limit_non_rotating 131072
zfs_vdev_async_read_max_active 3
zfs_vdev_async_read_min_active 1
zfs_vdev_async_write_active_max_dirty_percent 60
zfs_vdev_async_write_active_min_dirty_percent 30
zfs_vdev_async_write_max_active 10
zfs_vdev_async_write_min_active 2
zfs_vdev_cache_bshift 16
zfs_vdev_cache_max 16384
zfs_vdev_cache_size 0
zfs_vdev_default_ms_count 200
zfs_vdev_default_ms_shift 29
zfs_vdev_initializing_max_active 1
zfs_vdev_initializing_min_active 1
zfs_vdev_max_active 1000
zfs_vdev_max_auto_ashift 16
zfs_vdev_min_auto_ashift 9
zfs_vdev_min_ms_count 16
zfs_vdev_mirror_non_rotating_inc 0
zfs_vdev_mirror_non_rotating_seek_inc 1
zfs_vdev_mirror_rotating_inc 0
zfs_vdev_mirror_rotating_seek_inc 5
zfs_vdev_mirror_rotating_seek_offset 1048576
zfs_vdev_ms_count_limit 131072
zfs_vdev_nia_credit 5
zfs_vdev_nia_delay 5
zfs_vdev_queue_depth_pct 1000
zfs_vdev_raidz_impl cycle [fastest] original scalar sse2 ssse3 avx2 avx512f avx512bw
zfs_vdev_read_gap_limit 32768
zfs_vdev_rebuild_max_active 3
zfs_vdev_rebuild_min_active 1
zfs_vdev_removal_max_active 2
zfs_vdev_removal_min_active 1
zfs_vdev_scheduler unused
zfs_vdev_scrub_max_active 3
zfs_vdev_scrub_min_active 1
zfs_vdev_sync_read_max_active 10
zfs_vdev_sync_read_min_active 10
zfs_vdev_sync_write_max_active 10
zfs_vdev_sync_write_min_active 10
zfs_vdev_trim_max_active 2
zfs_vdev_trim_min_active 1
zfs_vdev_write_gap_limit 4096
zfs_vnops_read_chunk_size 1048576
zfs_zevent_len_max 512
zfs_zevent_retain_expire_secs 900
zfs_zevent_retain_max 2000
zfs_zil_clean_taskq_maxalloc 1048576
zfs_zil_clean_taskq_minalloc 1024
zfs_zil_clean_taskq_nthr_pct 100
zil_maxblocksize 131072
zil_nocacheflush 0
zil_replay_disable 0
zil_slog_bulk 786432
zio_deadman_log_all 0
zio_dva_throttle_enabled 1
zio_requeue_io_start_cut_in_line 1
zio_slow_io_ms 30000
zio_taskq_batch_pct 80
zio_taskq_batch_tpq 0
zvol_inhibit_dev 0
zvol_major 230
zvol_max_discard_blocks 16384
zvol_prefetch_bytes 131072
zvol_request_sync 0
zvol_threads 32
zvol_volmode 1
VDEV cache disabled, skipping section
ZIL committed transactions: 3.2G
Commit requests: 404.3M
Flushes to stable storage: 369.8M
Transactions to SLOG storage pool: 167.6 TiB 1.6G
Transactions to non-SLOG storage pool: 0 Bytes 0
```
#### zed logs
```
-- Logs begin at Wed 2022-01-12 17:47:33 UTC, end at Wed 2022-02-09 17:58:06 UTC. --
Jan 19 02:15:35 r8-n9 systemd[1]: Started ZFS Event Daemon (zed).
Jan 19 02:15:35 r8-n9 zed[246555]: ZFS Event Daemon 2.1.2-1 (PID 246555)
Jan 19 02:15:35 r8-n9 zed[246555]: Processing events since eid=0
Jan 19 02:16:16 r8-n9 zed[249150]: eid=38 class=config_sync pool='pod-10'
Jan 19 02:25:41 r8-n9 zed[256421]: eid=45 class=vdev_add pool='pod-10'
Feb 09 14:46:20 r8-n9 zed[1838695]: eid=58 class=delay pool='pod-10' vdev=35000c500ae95cc5b size=4096 offset=4740717096960 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1036122
Feb 09 14:46:20 r8-n9 zed[1838693]: eid=57 class=delay pool='pod-10' vdev=35000c500ae970e3b size=20480 offset=4785921613824 priority=0 err=0 flags=0x40080c80 delay=30174ms
Feb 09 14:46:20 r8-n9 zed[1838700]: eid=60 class=delay pool='pod-10' vdev=35000c500ae34d4c7 size=4096 offset=4731204771840 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1059557
Feb 09 14:46:20 r8-n9 zed[1838703]: eid=62 class=delay pool='pod-10' vdev=35000c500ae96a4a3 size=4096 offset=4785771388928 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1033656
Feb 09 14:46:21 r8-n9 zed[1838803]: eid=70 class=delay pool='pod-10' vdev=35000c500ae95d66f size=4096 offset=4774821298176 priority=0 err=0 flags=0x180980 delay=30547ms bookmark=269:0:0:1041464
Feb 09 14:46:21 r8-n9 zed[1838814]: eid=76 class=delay pool='pod-10' vdev=35000c500ae62c433 size=4096 offset=4721981607936 priority=0 err=0 flags=0x180980 delay=30159ms bookmark=269:0:0:1044753
Feb 09 14:46:22 r8-n9 zed[1838836]: eid=77 class=delay pool='pod-10' vdev=35000c500ae29d97f size=4096 offset=4781092179968 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1053826
Feb 09 14:46:22 r8-n9 zed[1838842]: eid=78 class=delay pool='pod-10' vdev=35000c500ae29a4bb size=4096 offset=4776777699328 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1031840
Feb 09 14:46:22 r8-n9 zed[1838846]: eid=82 class=delay pool='pod-10' vdev=35000c500ae96aa6f size=4096 offset=4786110545920 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1033105
Feb 09 14:46:22 r8-n9 zed[1838875]: eid=84 class=delay pool='pod-10' vdev=35000c500ae95555b size=53248 offset=4774958784512 priority=0 err=0 flags=0x40080c80 delay=30460ms
Feb 09 14:46:22 r8-n9 zed[1838878]: eid=83 class=delay pool='pod-10' vdev=35000c500ae957c3b size=4096 offset=4743807332352 priority=0 err=0 flags=0x180980 delay=30668ms bookmark=269:0:0:1054047
Feb 09 14:46:22 r8-n9 zed[1838884]: eid=87 class=delay pool='pod-10' vdev=35000c500ae96d793 size=4096 offset=4768750641152 priority=0 err=0 flags=0x180980 delay=30444ms bookmark=269:0:0:1042517
Feb 09 14:46:22 r8-n9 zed[1838935]: eid=89 class=delay pool='pod-10' vdev=35000c500ae972b2f size=4096 offset=4769971810304 priority=0 err=0 flags=0x180980 delay=30623ms bookmark=269:0:0:1033391
Feb 09 14:46:22 r8-n9 zed[1838942]: eid=91 class=delay pool='pod-10' vdev=35000c500ae9737c3 size=4096 offset=4728066199552 priority=0 err=0 flags=0x180980 delay=30623ms bookmark=269:0:0:1057873
Feb 09 14:46:22 r8-n9 zed[1838963]: eid=94 class=delay pool='pod-10' vdev=35000c500ae95d53b size=4096 offset=4712411389952 priority=0 err=0 flags=0x180980 delay=30831ms bookmark=269:0:0:1049636
Feb 09 14:46:22 r8-n9 zed[1838967]: eid=95 class=delay pool='pod-10' vdev=35000c500ae970cd7 size=4096 offset=4776225902592 priority=0 err=0 flags=0x180980 delay=30626ms bookmark=269:0:0:1046985
Feb 09 14:46:22 r8-n9 zed[1838972]: eid=97 class=delay pool='pod-10' vdev=35000c500ae2a6e9b size=4096 offset=4782662496256 priority=0 err=0 flags=0x180980 delay=30831ms bookmark=269:0:0:1048773
Feb 09 14:46:22 r8-n9 zed[1838969]: eid=96 class=delay pool='pod-10' vdev=35000c500ae955eef size=20480 offset=4774534471680 priority=0 err=0 flags=0x40080c80 delay=30626ms
```
#### vmstat
```
# cat /proc/vmstat
nr_free_pages 497162
nr_zone_inactive_anon 77677
nr_zone_active_anon 81958
nr_zone_inactive_file 16392
nr_zone_active_file 10007
nr_zone_unevictable 7960
nr_zone_write_pending 844
nr_mlock 7960
nr_page_table_pages 1963
nr_kernel_stack 19296
nr_bounce 0
nr_zspages 0
nr_free_cma 0
numa_hit 151121274140
numa_miss 652564522
numa_foreign 652564522
numa_interleave 96982
numa_local 133392770941
numa_other 18381067721
nr_inactive_anon 77677
nr_active_anon 81958
nr_inactive_file 16392
nr_active_file 10007
nr_unevictable 7960
nr_slab_reclaimable 1104575
nr_slab_unreclaimable 28741700
nr_isolated_anon 0
nr_isolated_file 0
workingset_nodes 1973
workingset_refault 78664
workingset_activate 26788
workingset_restore 14173
workingset_nodereclaim 816
nr_anon_pages 162434
nr_mapped 20118
nr_file_pages 32246
nr_dirty 844
nr_writeback 0
nr_writeback_temp 0
nr_shmem 369
nr_shmem_hugepages 0
nr_shmem_pmdmapped 0
nr_file_hugepages 0
nr_file_pmdmapped 0
nr_anon_transparent_hugepages 0
nr_unstable 0
nr_vmscan_write 68263
nr_vmscan_immediate_reclaim 641
nr_dirtied 3379059
nr_written 3113482
nr_kernel_misc_reclaimable 0
nr_dirty_threshold 73743
nr_dirty_background_threshold 36826
pgpgin 194613865828
pgpgout 594833359918
pswpin 4218
pswpout 68258
pgalloc_dma 0
pgalloc_dma32 324399540
pgalloc_normal 245349772461
pgalloc_movable 0
allocstall_dma 0
allocstall_dma32 0
allocstall_normal 1
allocstall_movable 2
pgskip_dma 0
pgskip_dma32 0
pgskip_normal 0
pgskip_movable 0
pgfree 245674776191
pgactivate 730915
pgdeactivate 358591
pglazyfree 6389
pgfault 243384077
pgmajfault 15967
pglazyfreed 0
pgrefill 382033
pgsteal_kswapd 292238
pgsteal_direct 1375
pgscan_kswapd 477828
pgscan_direct 1399
pgscan_direct_throttle 0
zone_reclaim_failed 0
pginodesteal 74
slabs_scanned 459729476
kswapd_inodesteal 26293
kswapd_low_wmark_hit_quickly 1
kswapd_high_wmark_hit_quickly 14
pageoutrun 277
pgrotated 69682
drop_pagecache 1
drop_slab 1
oom_kill 0
numa_pte_updates 4332995
numa_huge_pte_updates 12
numa_hint_faults 3863526
numa_hint_faults_local 3281256
numa_pages_migrated 327088
pgmigrate_success 411548
pgmigrate_fail 3419
compact_migrate_scanned 10896483
compact_free_scanned 460891
compact_isolated 170144
compact_stall 4
compact_fail 4
compact_success 0
compact_daemon_wake 255
compact_daemon_migrate_scanned 1507268
compact_daemon_free_scanned 317747
htlb_buddy_alloc_success 0
htlb_buddy_alloc_fail 0
unevictable_pgs_culled 81639
unevictable_pgs_scanned 0
unevictable_pgs_rescued 14620
unevictable_pgs_mlocked 26080
unevictable_pgs_munlocked 16730
unevictable_pgs_cleared 1390
unevictable_pgs_stranded 1390
thp_fault_alloc 10
thp_fault_fallback 0
thp_collapse_alloc 8
thp_collapse_alloc_failed 2
thp_file_alloc 0
thp_file_mapped 0
thp_split_page 0
thp_split_page_failed 0
thp_deferred_split_page 18
thp_split_pmd 9
thp_split_pud 0
thp_zero_page_alloc 0
thp_zero_page_alloc_failed 0
thp_swpout 0
thp_swpout_fallback 0
balloon_inflate 0
balloon_deflate 0
balloon_migrate 0
swap_ra 1519
swap_ra_hit 833
```
#### buddyinfo
```
Node: 0
Zone: DMA
Free KiB in zone: 15876.00
Fragment size Free fragments Total available KiB
4096 1 4.0
8192 0 0.0
16384 0 0.0
32768 0 0.0
65536 2 128.0
131072 1 128.0
262144 1 256.0
524288 0 0.0
1048576 1 1024.0
2097152 1 2048.0
4194304 3 12288.0
Zone: DMA32
Free KiB in zone: 252500.00
Fragment size Free fragments Total available KiB
4096 1923 7692.0
8192 5143 41144.0
16384 2581 41296.0
32768 142 4544.0
65536 26 1664.0
131072 16 2048.0
262144 78 19968.0
524288 70 35840.0
1048576 36 36864.0
2097152 22 45056.0
4194304 4 16384.0
Zone: Normal
Free KiB in zone: 107160.00
Fragment size Free fragments Total available KiB
4096 1586 6344.0
8192 8328 66624.0
16384 1987 31792.0
32768 27 864.0
65536 24 1536.0
131072 0 0.0
262144 0 0.0
524288 0 0.0
1048576 0 0.0
2097152 0 0.0
4194304 0 0.0
Node: 1
Zone: Normal
Free KiB in zone: 2152516.00
Fragment size Free fragments Total available KiB
4096 149157 596628.0
8192 107244 857952.0
16384 3545 56720.0
32768 4110 131520.0
65536 3156 201984.0
131072 1610 206080.0
262144 297 76032.0
524288 48 24576.0
1048576 1 1024.0
2097152 0 0.0
4194304 0 0.0
```
### Describe how to reproduce the problem
Uncertain but it has occurred on 2 separate servers so it is likely to happen again.
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
| defect | zfs arc stuck pruning evicting flatlines throughput thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name ubuntu distribution version lts kernel version generic architecture openzfs version zfs command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing zfs is flatlined on throughput with an arc evict and arc prune process spinning at the workload is kernel nfs server all clients with zfs built from source characterized by high cpu iowait and throughput flatlining ram graph img width alt screen shot at am src it s holding on to ram pretty hard this is where i begin to lose debugging expertise having discovered slabs yesterday top top up days user load average tasks total running sleeping stopped zombie cpu s us sy ni id wa hi si st mib mem total free used buff cache mib swap total free used avail mem pid user pr ni virt res shr s cpu mem time command root r arc evict root s arc prune root s zed root r top root rt s multipathd root d nfsd slabtop slabtop active total objects used active total slabs used active total caches used active total size used minimum average maximum object objs active use obj size slabs obj slab cache size name dnode t dmu buf impl t kmalloc kmalloc arc buf hdr t kmalloc kmalloc dentry zfs znode cache sa cache stateids filp cred jar lsm file cache files fsnotify mark connector acpi state khugepaged mm slot kmalloc rcl arc buf hdr t full nsproxy abd t radix tree node openowners kmalloc arc buf t zio buf comb kmalloc buffer head kernfs node cache anon vma chain inode cache kmalloc acpi operand vm area struct kmalloc anon vma kmalloc rcl zpool zpool status pod pool pod state online config name state read write cksum pod online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online logs mirror online wwn online wwn online cache wwn online wwn online wwn online spares avail avail avail avail avail avail avail errors no known data errors zfs fs zfs list name used avail refer mountpoint pod pod pod pod srv pod zfs get primarycache secondarycache pod pod name property value source pod pod primarycache metadata local pod pod secondarycache metadata local intention is for arc to be entirely metadata initial arcstats arcstats reported metadata usage above the limit arc size current gib target size adaptive gib min size hard limit gib max size high water gib most frequently used mfu cache size gib most recently used mru cache size gib metadata cache size hard limit gib metadata cache size current gib dnode cache size hard limit gib dnode cache size current gib initial adjustment attempted to remediate by increasing arc size by half of remaining ram and increasing the metadata allocation in the arc echo sys module zfs parameters zfs arc max echo sys module zfs parameters zfs arc meta limit percent echo sys module zfs parameters zfs arc dnode limit percent current issue this remediated the issue temporarily and the prune processes stopped but we are back stopped nfs server and it seems to be free ing memory albeit very slowly many dp sync taskq processes but removing the nfs server is very un ideal given they are both living in the kernel it s difficult for me personally to determine who is eating up the memory top up days users load average tasks total running sleeping stopped zombie cpu s us sy ni id wa hi si st mib mem total free used buff cache mib swap total free used avail mem pid user pr ni virt res shr s cpu mem time command root s txg sync root s dp sync taskq root d kworker events root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq arc summary zfs subsystem report wed feb linux generic machine arc status healthy memory throttle count arc size current gib target size adaptive gib min size hard limit gib max size high water gib most frequently used mfu cache size gib most recently used mru cache size gib metadata cache size hard limit gib metadata cache size current gib dnode cache size hard limit gib dnode cache size current gib arc hash breakdown elements max elements current collisions chain max chains arc misc deleted mutex misses eviction skips eviction skips due to writes cached evictions tib eligible evictions gib eligible mfu evictions gib eligible mru evictions gib ineligible evictions gib arc total accesses hits misses cache hit ratio cache miss ratio actual hit ratio mfu mru hits data demand efficiency data prefetch efficiency n a cache hits by cache type most frequently used mfu most recently used mru most frequently used mfu ghost most recently used mru ghost cache hits by data type demand data demand prefetch data demand metadata demand prefetch metadata cache misses by data type demand data demand prefetch data demand metadata demand prefetch metadata dmu prefetch efficiency hit ratio miss ratio status healthy low memory aborts free on write r w clashes bad checksums i o errors size adaptive tib compressed gib header size gib mfu allocated size gib mru allocated size gib prefetch allocated size mib data buffer content allocated size bytes metadata buffer content allocated size gib breakdown hit ratio miss ratio feeds writes writes sent evicts lock retries upon reading solaris porting layer spl spl hostid spl hostid path etc hostid spl kmem alloc max spl kmem alloc warn spl kmem cache kmem threads spl kmem cache magazine size spl kmem cache max size spl kmem cache obj per slab spl kmem cache reclaim spl kmem cache slab limit spl max show tasks spl panic halt spl schedule hrtimeout slack us spl taskq kick spl taskq thread bind spl taskq thread dynamic spl taskq thread priority spl taskq thread sequential tunables dbuf cache hiwater pct dbuf cache lowater pct dbuf cache max bytes dbuf cache shift dbuf metadata cache max bytes dbuf metadata cache shift dmu object alloc chunk shift dmu prefetch max ignore hole birth feed again feed min ms feed secs headroom headroom boost meta percent mfuonly noprefetch norw rebuild blocks min rebuild enabled trim ahead write boost write max metaslab aliquot metaslab bias enabled metaslab debug load metaslab debug unload metaslab df max search metaslab df use largest segment metaslab force ganging metaslab fragmentation factor enabled metaslab lba weighting enabled metaslab preload enabled metaslab unload delay metaslab unload delay ms send holes without birth time spa asize inflation spa config path etc zfs zpool cache spa load print vdev tree spa load verify data spa load verify metadata spa load verify shift spa slop shift vdev file logical ashift vdev file physical ashift vdev removal max span vdev validate skip zap iterate prefetch zfetch array rd sz zfetch max distance zfetch max idistance zfetch max streams zfetch min sec reap zfs abd scatter enabled zfs abd scatter max order zfs abd scatter min size zfs admin snapshot zfs allow redacted dataset mount zfs arc average blocksize zfs arc dnode limit zfs arc dnode limit percent zfs arc dnode reduce percent zfs arc evict batch limit zfs arc eviction pct zfs arc grow retry zfs arc lotsfree percent zfs arc max zfs arc meta adjust restarts zfs arc meta limit zfs arc meta limit percent zfs arc meta min zfs arc meta prune zfs arc meta strategy zfs arc min zfs arc min prefetch ms zfs arc min prescient prefetch ms zfs arc p dampener disable zfs arc p min shift zfs arc pc percent zfs arc shrink shift zfs arc shrinker limit zfs arc sys free zfs async block max blocks zfs autoimport disable zfs checksum events per second zfs commit timeout pct zfs compressed arc enabled zfs condense indirect commit entry delay ms zfs condense indirect obsolete pct zfs condense indirect vdevs enable zfs condense max obsolete bytes zfs condense min mapping bytes zfs dbgmsg enable zfs dbgmsg maxsize zfs dbuf state index zfs ddt data is special zfs deadman checktime ms zfs deadman enabled zfs deadman failmode wait zfs deadman synctime ms zfs deadman ziotime ms zfs dedup prefetch zfs delay min dirty percent zfs delay scale zfs delete blocks zfs dirty data max zfs dirty data max max zfs dirty data max max percent zfs dirty data max percent zfs dirty data sync percent zfs disable ivset guid check zfs dmu offset next sync zfs embedded slog min ms zfs expire snapshot zfs fallocate reserve percent zfs flags zfs free bpobj enabled zfs free leak on eio zfs free min time ms zfs history output max zfs immediate write sz zfs initialize chunk size zfs initialize value zfs keep log spacemaps at export zfs key max salt uses zfs livelist condense new alloc zfs livelist condense sync cancel zfs livelist condense sync pause zfs livelist condense zthr cancel zfs livelist condense zthr pause zfs livelist max entries zfs livelist min percent shared zfs lua max instrlimit zfs lua max memlimit zfs max async dedup frees zfs max log walking zfs max logsm summary length zfs max missing tvds zfs max nvlist src size zfs max recordsize zfs metaslab find max tries zfs metaslab fragmentation threshold zfs metaslab max size cache sec zfs metaslab mem limit zfs metaslab segment weight enabled zfs metaslab switch threshold zfs metaslab try hard before gang zfs mg fragmentation threshold zfs mg noalloc threshold zfs min metaslabs to flush zfs multihost fail intervals zfs multihost history zfs multihost import intervals zfs multihost interval zfs multilist num sublists zfs no scrub io zfs no scrub prefetch zfs nocacheflush zfs nopwrite enabled zfs object mutex size zfs obsolete min time ms zfs override estimate recordsize zfs pd bytes max zfs per txg dirty frees percent zfs prefetch disable zfs read history zfs read history hits zfs rebuild max segment zfs rebuild scrub enabled zfs rebuild vdev limit zfs reconstruct indirect combinations max zfs recover zfs recv queue ff zfs recv queue length zfs recv write batch size zfs removal ignore errors zfs removal suspend progress zfs remove max segment zfs resilver disable defer zfs resilver min time ms zfs scan checkpoint intval zfs scan fill weight zfs scan ignore errors zfs scan issue strategy zfs scan legacy zfs scan max ext gap zfs scan mem lim fact zfs scan mem lim soft fact zfs scan strict mem lim zfs scan suspend progress zfs scan vdev limit zfs scrub min time ms zfs send corrupt data zfs send no prefetch queue ff zfs send no prefetch queue length zfs send queue ff zfs send queue length zfs send unmodified spill blocks zfs slow io events per second zfs spa discard memory limit zfs special class metadata reserve pct zfs sync pass deferred free zfs sync pass dont compress zfs sync pass rewrite zfs sync taskq batch pct zfs traverse indirect prefetch limit zfs trim extent bytes max zfs trim extent bytes min zfs trim metaslab skip zfs trim queue limit zfs trim txg batch zfs txg history zfs txg timeout zfs unflushed log block max zfs unflushed log block min zfs unflushed log block pct zfs unflushed max mem amt zfs unflushed max mem ppm zfs unlink suspend progress zfs user indirect is special zfs vdev aggregate trim zfs vdev aggregation limit zfs vdev aggregation limit non rotating zfs vdev async read max active zfs vdev async read min active zfs vdev async write active max dirty percent zfs vdev async write active min dirty percent zfs vdev async write max active zfs vdev async write min active zfs vdev cache bshift zfs vdev cache max zfs vdev cache size zfs vdev default ms count zfs vdev default ms shift zfs vdev initializing max active zfs vdev initializing min active zfs vdev max active zfs vdev max auto ashift zfs vdev min auto ashift zfs vdev min ms count zfs vdev mirror non rotating inc zfs vdev mirror non rotating seek inc zfs vdev mirror rotating inc zfs vdev mirror rotating seek inc zfs vdev mirror rotating seek offset zfs vdev ms count limit zfs vdev nia credit zfs vdev nia delay zfs vdev queue depth pct zfs vdev raidz impl cycle original scalar zfs vdev read gap limit zfs vdev rebuild max active zfs vdev rebuild min active zfs vdev removal max active zfs vdev removal min active zfs vdev scheduler unused zfs vdev scrub max active zfs vdev scrub min active zfs vdev sync read max active zfs vdev sync read min active zfs vdev sync write max active zfs vdev sync write min active zfs vdev trim max active zfs vdev trim min active zfs vdev write gap limit zfs vnops read chunk size zfs zevent len max zfs zevent retain expire secs zfs zevent retain max zfs zil clean taskq maxalloc zfs zil clean taskq minalloc zfs zil clean taskq nthr pct zil maxblocksize zil nocacheflush zil replay disable zil slog bulk zio deadman log all zio dva throttle enabled zio requeue io start cut in line zio slow io ms zio taskq batch pct zio taskq batch tpq zvol inhibit dev zvol major zvol max discard blocks zvol prefetch bytes zvol request sync zvol threads zvol volmode vdev cache disabled skipping section zil committed transactions commit requests flushes to stable storage transactions to slog storage pool tib transactions to non slog storage pool bytes zed logs logs begin at wed utc end at wed utc jan systemd started zfs event daemon zed jan zed zfs event daemon pid jan zed processing events since eid jan zed eid class config sync pool pod jan zed eid class vdev add pool pod feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay vmstat cat proc vmstat nr free pages nr zone inactive anon nr zone active anon nr zone inactive file nr zone active file nr zone unevictable nr zone write pending nr mlock nr page table pages nr kernel stack nr bounce nr zspages nr free cma numa hit numa miss numa foreign numa interleave numa local numa other nr inactive anon nr active anon nr inactive file nr active file nr unevictable nr slab reclaimable nr slab unreclaimable nr isolated anon nr isolated file workingset nodes workingset refault workingset activate workingset restore workingset nodereclaim nr anon pages nr mapped nr file pages nr dirty nr writeback nr writeback temp nr shmem nr shmem hugepages nr shmem pmdmapped nr file hugepages nr file pmdmapped nr anon transparent hugepages nr unstable nr vmscan write nr vmscan immediate reclaim nr dirtied nr written nr kernel misc reclaimable nr dirty threshold nr dirty background threshold pgpgin pgpgout pswpin pswpout pgalloc dma pgalloc pgalloc normal pgalloc movable allocstall dma allocstall allocstall normal allocstall movable pgskip dma pgskip pgskip normal pgskip movable pgfree pgactivate pgdeactivate pglazyfree pgfault pgmajfault pglazyfreed pgrefill pgsteal kswapd pgsteal direct pgscan kswapd pgscan direct pgscan direct throttle zone reclaim failed pginodesteal slabs scanned kswapd inodesteal kswapd low wmark hit quickly kswapd high wmark hit quickly pageoutrun pgrotated drop pagecache drop slab oom kill numa pte updates numa huge pte updates numa hint faults numa hint faults local numa pages migrated pgmigrate success pgmigrate fail compact migrate scanned compact free scanned compact isolated compact stall compact fail compact success compact daemon wake compact daemon migrate scanned compact daemon free scanned htlb buddy alloc success htlb buddy alloc fail unevictable pgs culled unevictable pgs scanned unevictable pgs rescued unevictable pgs mlocked unevictable pgs munlocked unevictable pgs cleared unevictable pgs stranded thp fault alloc thp fault fallback thp collapse alloc thp collapse alloc failed thp file alloc thp file mapped thp split page thp split page failed thp deferred split page thp split pmd thp split pud thp zero page alloc thp zero page alloc failed thp swpout thp swpout fallback balloon inflate balloon deflate balloon migrate swap ra swap ra hit buddyinfo node zone dma free kib in zone fragment size free fragments total available kib zone free kib in zone fragment size free fragments total available kib zone normal free kib in zone fragment size free fragments total available kib node zone normal free kib in zone fragment size free fragments total available kib describe how to reproduce the problem uncertain but it has occurred on separate servers so it is likely to happen again include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with | 1 |
44,633 | 12,301,484,555 | IssuesEvent | 2020-05-11 15:28:43 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | Bad formatting when selecting NULL | C: Functionality E: All Editions P: Low R: Fixed T: Defect | Generating the formatted SQL for:
```java
var n = inline(null, INTEGER);
ctx.select(n, n);
```
Produces ill formatted SQL:
```sql
select
null,
null
```
The first column is indented too much. | 1.0 | Bad formatting when selecting NULL - Generating the formatted SQL for:
```java
var n = inline(null, INTEGER);
ctx.select(n, n);
```
Produces ill formatted SQL:
```sql
select
null,
null
```
The first column is indented too much. | defect | bad formatting when selecting null generating the formatted sql for java var n inline null integer ctx select n n produces ill formatted sql sql select null null the first column is indented too much | 1 |
52,243 | 13,211,411,442 | IssuesEvent | 2020-08-15 22:57:21 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | [vemcal] missing mysql breaks the build (Trac #1909) | Incomplete Migration Migrated from Trac combo reconstruction defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1909">https://code.icecube.wisc.edu/projects/icecube/ticket/1909</a>, reported by david.schultzand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-09-18T05:49:06",
"_ts": "1568785746891317",
"description": "On a standard py2-v2 cvmfs environment:\n\n{{{\n-- + vemcal\n-- +-- python [symlinks] \nCMake Error at cmake/tools.cmake:71 (message):\n Attempt to use tool 'mysql' which wasn't found\nCall Stack (most recent call first):\n cmake/tools.cmake:112 (use_tool)\n cmake/project.cmake:252 (use_tools)\n vemcal/CMakeLists.txt:39 (i3_add_library)\n}}}\n\nThis should probably be optional.",
"reporter": "david.schultz",
"cc": "",
"resolution": "worksforme",
"time": "2016-11-15T16:55:59",
"component": "combo reconstruction",
"summary": "[vemcal] missing mysql breaks the build",
"priority": "major",
"keywords": "",
"milestone": "Long-Term Future",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [vemcal] missing mysql breaks the build (Trac #1909) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1909">https://code.icecube.wisc.edu/projects/icecube/ticket/1909</a>, reported by david.schultzand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-09-18T05:49:06",
"_ts": "1568785746891317",
"description": "On a standard py2-v2 cvmfs environment:\n\n{{{\n-- + vemcal\n-- +-- python [symlinks] \nCMake Error at cmake/tools.cmake:71 (message):\n Attempt to use tool 'mysql' which wasn't found\nCall Stack (most recent call first):\n cmake/tools.cmake:112 (use_tool)\n cmake/project.cmake:252 (use_tools)\n vemcal/CMakeLists.txt:39 (i3_add_library)\n}}}\n\nThis should probably be optional.",
"reporter": "david.schultz",
"cc": "",
"resolution": "worksforme",
"time": "2016-11-15T16:55:59",
"component": "combo reconstruction",
"summary": "[vemcal] missing mysql breaks the build",
"priority": "major",
"keywords": "",
"milestone": "Long-Term Future",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | missing mysql breaks the build trac migrated from json status closed changetime ts description on a standard cvmfs environment n n n vemcal n python ncmake error at cmake tools cmake message n attempt to use tool mysql which wasn t found ncall stack most recent call first n cmake tools cmake use tool n cmake project cmake use tools n vemcal cmakelists txt add library n n nthis should probably be optional reporter david schultz cc resolution worksforme time component combo reconstruction summary missing mysql breaks the build priority major keywords milestone long term future owner nega type defect | 1 |
291,796 | 21,940,546,270 | IssuesEvent | 2022-05-23 17:36:59 | pharmaverse/admiral | https://api.github.com/repos/pharmaverse/admiral | closed | Documentation: Get Started has incorrect trt variables in mutate | documentation good first issue help wanted release 20-May-2022 contribution | ### Please select a category the issue is focused on?
User Guides
### Let us know where something needs a refresh or put your idea here!

Treatment variables are not correct. Should be TRT01P = ARM and TRT01A = ACTARM | 1.0 | Documentation: Get Started has incorrect trt variables in mutate - ### Please select a category the issue is focused on?
User Guides
### Let us know where something needs a refresh or put your idea here!

Treatment variables are not correct. Should be TRT01P = ARM and TRT01A = ACTARM | non_defect | documentation get started has incorrect trt variables in mutate please select a category the issue is focused on user guides let us know where something needs a refresh or put your idea here treatment variables are not correct should be arm and actarm | 0 |
25,976 | 4,539,518,146 | IssuesEvent | 2016-09-09 11:21:03 | netty/netty | https://api.github.com/repos/netty/netty | closed | HTTP/2: SETTINGS_HEADER_TABLE_SIZE should be an unsigned int | defect | The HTTP/2 spec demands that the max value for `SETTINGS_HEADER_TABLE_SIZE ` should be an unsigned 32-bit integer. However, it seems that some [limitations in HPACK](https://github.com/netty/netty/blob/4.1/codec-http2/src/main/java/io/netty/handler/codec/http2/Http2CodecUtil.java#L90) prevent us from supporting that.
h2spec test suite 6.5 fails due to this limitation. The exception is `Setting HEADER_TABLE_SIZE is invalid: 4294967295`.
```
× Sends a SETTINGS frame
- The endpoint MUST sends a SETTINGS frame with ACK.
Expected: SETTINGS frame (Flags: 1)
Actual: GOAWAY frame (Length: 56, Flags: 0, ErrorCode: PROTOCOL_ERROR)
``` | 1.0 | HTTP/2: SETTINGS_HEADER_TABLE_SIZE should be an unsigned int - The HTTP/2 spec demands that the max value for `SETTINGS_HEADER_TABLE_SIZE ` should be an unsigned 32-bit integer. However, it seems that some [limitations in HPACK](https://github.com/netty/netty/blob/4.1/codec-http2/src/main/java/io/netty/handler/codec/http2/Http2CodecUtil.java#L90) prevent us from supporting that.
h2spec test suite 6.5 fails due to this limitation. The exception is `Setting HEADER_TABLE_SIZE is invalid: 4294967295`.
```
× Sends a SETTINGS frame
- The endpoint MUST sends a SETTINGS frame with ACK.
Expected: SETTINGS frame (Flags: 1)
Actual: GOAWAY frame (Length: 56, Flags: 0, ErrorCode: PROTOCOL_ERROR)
``` | defect | http settings header table size should be an unsigned int the http spec demands that the max value for settings header table size should be an unsigned bit integer however it seems that some prevent us from supporting that test suite fails due to this limitation the exception is setting header table size is invalid × sends a settings frame the endpoint must sends a settings frame with ack expected settings frame flags actual goaway frame length flags errorcode protocol error | 1 |
13,422 | 5,358,982,495 | IssuesEvent | 2017-02-21 00:45:19 | elegantthemes/Divi-Beta | https://api.github.com/repos/elegantthemes/Divi-Beta | closed | Builder Sync :: Number Counter :: updating number and percent sign on BB causes console error when switching to VB | BUILDER SYNC DESIGN SIGNOFF QUALITY ASSURED READY FOR REVIEW | ### Problem:
Updating number and percent sign on BB causes console error when switching to VB:

### Steps To Reproduce:
1. Open a page on BB and VB . Open their console
2. Update number and percent sign on BB
3. Switch to VB
4. You'll see an error thrown
## Attached PR
- https://github.com/elegantthemes/submodule-builder/pull/1744 | 1.0 | Builder Sync :: Number Counter :: updating number and percent sign on BB causes console error when switching to VB - ### Problem:
Updating number and percent sign on BB causes console error when switching to VB:

### Steps To Reproduce:
1. Open a page on BB and VB . Open their console
2. Update number and percent sign on BB
3. Switch to VB
4. You'll see an error thrown
## Attached PR
- https://github.com/elegantthemes/submodule-builder/pull/1744 | non_defect | builder sync number counter updating number and percent sign on bb causes console error when switching to vb problem updating number and percent sign on bb causes console error when switching to vb steps to reproduce open a page on bb and vb open their console update number and percent sign on bb switch to vb you ll see an error thrown attached pr | 0 |
46,681 | 13,055,958,985 | IssuesEvent | 2020-07-30 03:14:17 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | IceHive - crashes during icetray-inspect (Trac #1708) | Incomplete Migration Migrated from Trac combo simulation defect | Migrated from https://code.icecube.wisc.edu/ticket/1708
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:58",
"description": "this is preventing the docs from being built\n\nmaybe `Limits::RingLimits::RingLimits()` is getting bum vector?\n\nroot-ified stack trace:\n\n{{{\n===========================================================\nThere was a crash.\nThis is the entire stack trace of all threads:\n===========================================================\n\nThread 5 (Thread 0x7ff746184700 (LWP 19476)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 4 (Thread 0x7ff745783700 (LWP 19477)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 3 (Thread 0x7ff744d82700 (LWP 19478)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 2 (Thread 0x7ff744381700 (LWP 19479)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 1 (Thread 0x7ff81190b700 (LWP 19361)):\n#0 0x00007ff81078d69d in waitpid () from /lib64/libc.so.6\n#1 0x00007ff81071f609 in do_system () from /lib64/libc.so.6\n#2 0x00007ff81071f940 in system () from /lib64/libc.so.6\n#3 0x00007ff7f2dfbad8 in TUnixSystem::StackTrace() () from /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/root-v5.34.18/lib/libCore.so\n#4 0x00007ff7f2dfa5d3 in TUnixSystem::DispatchSignals(ESignals) () from /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/root-v5.34.18/lib/libCore.so\n#5 <signal handler called>\n#6 0x00007ff742492f52 in std::_Construct<Limits::LimitPair, Limits::LimitPair> (__p=0x7ff73d7aa010, __value=...) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_construct.h:80\n#7 0x00007ff74249121c in std::__uninitialized_copy<false>::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#8 0x00007ff74248f3d5 in std::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#9 0x00007ff74248d1c8 in std::__uninitialized_copy_a<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*, Limits::LimitPair> (__first=Cannot access memory at address 0x1\n#10 0x00007ff74248c0bb in std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> >::vector (this=0x7fff1a3c9f00, __x=std::vector of length 4693611, capacity 0 = {...}) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_vector.h:243\n#11 0x00007ff742499af3 in Limits::RingLimits::RingLimits (this=0x7fff1a3c9f00) at ../../source/IceHive/public/IceHive/IceHiveHelpers.h:38\n#12 0x00007ff742092bd1 in I3HiveCluster<I3RecoPulse>::I3HiveCluster (this=0x483e450, context=...) at ../../source/IceHive/private/IceHive/I3HiveCluster.h:130\n#13 0x00007ff74209135c in StandardCreate<I3Module, I3HiveCluster<I3RecoPulse> >::Create (c=...) at ../../source/icetray/public/icetray/I3Factory.h:110\n#14 0x00007ff805c9d73b in boost::detail::function::function_invoker1<boost::shared_ptr<I3Module> (*)(I3Context const&), boost::shared_ptr<I3Module>, I3Context const&>::invoke (function_ptr=..., a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:94\n#15 0x00007ff805c0bc74 in boost::function1<boost::shared_ptr<I3Module>, I3Context const&>::operator() (this=0x7fff1a3ca2d0, a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:989\n#16 0x00007ff805dcd4ee in module_default_config (name=\"I3HiveCluster<I3RecoPulse>\") at ../../source/icetray/private/pybindings/I3ModuleFactory.cxx:36\n#17 0x00007ff805dd0ce9 in boost::python::detail::invoke<boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::result_converter, I3ConfigurationPtr (*)(const std::string&), boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::c_t0>(boost::python::detail::invoke_tag_<false, false>, const impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::result_converter &, I3ConfigurationPtr (*&)(const std::string &), impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::c_t0 &) (rc=..., f=\n0x1d22fb8, ac0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/invoke.hpp:75\n#18 0x00007ff805dd072a in impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::operator() (this=0x1d22fb8, args_=0x4037290) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/caller.hpp:223\n#19 0x00007ff805dd0051 in boost::python::objects::caller_py_function_impl<boost::python::detail::caller<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> > >::operator() (this=0x1d22fb0, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/object/py_function.hpp:38\n#20 0x00007ff804eb43c5 in boost::python::objects::py_function::operator() (this=0x1d23440, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/object/py_function.hpp:143\n#21 0x00007ff804eb1f7b in boost::python::objects::function::call (this=0x1d23430, args=0x4037290, keywords=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:226\n#22 0x00007ff804eb372f in boost::python::objects::(anonymous namespace)::bind_return::operator() (this=0x7fff1a3ca900) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:581\n#23 0x00007ff804eb40f7 in boost::detail::function::void_function_ref_invoker0<boost::python::objects::<unnamed>::bind_return, void>::invoke(boost::detail::function::function_buffer &) (function_obj_ptr=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:187\n#24 0x00007ff804eca39a in boost::function0<void>::operator() (this=0x7fff1a3ca8b0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#25 0x00007ff804ec9dfc in boost::python::detail::exception_handler::operator() (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:74\n#26 0x00007ff7fdc810d0 in boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>::operator() (this=0x2040c98, handler=..., f=..., translate=0x7ff7fdc59bec <translate(not_found_exception const&)>) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/translate_exception.hpp:46\n#27 0x00007ff7fdc7e3dd in boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> >::operator()<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list2<boost::python::detail::exception_handler const&, boost::function0<void> const&> > (this=0x2040ca0, f=..., a=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind.hpp:361\n#28 0x00007ff7fdc79dde in boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >::operator()<boost::python::detail::exception_handler, boost::function0<void> > (this=0x2040c98, a1=..., a2=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind/bind_template.hpp:102\n#29 0x00007ff7fdc7504c in boost::detail::function::function_obj_invoker2<boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >, bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::invoke (function_obj_ptr=..., a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:131\n#30 0x00007ff804eca315 in boost::function2<bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::operator() (this=0x2040c90, a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#31 0x00007ff804eca1a3 in boost::python::detail::exception_handler::handle (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/detail/exception_handler.hpp:41\n#32 0x00007ff804ec9b13 in boost::python::handle_exception_impl (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:24\n#33 0x00007ff804eb3dab in boost::python::handle_exception<boost::python::objects::<unnamed>::bind_return>(boost::python::objects::(anonymous namespace)::bind_return) (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/errors.hpp:29\n#34 0x00007ff804eb380f in boost::python::objects::function_call (func=0x1d23430, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:613\n#35 0x00007ff811363dd3 in PyObject_Call (func=0x1d23430, arg=<value optimized out>, kw=<value optimized out>) at Objects/abstract.c:2529\n#36 0x00007ff81141126d in do_call (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4239\n#37 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4044\n#38 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#39 0x00007ff81141281e in fast_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4107\n#40 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4042\n#41 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#42 0x00007ff81141391e in PyEval_EvalCodeEx (co=0x7ff8117a3b30, globals=<value optimized out>, locals=<value optimized out>, args=<value optimized out>, argcount=0, kws=0x0, kwcount=0, defs=0x0, defcount=0, closure=0x0) at Python/ceval.c:3253\n#43 0x00007ff811413a32 in PyEval_EvalCode (co=<value optimized out>, globals=<value optimized out>, locals=<value optimized out>) at Python/ceval.c:667\n#44 0x00007ff8114334d0 in run_mod (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1353\n#45 PyRun_FileExFlags (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1339\n#46 0x00007ff81143369c in PyRun_SimpleFileExFlags (fp=0x1a14a80, filename=0x7fff1a3cc2de \"/build/buildslave/kipple_cvmfs/docs/build/bin/icetray-inspect\", closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:943\n#47 0x00007ff8114490ec in Py_Main (argc=<value optimized out>, argv=<value optimized out>) at Modules/main.c:639\n#48 0x00007ff8106ffd5d in __libc_start_main () from /lib64/libc.so.6\n#49 0x0000000000400649 in _start ()\n===========================================================\n\n\nThe lines below might hint at the cause of the crash.\nIf they do not help you then please submit a bug report at\nhttp://root.cern.ch/bugs. Please post the ENTIRE stack trace\nfrom above as an attachment in addition to anything else\nthat might help us fixing this issue.\n===========================================================\n#6 0x00007ff742492f52 in std::_Construct<Limits::LimitPair, Limits::LimitPair> (__p=0x7ff73d7aa010, __value=...) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_construct.h:80\n#7 0x00007ff74249121c in std::__uninitialized_copy<false>::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#8 0x00007ff74248f3d5 in std::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#9 0x00007ff74248d1c8 in std::__uninitialized_copy_a<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*, Limits::LimitPair> (__first=Cannot access memory at address 0x1\n#10 0x00007ff74248c0bb in std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> >::vector (this=0x7fff1a3c9f00, __x=std::vector of length 4693611, capacity 0 = {...}) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_vector.h:243\n#11 0x00007ff742499af3 in Limits::RingLimits::RingLimits (this=0x7fff1a3c9f00) at ../../source/IceHive/public/IceHive/IceHiveHelpers.h:38\n#12 0x00007ff742092bd1 in I3HiveCluster<I3RecoPulse>::I3HiveCluster (this=0x483e450, context=...) at ../../source/IceHive/private/IceHive/I3HiveCluster.h:130\n#13 0x00007ff74209135c in StandardCreate<I3Module, I3HiveCluster<I3RecoPulse> >::Create (c=...) at ../../source/icetray/public/icetray/I3Factory.h:110\n#14 0x00007ff805c9d73b in boost::detail::function::function_invoker1<boost::shared_ptr<I3Module> (*)(I3Context const&), boost::shared_ptr<I3Module>, I3Context const&>::invoke (function_ptr=..., a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:94\n#15 0x00007ff805c0bc74 in boost::function1<boost::shared_ptr<I3Module>, I3Context const&>::operator() (this=0x7fff1a3ca2d0, a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:989\n#16 0x00007ff805dcd4ee in module_default_config (name=\"I3HiveCluster<I3RecoPulse>\") at ../../source/icetray/private/pybindings/I3ModuleFactory.cxx:36\n#17 0x00007ff805dd0ce9 in boost::python::detail::invoke<boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::result_converter, I3ConfigurationPtr (*)(const std::string&), boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::c_t0>(boost::python::detail::invoke_tag_<false, false>, const impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::result_converter &, I3ConfigurationPtr (*&)(const std::string &), impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::c_t0 &) (rc=..., f=\n0x1d22fb8, ac0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/invoke.hpp:75\n#18 0x00007ff805dd072a in impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::operator() (this=0x1d22fb8, args_=0x4037290) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/caller.hpp:223\n#19 0x00007ff805dd0051 in boost::python::objects::caller_py_function_impl<boost::python::detail::caller<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> > >::operator() (this=0x1d22fb0, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/object/py_function.hpp:38\n#20 0x00007ff804eb43c5 in boost::python::objects::py_function::operator() (this=0x1d23440, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/object/py_function.hpp:143\n#21 0x00007ff804eb1f7b in boost::python::objects::function::call (this=0x1d23430, args=0x4037290, keywords=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:226\n#22 0x00007ff804eb372f in boost::python::objects::(anonymous namespace)::bind_return::operator() (this=0x7fff1a3ca900) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:581\n#23 0x00007ff804eb40f7 in boost::detail::function::void_function_ref_invoker0<boost::python::objects::<unnamed>::bind_return, void>::invoke(boost::detail::function::function_buffer &) (function_obj_ptr=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:187\n#24 0x00007ff804eca39a in boost::function0<void>::operator() (this=0x7fff1a3ca8b0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#25 0x00007ff804ec9dfc in boost::python::detail::exception_handler::operator() (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:74\n#26 0x00007ff7fdc810d0 in boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>::operator() (this=0x2040c98, handler=..., f=..., translate=0x7ff7fdc59bec <translate(not_found_exception const&)>) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/translate_exception.hpp:46\n#27 0x00007ff7fdc7e3dd in boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> >::operator()<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list2<boost::python::detail::exception_handler const&, boost::function0<void> const&> > (this=0x2040ca0, f=..., a=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind.hpp:361\n#28 0x00007ff7fdc79dde in boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >::operator()<boost::python::detail::exception_handler, boost::function0<void> > (this=0x2040c98, a1=..., a2=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind/bind_template.hpp:102\n#29 0x00007ff7fdc7504c in boost::detail::function::function_obj_invoker2<boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >, bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::invoke (function_obj_ptr=..., a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:131\n#30 0x00007ff804eca315 in boost::function2<bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::operator() (this=0x2040c90, a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#31 0x00007ff804eca1a3 in boost::python::detail::exception_handler::handle (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/detail/exception_handler.hpp:41\n#32 0x00007ff804ec9b13 in boost::python::handle_exception_impl (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:24\n#33 0x00007ff804eb3dab in boost::python::handle_exception<boost::python::objects::<unnamed>::bind_return>(boost::python::objects::(anonymous namespace)::bind_return) (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/errors.hpp:29\n#34 0x00007ff804eb380f in boost::python::objects::function_call (func=0x1d23430, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:613\n#35 0x00007ff811363dd3 in PyObject_Call (func=0x1d23430, arg=<value optimized out>, kw=<value optimized out>) at Objects/abstract.c:2529\n#36 0x00007ff81141126d in do_call (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4239\n#37 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4044\n#38 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#39 0x00007ff81141281e in fast_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4107\n#40 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4042\n#41 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#42 0x00007ff81141391e in PyEval_EvalCodeEx (co=0x7ff8117a3b30, globals=<value optimized out>, locals=<value optimized out>, args=<value optimized out>, argcount=0, kws=0x0, kwcount=0, defs=0x0, defcount=0, closure=0x0) at Python/ceval.c:3253\n#43 0x00007ff811413a32 in PyEval_EvalCode (co=<value optimized out>, globals=<value optimized out>, locals=<value optimized out>) at Python/ceval.c:667\n#44 0x00007ff8114334d0 in run_mod (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1353\n#45 PyRun_FileExFlags (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1339\n#46 0x00007ff81143369c in PyRun_SimpleFileExFlags (fp=0x1a14a80, filename=0x7fff1a3cc2de \"/build/buildslave/kipple_cvmfs/docs/build/bin/icetray-inspect\", closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:943\n#47 0x00007ff8114490ec in Py_Main (argc=<value optimized out>, argv=<value optimized out>) at Modules/main.c:639\n#48 0x00007ff8106ffd5d in __libc_start_main () from /lib64/libc.so.6\n#49 0x0000000000400649 in _start ()\n===========================================================\n\n\nmake[3]: *** [CMakeFiles/inspect] Error 1\nmake[2]: *** [CMakeFiles/inspect.dir/all] Error 2\nmake[1]: *** [CMakeFiles/docs.dir/rule] Error 2\nmake: *** [docs] Error 2\n(virtenv) [buildslave@kipple build]$\n}}}",
"reporter": "nega",
"cc": "olivas",
"resolution": "wontfix",
"_ts": "1550067178841456",
"component": "combo simulation",
"summary": "IceHive - crashes during icetray-inspect",
"priority": "major",
"keywords": "icetray-inspect icehive documentation",
"time": "2016-05-17T17:07:39",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| 1.0 | IceHive - crashes during icetray-inspect (Trac #1708) - Migrated from https://code.icecube.wisc.edu/ticket/1708
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:58",
"description": "this is preventing the docs from being built\n\nmaybe `Limits::RingLimits::RingLimits()` is getting bum vector?\n\nroot-ified stack trace:\n\n{{{\n===========================================================\nThere was a crash.\nThis is the entire stack trace of all threads:\n===========================================================\n\nThread 5 (Thread 0x7ff746184700 (LWP 19476)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 4 (Thread 0x7ff745783700 (LWP 19477)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 3 (Thread 0x7ff744d82700 (LWP 19478)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 2 (Thread 0x7ff744381700 (LWP 19479)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 1 (Thread 0x7ff81190b700 (LWP 19361)):\n#0 0x00007ff81078d69d in waitpid () from /lib64/libc.so.6\n#1 0x00007ff81071f609 in do_system () from /lib64/libc.so.6\n#2 0x00007ff81071f940 in system () from /lib64/libc.so.6\n#3 0x00007ff7f2dfbad8 in TUnixSystem::StackTrace() () from /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/root-v5.34.18/lib/libCore.so\n#4 0x00007ff7f2dfa5d3 in TUnixSystem::DispatchSignals(ESignals) () from /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/root-v5.34.18/lib/libCore.so\n#5 <signal handler called>\n#6 0x00007ff742492f52 in std::_Construct<Limits::LimitPair, Limits::LimitPair> (__p=0x7ff73d7aa010, __value=...) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_construct.h:80\n#7 0x00007ff74249121c in std::__uninitialized_copy<false>::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#8 0x00007ff74248f3d5 in std::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#9 0x00007ff74248d1c8 in std::__uninitialized_copy_a<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*, Limits::LimitPair> (__first=Cannot access memory at address 0x1\n#10 0x00007ff74248c0bb in std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> >::vector (this=0x7fff1a3c9f00, __x=std::vector of length 4693611, capacity 0 = {...}) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_vector.h:243\n#11 0x00007ff742499af3 in Limits::RingLimits::RingLimits (this=0x7fff1a3c9f00) at ../../source/IceHive/public/IceHive/IceHiveHelpers.h:38\n#12 0x00007ff742092bd1 in I3HiveCluster<I3RecoPulse>::I3HiveCluster (this=0x483e450, context=...) at ../../source/IceHive/private/IceHive/I3HiveCluster.h:130\n#13 0x00007ff74209135c in StandardCreate<I3Module, I3HiveCluster<I3RecoPulse> >::Create (c=...) at ../../source/icetray/public/icetray/I3Factory.h:110\n#14 0x00007ff805c9d73b in boost::detail::function::function_invoker1<boost::shared_ptr<I3Module> (*)(I3Context const&), boost::shared_ptr<I3Module>, I3Context const&>::invoke (function_ptr=..., a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:94\n#15 0x00007ff805c0bc74 in boost::function1<boost::shared_ptr<I3Module>, I3Context const&>::operator() (this=0x7fff1a3ca2d0, a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:989\n#16 0x00007ff805dcd4ee in module_default_config (name=\"I3HiveCluster<I3RecoPulse>\") at ../../source/icetray/private/pybindings/I3ModuleFactory.cxx:36\n#17 0x00007ff805dd0ce9 in boost::python::detail::invoke<boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::result_converter, I3ConfigurationPtr (*)(const std::string&), boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::c_t0>(boost::python::detail::invoke_tag_<false, false>, const impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::result_converter &, I3ConfigurationPtr (*&)(const std::string &), impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::c_t0 &) (rc=..., f=\n0x1d22fb8, ac0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/invoke.hpp:75\n#18 0x00007ff805dd072a in impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::operator() (this=0x1d22fb8, args_=0x4037290) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/caller.hpp:223\n#19 0x00007ff805dd0051 in boost::python::objects::caller_py_function_impl<boost::python::detail::caller<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> > >::operator() (this=0x1d22fb0, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/object/py_function.hpp:38\n#20 0x00007ff804eb43c5 in boost::python::objects::py_function::operator() (this=0x1d23440, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/object/py_function.hpp:143\n#21 0x00007ff804eb1f7b in boost::python::objects::function::call (this=0x1d23430, args=0x4037290, keywords=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:226\n#22 0x00007ff804eb372f in boost::python::objects::(anonymous namespace)::bind_return::operator() (this=0x7fff1a3ca900) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:581\n#23 0x00007ff804eb40f7 in boost::detail::function::void_function_ref_invoker0<boost::python::objects::<unnamed>::bind_return, void>::invoke(boost::detail::function::function_buffer &) (function_obj_ptr=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:187\n#24 0x00007ff804eca39a in boost::function0<void>::operator() (this=0x7fff1a3ca8b0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#25 0x00007ff804ec9dfc in boost::python::detail::exception_handler::operator() (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:74\n#26 0x00007ff7fdc810d0 in boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>::operator() (this=0x2040c98, handler=..., f=..., translate=0x7ff7fdc59bec <translate(not_found_exception const&)>) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/translate_exception.hpp:46\n#27 0x00007ff7fdc7e3dd in boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> >::operator()<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list2<boost::python::detail::exception_handler const&, boost::function0<void> const&> > (this=0x2040ca0, f=..., a=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind.hpp:361\n#28 0x00007ff7fdc79dde in boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >::operator()<boost::python::detail::exception_handler, boost::function0<void> > (this=0x2040c98, a1=..., a2=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind/bind_template.hpp:102\n#29 0x00007ff7fdc7504c in boost::detail::function::function_obj_invoker2<boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >, bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::invoke (function_obj_ptr=..., a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:131\n#30 0x00007ff804eca315 in boost::function2<bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::operator() (this=0x2040c90, a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#31 0x00007ff804eca1a3 in boost::python::detail::exception_handler::handle (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/detail/exception_handler.hpp:41\n#32 0x00007ff804ec9b13 in boost::python::handle_exception_impl (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:24\n#33 0x00007ff804eb3dab in boost::python::handle_exception<boost::python::objects::<unnamed>::bind_return>(boost::python::objects::(anonymous namespace)::bind_return) (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/errors.hpp:29\n#34 0x00007ff804eb380f in boost::python::objects::function_call (func=0x1d23430, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:613\n#35 0x00007ff811363dd3 in PyObject_Call (func=0x1d23430, arg=<value optimized out>, kw=<value optimized out>) at Objects/abstract.c:2529\n#36 0x00007ff81141126d in do_call (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4239\n#37 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4044\n#38 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#39 0x00007ff81141281e in fast_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4107\n#40 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4042\n#41 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#42 0x00007ff81141391e in PyEval_EvalCodeEx (co=0x7ff8117a3b30, globals=<value optimized out>, locals=<value optimized out>, args=<value optimized out>, argcount=0, kws=0x0, kwcount=0, defs=0x0, defcount=0, closure=0x0) at Python/ceval.c:3253\n#43 0x00007ff811413a32 in PyEval_EvalCode (co=<value optimized out>, globals=<value optimized out>, locals=<value optimized out>) at Python/ceval.c:667\n#44 0x00007ff8114334d0 in run_mod (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1353\n#45 PyRun_FileExFlags (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1339\n#46 0x00007ff81143369c in PyRun_SimpleFileExFlags (fp=0x1a14a80, filename=0x7fff1a3cc2de \"/build/buildslave/kipple_cvmfs/docs/build/bin/icetray-inspect\", closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:943\n#47 0x00007ff8114490ec in Py_Main (argc=<value optimized out>, argv=<value optimized out>) at Modules/main.c:639\n#48 0x00007ff8106ffd5d in __libc_start_main () from /lib64/libc.so.6\n#49 0x0000000000400649 in _start ()\n===========================================================\n\n\nThe lines below might hint at the cause of the crash.\nIf they do not help you then please submit a bug report at\nhttp://root.cern.ch/bugs. Please post the ENTIRE stack trace\nfrom above as an attachment in addition to anything else\nthat might help us fixing this issue.\n===========================================================\n#6 0x00007ff742492f52 in std::_Construct<Limits::LimitPair, Limits::LimitPair> (__p=0x7ff73d7aa010, __value=...) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_construct.h:80\n#7 0x00007ff74249121c in std::__uninitialized_copy<false>::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#8 0x00007ff74248f3d5 in std::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#9 0x00007ff74248d1c8 in std::__uninitialized_copy_a<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*, Limits::LimitPair> (__first=Cannot access memory at address 0x1\n#10 0x00007ff74248c0bb in std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> >::vector (this=0x7fff1a3c9f00, __x=std::vector of length 4693611, capacity 0 = {...}) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_vector.h:243\n#11 0x00007ff742499af3 in Limits::RingLimits::RingLimits (this=0x7fff1a3c9f00) at ../../source/IceHive/public/IceHive/IceHiveHelpers.h:38\n#12 0x00007ff742092bd1 in I3HiveCluster<I3RecoPulse>::I3HiveCluster (this=0x483e450, context=...) at ../../source/IceHive/private/IceHive/I3HiveCluster.h:130\n#13 0x00007ff74209135c in StandardCreate<I3Module, I3HiveCluster<I3RecoPulse> >::Create (c=...) at ../../source/icetray/public/icetray/I3Factory.h:110\n#14 0x00007ff805c9d73b in boost::detail::function::function_invoker1<boost::shared_ptr<I3Module> (*)(I3Context const&), boost::shared_ptr<I3Module>, I3Context const&>::invoke (function_ptr=..., a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:94\n#15 0x00007ff805c0bc74 in boost::function1<boost::shared_ptr<I3Module>, I3Context const&>::operator() (this=0x7fff1a3ca2d0, a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:989\n#16 0x00007ff805dcd4ee in module_default_config (name=\"I3HiveCluster<I3RecoPulse>\") at ../../source/icetray/private/pybindings/I3ModuleFactory.cxx:36\n#17 0x00007ff805dd0ce9 in boost::python::detail::invoke<boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::result_converter, I3ConfigurationPtr (*)(const std::string&), boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::c_t0>(boost::python::detail::invoke_tag_<false, false>, const impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::result_converter &, I3ConfigurationPtr (*&)(const std::string &), impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::c_t0 &) (rc=..., f=\n0x1d22fb8, ac0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/invoke.hpp:75\n#18 0x00007ff805dd072a in impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::operator() (this=0x1d22fb8, args_=0x4037290) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/caller.hpp:223\n#19 0x00007ff805dd0051 in boost::python::objects::caller_py_function_impl<boost::python::detail::caller<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> > >::operator() (this=0x1d22fb0, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/object/py_function.hpp:38\n#20 0x00007ff804eb43c5 in boost::python::objects::py_function::operator() (this=0x1d23440, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/object/py_function.hpp:143\n#21 0x00007ff804eb1f7b in boost::python::objects::function::call (this=0x1d23430, args=0x4037290, keywords=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:226\n#22 0x00007ff804eb372f in boost::python::objects::(anonymous namespace)::bind_return::operator() (this=0x7fff1a3ca900) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:581\n#23 0x00007ff804eb40f7 in boost::detail::function::void_function_ref_invoker0<boost::python::objects::<unnamed>::bind_return, void>::invoke(boost::detail::function::function_buffer &) (function_obj_ptr=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:187\n#24 0x00007ff804eca39a in boost::function0<void>::operator() (this=0x7fff1a3ca8b0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#25 0x00007ff804ec9dfc in boost::python::detail::exception_handler::operator() (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:74\n#26 0x00007ff7fdc810d0 in boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>::operator() (this=0x2040c98, handler=..., f=..., translate=0x7ff7fdc59bec <translate(not_found_exception const&)>) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/translate_exception.hpp:46\n#27 0x00007ff7fdc7e3dd in boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> >::operator()<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list2<boost::python::detail::exception_handler const&, boost::function0<void> const&> > (this=0x2040ca0, f=..., a=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind.hpp:361\n#28 0x00007ff7fdc79dde in boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >::operator()<boost::python::detail::exception_handler, boost::function0<void> > (this=0x2040c98, a1=..., a2=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind/bind_template.hpp:102\n#29 0x00007ff7fdc7504c in boost::detail::function::function_obj_invoker2<boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >, bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::invoke (function_obj_ptr=..., a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:131\n#30 0x00007ff804eca315 in boost::function2<bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::operator() (this=0x2040c90, a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#31 0x00007ff804eca1a3 in boost::python::detail::exception_handler::handle (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/detail/exception_handler.hpp:41\n#32 0x00007ff804ec9b13 in boost::python::handle_exception_impl (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:24\n#33 0x00007ff804eb3dab in boost::python::handle_exception<boost::python::objects::<unnamed>::bind_return>(boost::python::objects::(anonymous namespace)::bind_return) (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/errors.hpp:29\n#34 0x00007ff804eb380f in boost::python::objects::function_call (func=0x1d23430, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:613\n#35 0x00007ff811363dd3 in PyObject_Call (func=0x1d23430, arg=<value optimized out>, kw=<value optimized out>) at Objects/abstract.c:2529\n#36 0x00007ff81141126d in do_call (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4239\n#37 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4044\n#38 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#39 0x00007ff81141281e in fast_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4107\n#40 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4042\n#41 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#42 0x00007ff81141391e in PyEval_EvalCodeEx (co=0x7ff8117a3b30, globals=<value optimized out>, locals=<value optimized out>, args=<value optimized out>, argcount=0, kws=0x0, kwcount=0, defs=0x0, defcount=0, closure=0x0) at Python/ceval.c:3253\n#43 0x00007ff811413a32 in PyEval_EvalCode (co=<value optimized out>, globals=<value optimized out>, locals=<value optimized out>) at Python/ceval.c:667\n#44 0x00007ff8114334d0 in run_mod (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1353\n#45 PyRun_FileExFlags (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1339\n#46 0x00007ff81143369c in PyRun_SimpleFileExFlags (fp=0x1a14a80, filename=0x7fff1a3cc2de \"/build/buildslave/kipple_cvmfs/docs/build/bin/icetray-inspect\", closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:943\n#47 0x00007ff8114490ec in Py_Main (argc=<value optimized out>, argv=<value optimized out>) at Modules/main.c:639\n#48 0x00007ff8106ffd5d in __libc_start_main () from /lib64/libc.so.6\n#49 0x0000000000400649 in _start ()\n===========================================================\n\n\nmake[3]: *** [CMakeFiles/inspect] Error 1\nmake[2]: *** [CMakeFiles/inspect.dir/all] Error 2\nmake[1]: *** [CMakeFiles/docs.dir/rule] Error 2\nmake: *** [docs] Error 2\n(virtenv) [buildslave@kipple build]$\n}}}",
"reporter": "nega",
"cc": "olivas",
"resolution": "wontfix",
"_ts": "1550067178841456",
"component": "combo simulation",
"summary": "IceHive - crashes during icetray-inspect",
"priority": "major",
"keywords": "icetray-inspect icehive documentation",
"time": "2016-05-17T17:07:39",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| defect | icehive crashes during icetray inspect trac migrated from json status closed changetime description this is preventing the docs from being built n nmaybe limits ringlimits ringlimits is getting bum vector n nroot ified stack trace n n n nthere was a crash nthis is the entire stack trace of all threads n n nthread thread lwp n in pthread cond wait n nglibc from libpthread so n in th worker tidptr at numexpr module cpp n in start thread from libpthread so n in clone from libc so n nthread thread lwp n in pthread cond wait n nglibc from libpthread so n in th worker tidptr at numexpr module cpp n in start thread from libpthread so n in clone from libc so n nthread thread lwp n in pthread cond wait n nglibc from libpthread so n in th worker tidptr at numexpr module cpp n in start thread from libpthread so n in clone from libc so n nthread thread lwp n in pthread cond wait n nglibc from libpthread so n in th worker tidptr at numexpr module cpp n in start thread from libpthread so n in clone from libc so n nthread thread lwp n in waitpid from libc so n in do system from libc so n in system from libc so n in tunixsystem stacktrace from cvmfs icecube opensciencegrid org rhel root lib libcore so n in tunixsystem dispatchsignals esignals from cvmfs icecube opensciencegrid org rhel root lib libcore so n n in std construct p value at usr lib gcc redhat linux include c bits stl construct h n in std uninitialized copy uninitialized copy limits limitpair first cannot access memory at address n in std uninitialized copy limits limitpair first cannot access memory at address n in std uninitialized copy a limits limitpair limits limitpair first cannot access memory at address n in std vector vector this x std vector of length capacity at usr lib gcc redhat linux include c bits stl vector h n in limits ringlimits ringlimits this at source icehive public icehive icehivehelpers h n in this context at source icehive private icehive h n in standardcreate create c at source icetray public icetray h n in boost detail function function const boost shared ptr const invoke function ptr at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in boost const operator this at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in module default config name at source icetray private pybindings cxx n in boost python detail invoke impl operator pyobject pyobject result converter const std string boost python detail caller arity impl operator pyobject pyobject c boost python detail invoke tag const impl std string const result converter const std string impl std string const c rc f at cvmfs icecube opensciencegrid org rhel include boost boost python detail invoke hpp n in impl std string const operator this args at cvmfs icecube opensciencegrid org rhel include boost boost python detail caller hpp n in boost python objects caller py function impl std string const operator this args kw at cvmfs icecube opensciencegrid org rhel include boost boost python object py function hpp n in boost python objects py function operator this args kw at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python object py function hpp n in boost python objects function call this args keywords at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in boost python objects anonymous namespace bind return operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in boost detail function void function ref bind return void invoke boost detail function function buffer function obj ptr at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost python detail exception handler operator this f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src errors cpp n in boost python detail translate exception operator this handler f translate at cvmfs icecube opensciencegrid org rhel include boost boost python detail translate exception hpp n in boost bi boost arg boost bi value operator boost bi const this f a at cvmfs icecube opensciencegrid org rhel include boost boost bind hpp n in boost bi bind t boost bi boost arg boost bi value operator this at cvmfs icecube opensciencegrid org rhel include boost boost bind bind template hpp n in boost detail function function obj boost bi boost arg boost bi value bool boost python detail exception handler const boost const invoke function obj ptr at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in boost const operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost python detail exception handler handle this f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python detail exception handler hpp n in boost python handle exception impl f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src errors cpp n in boost python handle exception bind return boost python objects anonymous namespace bind return f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python errors hpp n in boost python objects function call func args kw at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in pyobject call func arg kw at objects abstract c n in do call f throwflag at python ceval c n call function f throwflag at python ceval c n pyeval evalframeex f throwflag at python ceval c n in fast function f throwflag at python ceval c n call function f throwflag at python ceval c n pyeval evalframeex f throwflag at python ceval c n in pyeval evalcodeex co globals locals args argcount kws kwcount defs defcount closure at python ceval c n in pyeval evalcode co globals locals at python ceval c n in run mod fp filename start globals locals closeit flags at python pythonrun c n pyrun fileexflags fp filename start globals locals closeit flags at python pythonrun c n in pyrun simplefileexflags fp filename build buildslave kipple cvmfs docs build bin icetray inspect closeit flags at python pythonrun c n in py main argc argv at modules main c n in libc start main from libc so n in start n n n nthe lines below might hint at the cause of the crash nif they do not help you then please submit a bug report at n please post the entire stack trace nfrom above as an attachment in addition to anything else nthat might help us fixing this issue n n in std construct p value at usr lib gcc redhat linux include c bits stl construct h n in std uninitialized copy uninitialized copy limits limitpair first cannot access memory at address n in std uninitialized copy limits limitpair first cannot access memory at address n in std uninitialized copy a limits limitpair limits limitpair first cannot access memory at address n in std vector vector this x std vector of length capacity at usr lib gcc redhat linux include c bits stl vector h n in limits ringlimits ringlimits this at source icehive public icehive icehivehelpers h n in this context at source icehive private icehive h n in standardcreate create c at source icetray public icetray h n in boost detail function function const boost shared ptr const invoke function ptr at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in boost const operator this at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in module default config name at source icetray private pybindings cxx n in boost python detail invoke impl operator pyobject pyobject result converter const std string boost python detail caller arity impl operator pyobject pyobject c boost python detail invoke tag const impl std string const result converter const std string impl std string const c rc f at cvmfs icecube opensciencegrid org rhel include boost boost python detail invoke hpp n in impl std string const operator this args at cvmfs icecube opensciencegrid org rhel include boost boost python detail caller hpp n in boost python objects caller py function impl std string const operator this args kw at cvmfs icecube opensciencegrid org rhel include boost boost python object py function hpp n in boost python objects py function operator this args kw at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python object py function hpp n in boost python objects function call this args keywords at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in boost python objects anonymous namespace bind return operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in boost detail function void function ref bind return void invoke boost detail function function buffer function obj ptr at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost python detail exception handler operator this f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src errors cpp n in boost python detail translate exception operator this handler f translate at cvmfs icecube opensciencegrid org rhel include boost boost python detail translate exception hpp n in boost bi boost arg boost bi value operator boost bi const this f a at cvmfs icecube opensciencegrid org rhel include boost boost bind hpp n in boost bi bind t boost bi boost arg boost bi value operator this at cvmfs icecube opensciencegrid org rhel include boost boost bind bind template hpp n in boost detail function function obj boost bi boost arg boost bi value bool boost python detail exception handler const boost const invoke function obj ptr at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in boost const operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost python detail exception handler handle this f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python detail exception handler hpp n in boost python handle exception impl f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src errors cpp n in boost python handle exception bind return boost python objects anonymous namespace bind return f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python errors hpp n in boost python objects function call func args kw at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in pyobject call func arg kw at objects abstract c n in do call f throwflag at python ceval c n call function f throwflag at python ceval c n pyeval evalframeex f throwflag at python ceval c n in fast function f throwflag at python ceval c n call function f throwflag at python ceval c n pyeval evalframeex f throwflag at python ceval c n in pyeval evalcodeex co globals locals args argcount kws kwcount defs defcount closure at python ceval c n in pyeval evalcode co globals locals at python ceval c n in run mod fp filename start globals locals closeit flags at python pythonrun c n pyrun fileexflags fp filename start globals locals closeit flags at python pythonrun c n in pyrun simplefileexflags fp filename build buildslave kipple cvmfs docs build bin icetray inspect closeit flags at python pythonrun c n in py main argc argv at modules main c n in libc start main from libc so n in start n n n nmake error nmake error nmake error nmake error n virtenv n reporter nega cc olivas resolution wontfix ts component combo simulation summary icehive crashes during icetray inspect priority major keywords icetray inspect icehive documentation time milestone owner nega type defect | 1 |
80,421 | 30,282,624,431 | IssuesEvent | 2023-07-08 09:00:24 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | zdb -R shows incorrect data | Type: Defect | ### System information
Type | Version/Name
--- | ---
Distribution Name | ubuntu
Distribution Version | 22.4
Kernel Version | ubuntu 5.15.0-1033-raspi
Architecture | aarch64 aarch64 aarch64 GNU/Linux
OpenZFS Version | zfs-2.2.99-1, zfs-kmod-2.1.5-1ubuntu6~22.04.1
### Describe the problem you're observing
## Background:
I have a 4disk raid z2 pool that had one disk failed. As I can't import the pool any longer ( crashes on import on macos, freebsd, spews out errors on linux ). Importing it with readonly works, but anything else but using zpool status will crash the machine (freebsd,macos) or hangs zfs (linux). So now I'm in the process of getting my data back with zdb when I encountered this bug.
```
root@ubuntu:/home/ubuntu# zpool status
pool: ocean
state: DEGRADED
status: One or more devices is currently being resilvered. The pool will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
scan: resilver in progress since Sat Jul 1 10:01:24 2023
0B / 11.6T scanned, 0B / 11.6T issued
0B resilvered, 0.00% done, no estimated completion time
config:
NAME STATE READ WRITE CKSUM
ocean DEGRADED 0 0 0
raidz2-0 DEGRADED 0 0 0
18276338438250889807 FAULTED 0 0 0 was /dev/sdd2
sdc1 ONLINE 0 0 0
sdb2 ONLINE 0 0 0
sdd2 ONLINE 0 0 0
errors: 3962 data errors, use '-v' for a list
root@ubuntu:/home/ubuntu# zdb -u -e ocean
Uberblock:
magic = 0000000000bab10c
version = 5000
txg = 18360509
guid_sum = 15765822683839940218
timestamp = 1688241822 UTC = Sat Jul 1 20:03:42 2023
mmp_magic = 00000000a11cea11
mmp_delay = 0
mmp_valid = 0
checkpoint_txg = 0
root@ubuntu:/home/ubuntu# zdb -dP -e ocean
Dataset mos [META], ID 0, cr_txg 4, 1092430080, 6633 objects
Dataset ocean/tmtinkerbell@2023-06-24T00:00:01 [ZPL], ID 5114, cr_txg 18302756, 1111802459136, 24930 objects
...
Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 4065675030144, 9979 objects
...
dmu_bonus_hold(os, object, dl, &dl->dl_dbuf) == 0 (0x34 == 0)
ASSERT at module/zfs/dsl_deadlist.c:308:dsl_deadlist_open()Aborted (core dumped)
root@ubuntu:/home/ubuntu# zdb -dd -e ocean/movies
Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 3.70T, 9979 objects
ZIL header: claim_txg 18360433, claim_blk_seq 1, claim_lr_seq 0 replay_seq 0, flags 0x2
Object lvl iblk dblk dsize dnsize lsize %full type
0 6 128K 16K 14.7M 512 96.9M 5.03 DMU dnode
-1 1 128K 512 11.5K 512 512 100.00 ZFS user/group/project used
-2 1 128K 512 11.5K 512 512 100.00 ZFS user/group/project used
1 1 128K 1K 11.5K 512 1K 100.00 ZFS master node
2 1 128K 512 11.5K 512 512 100.00 SA master node
3 1 128K 21.5K 11.5K 512 21.5K 100.00 ZFS delete queue
4 2 128K 16K 35K 512 32K 100.00 ZFS directory
5 1 128K 2K 11.5K 512 2K 100.00 SA attr registration
...
```
And looking at one file:
```
root@ubuntu:~# zdb -ddddd -K $pass -e ocean/movies 5268
Unlocked encryption root: ocean/movies
Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 3.70T, 9979 objects, rootbp DVA[0]=<0:2400014a000:3000> DVA[1]=<0:f20081e4000:3000> [L0 DMU objset] fletcher4 uncompressed authenticated LE contiguous unique double size=800L/800P birth=18360433L/18360433P fill=9979 cksum=00000021c040a809:000025b93e7be9da:00190442ec3d9de4:0c1fdfcfea494199
Object lvl iblk dblk dsize dnsize lsize %full type
5268 1 128K 105K 40.5K 512 105K 100.00 ZFS plain file
168 bonus System attributes
dnode flags: USED_BYTES USERUSED_ACCOUNTED
dnode maxblkid: 0
path /Pictures/Testbilder/NecPattern.png
uid 1000
gid 1000
atime Tue Dec 31 08:24:53 2019
mtime Thu Feb 11 09:05:10 2016
ctime Wed Jan 27 17:59:02 2021
crtime Tue Jan 26 20:40:27 2021
gen 4641518
mode 100644
size 107030
parent 5267
links 1
pflags 40800000004
Indirect blocks:
0 L0 0:10a1263d2000:15000 1a400L/9000P F=1 B=4641518/4641518 cksum=246684c5894cb36e:ecb92ef5e414d48a:13123f3d0f3e78df:9619bdcef80cd664
segment [0000000000000000, 000000000001a400) size 105K
```
## The Issue
Using **zdb -R** on an unencrypted pool works and gives me back the data, but on encrypted pools that does not work as shown here there is no PNG header as it should ( described in the how to replicate section)
```
root@ubuntu:/home/ubuntu# zdb -R -K $pass $disks -e ocean/movies 0:10a1263d2000:1a400 |head
Found vdev type: raidz
0:10a1263d2000:1a400
0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef
000000: 9c042f5028cc72d6 1eb5378942ca849d .r.(P/.....B.7..
000010: 39d46c5bb5ef339d 9aeb9cbf22549f9a .3..[l.9..T"....
000020: c4bcb91876bfb295 9b297424b1886cdd ...v.....l..$t).
000030: dab8c9b2e13c8f06 f9517b44e365816a ..<.....j.e.D{Q.
000040: f63da5f46b496aa5 d2a3744d20577b5a .jIk..=.Z{W Mt..
...
```
### Describe how to reproduce the problem
Create a pool with an ecrypted pool inside and copy an image.png into it.
```
mkfile 100m sda sdb sdc sdd
pass='testtest'
disks='-p /Users/jolly/Desktop/zfsdebug2/sda -p /Users/jolly/Desktop/zfsdebug2/sdb -p /Users/jolly/Desktop/zfsdebug2/sdc -p /Users/jolly/Desktop/zfsdebug2/sdd'
# zpool create test raidz2 $disks
# create -o encryption=on -o keylocation=prompt -o keyformat=passphrase test/movies
# cp Image.png /Volumes/test/movies/
```
look for the image size in
```
# zdb -dd $disks -e test/movies
physmem = 2097152 pages (32.00 GB)
looking up movies in obj34
Dataset test/movies [ZPL], ID 388, cr_txg 32, 2.08M, 104 objects
ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0
Object lvl iblk dblk dsize dnsize lsize %full type
0 6 128K 16K 280K 512 288K 18.06 DMU dnode
-1 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used
-2 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used
-3 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used
1 1 128K 1K 1.50K 512 1K 100.00 ZFS master node
2 2 128K 128K 1.66M 512 1.75M 100.00 ZFS plain file
3 1 128K 512 1.50K 512 512 100.00 ZFS directory
4 1 128K 512 512 512 512 100.00 ZFS plain file
5 1 128K 6.50K 4.50K 512 6.50K 100.00 ZFS directory
6 1 128K 512 0 512 512 0.00 ZFS plain file
7 1 128K 512 1.50K 512 512 100.00 ZFS directory
9 1 128K 105K 33K 512 105K 100.00 ZFS plain file
10 1 128K 512 1.50K 512 512 100.00 ZFS directory
...
```
I my case 105k Bytes - so it's object number 9
Then find out the blocks
```
# zdb -ddddd $disks -e test/movies 9
physmem = 2097152 pages (32.00 GB)
looking up movies in obj34
Dataset test/movies [ZPL], ID 388, cr_txg 32, 2.08M, 104 objects, rootbp DVA[0]=<0:ce95800:2400> DVA[1]=<0:41f1400:2400> [L0 DMU objset] fletcher4 uncompressed authenticated LE contiguous unique double size=1000L/1000P birth=43L/43P fill=104 cksum=00000022e10e3971:000064b3a2f91165:009866b8a335c5af:9fae33cbc48176dc
Object lvl iblk dblk dsize dnsize lsize %full type
9 1 128K 105K 33K 512 105K 100.00 ZFS plain file
176 bonus System attributes
dnode flags: USED_BYTES USERUSED_ACCOUNTED USEROBJUSED_ACCOUNTED
dnode maxblkid: 0
(bonus encrypted)
(object encrypted)
Indirect blocks:
0 L0 0:c969600:10800 1a400L/8200P F=1 B=40/40 cksum=3243e353334e32fb:6d3c14cd8a3ce5d2:31f1bac457bf51ec:413b385e911ad9e2
segment [0000000000000000, 000000000001a400) size 105K
```
and now finally try to get the image data at 0:c969600:1a400
```
bash-3.2# zdb -R -K $pass $disks -e test/movies 0:c969600:1a400 |head
Found vdev type: raidz
physmem = 2097152 pages (32.00 GB)
0:c969600:1a400L
0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef
000000: 5727568bba70a4aa 13faac8aec9d2530 W'V..p........%0
000010: 4e83318bd6144700 c1711eed6c4bff06 N.1...G..q..lK..
000020: 73bc77792344f8c6 071924efb202fa5a s.wy#D....$....Z
000030: 48e5991759d8d5e9 5b647ef25695d182 H...Y...[d~.V...
000040: f87618227293d53f 958eb1815f8df46b .v."r..?...._..k
```
those do not contain the PNG header as on an unencrypted pool:
```
bash-3.2# zdb -R -p /Users/jolly/Desktop/zfsdebug2/sda -p /Users/jolly/Desktop/zfsdebug2/sdb -p /Users/jolly/Desktop/zfsdebug2/sdc -e test 0:4604200:10800 |head
Found vdev type: raidz
physmem = 2097152 pages (32.00 GB)
0:4604200:10800
0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef
000000: 0000803ff0698950 4e470d0a1a0a0000 ...?.i.PNG......
000010: 000d494844520000 0780000004380802 ..IHDR.......8..
000020: 00000067b1561400 00002c7445587443 ...g.V....,tEXtC
000030: 72656174696f6e20 54696d6500536174 reation Time.Sat
```
| 1.0 | zdb -R shows incorrect data - ### System information
Type | Version/Name
--- | ---
Distribution Name | ubuntu
Distribution Version | 22.4
Kernel Version | ubuntu 5.15.0-1033-raspi
Architecture | aarch64 aarch64 aarch64 GNU/Linux
OpenZFS Version | zfs-2.2.99-1, zfs-kmod-2.1.5-1ubuntu6~22.04.1
### Describe the problem you're observing
## Background:
I have a 4disk raid z2 pool that had one disk failed. As I can't import the pool any longer ( crashes on import on macos, freebsd, spews out errors on linux ). Importing it with readonly works, but anything else but using zpool status will crash the machine (freebsd,macos) or hangs zfs (linux). So now I'm in the process of getting my data back with zdb when I encountered this bug.
```
root@ubuntu:/home/ubuntu# zpool status
pool: ocean
state: DEGRADED
status: One or more devices is currently being resilvered. The pool will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
scan: resilver in progress since Sat Jul 1 10:01:24 2023
0B / 11.6T scanned, 0B / 11.6T issued
0B resilvered, 0.00% done, no estimated completion time
config:
NAME STATE READ WRITE CKSUM
ocean DEGRADED 0 0 0
raidz2-0 DEGRADED 0 0 0
18276338438250889807 FAULTED 0 0 0 was /dev/sdd2
sdc1 ONLINE 0 0 0
sdb2 ONLINE 0 0 0
sdd2 ONLINE 0 0 0
errors: 3962 data errors, use '-v' for a list
root@ubuntu:/home/ubuntu# zdb -u -e ocean
Uberblock:
magic = 0000000000bab10c
version = 5000
txg = 18360509
guid_sum = 15765822683839940218
timestamp = 1688241822 UTC = Sat Jul 1 20:03:42 2023
mmp_magic = 00000000a11cea11
mmp_delay = 0
mmp_valid = 0
checkpoint_txg = 0
root@ubuntu:/home/ubuntu# zdb -dP -e ocean
Dataset mos [META], ID 0, cr_txg 4, 1092430080, 6633 objects
Dataset ocean/tmtinkerbell@2023-06-24T00:00:01 [ZPL], ID 5114, cr_txg 18302756, 1111802459136, 24930 objects
...
Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 4065675030144, 9979 objects
...
dmu_bonus_hold(os, object, dl, &dl->dl_dbuf) == 0 (0x34 == 0)
ASSERT at module/zfs/dsl_deadlist.c:308:dsl_deadlist_open()Aborted (core dumped)
root@ubuntu:/home/ubuntu# zdb -dd -e ocean/movies
Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 3.70T, 9979 objects
ZIL header: claim_txg 18360433, claim_blk_seq 1, claim_lr_seq 0 replay_seq 0, flags 0x2
Object lvl iblk dblk dsize dnsize lsize %full type
0 6 128K 16K 14.7M 512 96.9M 5.03 DMU dnode
-1 1 128K 512 11.5K 512 512 100.00 ZFS user/group/project used
-2 1 128K 512 11.5K 512 512 100.00 ZFS user/group/project used
1 1 128K 1K 11.5K 512 1K 100.00 ZFS master node
2 1 128K 512 11.5K 512 512 100.00 SA master node
3 1 128K 21.5K 11.5K 512 21.5K 100.00 ZFS delete queue
4 2 128K 16K 35K 512 32K 100.00 ZFS directory
5 1 128K 2K 11.5K 512 2K 100.00 SA attr registration
...
```
And looking at one file:
```
root@ubuntu:~# zdb -ddddd -K $pass -e ocean/movies 5268
Unlocked encryption root: ocean/movies
Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 3.70T, 9979 objects, rootbp DVA[0]=<0:2400014a000:3000> DVA[1]=<0:f20081e4000:3000> [L0 DMU objset] fletcher4 uncompressed authenticated LE contiguous unique double size=800L/800P birth=18360433L/18360433P fill=9979 cksum=00000021c040a809:000025b93e7be9da:00190442ec3d9de4:0c1fdfcfea494199
Object lvl iblk dblk dsize dnsize lsize %full type
5268 1 128K 105K 40.5K 512 105K 100.00 ZFS plain file
168 bonus System attributes
dnode flags: USED_BYTES USERUSED_ACCOUNTED
dnode maxblkid: 0
path /Pictures/Testbilder/NecPattern.png
uid 1000
gid 1000
atime Tue Dec 31 08:24:53 2019
mtime Thu Feb 11 09:05:10 2016
ctime Wed Jan 27 17:59:02 2021
crtime Tue Jan 26 20:40:27 2021
gen 4641518
mode 100644
size 107030
parent 5267
links 1
pflags 40800000004
Indirect blocks:
0 L0 0:10a1263d2000:15000 1a400L/9000P F=1 B=4641518/4641518 cksum=246684c5894cb36e:ecb92ef5e414d48a:13123f3d0f3e78df:9619bdcef80cd664
segment [0000000000000000, 000000000001a400) size 105K
```
## The Issue
Using **zdb -R** on an unencrypted pool works and gives me back the data, but on encrypted pools that does not work as shown here there is no PNG header as it should ( described in the how to replicate section)
```
root@ubuntu:/home/ubuntu# zdb -R -K $pass $disks -e ocean/movies 0:10a1263d2000:1a400 |head
Found vdev type: raidz
0:10a1263d2000:1a400
0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef
000000: 9c042f5028cc72d6 1eb5378942ca849d .r.(P/.....B.7..
000010: 39d46c5bb5ef339d 9aeb9cbf22549f9a .3..[l.9..T"....
000020: c4bcb91876bfb295 9b297424b1886cdd ...v.....l..$t).
000030: dab8c9b2e13c8f06 f9517b44e365816a ..<.....j.e.D{Q.
000040: f63da5f46b496aa5 d2a3744d20577b5a .jIk..=.Z{W Mt..
...
```
### Describe how to reproduce the problem
Create a pool with an ecrypted pool inside and copy an image.png into it.
```
mkfile 100m sda sdb sdc sdd
pass='testtest'
disks='-p /Users/jolly/Desktop/zfsdebug2/sda -p /Users/jolly/Desktop/zfsdebug2/sdb -p /Users/jolly/Desktop/zfsdebug2/sdc -p /Users/jolly/Desktop/zfsdebug2/sdd'
# zpool create test raidz2 $disks
# create -o encryption=on -o keylocation=prompt -o keyformat=passphrase test/movies
# cp Image.png /Volumes/test/movies/
```
look for the image size in
```
# zdb -dd $disks -e test/movies
physmem = 2097152 pages (32.00 GB)
looking up movies in obj34
Dataset test/movies [ZPL], ID 388, cr_txg 32, 2.08M, 104 objects
ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0
Object lvl iblk dblk dsize dnsize lsize %full type
0 6 128K 16K 280K 512 288K 18.06 DMU dnode
-1 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used
-2 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used
-3 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used
1 1 128K 1K 1.50K 512 1K 100.00 ZFS master node
2 2 128K 128K 1.66M 512 1.75M 100.00 ZFS plain file
3 1 128K 512 1.50K 512 512 100.00 ZFS directory
4 1 128K 512 512 512 512 100.00 ZFS plain file
5 1 128K 6.50K 4.50K 512 6.50K 100.00 ZFS directory
6 1 128K 512 0 512 512 0.00 ZFS plain file
7 1 128K 512 1.50K 512 512 100.00 ZFS directory
9 1 128K 105K 33K 512 105K 100.00 ZFS plain file
10 1 128K 512 1.50K 512 512 100.00 ZFS directory
...
```
I my case 105k Bytes - so it's object number 9
Then find out the blocks
```
# zdb -ddddd $disks -e test/movies 9
physmem = 2097152 pages (32.00 GB)
looking up movies in obj34
Dataset test/movies [ZPL], ID 388, cr_txg 32, 2.08M, 104 objects, rootbp DVA[0]=<0:ce95800:2400> DVA[1]=<0:41f1400:2400> [L0 DMU objset] fletcher4 uncompressed authenticated LE contiguous unique double size=1000L/1000P birth=43L/43P fill=104 cksum=00000022e10e3971:000064b3a2f91165:009866b8a335c5af:9fae33cbc48176dc
Object lvl iblk dblk dsize dnsize lsize %full type
9 1 128K 105K 33K 512 105K 100.00 ZFS plain file
176 bonus System attributes
dnode flags: USED_BYTES USERUSED_ACCOUNTED USEROBJUSED_ACCOUNTED
dnode maxblkid: 0
(bonus encrypted)
(object encrypted)
Indirect blocks:
0 L0 0:c969600:10800 1a400L/8200P F=1 B=40/40 cksum=3243e353334e32fb:6d3c14cd8a3ce5d2:31f1bac457bf51ec:413b385e911ad9e2
segment [0000000000000000, 000000000001a400) size 105K
```
and now finally try to get the image data at 0:c969600:1a400
```
bash-3.2# zdb -R -K $pass $disks -e test/movies 0:c969600:1a400 |head
Found vdev type: raidz
physmem = 2097152 pages (32.00 GB)
0:c969600:1a400L
0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef
000000: 5727568bba70a4aa 13faac8aec9d2530 W'V..p........%0
000010: 4e83318bd6144700 c1711eed6c4bff06 N.1...G..q..lK..
000020: 73bc77792344f8c6 071924efb202fa5a s.wy#D....$....Z
000030: 48e5991759d8d5e9 5b647ef25695d182 H...Y...[d~.V...
000040: f87618227293d53f 958eb1815f8df46b .v."r..?...._..k
```
those do not contain the PNG header as on an unencrypted pool:
```
bash-3.2# zdb -R -p /Users/jolly/Desktop/zfsdebug2/sda -p /Users/jolly/Desktop/zfsdebug2/sdb -p /Users/jolly/Desktop/zfsdebug2/sdc -e test 0:4604200:10800 |head
Found vdev type: raidz
physmem = 2097152 pages (32.00 GB)
0:4604200:10800
0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef
000000: 0000803ff0698950 4e470d0a1a0a0000 ...?.i.PNG......
000010: 000d494844520000 0780000004380802 ..IHDR.......8..
000020: 00000067b1561400 00002c7445587443 ...g.V....,tEXtC
000030: 72656174696f6e20 54696d6500536174 reation Time.Sat
```
| defect | zdb r shows incorrect data system information type version name distribution name ubuntu distribution version kernel version ubuntu raspi architecture gnu linux openzfs version zfs zfs kmod describe the problem you re observing background i have a raid pool that had one disk failed as i can t import the pool any longer crashes on import on macos freebsd spews out errors on linux importing it with readonly works but anything else but using zpool status will crash the machine freebsd macos or hangs zfs linux so now i m in the process of getting my data back with zdb when i encountered this bug root ubuntu home ubuntu zpool status pool ocean state degraded status one or more devices is currently being resilvered the pool will continue to function possibly in a degraded state action wait for the resilver to complete scan resilver in progress since sat jul scanned issued resilvered done no estimated completion time config name state read write cksum ocean degraded degraded faulted was dev online online online errors data errors use v for a list root ubuntu home ubuntu zdb u e ocean uberblock magic version txg guid sum timestamp utc sat jul mmp magic mmp delay mmp valid checkpoint txg root ubuntu home ubuntu zdb dp e ocean dataset mos id cr txg objects dataset ocean tmtinkerbell id cr txg objects dataset ocean movies id cr txg objects dmu bonus hold os object dl dl dl dbuf assert at module zfs dsl deadlist c dsl deadlist open aborted core dumped root ubuntu home ubuntu zdb dd e ocean movies dataset ocean movies id cr txg objects zil header claim txg claim blk seq claim lr seq replay seq flags object lvl iblk dblk dsize dnsize lsize full type dmu dnode zfs user group project used zfs user group project used zfs master node sa master node zfs delete queue zfs directory sa attr registration and looking at one file root ubuntu zdb ddddd k pass e ocean movies unlocked encryption root ocean movies dataset ocean movies id cr txg objects rootbp dva dva uncompressed authenticated le contiguous unique double size birth fill cksum object lvl iblk dblk dsize dnsize lsize full type zfs plain file bonus system attributes dnode flags used bytes userused accounted dnode maxblkid path pictures testbilder necpattern png uid gid atime tue dec mtime thu feb ctime wed jan crtime tue jan gen mode size parent links pflags indirect blocks f b cksum segment size the issue using zdb r on an unencrypted pool works and gives me back the data but on encrypted pools that does not work as shown here there is no png header as it should described in the how to replicate section root ubuntu home ubuntu zdb r k pass disks e ocean movies head found vdev type raidz a b c d e f r p b l t v l t j e d q jik z w mt describe how to reproduce the problem create a pool with an ecrypted pool inside and copy an image png into it mkfile sda sdb sdc sdd pass testtest disks p users jolly desktop sda p users jolly desktop sdb p users jolly desktop sdc p users jolly desktop sdd zpool create test disks create o encryption on o keylocation prompt o keyformat passphrase test movies cp image png volumes test movies look for the image size in zdb dd disks e test movies physmem pages gb looking up movies in dataset test movies id cr txg objects zil header claim txg claim blk seq claim lr seq replay seq flags object lvl iblk dblk dsize dnsize lsize full type dmu dnode zfs user group project used zfs user group project used zfs user group project used zfs master node zfs plain file zfs directory zfs plain file zfs directory zfs plain file zfs directory zfs plain file zfs directory i my case bytes so it s object number then find out the blocks zdb ddddd disks e test movies physmem pages gb looking up movies in dataset test movies id cr txg objects rootbp dva dva uncompressed authenticated le contiguous unique double size birth fill cksum object lvl iblk dblk dsize dnsize lsize full type zfs plain file bonus system attributes dnode flags used bytes userused accounted userobjused accounted dnode maxblkid bonus encrypted object encrypted indirect blocks f b cksum segment size and now finally try to get the image data at bash zdb r k pass disks e test movies head found vdev type raidz physmem pages gb a b c d e f w v p n g q lk s wy d z h y d v v r k those do not contain the png header as on an unencrypted pool bash zdb r p users jolly desktop sda p users jolly desktop sdb p users jolly desktop sdc e test head found vdev type raidz physmem pages gb a b c d e f i png ihdr g v textc reation time sat | 1 |
487,915 | 14,061,255,102 | IssuesEvent | 2020-11-03 07:48:55 | mobigen/IRIS-BigData-Platform | https://api.github.com/repos/mobigen/IRIS-BigData-Platform | closed | [LH빅데이터 플랫폼] 알티베이스 연동 I/F 추가 | #Discovery Priority: P0 R: 20.10.30 Status: Backlog | ## 기능 요청 ##
LH빅데이터 플랫폼 과제 수행을 위해 **알티베이스 연동 I/F 추가 필요**
## 원하는 솔루션 설명 ##
IRIS Discovery 메뉴의 "연결정보" 설정 기능에 알티베이스 추가

IRIS Studio에서 DBMS 연결유형에 알티베이스 추가

## 고려한 다른 대안 ##
## 기타 ##
| 1.0 | [LH빅데이터 플랫폼] 알티베이스 연동 I/F 추가 - ## 기능 요청 ##
LH빅데이터 플랫폼 과제 수행을 위해 **알티베이스 연동 I/F 추가 필요**
## 원하는 솔루션 설명 ##
IRIS Discovery 메뉴의 "연결정보" 설정 기능에 알티베이스 추가

IRIS Studio에서 DBMS 연결유형에 알티베이스 추가

## 고려한 다른 대안 ##
## 기타 ##
| non_defect | 알티베이스 연동 i f 추가 기능 요청 lh빅데이터 플랫폼 과제 수행을 위해 알티베이스 연동 i f 추가 필요 원하는 솔루션 설명 iris discovery 메뉴의 연결정보 설정 기능에 알티베이스 추가 iris studio에서 dbms 연결유형에 알티베이스 추가 고려한 다른 대안 기타 | 0 |
39,051 | 9,186,855,205 | IssuesEvent | 2019-03-06 00:23:03 | vmware/vic | https://api.github.com/repos/vmware/vic | closed | share templates across VCHs | area/ux component/install component/isos component/portlayer/execution kind/defect/performance severity/3-moderate | When using vmfork enabled VCHs as a tenancy mechanism in large numbers it's necessary to share the vmfork templates between those VCHs using the same bootstrap versions. This allows amortization of the template overhead across all containerVMs and not just those within a single VCH.
This is a core scenario for service provider usage and for on-prem dev/test.
This requires a mechanism for discovering templates and performing reference counting of some kind. As an initial implementation it's acceptable to have the location of templates specified by the admin, however we should still manage template versions within that location or the upgrade paths become problematic.
| 1.0 | share templates across VCHs - When using vmfork enabled VCHs as a tenancy mechanism in large numbers it's necessary to share the vmfork templates between those VCHs using the same bootstrap versions. This allows amortization of the template overhead across all containerVMs and not just those within a single VCH.
This is a core scenario for service provider usage and for on-prem dev/test.
This requires a mechanism for discovering templates and performing reference counting of some kind. As an initial implementation it's acceptable to have the location of templates specified by the admin, however we should still manage template versions within that location or the upgrade paths become problematic.
| defect | share templates across vchs when using vmfork enabled vchs as a tenancy mechanism in large numbers it s necessary to share the vmfork templates between those vchs using the same bootstrap versions this allows amortization of the template overhead across all containervms and not just those within a single vch this is a core scenario for service provider usage and for on prem dev test this requires a mechanism for discovering templates and performing reference counting of some kind as an initial implementation it s acceptable to have the location of templates specified by the admin however we should still manage template versions within that location or the upgrade paths become problematic | 1 |
40,429 | 9,992,287,876 | IssuesEvent | 2019-07-11 13:09:35 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | Review correctness of jOOQ's CURRENT_TIME, CURRENT_TIMESTAMP functions | C: Functionality E: All Editions P: Medium T: Defect T: Incompatible change | The SQL standard specifies that
- `CURRENT_TIME` produces a `TIME WITH TIME ZONE` value (Java `OffsetTime`)
- `CURRENT_TIMESTAMP` produces a `TIMESTAMP WITH TIME ZONE` value (Java `OffsetDateTime`)
- `LOCALTIME` produces a `TIME WITHOUT TIME ZONE` value (Java `LocalTime` or `java.sql.Time`)
- `LOCALTIMESTAMP` produces a `TIMESTAMP WITHOUT TIME ZONE` value (Java `LocalDateTime` or `java.sql.Timestamp`)
This is not what the current `currentTimestamp()` and other functions suggest. Historically, as jOOQ hasn't always supported `WITH TIME ZONE` types, all of these functions map to the `WITHOUT TIME ZONE` counterparts.
This may produce different, inconsistent results depending on the dialect.
It is difficult to see a solution here that doesn't break backwards compatibility of behaviour. | 1.0 | Review correctness of jOOQ's CURRENT_TIME, CURRENT_TIMESTAMP functions - The SQL standard specifies that
- `CURRENT_TIME` produces a `TIME WITH TIME ZONE` value (Java `OffsetTime`)
- `CURRENT_TIMESTAMP` produces a `TIMESTAMP WITH TIME ZONE` value (Java `OffsetDateTime`)
- `LOCALTIME` produces a `TIME WITHOUT TIME ZONE` value (Java `LocalTime` or `java.sql.Time`)
- `LOCALTIMESTAMP` produces a `TIMESTAMP WITHOUT TIME ZONE` value (Java `LocalDateTime` or `java.sql.Timestamp`)
This is not what the current `currentTimestamp()` and other functions suggest. Historically, as jOOQ hasn't always supported `WITH TIME ZONE` types, all of these functions map to the `WITHOUT TIME ZONE` counterparts.
This may produce different, inconsistent results depending on the dialect.
It is difficult to see a solution here that doesn't break backwards compatibility of behaviour. | defect | review correctness of jooq s current time current timestamp functions the sql standard specifies that current time produces a time with time zone value java offsettime current timestamp produces a timestamp with time zone value java offsetdatetime localtime produces a time without time zone value java localtime or java sql time localtimestamp produces a timestamp without time zone value java localdatetime or java sql timestamp this is not what the current currenttimestamp and other functions suggest historically as jooq hasn t always supported with time zone types all of these functions map to the without time zone counterparts this may produce different inconsistent results depending on the dialect it is difficult to see a solution here that doesn t break backwards compatibility of behaviour | 1 |
28,190 | 5,217,162,160 | IssuesEvent | 2017-01-26 13:00:11 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | Offset when opening up menu | defect |
**I'm submitting a ...** (check one with "x")
```
[x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
Please fork the plunkr below and create a case demonstrating your bug report. Issues without a plunkr have much less possibility to be reviewed.
http://plnkr.co/edit/NtWWnN
**Current behavior**
if an anchor tag contains another element, the menu attaches to that element
**Expected behavior**
the menu should always attach to the anchor tag
**Minimal reproduction of the problem with instructions**
<!--
If the current behavior is a bug or you can illustrate your feature request better with an example,
please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via
https://plnkr.co or similar (you can use this template as a starting point: http://plnkr.co/edit/tpl:AvJOMERrnz94ekVua0u5).
-->
**What is the motivation / use case for changing the behavior?**
<!-- Describe the motivation or the concrete use case -->
**Please tell us about your environment:**
<!-- Operating system, IDE, package manager, HTTP server, ... -->
* **Angular version:** 2.0.X
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 2.0.X
<!-- Check whether this is still an issue in the most recent Angular version -->
* **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ]
<!-- All browsers where this could be reproduced -->
* **Language:** [all | TypeScript X.X | ES6/7 | ES5]
* **Node (for AoT issues):** `node --version` =
| 1.0 | Offset when opening up menu -
**I'm submitting a ...** (check one with "x")
```
[x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
Please fork the plunkr below and create a case demonstrating your bug report. Issues without a plunkr have much less possibility to be reviewed.
http://plnkr.co/edit/NtWWnN
**Current behavior**
if an anchor tag contains another element, the menu attaches to that element
**Expected behavior**
the menu should always attach to the anchor tag
**Minimal reproduction of the problem with instructions**
<!--
If the current behavior is a bug or you can illustrate your feature request better with an example,
please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via
https://plnkr.co or similar (you can use this template as a starting point: http://plnkr.co/edit/tpl:AvJOMERrnz94ekVua0u5).
-->
**What is the motivation / use case for changing the behavior?**
<!-- Describe the motivation or the concrete use case -->
**Please tell us about your environment:**
<!-- Operating system, IDE, package manager, HTTP server, ... -->
* **Angular version:** 2.0.X
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 2.0.X
<!-- Check whether this is still an issue in the most recent Angular version -->
* **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ]
<!-- All browsers where this could be reproduced -->
* **Language:** [all | TypeScript X.X | ES6/7 | ES5]
* **Node (for AoT issues):** `node --version` =
| defect | offset when opening up menu i m submitting a check one with x bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see plunkr case bug reports please fork the plunkr below and create a case demonstrating your bug report issues without a plunkr have much less possibility to be reviewed current behavior if an anchor tag contains another element the menu attaches to that element expected behavior the menu should always attach to the anchor tag minimal reproduction of the problem with instructions if the current behavior is a bug or you can illustrate your feature request better with an example please provide the steps to reproduce and if possible a minimal demo of the problem via or similar you can use this template as a starting point what is the motivation use case for changing the behavior please tell us about your environment angular version x primeng version x browser language node for aot issues node version | 1 |
41,920 | 10,707,853,738 | IssuesEvent | 2019-10-24 18:23:02 | STEllAR-GROUP/phylanx | https://api.github.com/repos/STEllAR-GROUP/phylanx | closed | `fold_left`, `fold_right` and `fmap` do not work with a lazy function | category: @Phylanx type: compatibility issue type: defect | Having the following code:
```py
import numpy as np
from phylanx import Phylanx, PhylanxSession, execution_tree
PhylanxSession.init(1)
def variable(value, dtype=None, name=None, constraint=None):
if dtype is None:
dtype = "float32"
if constraint is not None:
raise TypeError("Constraint is the projection function to be "
"applied to the variable after an optimizer update")
from phylanx.ast.physl import PhySL
if isinstance(value, PhySL.eval_wrapper):
return execution_tree.variable(value.code(), dtype)
if isinstance(value, execution_tree.variable):
return value
return execution_tree.variable(value, dtype=dtype, name=name)
def eval(func):
return func.eval()
@Phylanx
def map_fn_eager(fn, elems, dtype):
return fmap(fn, elems)
def map_fn(fn, elems, name=None, dtype=None):
return map_fn_eager.lazy(fn, elems, dtype)
@Phylanx
def sum_eager(x, axis=None, keepdims=False):
return np.sum(x, axis, keepdims)
def sum(x, axis=None, keepdims=False):
return sum_eager.lazy(x, axis, keepdims)
def test_map():
x = [1,2,3]
vx = variable(x)
kx = eval(map_fn(sum, vx))
return kx
test_map()
```
results in:
```pytb
Traceback (most recent call last):
File "test51.py", line 63, in <module>
print(test_map())
File "test51.py", line 47, in test_map
kx = eval(foldl(sum, vx))
File "test51.py", line 41, in foldl
return foldl_eager.lazy(fn, elems, initializer)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\transducer.py", line 123, in lazy
return self.backend.lazy(map(self.map_decorated, args))
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 526, in lazy
return self.eval_wrapper(self, tuple(map(self.map_wrapped, args)))
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\transducer.py", line 111, in map_decorated
fn_physl = PhySL(val, fn_ast, {})
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 387, in __init__
self.ir = self.apply_rule(tree.body[0])
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule
return eval('self._%s' % node_name)(node)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 951, in _FunctionDef
body = self.block(node.body)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 440, in block
block = tuple(map(self.apply_rule, node))
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule
return eval('self._%s' % node_name)(node)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 1144, in _Return
value = self.apply_rule(node.value)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule
return eval('self._%s' % node_name)(node)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 693, in _Call
symbol = self.apply_rule(node.func)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule
return eval('self._%s' % node_name)(node)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 654, in _Attribute
'Cannot transform: %s' % attr)
NotImplementedError: Phylanx does not support non-NumPy member functions.Cannot transform: sum_eager.lazy
``` | 1.0 | `fold_left`, `fold_right` and `fmap` do not work with a lazy function - Having the following code:
```py
import numpy as np
from phylanx import Phylanx, PhylanxSession, execution_tree
PhylanxSession.init(1)
def variable(value, dtype=None, name=None, constraint=None):
if dtype is None:
dtype = "float32"
if constraint is not None:
raise TypeError("Constraint is the projection function to be "
"applied to the variable after an optimizer update")
from phylanx.ast.physl import PhySL
if isinstance(value, PhySL.eval_wrapper):
return execution_tree.variable(value.code(), dtype)
if isinstance(value, execution_tree.variable):
return value
return execution_tree.variable(value, dtype=dtype, name=name)
def eval(func):
return func.eval()
@Phylanx
def map_fn_eager(fn, elems, dtype):
return fmap(fn, elems)
def map_fn(fn, elems, name=None, dtype=None):
return map_fn_eager.lazy(fn, elems, dtype)
@Phylanx
def sum_eager(x, axis=None, keepdims=False):
return np.sum(x, axis, keepdims)
def sum(x, axis=None, keepdims=False):
return sum_eager.lazy(x, axis, keepdims)
def test_map():
x = [1,2,3]
vx = variable(x)
kx = eval(map_fn(sum, vx))
return kx
test_map()
```
results in:
```pytb
Traceback (most recent call last):
File "test51.py", line 63, in <module>
print(test_map())
File "test51.py", line 47, in test_map
kx = eval(foldl(sum, vx))
File "test51.py", line 41, in foldl
return foldl_eager.lazy(fn, elems, initializer)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\transducer.py", line 123, in lazy
return self.backend.lazy(map(self.map_decorated, args))
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 526, in lazy
return self.eval_wrapper(self, tuple(map(self.map_wrapped, args)))
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\transducer.py", line 111, in map_decorated
fn_physl = PhySL(val, fn_ast, {})
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 387, in __init__
self.ir = self.apply_rule(tree.body[0])
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule
return eval('self._%s' % node_name)(node)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 951, in _FunctionDef
body = self.block(node.body)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 440, in block
block = tuple(map(self.apply_rule, node))
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule
return eval('self._%s' % node_name)(node)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 1144, in _Return
value = self.apply_rule(node.value)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule
return eval('self._%s' % node_name)(node)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 693, in _Call
symbol = self.apply_rule(node.func)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule
return eval('self._%s' % node_name)(node)
File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 654, in _Attribute
'Cannot transform: %s' % attr)
NotImplementedError: Phylanx does not support non-NumPy member functions.Cannot transform: sum_eager.lazy
``` | defect | fold left fold right and fmap do not work with a lazy function having the following code py import numpy as np from phylanx import phylanx phylanxsession execution tree phylanxsession init def variable value dtype none name none constraint none if dtype is none dtype if constraint is not none raise typeerror constraint is the projection function to be applied to the variable after an optimizer update from phylanx ast physl import physl if isinstance value physl eval wrapper return execution tree variable value code dtype if isinstance value execution tree variable return value return execution tree variable value dtype dtype name name def eval func return func eval phylanx def map fn eager fn elems dtype return fmap fn elems def map fn fn elems name none dtype none return map fn eager lazy fn elems dtype phylanx def sum eager x axis none keepdims false return np sum x axis keepdims def sum x axis none keepdims false return sum eager lazy x axis keepdims def test map x vx variable x kx eval map fn sum vx return kx test map results in pytb traceback most recent call last file py line in print test map file py line in test map kx eval foldl sum vx file py line in foldl return foldl eager lazy fn elems initializer file c repos phylanx cmake build debug python build lib win phylanx ast transducer py line in lazy return self backend lazy map self map decorated args file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in lazy return self eval wrapper self tuple map self map wrapped args file c repos phylanx cmake build debug python build lib win phylanx ast transducer py line in map decorated fn physl physl val fn ast file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in init self ir self apply rule tree body file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in apply rule return eval self s node name node file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in functiondef body self block node body file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in block block tuple map self apply rule node file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in apply rule return eval self s node name node file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in return value self apply rule node value file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in apply rule return eval self s node name node file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in call symbol self apply rule node func file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in apply rule return eval self s node name node file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in attribute cannot transform s attr notimplementederror phylanx does not support non numpy member functions cannot transform sum eager lazy | 1 |
24,697 | 4,074,930,861 | IssuesEvent | 2016-05-28 20:25:14 | Nuand/bladeRF | https://api.github.com/repos/Nuand/bladeRF | closed | [libbladeRF] Clarify required SMB clock port config prior to trigger API usage | documentation defect | [This documentation](https://nuand.com/libbladeRF-doc/v1.7.0/group___f_n___t_r_i_g.html) does not explicitly note the requirement to have the SMB clock configuration completed prior to using these trigger features.
This is required and should be explicitly be noted as a prerequisite. | 1.0 | [libbladeRF] Clarify required SMB clock port config prior to trigger API usage - [This documentation](https://nuand.com/libbladeRF-doc/v1.7.0/group___f_n___t_r_i_g.html) does not explicitly note the requirement to have the SMB clock configuration completed prior to using these trigger features.
This is required and should be explicitly be noted as a prerequisite. | defect | clarify required smb clock port config prior to trigger api usage does not explicitly note the requirement to have the smb clock configuration completed prior to using these trigger features this is required and should be explicitly be noted as a prerequisite | 1 |
292,021 | 21,946,706,960 | IssuesEvent | 2022-05-24 01:56:29 | AndyBoyee604/Taskinator | https://api.github.com/repos/AndyBoyee604/Taskinator | opened | Initial Setup | documentation | ## Requirements
* Create the task tracking HTML page that needs a:
* Header
* Main Content area for the task list
* footer
* Use the style sheet provided
* Add functionality to the button to add tasks to the list | 1.0 | Initial Setup - ## Requirements
* Create the task tracking HTML page that needs a:
* Header
* Main Content area for the task list
* footer
* Use the style sheet provided
* Add functionality to the button to add tasks to the list | non_defect | initial setup requirements create the task tracking html page that needs a header main content area for the task list footer use the style sheet provided add functionality to the button to add tasks to the list | 0 |
269,648 | 8,441,481,436 | IssuesEvent | 2018-10-18 10:20:47 | fac-14/sleuth-coop | https://api.github.com/repos/fac-14/sleuth-coop | opened | Password validation consistency | T25m bug priority-1 | The validation on the password for the sign up form and the login form is different. You can use a 3 character password on sign up but your password has to be min 4 characters on the log in! | 1.0 | Password validation consistency - The validation on the password for the sign up form and the login form is different. You can use a 3 character password on sign up but your password has to be min 4 characters on the log in! | non_defect | password validation consistency the validation on the password for the sign up form and the login form is different you can use a character password on sign up but your password has to be min characters on the log in | 0 |
77,906 | 27,229,151,217 | IssuesEvent | 2023-02-21 11:56:28 | scoutplan/scoutplan | https://api.github.com/repos/scoutplan/scoutplan | closed | [Scoutplan Production/production] Net::OpenTimeout: Timeout to open TCP connection to smtp.mailgun.org:587 (exceeds 5 seconds) | defect | ## Backtrace
[View full backtrace and more info at honeybadger.io](https://app.honeybadger.io/projects/97676/faults/93912390) | 1.0 | [Scoutplan Production/production] Net::OpenTimeout: Timeout to open TCP connection to smtp.mailgun.org:587 (exceeds 5 seconds) - ## Backtrace
[View full backtrace and more info at honeybadger.io](https://app.honeybadger.io/projects/97676/faults/93912390) | defect | net opentimeout timeout to open tcp connection to smtp mailgun org exceeds seconds backtrace | 1 |
57,385 | 15,761,689,092 | IssuesEvent | 2021-03-31 10:15:01 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | [COGNITION]: Actionable links should have visual cues | 508-defect-2 508/Accessibility direct deposit vsa-authenticated-exp | ## Description
[COGNITION]: Actionable links should have visual cues
<img width="788" alt="direct-deposit-edit-links-issue" src="https://user-images.githubusercontent.com/57469/68237941-2cba4580-ffd6-11e9-823c-5627a2f09ed2.png">
Actionable elements must be identified visually, by convention, and by information provided to assistive technologies. The Edit links have a color difference from the text, but the visual cue is not sufficiently clear that it is actionable.
## Point of Contact
Jennifer
## Acceptance Criteria
As a user of VA.gov, I want to open the Direct Deposit information update functionality. These action links should be clearly visible as actionable links.
## Environment
* Operating System: all
* Browser: all
## Steps to Recreate
1. Log into staging.va.gov with user 1
2. Click on the person menu in the top right corner
3. Click "Profile" from that menu. Your profile view will load.
4. Click the "Direct deposit information" link to jump to the direct deposit section
5. See any of the Edit links next to Bank name, Account number, or Account type to open the DD modal. They are currently coded as buttons, but styled as links, and need to follow the link styling from [https://design.va.gov/design/typography#links](https://design.va.gov/design/typography#links).
| 1.0 | [COGNITION]: Actionable links should have visual cues - ## Description
[COGNITION]: Actionable links should have visual cues
<img width="788" alt="direct-deposit-edit-links-issue" src="https://user-images.githubusercontent.com/57469/68237941-2cba4580-ffd6-11e9-823c-5627a2f09ed2.png">
Actionable elements must be identified visually, by convention, and by information provided to assistive technologies. The Edit links have a color difference from the text, but the visual cue is not sufficiently clear that it is actionable.
## Point of Contact
Jennifer
## Acceptance Criteria
As a user of VA.gov, I want to open the Direct Deposit information update functionality. These action links should be clearly visible as actionable links.
## Environment
* Operating System: all
* Browser: all
## Steps to Recreate
1. Log into staging.va.gov with user 1
2. Click on the person menu in the top right corner
3. Click "Profile" from that menu. Your profile view will load.
4. Click the "Direct deposit information" link to jump to the direct deposit section
5. See any of the Edit links next to Bank name, Account number, or Account type to open the DD modal. They are currently coded as buttons, but styled as links, and need to follow the link styling from [https://design.va.gov/design/typography#links](https://design.va.gov/design/typography#links).
| defect | actionable links should have visual cues description actionable links should have visual cues img width alt direct deposit edit links issue src actionable elements must be identified visually by convention and by information provided to assistive technologies the edit links have a color difference from the text but the visual cue is not sufficiently clear that it is actionable point of contact jennifer acceptance criteria as a user of va gov i want to open the direct deposit information update functionality these action links should be clearly visible as actionable links environment operating system all browser all steps to recreate log into staging va gov with user click on the person menu in the top right corner click profile from that menu your profile view will load click the direct deposit information link to jump to the direct deposit section see any of the edit links next to bank name account number or account type to open the dd modal they are currently coded as buttons but styled as links and need to follow the link styling from | 1 |
63,039 | 17,355,452,752 | IssuesEvent | 2021-07-29 13:56:25 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | NPE from SQL | Type: Defect | Using 5.0-BETA-1, the following code throws a NullPointerException
```
public static void main(String[] args) throws Exception {
Config config = new Config();
HazelcastInstance hazelcastInstance = Hazelcast.newHazelcastInstance(config);
String sql = "CREATE OR REPLACE MAPPING \"neil\" "
+ "TYPE IMap "
+ " OPTIONS ( "
+ " 'keyFormat' = 'java',"
+ " 'keyJavaClass' = '" + Integer.class.getCanonicalName() + "',"
+ " 'valueFormat' = 'java',"
+ " 'valueJavaClass' = '" + Integer.class.getCanonicalName() + "'"
+ " )";
hazelcastInstance.getSql().execute(sql);
hazelcastInstance.shutdown();
}
```
classpath is
```
[INFO] +- com.hazelcast:hazelcast:jar:5.0-BETA-1:compile
[INFO] \- com.hazelcast:hazelcast-sql:jar:5.0-BETA-1:compile
```
With 5.0-BETA-1 I get
```
Exception in thread "main" com.hazelcast.sql.HazelcastSqlException
at com.hazelcast.sql.impl.QueryUtils.toPublicException(QueryUtils.java:79)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:233)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:199)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:195)
at com.hazelcast.sql.SqlService.execute(SqlService.java:146)
at neil.Application.main(Application.java:20)
Caused by: com.hazelcast.sql.impl.QueryException
at com.hazelcast.sql.impl.QueryException.error(QueryException.java:60)
at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:86)
at com.hazelcast.sql.impl.calcite.OptimizerContext.parse(OptimizerContext.java:136)
at com.hazelcast.sql.impl.calcite.CalciteSqlOptimizer.prepare(CalciteSqlOptimizer.java:127)
at com.hazelcast.sql.impl.SqlServiceImpl.prepare(SqlServiceImpl.java:287)
at com.hazelcast.sql.impl.SqlServiceImpl.query0(SqlServiceImpl.java:268)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:220)
... 4 more
Caused by: java.lang.NullPointerException
at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:91)
at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:76)
... 9 more
```
5.0-SNAPSHOT (`20210729 - 12d65bd`) gives a different error
```
Exception in thread "main" com.hazelcast.sql.HazelcastSqlException: Encountered "CREATE" at line 1, column 1.
at com.hazelcast.sql.impl.QueryUtils.toPublicException(QueryUtils.java:79)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:238)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:200)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:196)
at com.hazelcast.sql.SqlService.execute(SqlService.java:146)
at neil.Application.main(Application.java:20)
Caused by: com.hazelcast.sql.impl.QueryException: Encountered "CREATE" at line 1, column 1.
at com.hazelcast.sql.impl.QueryException.error(QueryException.java:60)
at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:95)
at com.hazelcast.sql.impl.calcite.OptimizerContext.parse(OptimizerContext.java:140)
at com.hazelcast.sql.impl.calcite.CalciteSqlOptimizer.prepare(CalciteSqlOptimizer.java:127)
at com.hazelcast.sql.impl.SqlServiceImpl.prepare(SqlServiceImpl.java:292)
at com.hazelcast.sql.impl.SqlServiceImpl.query0(SqlServiceImpl.java:273)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:225)
... 4 more
Caused by: com.hazelcast.org.apache.calcite.sql.parser.SqlParseException: Encountered "CREATE" at line 1, column 1.
Was expecting one of:
"ABS" ...
"ALTER" ...
"ARRAY" ...
"AVG" ...
"CALL" ...
"CARDINALITY" ...
"CASE" ...
"CAST" ...
"CEIL" ...
"CEILING" ...
"CHAR_LENGTH" ...
"CHARACTER_LENGTH" ...
"CLASSIFIER" ...
"COALESCE" ...
"COLLECT" ...
"CONVERT" ...
"COUNT" ...
"COVAR_POP" ...
"COVAR_SAMP" ...
"CUME_DIST" ...
```
`CREATE` isn't listed, so not sure if this is a true defect or due to beta changes | 1.0 | NPE from SQL - Using 5.0-BETA-1, the following code throws a NullPointerException
```
public static void main(String[] args) throws Exception {
Config config = new Config();
HazelcastInstance hazelcastInstance = Hazelcast.newHazelcastInstance(config);
String sql = "CREATE OR REPLACE MAPPING \"neil\" "
+ "TYPE IMap "
+ " OPTIONS ( "
+ " 'keyFormat' = 'java',"
+ " 'keyJavaClass' = '" + Integer.class.getCanonicalName() + "',"
+ " 'valueFormat' = 'java',"
+ " 'valueJavaClass' = '" + Integer.class.getCanonicalName() + "'"
+ " )";
hazelcastInstance.getSql().execute(sql);
hazelcastInstance.shutdown();
}
```
classpath is
```
[INFO] +- com.hazelcast:hazelcast:jar:5.0-BETA-1:compile
[INFO] \- com.hazelcast:hazelcast-sql:jar:5.0-BETA-1:compile
```
With 5.0-BETA-1 I get
```
Exception in thread "main" com.hazelcast.sql.HazelcastSqlException
at com.hazelcast.sql.impl.QueryUtils.toPublicException(QueryUtils.java:79)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:233)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:199)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:195)
at com.hazelcast.sql.SqlService.execute(SqlService.java:146)
at neil.Application.main(Application.java:20)
Caused by: com.hazelcast.sql.impl.QueryException
at com.hazelcast.sql.impl.QueryException.error(QueryException.java:60)
at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:86)
at com.hazelcast.sql.impl.calcite.OptimizerContext.parse(OptimizerContext.java:136)
at com.hazelcast.sql.impl.calcite.CalciteSqlOptimizer.prepare(CalciteSqlOptimizer.java:127)
at com.hazelcast.sql.impl.SqlServiceImpl.prepare(SqlServiceImpl.java:287)
at com.hazelcast.sql.impl.SqlServiceImpl.query0(SqlServiceImpl.java:268)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:220)
... 4 more
Caused by: java.lang.NullPointerException
at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:91)
at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:76)
... 9 more
```
5.0-SNAPSHOT (`20210729 - 12d65bd`) gives a different error
```
Exception in thread "main" com.hazelcast.sql.HazelcastSqlException: Encountered "CREATE" at line 1, column 1.
at com.hazelcast.sql.impl.QueryUtils.toPublicException(QueryUtils.java:79)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:238)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:200)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:196)
at com.hazelcast.sql.SqlService.execute(SqlService.java:146)
at neil.Application.main(Application.java:20)
Caused by: com.hazelcast.sql.impl.QueryException: Encountered "CREATE" at line 1, column 1.
at com.hazelcast.sql.impl.QueryException.error(QueryException.java:60)
at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:95)
at com.hazelcast.sql.impl.calcite.OptimizerContext.parse(OptimizerContext.java:140)
at com.hazelcast.sql.impl.calcite.CalciteSqlOptimizer.prepare(CalciteSqlOptimizer.java:127)
at com.hazelcast.sql.impl.SqlServiceImpl.prepare(SqlServiceImpl.java:292)
at com.hazelcast.sql.impl.SqlServiceImpl.query0(SqlServiceImpl.java:273)
at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:225)
... 4 more
Caused by: com.hazelcast.org.apache.calcite.sql.parser.SqlParseException: Encountered "CREATE" at line 1, column 1.
Was expecting one of:
"ABS" ...
"ALTER" ...
"ARRAY" ...
"AVG" ...
"CALL" ...
"CARDINALITY" ...
"CASE" ...
"CAST" ...
"CEIL" ...
"CEILING" ...
"CHAR_LENGTH" ...
"CHARACTER_LENGTH" ...
"CLASSIFIER" ...
"COALESCE" ...
"COLLECT" ...
"CONVERT" ...
"COUNT" ...
"COVAR_POP" ...
"COVAR_SAMP" ...
"CUME_DIST" ...
```
`CREATE` isn't listed, so not sure if this is a true defect or due to beta changes | defect | npe from sql using beta the following code throws a nullpointerexception public static void main string args throws exception config config new config hazelcastinstance hazelcastinstance hazelcast newhazelcastinstance config string sql create or replace mapping neil type imap options keyformat java keyjavaclass integer class getcanonicalname valueformat java valuejavaclass integer class getcanonicalname hazelcastinstance getsql execute sql hazelcastinstance shutdown classpath is com hazelcast hazelcast jar beta compile com hazelcast hazelcast sql jar beta compile with beta i get exception in thread main com hazelcast sql hazelcastsqlexception at com hazelcast sql impl queryutils topublicexception queryutils java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql sqlservice execute sqlservice java at neil application main application java caused by com hazelcast sql impl queryexception at com hazelcast sql impl queryexception error queryexception java at com hazelcast sql impl calcite parse queryparser parse queryparser java at com hazelcast sql impl calcite optimizercontext parse optimizercontext java at com hazelcast sql impl calcite calcitesqloptimizer prepare calcitesqloptimizer java at com hazelcast sql impl sqlserviceimpl prepare sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java more caused by java lang nullpointerexception at com hazelcast sql impl calcite parse queryparser parse queryparser java at com hazelcast sql impl calcite parse queryparser parse queryparser java more snapshot gives a different error exception in thread main com hazelcast sql hazelcastsqlexception encountered create at line column at com hazelcast sql impl queryutils topublicexception queryutils java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql sqlservice execute sqlservice java at neil application main application java caused by com hazelcast sql impl queryexception encountered create at line column at com hazelcast sql impl queryexception error queryexception java at com hazelcast sql impl calcite parse queryparser parse queryparser java at com hazelcast sql impl calcite optimizercontext parse optimizercontext java at com hazelcast sql impl calcite calcitesqloptimizer prepare calcitesqloptimizer java at com hazelcast sql impl sqlserviceimpl prepare sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java more caused by com hazelcast org apache calcite sql parser sqlparseexception encountered create at line column was expecting one of abs alter array avg call cardinality case cast ceil ceiling char length character length classifier coalesce collect convert count covar pop covar samp cume dist create isn t listed so not sure if this is a true defect or due to beta changes | 1 |
21,192 | 16,616,407,692 | IssuesEvent | 2021-06-02 17:15:46 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | opened | About access restricting to system database . It's kinda cumbersome. | usability | ```sql
create user foo identified by '123';
revoke all on *.* from foo;
create role RO;
CREATE ROW POLICY ro_query_log_filter ON system.query_log USING 1 AS RESTRICTIVE TO RO;
CREATE ROW POLICY ro_part_log_filter ON system.part_log USING 1 AS RESTRICTIVE TO RO;
CREATE ROW POLICY ro_trace_log_filter ON system.trace_log USING 1 AS RESTRICTIVE TO RO;
CREATE ROW POLICY ro_processes_filter ON system.processes USING 1 AS RESTRICTIVE TO RO;
grant RO to foo;
ALTER USER foo DEFAULT ROLE RO SETTINGS NONE;
clickhouse-client -u foo --password=123
select count() from system.trace_log;
0 rows in set.
select count() from system.query_log;
0 rows in set.
select count() from system.tables;
0 rows in set.
select query from system.processes;
0 rows in set
```
Also this is a bug:
```
use system;
DB::Exception: foo: Not enough privileges.
```
Granting access to own rows in query_log/trace_log is even more cumbersome:
```
CREATE POLICY ${user_name}_query_log_filter ON system.query_log FOR SELECT USING initial_user='${user_name}' TO ${user_name};
```
| True | About access restricting to system database . It's kinda cumbersome. - ```sql
create user foo identified by '123';
revoke all on *.* from foo;
create role RO;
CREATE ROW POLICY ro_query_log_filter ON system.query_log USING 1 AS RESTRICTIVE TO RO;
CREATE ROW POLICY ro_part_log_filter ON system.part_log USING 1 AS RESTRICTIVE TO RO;
CREATE ROW POLICY ro_trace_log_filter ON system.trace_log USING 1 AS RESTRICTIVE TO RO;
CREATE ROW POLICY ro_processes_filter ON system.processes USING 1 AS RESTRICTIVE TO RO;
grant RO to foo;
ALTER USER foo DEFAULT ROLE RO SETTINGS NONE;
clickhouse-client -u foo --password=123
select count() from system.trace_log;
0 rows in set.
select count() from system.query_log;
0 rows in set.
select count() from system.tables;
0 rows in set.
select query from system.processes;
0 rows in set
```
Also this is a bug:
```
use system;
DB::Exception: foo: Not enough privileges.
```
Granting access to own rows in query_log/trace_log is even more cumbersome:
```
CREATE POLICY ${user_name}_query_log_filter ON system.query_log FOR SELECT USING initial_user='${user_name}' TO ${user_name};
```
| non_defect | about access restricting to system database it s kinda cumbersome sql create user foo identified by revoke all on from foo create role ro create row policy ro query log filter on system query log using as restrictive to ro create row policy ro part log filter on system part log using as restrictive to ro create row policy ro trace log filter on system trace log using as restrictive to ro create row policy ro processes filter on system processes using as restrictive to ro grant ro to foo alter user foo default role ro settings none clickhouse client u foo password select count from system trace log rows in set select count from system query log rows in set select count from system tables rows in set select query from system processes rows in set also this is a bug use system db exception foo not enough privileges granting access to own rows in query log trace log is even more cumbersome create policy user name query log filter on system query log for select using initial user user name to user name | 0 |
232,613 | 17,788,731,710 | IssuesEvent | 2021-08-31 14:02:08 | pythonarcade/arcade | https://api.github.com/repos/pythonarcade/arcade | closed | [DOC] Example for UI event register | enhancement documentation | ## Documentation request:
Hello! I'm a newcomer from Qt UI designing.
In PyQt5 or PySide2, there's a "signal & slot" mechanism in all event handler, which is relative to `EventDispatcher` in Pyglet.
In Qt:
```python
# Slot function (an optional marked callable object)
@Slot()
def close(self):
...
# Signal connect to callable
exit_btn.clicked.connect(lambda: print("Exit!"))
exit_btn.clicked.connect(self.close)
```
In `arcade.gui`:
```python
exit_btn.push_handlers(on_click=lambda: print("Exit!"))
exit_btn.push_handlers(on_click=close_window)
```
This is also a better design instead of override event function directly when creating a lot UI elements.
Not yet explained in documentation after my searching.
```python
# Directly assignment or create new inheritance
# This looks like a monkey patch to me
exit_btn.on_click = close_window
```
### Where is it located?
The API of `arcade.gui`.
### What is wrong with it? How can it be improved?
There is still no example for `arcade.gui` package.
Add a little example code for `EventDispatcher`. (like above)
Or a bigger one, like a kind of extensible option menu before the game.
```python
from arcade import set_background_color, start_render, run
from arcade.application import Window, View
from arcade.csscolor import CORNFLOWER_BLUE, GREENYELLOW, BLACK
from arcade.text import draw_text
from arcade.gui import UIManager, UIFlatButton
class MainView(View):
def __init__(self):
super(MainView, self).__init__()
self.title = draw_text('Game', self.window.width // 2,
self.window.height // 2,
BLACK, font_size=72, anchor_x='center')
def on_draw(self):
start_render()
set_background_color(GREENYELLOW)
self.title.draw()
class OptionView(View):
def __init__(self):
super(OptionView, self).__init__()
self.title = draw_text('Option', self.window.width // 2,
self.window.height // 2,
BLACK, font_size=72, anchor_x='center')
def on_draw(self):
start_render()
set_background_color(CORNFLOWER_BLUE)
self.title.draw()
class Main(Window):
def __init__(self):
super(Main, self).__init__()
self.ui = UIManager(self)
def setup(self):
self.show_main_ui()
def show_main_ui(self):
self.ui.purge_ui_elements()
opt_btn = UIFlatButton('Option', self.width // 2, self.height // 2 - 50,
300, 40)
opt_btn.push_handlers(on_click=self.show_opt_ui)
self.ui.add_ui_element(opt_btn)
self.show_view(MainView())
def show_opt_ui(self):
self.ui.purge_ui_elements()
back_btn = UIFlatButton('Back', self.width // 2, self.height // 2 - 50,
300, 40)
back_btn.push_handlers(on_click=self.show_main_ui)
self.ui.add_ui_element(back_btn)
self.show_view(OptionView())
def on_draw(self):
start_render()
self.current_view.on_draw()
self.ui.on_draw()
def main():
w = Main()
w.setup()
run()
if __name__ == '__main__':
main()
``` | 1.0 | [DOC] Example for UI event register - ## Documentation request:
Hello! I'm a newcomer from Qt UI designing.
In PyQt5 or PySide2, there's a "signal & slot" mechanism in all event handler, which is relative to `EventDispatcher` in Pyglet.
In Qt:
```python
# Slot function (an optional marked callable object)
@Slot()
def close(self):
...
# Signal connect to callable
exit_btn.clicked.connect(lambda: print("Exit!"))
exit_btn.clicked.connect(self.close)
```
In `arcade.gui`:
```python
exit_btn.push_handlers(on_click=lambda: print("Exit!"))
exit_btn.push_handlers(on_click=close_window)
```
This is also a better design instead of override event function directly when creating a lot UI elements.
Not yet explained in documentation after my searching.
```python
# Directly assignment or create new inheritance
# This looks like a monkey patch to me
exit_btn.on_click = close_window
```
### Where is it located?
The API of `arcade.gui`.
### What is wrong with it? How can it be improved?
There is still no example for `arcade.gui` package.
Add a little example code for `EventDispatcher`. (like above)
Or a bigger one, like a kind of extensible option menu before the game.
```python
from arcade import set_background_color, start_render, run
from arcade.application import Window, View
from arcade.csscolor import CORNFLOWER_BLUE, GREENYELLOW, BLACK
from arcade.text import draw_text
from arcade.gui import UIManager, UIFlatButton
class MainView(View):
def __init__(self):
super(MainView, self).__init__()
self.title = draw_text('Game', self.window.width // 2,
self.window.height // 2,
BLACK, font_size=72, anchor_x='center')
def on_draw(self):
start_render()
set_background_color(GREENYELLOW)
self.title.draw()
class OptionView(View):
def __init__(self):
super(OptionView, self).__init__()
self.title = draw_text('Option', self.window.width // 2,
self.window.height // 2,
BLACK, font_size=72, anchor_x='center')
def on_draw(self):
start_render()
set_background_color(CORNFLOWER_BLUE)
self.title.draw()
class Main(Window):
def __init__(self):
super(Main, self).__init__()
self.ui = UIManager(self)
def setup(self):
self.show_main_ui()
def show_main_ui(self):
self.ui.purge_ui_elements()
opt_btn = UIFlatButton('Option', self.width // 2, self.height // 2 - 50,
300, 40)
opt_btn.push_handlers(on_click=self.show_opt_ui)
self.ui.add_ui_element(opt_btn)
self.show_view(MainView())
def show_opt_ui(self):
self.ui.purge_ui_elements()
back_btn = UIFlatButton('Back', self.width // 2, self.height // 2 - 50,
300, 40)
back_btn.push_handlers(on_click=self.show_main_ui)
self.ui.add_ui_element(back_btn)
self.show_view(OptionView())
def on_draw(self):
start_render()
self.current_view.on_draw()
self.ui.on_draw()
def main():
w = Main()
w.setup()
run()
if __name__ == '__main__':
main()
``` | non_defect | example for ui event register documentation request hello i m a newcomer from qt ui designing in or there s a signal slot mechanism in all event handler which is relative to eventdispatcher in pyglet in qt python slot function an optional marked callable object slot def close self signal connect to callable exit btn clicked connect lambda print exit exit btn clicked connect self close in arcade gui python exit btn push handlers on click lambda print exit exit btn push handlers on click close window this is also a better design instead of override event function directly when creating a lot ui elements not yet explained in documentation after my searching python directly assignment or create new inheritance this looks like a monkey patch to me exit btn on click close window where is it located the api of arcade gui what is wrong with it how can it be improved there is still no example for arcade gui package add a little example code for eventdispatcher like above or a bigger one like a kind of extensible option menu before the game python from arcade import set background color start render run from arcade application import window view from arcade csscolor import cornflower blue greenyellow black from arcade text import draw text from arcade gui import uimanager uiflatbutton class mainview view def init self super mainview self init self title draw text game self window width self window height black font size anchor x center def on draw self start render set background color greenyellow self title draw class optionview view def init self super optionview self init self title draw text option self window width self window height black font size anchor x center def on draw self start render set background color cornflower blue self title draw class main window def init self super main self init self ui uimanager self def setup self self show main ui def show main ui self self ui purge ui elements opt btn uiflatbutton option self width self height opt btn push handlers on click self show opt ui self ui add ui element opt btn self show view mainview def show opt ui self self ui purge ui elements back btn uiflatbutton back self width self height back btn push handlers on click self show main ui self ui add ui element back btn self show view optionview def on draw self start render self current view on draw self ui on draw def main w main w setup run if name main main | 0 |
21,638 | 3,737,403,224 | IssuesEvent | 2016-03-08 19:09:38 | ELENA-LANG/elena-lang | https://api.github.com/repos/ELENA-LANG/elena-lang | closed | shorthand property settings | Design Idea Discussion | In the following code
#class MyClass
{
#field theField1.
#method Field1 = theField1.
#method set &Field1:val [ theField1 := val. ]
}
the following shorthand syntax is proposed
myObj Field1 := 2.
which is compiled as
myObject set &Field1:2. | 1.0 | shorthand property settings - In the following code
#class MyClass
{
#field theField1.
#method Field1 = theField1.
#method set &Field1:val [ theField1 := val. ]
}
the following shorthand syntax is proposed
myObj Field1 := 2.
which is compiled as
myObject set &Field1:2. | non_defect | shorthand property settings in the following code class myclass field method method set val the following shorthand syntax is proposed myobj which is compiled as myobject set | 0 |
11,401 | 9,344,868,968 | IssuesEvent | 2019-03-30 01:42:43 | microsoftgraph/msgraph-sdk-dotnet | https://api.github.com/repos/microsoftgraph/msgraph-sdk-dotnet | closed | FR: Enable parallel/out-of-sequence large file uploads in session | enhancement (service) | Currently it appears that a [file] UploadSession only supports receiving content chunks in the exact order the file needs to be constructed in (exact sequence). This defeats a straightforward method to improve large file upload performance by orders of magnitude.
If we can put content chunks in any sequence, we can leverage .Net parallel tasks and substantially improve customer experience.
Since the UploadSession (and the library's helper-class ChunkedUploadProvider) already keep track of which byte-ranges are missing, it appears to be a reasonably simple update for the service to allow this (and simply collate all the chunks and append them when all ranges have been received).
I have sample/prototype ParallelUpload code to offer the project if/when the service can support it (because I didn't read the fine-print before I wrote it, only to track down ServiceExceptions when testing).
Thanks for considering it. | 1.0 | FR: Enable parallel/out-of-sequence large file uploads in session - Currently it appears that a [file] UploadSession only supports receiving content chunks in the exact order the file needs to be constructed in (exact sequence). This defeats a straightforward method to improve large file upload performance by orders of magnitude.
If we can put content chunks in any sequence, we can leverage .Net parallel tasks and substantially improve customer experience.
Since the UploadSession (and the library's helper-class ChunkedUploadProvider) already keep track of which byte-ranges are missing, it appears to be a reasonably simple update for the service to allow this (and simply collate all the chunks and append them when all ranges have been received).
I have sample/prototype ParallelUpload code to offer the project if/when the service can support it (because I didn't read the fine-print before I wrote it, only to track down ServiceExceptions when testing).
Thanks for considering it. | non_defect | fr enable parallel out of sequence large file uploads in session currently it appears that a uploadsession only supports receiving content chunks in the exact order the file needs to be constructed in exact sequence this defeats a straightforward method to improve large file upload performance by orders of magnitude if we can put content chunks in any sequence we can leverage net parallel tasks and substantially improve customer experience since the uploadsession and the library s helper class chunkeduploadprovider already keep track of which byte ranges are missing it appears to be a reasonably simple update for the service to allow this and simply collate all the chunks and append them when all ranges have been received i have sample prototype parallelupload code to offer the project if when the service can support it because i didn t read the fine print before i wrote it only to track down serviceexceptions when testing thanks for considering it | 0 |
22,668 | 3,681,675,469 | IssuesEvent | 2016-02-24 05:03:10 | sccn/labstreaminglayer | https://api.github.com/repos/sccn/labstreaminglayer | closed | Fix for vis_stream does not work with Matlab R2014b | auto-migrated Priority-Medium Type-Defect | ```
It seems Matlab's new graphics system has broken vis_stream. The only problem
is in supergui line 109, finputcheck requires fig to be numeric but it is now
an object. I just emptied the type column (i.e. '') and it works again.
```
Original issue reported on code.google.com by `chadwick...@gmail.com` on 3 Dec 2014 at 6:47 | 1.0 | Fix for vis_stream does not work with Matlab R2014b - ```
It seems Matlab's new graphics system has broken vis_stream. The only problem
is in supergui line 109, finputcheck requires fig to be numeric but it is now
an object. I just emptied the type column (i.e. '') and it works again.
```
Original issue reported on code.google.com by `chadwick...@gmail.com` on 3 Dec 2014 at 6:47 | defect | fix for vis stream does not work with matlab it seems matlab s new graphics system has broken vis stream the only problem is in supergui line finputcheck requires fig to be numeric but it is now an object i just emptied the type column i e and it works again original issue reported on code google com by chadwick gmail com on dec at | 1 |
71,879 | 23,839,014,148 | IssuesEvent | 2022-09-06 08:43:12 | vector-im/element-android | https://api.github.com/repos/vector-im/element-android | opened | App Layout: Slow to load filter lists | T-Defect Team: Delight Z-AppLayout | ### Steps to reproduce
When the user switches between filters on the home page it takes several moments to load the list (even if it's a short list)
### Outcome
#### What did you expect?
#### What happened instead?
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
_No response_
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
No | 1.0 | App Layout: Slow to load filter lists - ### Steps to reproduce
When the user switches between filters on the home page it takes several moments to load the list (even if it's a short list)
### Outcome
#### What did you expect?
#### What happened instead?
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
_No response_
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
No | defect | app layout slow to load filter lists steps to reproduce when the user switches between filters on the home page it takes several moments to load the list even if it s a short list outcome what did you expect what happened instead your phone model no response operating system version no response application version and app store no response homeserver no response will you send logs no are you willing to provide a pr no | 1 |
68,803 | 21,910,850,580 | IssuesEvent | 2022-05-21 02:59:30 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | closed | PANIC at zfs_znode.c zfs_znode_sa_init() - Regression in closed issue #10971 | Type: Defect Component: Encryption | ### System information
Type | Version/Name
--- | ---
Distribution Name | Pop!_OS |
Distribution Version | 20.04 & 22.04 |
Kernel Version | 5.16.15 & 5.17.5 |
Architecture | x86_64 |
OpenZFS Version | zfs-2.1.4-1pop1~1650059719~20.04~9934f01 |
### Describe the problem you're observing
Regression of bug in #10971
On 20.04 it used to work very well for many months, days ago I upgraded to 22.04 and this started popping up whenever pods in my k8s cluster access files in the raid array. There was a firmware update for my MOBO so I did that. To downgrade the OS I did a clean install, and `zpool import -f`. Would deleting the pool and recover from backup data help? It's many TBs so I would like to avoid this.
### Describe how to reproduce the problem
Use Pop!_OS 22.04 on an encrypted pool
### Include any warning/errors/backtraces from the system logs
```
[70099.854321] CPU: 3 PID: 3404665 Comm: .NET ThreadPool Tainted: P OE 5.17.5-76051705-generic #202204271406~1651504840~22.04~63e51bd
[70099.854323] Hardware name: To Be Filled By O.E.M. To Be Filled By O.E.M./Z590 Steel Legend WiFi 6E, BIOS P1.80 07/22/2021
[70099.854324] Call Trace:
[70099.854325] <TASK>
[70099.854327] show_stack+0x52/0x59
[70099.854330] dump_stack_lvl+0x4c/0x64
[70099.854333] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
[70099.854334] dump_stack+0x10/0x13
[70099.854335] PANIC at zfs_znode.c:339:zfs_znode_sa_init()
[70099.854336] Showing stack for process 3417543
[70099.854335] spl_dumpstack+0x29/0x2c [spl]
[70099.854341] spl_panic+0xd1/0xe9 [spl]
[70099.854345] ? sa_cache_constructor+0x27/0x50 [zfs]
[70099.854410] ? sa_build_index+0x2c1/0x310 [zfs]
[70099.854451] ? sa_build_index+0x25d/0x310 [zfs]
[70099.854491] ? dmu_buf_set_user_ie+0x62/0x80 [zfs]
[70099.854518] zfs_znode_sa_init+0xdf/0xf0 [zfs]
[70099.854569] zfs_znode_alloc+0x101/0x750 [zfs]
[70099.854617] ? __kmalloc_node+0x1c4/0x3e0
[70099.854620] ? arc_untransform+0x23/0x80 [zfs]
[70099.854644] ? aggsum_add+0x1a2/0x1c0 [zfs]
[70099.854666] ? dmu_buf_unlock_parent+0x39/0x80 [zfs]
[70099.854691] ? dbuf_read_impl.constprop.0+0x5a1/0x7f0 [zfs]
[70099.854717] ? spl_kmem_cache_alloc+0x9d/0x780 [spl]
[70099.854720] ? spl_kmem_cache_alloc+0x9d/0x780 [spl]
[70099.854723] ? aggsum_add+0x1a2/0x1c0 [zfs]
[70099.854745] ? aggsum_add+0x1a2/0x1c0 [zfs]
[70099.854768] ? arc_space_consume+0x54/0x120 [zfs]
[70099.854791] ? dbuf_create+0x5aa/0x5e0 [zfs]
[70099.854816] ? dbuf_read+0x11b/0x620 [zfs]
[70099.854843] ? dnode_rele_and_unlock+0x68/0xf0 [zfs]
[70099.854875] ? dmu_object_info_from_dnode+0x82/0xb0 [zfs]
[70099.854902] zfs_zget+0x249/0x290 [zfs]
[70099.854949] zfs_dirent_lock+0x4a7/0x670 [zfs]
[70099.854996] zfs_dirlook+0x91/0x2e0 [zfs]
[70099.855044] ? zfs_zaccess+0x216/0x420 [zfs]
[70099.855089] zfs_lookup+0x25b/0x410 [zfs]
[70099.855132] zpl_lookup+0xcf/0x1e0 [zfs]
[70099.855175] __lookup_slow+0x81/0x150
[70099.855178] walk_component+0x142/0x1c0
[70099.855179] link_path_walk.part.0.constprop.0+0x24b/0x3d0
[70099.855180] ? path_init+0x2c2/0x3f0
[70099.855181] path_lookupat+0x3e/0x1b0
[70099.855182] filename_lookup+0xcf/0x1d0
[70099.855183] ? __check_object_size+0x1a/0x20
[70099.855185] ? strncpy_from_user+0x44/0x140
[70099.855187] ? getname_flags.part.0+0x4c/0x1b0
[70099.855188] user_path_at_empty+0x3f/0x60
[70099.855189] vfs_statx+0x7a/0x120
[70099.855191] __do_sys_newlstat+0x3e/0x70
[70099.855192] ? restore_fpregs_from_fpstate+0x47/0xb0
[70099.855194] ? switch_fpu_return+0x4e/0xd0
[70099.855196] ? exit_to_user_mode_prepare+0x37/0xb0
[70099.855197] __x64_sys_newlstat+0x16/0x20
[70099.855199] do_syscall_64+0x59/0xc0
[70099.855200] ? exit_to_user_mode_prepare+0x37/0xb0
[70099.855201] ? syscall_exit_to_user_mode+0x27/0x50
[70099.855203] ? __do_sys_geteuid+0x28/0x30
[70099.855204] ? do_syscall_64+0x69/0xc0
[70099.855204] ? exit_to_user_mode_prepare+0x37/0xb0
[70099.855206] ? syscall_exit_to_user_mode+0x27/0x50
[70099.855207] ? do_syscall_64+0x69/0xc0
[70099.855208] ? do_syscall_64+0x69/0xc0
[70099.855208] entry_SYSCALL_64_after_hwframe+0x44/0xae
[70099.855210] RIP: 0033:0x7f7ce4111b45
[70099.855211] Code: 51 13 0d 00 64 c7 00 16 00 00 00 b8 ff ff ff ff c3 0f 1f 40 00 48 89 f0 83 ff 01 77 30 48 89 c7 48 89 d6 b8 06 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 03 c3 66 90 48 8b 15 19 13 0d 00 f7 d8 64 89
[70099.855212] RSP: 002b:00007f7bebffdc68 EFLAGS: 00000246 ORIG_RAX: 0000000000000006
[70099.855214] RAX: ffffffffffffffda RBX: 00007f7c35f786c8 RCX: 00007f7ce4111b45
[70099.855215] RDX: 00007f7bebffdc70 RSI: 00007f7bebffdc70 RDI: 00007f7bebffddc0
[70099.855215] RBP: 00007f7bebffddb0 R08: 00007f7bebffdd10 R09: 0000000000000047
[70099.855216] R10: 00007f7c6ab09538 R11: 0000000000000246 R12: 0000000000000000
[70099.855217] R13: 0000000000000001 R14: 00007f7c35f786c8 R15: 00007f7be0001770
[70099.855218] </TASK>
```
After downgrading back to 20.04
```
[28502.940773] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
[28502.940781] PANIC at zfs_znode.c:339:zfs_znode_sa_init()
[28502.940784] Showing stack for process 2986375
[28502.940787] CPU: 10 PID: 2986375 Comm: .NET ThreadPool Tainted: P OE 5.16.15-76051615-generic #202203161444~1647964027~20.04~e706226-Ubuntu
[28502.940789] Hardware name: To Be Filled By O.E.M. Z590 Steel Legend WiFi 6E/Z590 Steel Legend WiFi 6E, BIOS P2.00 10/21/2021
[28502.940791] Call Trace:
[28502.940794] <TASK>
[28502.940800] dump_stack_lvl+0x4c/0x63
[28502.940811] dump_stack+0x10/0x12
[28502.940816] spl_dumpstack+0x29/0x2b [spl]
[28502.940828] spl_panic+0xd1/0xe9 [spl]
[28502.940833] ? sa_cache_constructor+0x27/0x50 [zfs]
[28502.940976] ? spl_kmem_cache_alloc+0xa3/0x7e0 [spl]
[28502.940986] ? dmu_buf_replace_user+0x60/0x80 [zfs]
[28502.941041] ? dmu_buf_set_user_ie+0x1a/0x20 [zfs]
[28502.941094] zfs_znode_sa_init.isra.0+0xdd/0xf0 [zfs]
[28502.941175] zfs_znode_alloc+0x102/0x720 [zfs]
[28502.941273] ? __kmalloc_node+0x1db/0x540
[28502.941281] ? aggsum_add+0x1aa/0x1c0 [zfs]
[28502.941351] ? dmu_buf_unlock_parent+0x38/0x80 [zfs]
[28502.941412] ? dbuf_read_impl.constprop.0+0x654/0x760 [zfs]
[28502.941464] ? spl_kmem_cache_alloc+0xa3/0x7e0 [spl]
[28502.941471] ? arc_buf_access+0x266/0x280 [zfs]
[28502.941520] ? aggsum_add+0x1aa/0x1c0 [zfs]
[28502.941567] ? arc_space_consume+0x54/0x120 [zfs]
[28502.941615] ? dbuf_read+0x1ed/0x5f0 [zfs]
[28502.941664] ? dnode_rele_and_unlock+0x68/0xf0 [zfs]
[28502.941718] ? dmu_object_info_from_dnode+0x82/0xb0 [zfs]
[28502.941769] zfs_zget+0x24a/0x290 [zfs]
[28502.941838] zfs_dirent_lock+0x347/0x680 [zfs]
[28502.941906] zfs_dirlook+0x90/0x2e0 [zfs]
[28502.941972] ? zfs_zaccess+0x224/0x410 [zfs]
[28502.942039] zfs_lookup+0x25c/0x410 [zfs]
[28502.942104] zpl_lookup+0xc9/0x1e0 [zfs]
[28502.942173] ? __legitimize_path.isra.0+0x31/0x70
[28502.942177] __lookup_slow+0x81/0x140
[28502.942178] walk_component+0x141/0x1b0
[28502.942181] link_path_walk.part.0+0x22d/0x360
[28502.942182] ? path_init+0x2ab/0x390
[28502.942184] path_lookupat.isra.0+0x3e/0x150
[28502.942185] ? raw_spin_rq_unlock+0x10/0x30
[28502.942191] filename_lookup+0xcf/0x1a0
[28502.942192] ? __check_object_size+0x13f/0x150
[28502.942196] ? strncpy_from_user+0x44/0x150
[28502.942201] ? getname_flags+0x6f/0x1f0
[28502.942203] user_path_at_empty+0x3f/0x60
[28502.942204] vfs_statx+0x73/0x110
[28502.942207] __do_sys_newstat+0x3e/0x80
[28502.942209] ? __x64_sys_futex+0x7b/0x1b0
[28502.942212] ? exit_to_user_mode_prepare+0x3d/0x1c0
[28502.942218] ? syscall_exit_to_user_mode+0x27/0x50
[28502.942222] ? do_syscall_64+0x69/0xc0
[28502.942226] ? syscall_exit_to_user_mode+0x27/0x50
[28502.942228] __x64_sys_newstat+0x16/0x20
[28502.942229] do_syscall_64+0x59/0xc0
[28502.942231] ? exit_to_user_mode_prepare+0x3d/0x1c0
[28502.942233] ? schedule+0x55/0xd0
[28502.942236] ? syscall_exit_to_user_mode+0x27/0x50
[28502.942238] ? __do_sys_sched_yield+0xe/0x20
[28502.942240] ? do_syscall_64+0x69/0xc0
[28502.942242] ? do_syscall_64+0x69/0xc0
[28502.942244] ? do_syscall_64+0x69/0xc0
[28502.942245] ? do_syscall_64+0x69/0xc0
[28502.942247] ? do_syscall_64+0x69/0xc0
[28502.942248] ? asm_sysvec_apic_timer_interrupt+0xa/0x20
[28502.942250] entry_SYSCALL_64_after_hwframe+0x44/0xae
[28502.942252] RIP: 0033:0x7fdb95a9baa5
[28502.942255] Code: 00 00 00 75 05 48 83 c4 18 c3 e8 26 0d 02 00 66 0f 1f 44 00 00 48 89 f0 83 ff 01 77 30 48 89 c7 48 89 d6 b8 04 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 03 c3 66 90 48 8b 15 b9 13 0d 00 f7 d8 64 89
[28502.942257] RSP: 002b:00007fda6cff54b8 EFLAGS: 00000246 ORIG_RAX: 0000000000000004
[28502.942259] RAX: ffffffffffffffda RBX: 00007fda6cff5620 RCX: 00007fdb95a9baa5
[28502.942260] RDX: 00007fda6cff54c0 RSI: 00007fda6cff54c0 RDI: 00007fda6cff5620
[28502.942261] RBP: 00007fda6cff5610 R08: 00007fda6cff5570 R09: 0000000000000047
[28502.942262] R10: 00007fdb1c4794d8 R11: 0000000000000246 R12: 00007fdae82c30c0
[28502.942263] R13: 0000000000000001 R14: 00007fda6cff5780 R15: 00007fda6cff54c0
[28502.942265] </TASK>
``` | 1.0 | PANIC at zfs_znode.c zfs_znode_sa_init() - Regression in closed issue #10971 - ### System information
Type | Version/Name
--- | ---
Distribution Name | Pop!_OS |
Distribution Version | 20.04 & 22.04 |
Kernel Version | 5.16.15 & 5.17.5 |
Architecture | x86_64 |
OpenZFS Version | zfs-2.1.4-1pop1~1650059719~20.04~9934f01 |
### Describe the problem you're observing
Regression of bug in #10971
On 20.04 it used to work very well for many months, days ago I upgraded to 22.04 and this started popping up whenever pods in my k8s cluster access files in the raid array. There was a firmware update for my MOBO so I did that. To downgrade the OS I did a clean install, and `zpool import -f`. Would deleting the pool and recover from backup data help? It's many TBs so I would like to avoid this.
### Describe how to reproduce the problem
Use Pop!_OS 22.04 on an encrypted pool
### Include any warning/errors/backtraces from the system logs
```
[70099.854321] CPU: 3 PID: 3404665 Comm: .NET ThreadPool Tainted: P OE 5.17.5-76051705-generic #202204271406~1651504840~22.04~63e51bd
[70099.854323] Hardware name: To Be Filled By O.E.M. To Be Filled By O.E.M./Z590 Steel Legend WiFi 6E, BIOS P1.80 07/22/2021
[70099.854324] Call Trace:
[70099.854325] <TASK>
[70099.854327] show_stack+0x52/0x59
[70099.854330] dump_stack_lvl+0x4c/0x64
[70099.854333] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
[70099.854334] dump_stack+0x10/0x13
[70099.854335] PANIC at zfs_znode.c:339:zfs_znode_sa_init()
[70099.854336] Showing stack for process 3417543
[70099.854335] spl_dumpstack+0x29/0x2c [spl]
[70099.854341] spl_panic+0xd1/0xe9 [spl]
[70099.854345] ? sa_cache_constructor+0x27/0x50 [zfs]
[70099.854410] ? sa_build_index+0x2c1/0x310 [zfs]
[70099.854451] ? sa_build_index+0x25d/0x310 [zfs]
[70099.854491] ? dmu_buf_set_user_ie+0x62/0x80 [zfs]
[70099.854518] zfs_znode_sa_init+0xdf/0xf0 [zfs]
[70099.854569] zfs_znode_alloc+0x101/0x750 [zfs]
[70099.854617] ? __kmalloc_node+0x1c4/0x3e0
[70099.854620] ? arc_untransform+0x23/0x80 [zfs]
[70099.854644] ? aggsum_add+0x1a2/0x1c0 [zfs]
[70099.854666] ? dmu_buf_unlock_parent+0x39/0x80 [zfs]
[70099.854691] ? dbuf_read_impl.constprop.0+0x5a1/0x7f0 [zfs]
[70099.854717] ? spl_kmem_cache_alloc+0x9d/0x780 [spl]
[70099.854720] ? spl_kmem_cache_alloc+0x9d/0x780 [spl]
[70099.854723] ? aggsum_add+0x1a2/0x1c0 [zfs]
[70099.854745] ? aggsum_add+0x1a2/0x1c0 [zfs]
[70099.854768] ? arc_space_consume+0x54/0x120 [zfs]
[70099.854791] ? dbuf_create+0x5aa/0x5e0 [zfs]
[70099.854816] ? dbuf_read+0x11b/0x620 [zfs]
[70099.854843] ? dnode_rele_and_unlock+0x68/0xf0 [zfs]
[70099.854875] ? dmu_object_info_from_dnode+0x82/0xb0 [zfs]
[70099.854902] zfs_zget+0x249/0x290 [zfs]
[70099.854949] zfs_dirent_lock+0x4a7/0x670 [zfs]
[70099.854996] zfs_dirlook+0x91/0x2e0 [zfs]
[70099.855044] ? zfs_zaccess+0x216/0x420 [zfs]
[70099.855089] zfs_lookup+0x25b/0x410 [zfs]
[70099.855132] zpl_lookup+0xcf/0x1e0 [zfs]
[70099.855175] __lookup_slow+0x81/0x150
[70099.855178] walk_component+0x142/0x1c0
[70099.855179] link_path_walk.part.0.constprop.0+0x24b/0x3d0
[70099.855180] ? path_init+0x2c2/0x3f0
[70099.855181] path_lookupat+0x3e/0x1b0
[70099.855182] filename_lookup+0xcf/0x1d0
[70099.855183] ? __check_object_size+0x1a/0x20
[70099.855185] ? strncpy_from_user+0x44/0x140
[70099.855187] ? getname_flags.part.0+0x4c/0x1b0
[70099.855188] user_path_at_empty+0x3f/0x60
[70099.855189] vfs_statx+0x7a/0x120
[70099.855191] __do_sys_newlstat+0x3e/0x70
[70099.855192] ? restore_fpregs_from_fpstate+0x47/0xb0
[70099.855194] ? switch_fpu_return+0x4e/0xd0
[70099.855196] ? exit_to_user_mode_prepare+0x37/0xb0
[70099.855197] __x64_sys_newlstat+0x16/0x20
[70099.855199] do_syscall_64+0x59/0xc0
[70099.855200] ? exit_to_user_mode_prepare+0x37/0xb0
[70099.855201] ? syscall_exit_to_user_mode+0x27/0x50
[70099.855203] ? __do_sys_geteuid+0x28/0x30
[70099.855204] ? do_syscall_64+0x69/0xc0
[70099.855204] ? exit_to_user_mode_prepare+0x37/0xb0
[70099.855206] ? syscall_exit_to_user_mode+0x27/0x50
[70099.855207] ? do_syscall_64+0x69/0xc0
[70099.855208] ? do_syscall_64+0x69/0xc0
[70099.855208] entry_SYSCALL_64_after_hwframe+0x44/0xae
[70099.855210] RIP: 0033:0x7f7ce4111b45
[70099.855211] Code: 51 13 0d 00 64 c7 00 16 00 00 00 b8 ff ff ff ff c3 0f 1f 40 00 48 89 f0 83 ff 01 77 30 48 89 c7 48 89 d6 b8 06 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 03 c3 66 90 48 8b 15 19 13 0d 00 f7 d8 64 89
[70099.855212] RSP: 002b:00007f7bebffdc68 EFLAGS: 00000246 ORIG_RAX: 0000000000000006
[70099.855214] RAX: ffffffffffffffda RBX: 00007f7c35f786c8 RCX: 00007f7ce4111b45
[70099.855215] RDX: 00007f7bebffdc70 RSI: 00007f7bebffdc70 RDI: 00007f7bebffddc0
[70099.855215] RBP: 00007f7bebffddb0 R08: 00007f7bebffdd10 R09: 0000000000000047
[70099.855216] R10: 00007f7c6ab09538 R11: 0000000000000246 R12: 0000000000000000
[70099.855217] R13: 0000000000000001 R14: 00007f7c35f786c8 R15: 00007f7be0001770
[70099.855218] </TASK>
```
After downgrading back to 20.04
```
[28502.940773] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed
[28502.940781] PANIC at zfs_znode.c:339:zfs_znode_sa_init()
[28502.940784] Showing stack for process 2986375
[28502.940787] CPU: 10 PID: 2986375 Comm: .NET ThreadPool Tainted: P OE 5.16.15-76051615-generic #202203161444~1647964027~20.04~e706226-Ubuntu
[28502.940789] Hardware name: To Be Filled By O.E.M. Z590 Steel Legend WiFi 6E/Z590 Steel Legend WiFi 6E, BIOS P2.00 10/21/2021
[28502.940791] Call Trace:
[28502.940794] <TASK>
[28502.940800] dump_stack_lvl+0x4c/0x63
[28502.940811] dump_stack+0x10/0x12
[28502.940816] spl_dumpstack+0x29/0x2b [spl]
[28502.940828] spl_panic+0xd1/0xe9 [spl]
[28502.940833] ? sa_cache_constructor+0x27/0x50 [zfs]
[28502.940976] ? spl_kmem_cache_alloc+0xa3/0x7e0 [spl]
[28502.940986] ? dmu_buf_replace_user+0x60/0x80 [zfs]
[28502.941041] ? dmu_buf_set_user_ie+0x1a/0x20 [zfs]
[28502.941094] zfs_znode_sa_init.isra.0+0xdd/0xf0 [zfs]
[28502.941175] zfs_znode_alloc+0x102/0x720 [zfs]
[28502.941273] ? __kmalloc_node+0x1db/0x540
[28502.941281] ? aggsum_add+0x1aa/0x1c0 [zfs]
[28502.941351] ? dmu_buf_unlock_parent+0x38/0x80 [zfs]
[28502.941412] ? dbuf_read_impl.constprop.0+0x654/0x760 [zfs]
[28502.941464] ? spl_kmem_cache_alloc+0xa3/0x7e0 [spl]
[28502.941471] ? arc_buf_access+0x266/0x280 [zfs]
[28502.941520] ? aggsum_add+0x1aa/0x1c0 [zfs]
[28502.941567] ? arc_space_consume+0x54/0x120 [zfs]
[28502.941615] ? dbuf_read+0x1ed/0x5f0 [zfs]
[28502.941664] ? dnode_rele_and_unlock+0x68/0xf0 [zfs]
[28502.941718] ? dmu_object_info_from_dnode+0x82/0xb0 [zfs]
[28502.941769] zfs_zget+0x24a/0x290 [zfs]
[28502.941838] zfs_dirent_lock+0x347/0x680 [zfs]
[28502.941906] zfs_dirlook+0x90/0x2e0 [zfs]
[28502.941972] ? zfs_zaccess+0x224/0x410 [zfs]
[28502.942039] zfs_lookup+0x25c/0x410 [zfs]
[28502.942104] zpl_lookup+0xc9/0x1e0 [zfs]
[28502.942173] ? __legitimize_path.isra.0+0x31/0x70
[28502.942177] __lookup_slow+0x81/0x140
[28502.942178] walk_component+0x141/0x1b0
[28502.942181] link_path_walk.part.0+0x22d/0x360
[28502.942182] ? path_init+0x2ab/0x390
[28502.942184] path_lookupat.isra.0+0x3e/0x150
[28502.942185] ? raw_spin_rq_unlock+0x10/0x30
[28502.942191] filename_lookup+0xcf/0x1a0
[28502.942192] ? __check_object_size+0x13f/0x150
[28502.942196] ? strncpy_from_user+0x44/0x150
[28502.942201] ? getname_flags+0x6f/0x1f0
[28502.942203] user_path_at_empty+0x3f/0x60
[28502.942204] vfs_statx+0x73/0x110
[28502.942207] __do_sys_newstat+0x3e/0x80
[28502.942209] ? __x64_sys_futex+0x7b/0x1b0
[28502.942212] ? exit_to_user_mode_prepare+0x3d/0x1c0
[28502.942218] ? syscall_exit_to_user_mode+0x27/0x50
[28502.942222] ? do_syscall_64+0x69/0xc0
[28502.942226] ? syscall_exit_to_user_mode+0x27/0x50
[28502.942228] __x64_sys_newstat+0x16/0x20
[28502.942229] do_syscall_64+0x59/0xc0
[28502.942231] ? exit_to_user_mode_prepare+0x3d/0x1c0
[28502.942233] ? schedule+0x55/0xd0
[28502.942236] ? syscall_exit_to_user_mode+0x27/0x50
[28502.942238] ? __do_sys_sched_yield+0xe/0x20
[28502.942240] ? do_syscall_64+0x69/0xc0
[28502.942242] ? do_syscall_64+0x69/0xc0
[28502.942244] ? do_syscall_64+0x69/0xc0
[28502.942245] ? do_syscall_64+0x69/0xc0
[28502.942247] ? do_syscall_64+0x69/0xc0
[28502.942248] ? asm_sysvec_apic_timer_interrupt+0xa/0x20
[28502.942250] entry_SYSCALL_64_after_hwframe+0x44/0xae
[28502.942252] RIP: 0033:0x7fdb95a9baa5
[28502.942255] Code: 00 00 00 75 05 48 83 c4 18 c3 e8 26 0d 02 00 66 0f 1f 44 00 00 48 89 f0 83 ff 01 77 30 48 89 c7 48 89 d6 b8 04 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 03 c3 66 90 48 8b 15 b9 13 0d 00 f7 d8 64 89
[28502.942257] RSP: 002b:00007fda6cff54b8 EFLAGS: 00000246 ORIG_RAX: 0000000000000004
[28502.942259] RAX: ffffffffffffffda RBX: 00007fda6cff5620 RCX: 00007fdb95a9baa5
[28502.942260] RDX: 00007fda6cff54c0 RSI: 00007fda6cff54c0 RDI: 00007fda6cff5620
[28502.942261] RBP: 00007fda6cff5610 R08: 00007fda6cff5570 R09: 0000000000000047
[28502.942262] R10: 00007fdb1c4794d8 R11: 0000000000000246 R12: 00007fdae82c30c0
[28502.942263] R13: 0000000000000001 R14: 00007fda6cff5780 R15: 00007fda6cff54c0
[28502.942265] </TASK>
``` | defect | panic at zfs znode c zfs znode sa init regression in closed issue system information type version name distribution name pop os distribution version kernel version architecture openzfs version zfs describe the problem you re observing regression of bug in on it used to work very well for many months days ago i upgraded to and this started popping up whenever pods in my cluster access files in the raid array there was a firmware update for my mobo so i did that to downgrade the os i did a clean install and zpool import f would deleting the pool and recover from backup data help it s many tbs so i would like to avoid this describe how to reproduce the problem use pop os on an encrypted pool include any warning errors backtraces from the system logs cpu pid comm net threadpool tainted p oe generic hardware name to be filled by o e m to be filled by o e m steel legend wifi bios call trace show stack dump stack lvl verify sa handle get from db zfsvfs z os db zp sa hdl shared zp z sa hdl failed dump stack panic at zfs znode c zfs znode sa init showing stack for process spl dumpstack spl panic sa cache constructor sa build index sa build index dmu buf set user ie zfs znode sa init zfs znode alloc kmalloc node arc untransform aggsum add dmu buf unlock parent dbuf read impl constprop spl kmem cache alloc spl kmem cache alloc aggsum add aggsum add arc space consume dbuf create dbuf read dnode rele and unlock dmu object info from dnode zfs zget zfs dirent lock zfs dirlook zfs zaccess zfs lookup zpl lookup lookup slow walk component link path walk part constprop path init path lookupat filename lookup check object size strncpy from user getname flags part user path at empty vfs statx do sys newlstat restore fpregs from fpstate switch fpu return exit to user mode prepare sys newlstat do syscall exit to user mode prepare syscall exit to user mode do sys geteuid do syscall exit to user mode prepare syscall exit to user mode do syscall do syscall entry syscall after hwframe rip code ff ff ff ff ff ff ff rsp eflags orig rax rax ffffffffffffffda rbx rcx rdx rsi rdi rbp after downgrading back to verify sa handle get from db zfsvfs z os db zp sa hdl shared zp z sa hdl failed panic at zfs znode c zfs znode sa init showing stack for process cpu pid comm net threadpool tainted p oe generic ubuntu hardware name to be filled by o e m steel legend wifi steel legend wifi bios call trace dump stack lvl dump stack spl dumpstack spl panic sa cache constructor spl kmem cache alloc dmu buf replace user dmu buf set user ie zfs znode sa init isra zfs znode alloc kmalloc node aggsum add dmu buf unlock parent dbuf read impl constprop spl kmem cache alloc arc buf access aggsum add arc space consume dbuf read dnode rele and unlock dmu object info from dnode zfs zget zfs dirent lock zfs dirlook zfs zaccess zfs lookup zpl lookup legitimize path isra lookup slow walk component link path walk part path init path lookupat isra raw spin rq unlock filename lookup check object size strncpy from user getname flags user path at empty vfs statx do sys newstat sys futex exit to user mode prepare syscall exit to user mode do syscall syscall exit to user mode sys newstat do syscall exit to user mode prepare schedule syscall exit to user mode do sys sched yield do syscall do syscall do syscall do syscall do syscall asm sysvec apic timer interrupt entry syscall after hwframe rip code ff ff ff rsp eflags orig rax rax ffffffffffffffda rbx rcx rdx rsi rdi rbp | 1 |
292,666 | 25,228,368,838 | IssuesEvent | 2022-11-14 17:39:27 | mxsdev/ts-type-explorer | https://api.github.com/repos/mxsdev/ts-type-explorer | closed | Test Errors | bug tests | As of now running tests on the lib will always result in 2 errors.
This errors come from the `module.ts` file and are a result of how the tests are writen, i dont think this is a huge issue but at least one that should be mentioned on the contributing.md file.
Error:

| 1.0 | Test Errors - As of now running tests on the lib will always result in 2 errors.
This errors come from the `module.ts` file and are a result of how the tests are writen, i dont think this is a huge issue but at least one that should be mentioned on the contributing.md file.
Error:

| non_defect | test errors as of now running tests on the lib will always result in errors this errors come from the module ts file and are a result of how the tests are writen i dont think this is a huge issue but at least one that should be mentioned on the contributing md file error | 0 |
51,728 | 7,725,511,610 | IssuesEvent | 2018-05-24 18:12:25 | smartdevicelink/sdl_server | https://api.github.com/repos/smartdevicelink/sdl_server | closed | Allow More Flexible States for App Approvals | documentation enhancement | ### Enhancement
Allow more options for whether the app is approved or denied permissions in both staging and production modes. Update documentation to reflect new changes. | 1.0 | Allow More Flexible States for App Approvals - ### Enhancement
Allow more options for whether the app is approved or denied permissions in both staging and production modes. Update documentation to reflect new changes. | non_defect | allow more flexible states for app approvals enhancement allow more options for whether the app is approved or denied permissions in both staging and production modes update documentation to reflect new changes | 0 |
31,006 | 6,396,123,000 | IssuesEvent | 2017-08-04 14:49:36 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | opened | __dx() throws an IntlException for strings containing {{...}} | Defect i18n | This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.4
* Platform and Target: Windows 10, Nginx
### What you did
```php
__dx('wa', 'time_span', '{{startDate}} ... {{endDate}}');
```
### What happened
> IntlException: Constructor failed
It works just fine when I'm using `__d(`)` or `__dx()` without the curly brackets.
### What you expected to happen
Not getting an exception. | 1.0 | __dx() throws an IntlException for strings containing {{...}} - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.4
* Platform and Target: Windows 10, Nginx
### What you did
```php
__dx('wa', 'time_span', '{{startDate}} ... {{endDate}}');
```
### What happened
> IntlException: Constructor failed
It works just fine when I'm using `__d(`)` or `__dx()` without the curly brackets.
### What you expected to happen
Not getting an exception. | defect | dx throws an intlexception for strings containing this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target windows nginx what you did php dx wa time span startdate enddate what happened intlexception constructor failed it works just fine when i m using d or dx without the curly brackets what you expected to happen not getting an exception | 1 |
24,149 | 4,059,433,763 | IssuesEvent | 2016-05-25 09:33:18 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | opened | EntryListenerOnReconnectTest.testListenersSmartRoutingMultipleServer | Team: Core Type: Test-Failure | ```
org.junit.runners.model.TestTimedOutException: test timed out after 120000 milliseconds
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:315)
at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:156)
at com.hazelcast.client.spi.ClientProxy.invokeOnPartition(ClientProxy.java:153)
at com.hazelcast.client.spi.ClientProxy.invoke(ClientProxy.java:147)
at com.hazelcast.client.proxy.ClientMapProxy.putInternal(ClientMapProxy.java:452)
at com.hazelcast.client.proxy.ClientMapProxy.put(ClientMapProxy.java:446)
at com.hazelcast.client.proxy.ClientMapProxy.put(ClientMapProxy.java:247)
at com.hazelcast.client.listeners.EntryListenerOnReconnectTest.produceEvent(EntryListenerOnReconnectTest.java:49)
```
https://hazelcast-l337.ci.cloudbees.com/view/Official%20Builds/job/Hazelcast-3.x-IbmJDK1.8/com.hazelcast$hazelcast-client/228/testReport/junit/com.hazelcast.client.listeners/EntryListenerOnReconnectTest/testListenersSmartRoutingMultipleServer/ | 1.0 | EntryListenerOnReconnectTest.testListenersSmartRoutingMultipleServer - ```
org.junit.runners.model.TestTimedOutException: test timed out after 120000 milliseconds
at sun.misc.Unsafe.park(Native Method)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:315)
at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:156)
at com.hazelcast.client.spi.ClientProxy.invokeOnPartition(ClientProxy.java:153)
at com.hazelcast.client.spi.ClientProxy.invoke(ClientProxy.java:147)
at com.hazelcast.client.proxy.ClientMapProxy.putInternal(ClientMapProxy.java:452)
at com.hazelcast.client.proxy.ClientMapProxy.put(ClientMapProxy.java:446)
at com.hazelcast.client.proxy.ClientMapProxy.put(ClientMapProxy.java:247)
at com.hazelcast.client.listeners.EntryListenerOnReconnectTest.produceEvent(EntryListenerOnReconnectTest.java:49)
```
https://hazelcast-l337.ci.cloudbees.com/view/Official%20Builds/job/Hazelcast-3.x-IbmJDK1.8/com.hazelcast$hazelcast-client/228/testReport/junit/com.hazelcast.client.listeners/EntryListenerOnReconnectTest/testListenersSmartRoutingMultipleServer/ | non_defect | entrylisteneronreconnecttest testlistenerssmartroutingmultipleserver org junit runners model testtimedoutexception test timed out after milliseconds at sun misc unsafe park native method at java util concurrent locks locksupport park locksupport java at com hazelcast spi impl abstractinvocationfuture get abstractinvocationfuture java at com hazelcast client spi clientproxy invokeonpartition clientproxy java at com hazelcast client spi clientproxy invoke clientproxy java at com hazelcast client proxy clientmapproxy putinternal clientmapproxy java at com hazelcast client proxy clientmapproxy put clientmapproxy java at com hazelcast client proxy clientmapproxy put clientmapproxy java at com hazelcast client listeners entrylisteneronreconnecttest produceevent entrylisteneronreconnecttest java | 0 |
28,613 | 5,310,725,772 | IssuesEvent | 2017-02-12 22:23:37 | networkx/networkx | https://api.github.com/repos/networkx/networkx | closed | eulerian_circuit() is not compatible with MultiGraph | Defect | When calculating the Eulerian circuit on a MultiGraph (which is necessary when you convert a non-Eulerian graph to an Eulerian graph), the edges are returned as a 2-tuple, even if there is a key. This means all the data related to the edges is lost | 1.0 | eulerian_circuit() is not compatible with MultiGraph - When calculating the Eulerian circuit on a MultiGraph (which is necessary when you convert a non-Eulerian graph to an Eulerian graph), the edges are returned as a 2-tuple, even if there is a key. This means all the data related to the edges is lost | defect | eulerian circuit is not compatible with multigraph when calculating the eulerian circuit on a multigraph which is necessary when you convert a non eulerian graph to an eulerian graph the edges are returned as a tuple even if there is a key this means all the data related to the edges is lost | 1 |
66,682 | 20,514,660,307 | IssuesEvent | 2022-03-01 10:27:53 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Various rooms seem to repeat chains of state events and read receipts | T-Defect A-Timeline A-Read-Receipts | ### Steps to reproduce
1. Look at some rooms, perhaps scroll up a bit (I am not quite sure what it takes)
### Outcome
#### What did you expect?
State events and read receipts should only appear in one place in the timeline. Event level "..." menus should open when clicked.
#### What happened instead?
As seen below, some state events seem to repeat in a loop, as well as read receipts down the side. Also, the "..." menu on the state events does not want to open.
<img width="577" alt="image" src="https://user-images.githubusercontent.com/279572/156151985-50ad1f52-13cb-4d94-aa6e-c3619c6247ba.png">
### Operating system
macOS
### Browser information
Firefox 99
### URL for webapp
develop.element.io
### Application version
Element version: de9e65de2b18-react-36adba101caf-js-54e815085f4d, Olm version: 3.2.8
### Homeserver
matrix.org
### Will you send logs?
Yes | 1.0 | Various rooms seem to repeat chains of state events and read receipts - ### Steps to reproduce
1. Look at some rooms, perhaps scroll up a bit (I am not quite sure what it takes)
### Outcome
#### What did you expect?
State events and read receipts should only appear in one place in the timeline. Event level "..." menus should open when clicked.
#### What happened instead?
As seen below, some state events seem to repeat in a loop, as well as read receipts down the side. Also, the "..." menu on the state events does not want to open.
<img width="577" alt="image" src="https://user-images.githubusercontent.com/279572/156151985-50ad1f52-13cb-4d94-aa6e-c3619c6247ba.png">
### Operating system
macOS
### Browser information
Firefox 99
### URL for webapp
develop.element.io
### Application version
Element version: de9e65de2b18-react-36adba101caf-js-54e815085f4d, Olm version: 3.2.8
### Homeserver
matrix.org
### Will you send logs?
Yes | defect | various rooms seem to repeat chains of state events and read receipts steps to reproduce look at some rooms perhaps scroll up a bit i am not quite sure what it takes outcome what did you expect state events and read receipts should only appear in one place in the timeline event level menus should open when clicked what happened instead as seen below some state events seem to repeat in a loop as well as read receipts down the side also the menu on the state events does not want to open img width alt image src operating system macos browser information firefox url for webapp develop element io application version element version react js olm version homeserver matrix org will you send logs yes | 1 |
5,794 | 2,610,215,569 | IssuesEvent | 2015-02-26 19:08:46 | chrsmith/somefinders | https://api.github.com/repos/chrsmith/somefinders | opened | решебник по русскому языку 6 класс | auto-migrated Priority-Medium Type-Defect | ```
'''Антуан Блинов'''
Привет всем не подскажите где можно найти
.решебник по русскому языку 6 класс. как то
выкладывали уже
'''Авдей Петров'''
Качай тут http://bit.ly/16CYUOP
'''Гайдар Горшков'''
Спасибо вроде то но просит телефон вводить
'''Арий Павлов'''
Неа все ок у меня ничего не списало
'''Адам Кошелев'''
Не это не влияет на баланс
Информация о файле: решебник по русскому
языку 6 класс
Загружен: В этом месяце
Скачан раз: 1002
Рейтинг: 508
Средняя скорость скачивания: 947
Похожих файлов: 12
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 4:53 | 1.0 | решебник по русскому языку 6 класс - ```
'''Антуан Блинов'''
Привет всем не подскажите где можно найти
.решебник по русскому языку 6 класс. как то
выкладывали уже
'''Авдей Петров'''
Качай тут http://bit.ly/16CYUOP
'''Гайдар Горшков'''
Спасибо вроде то но просит телефон вводить
'''Арий Павлов'''
Неа все ок у меня ничего не списало
'''Адам Кошелев'''
Не это не влияет на баланс
Информация о файле: решебник по русскому
языку 6 класс
Загружен: В этом месяце
Скачан раз: 1002
Рейтинг: 508
Средняя скорость скачивания: 947
Похожих файлов: 12
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 4:53 | defect | решебник по русскому языку класс антуан блинов привет всем не подскажите где можно найти решебник по русскому языку класс как то выкладывали уже авдей петров качай тут гайдар горшков спасибо вроде то но просит телефон вводить арий павлов неа все ок у меня ничего не списало адам кошелев не это не влияет на баланс информация о файле решебник по русскому языку класс загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at | 1 |
65,416 | 19,488,000,663 | IssuesEvent | 2021-12-26 19:29:08 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | Test failures on macOS Big Sur (Version 11.4) with Apple M1 chip | defect scipy.special | <!--
Thank you for taking the time to file a bug report.
Please fill in the fields below, deleting the sections that
don't apply to your issue. You can view the final output
by clicking the preview button above.
Note: This is a comment, and won't appear in the output.
-->
My issue is about ...
#### Reproducing code example:
<!--
If you place your code between the triple backticks below,
it will be rendered as a code block.
-->
```python
import scipy
scipy.test()
```
#### Error message:
<!-- If any, paste the *full* error message inside a code block
as above (starting from line Traceback)
-->
```
=================================== FAILURES ===================================
_________________________ TestSmirnovi.test_round_trip _________________________
scipy/special/tests/test_kolmogorov.py:169: in test_round_trip
FuncData(_sm_smi, dataset, (0, 1), 2, rtol=_rtol).check(dtypes=[int, float, float])
E AssertionError:
E Max |adiff|: 0.484375
E Max |rdiff|: 1337.21
E Bad results (2 out of 7) for the following points (in output 0):
E 3 0.125 => 0.609375 != 0.125 (rdiff 3.875)
E 10 0.0001 => 0.13382053040755335 != 0.0001 (rdiff 1337.2053040755336)
_sm_smi = <function TestSmirnovi.test_round_trip.<locals>._sm_smi at 0x17ab99d30>
_smc_smci = <function TestSmirnovi.test_round_trip.<locals>._smc_smci at 0x17ab99820>
dataset = array([[1.00e+00, 4.00e-01, 4.00e-01],
[1.00e+00, 6.00e-01, 6.00e-01],
[2.00e+00, 8.75e-01, 8.75e-01],
...,
[3.00e+00, 1.25e-01, 1.25e-01],
[1.00e+01, 9.99e-01, 9.99e-01],
[1.00e+01, 1.00e-04, 1.00e-04]])
self = <scipy.special.tests.test_kolmogorov.TestSmirnovi object at 0x17ab86b20>
______________________ TestSmirnovi.test_x_equals_0point5 ______________________
scipy/special/tests/test_kolmogorov.py:186: in test_x_equals_0point5
FuncData(smirnovi, dataset, (0, 1), 2, rtol=_rtol).check(dtypes=[int, float, float])
E AssertionError:
E Max |adiff|: 0.13456
E Max |rdiff|: 0.708588
E Bad results (7 out of 11) for the following points (in output 0):
E 4 0.5 => 0.125 != 0.255520481121 (rdiff 0.5108024239324789)
E 5 0.5 => 0.1 != 0.234559536069 (rdiff 0.5736690067012106)
E 6 0.5 => 0.08333333333333333 != 0.21715965898 (rdiff 0.6162577629530716)
E 7 0.5 => 0.07142857142857142 != 0.202722580034 (rdiff 0.647653599236002)
E 8 0.5 => 0.0625 != 0.190621765256 (rdiff 0.67212558379121)
E 9 0.5 => 0.05555555555555555 != 0.180363501362 (rdiff 0.6919800561863548)
E 10 0.5 => 0.05 != 0.17157867006 (rdiff 0.7085884860716352)
dataset = array([[ 1. , 0.5 , 0.5 ],
[ 2. , 0.5 , 0.3660254 ],
[ 2. , ... , 0.19062177],
[ 9. , 0.5 , 0.1803635 ],
[10. , 0.5 , 0.17157867]])
self = <scipy.special.tests.test_kolmogorov.TestSmirnovi object at 0x17ac6fe50>
=========================== short test summary info ============================
FAILED scipy/special/tests/test_kolmogorov.py::TestSmirnovi::test_round_trip
FAILED scipy/special/tests/test_kolmogorov.py::TestSmirnovi::test_x_equals_0point5
```
#### Scipy/Numpy/Python version information:
```
1.8.0.dev0+1551.d92e048 1.21.1 sys.version_info(major=3, minor=9, micro=6, releaselevel='final', serial=0)
```
I am using macOS Big Sur (Version 11.4) with Apple M1 chip. Though the conda environment is having Intel software and hence using Rosetta to run on the hardware. | 1.0 | Test failures on macOS Big Sur (Version 11.4) with Apple M1 chip - <!--
Thank you for taking the time to file a bug report.
Please fill in the fields below, deleting the sections that
don't apply to your issue. You can view the final output
by clicking the preview button above.
Note: This is a comment, and won't appear in the output.
-->
My issue is about ...
#### Reproducing code example:
<!--
If you place your code between the triple backticks below,
it will be rendered as a code block.
-->
```python
import scipy
scipy.test()
```
#### Error message:
<!-- If any, paste the *full* error message inside a code block
as above (starting from line Traceback)
-->
```
=================================== FAILURES ===================================
_________________________ TestSmirnovi.test_round_trip _________________________
scipy/special/tests/test_kolmogorov.py:169: in test_round_trip
FuncData(_sm_smi, dataset, (0, 1), 2, rtol=_rtol).check(dtypes=[int, float, float])
E AssertionError:
E Max |adiff|: 0.484375
E Max |rdiff|: 1337.21
E Bad results (2 out of 7) for the following points (in output 0):
E 3 0.125 => 0.609375 != 0.125 (rdiff 3.875)
E 10 0.0001 => 0.13382053040755335 != 0.0001 (rdiff 1337.2053040755336)
_sm_smi = <function TestSmirnovi.test_round_trip.<locals>._sm_smi at 0x17ab99d30>
_smc_smci = <function TestSmirnovi.test_round_trip.<locals>._smc_smci at 0x17ab99820>
dataset = array([[1.00e+00, 4.00e-01, 4.00e-01],
[1.00e+00, 6.00e-01, 6.00e-01],
[2.00e+00, 8.75e-01, 8.75e-01],
...,
[3.00e+00, 1.25e-01, 1.25e-01],
[1.00e+01, 9.99e-01, 9.99e-01],
[1.00e+01, 1.00e-04, 1.00e-04]])
self = <scipy.special.tests.test_kolmogorov.TestSmirnovi object at 0x17ab86b20>
______________________ TestSmirnovi.test_x_equals_0point5 ______________________
scipy/special/tests/test_kolmogorov.py:186: in test_x_equals_0point5
FuncData(smirnovi, dataset, (0, 1), 2, rtol=_rtol).check(dtypes=[int, float, float])
E AssertionError:
E Max |adiff|: 0.13456
E Max |rdiff|: 0.708588
E Bad results (7 out of 11) for the following points (in output 0):
E 4 0.5 => 0.125 != 0.255520481121 (rdiff 0.5108024239324789)
E 5 0.5 => 0.1 != 0.234559536069 (rdiff 0.5736690067012106)
E 6 0.5 => 0.08333333333333333 != 0.21715965898 (rdiff 0.6162577629530716)
E 7 0.5 => 0.07142857142857142 != 0.202722580034 (rdiff 0.647653599236002)
E 8 0.5 => 0.0625 != 0.190621765256 (rdiff 0.67212558379121)
E 9 0.5 => 0.05555555555555555 != 0.180363501362 (rdiff 0.6919800561863548)
E 10 0.5 => 0.05 != 0.17157867006 (rdiff 0.7085884860716352)
dataset = array([[ 1. , 0.5 , 0.5 ],
[ 2. , 0.5 , 0.3660254 ],
[ 2. , ... , 0.19062177],
[ 9. , 0.5 , 0.1803635 ],
[10. , 0.5 , 0.17157867]])
self = <scipy.special.tests.test_kolmogorov.TestSmirnovi object at 0x17ac6fe50>
=========================== short test summary info ============================
FAILED scipy/special/tests/test_kolmogorov.py::TestSmirnovi::test_round_trip
FAILED scipy/special/tests/test_kolmogorov.py::TestSmirnovi::test_x_equals_0point5
```
#### Scipy/Numpy/Python version information:
```
1.8.0.dev0+1551.d92e048 1.21.1 sys.version_info(major=3, minor=9, micro=6, releaselevel='final', serial=0)
```
I am using macOS Big Sur (Version 11.4) with Apple M1 chip. Though the conda environment is having Intel software and hence using Rosetta to run on the hardware. | defect | test failures on macos big sur version with apple chip thank you for taking the time to file a bug report please fill in the fields below deleting the sections that don t apply to your issue you can view the final output by clicking the preview button above note this is a comment and won t appear in the output my issue is about reproducing code example if you place your code between the triple backticks below it will be rendered as a code block python import scipy scipy test error message if any paste the full error message inside a code block as above starting from line traceback failures testsmirnovi test round trip scipy special tests test kolmogorov py in test round trip funcdata sm smi dataset rtol rtol check dtypes e assertionerror e max adiff e max rdiff e bad results out of for the following points in output e rdiff e rdiff sm smi sm smi at smc smci smc smci at dataset array self testsmirnovi test x equals scipy special tests test kolmogorov py in test x equals funcdata smirnovi dataset rtol rtol check dtypes e assertionerror e max adiff e max rdiff e bad results out of for the following points in output e rdiff e rdiff e rdiff e rdiff e rdiff e rdiff e rdiff dataset array self short test summary info failed scipy special tests test kolmogorov py testsmirnovi test round trip failed scipy special tests test kolmogorov py testsmirnovi test x equals scipy numpy python version information sys version info major minor micro releaselevel final serial i am using macos big sur version with apple chip though the conda environment is having intel software and hence using rosetta to run on the hardware | 1 |
181,546 | 21,664,385,954 | IssuesEvent | 2022-05-07 01:11:28 | rgordon95/ecommerce-react-redux-saga-demo | https://api.github.com/repos/rgordon95/ecommerce-react-redux-saga-demo | opened | CVE-2022-29167 (High) detected in hawk-3.1.3.tgz | security vulnerability | ## CVE-2022-29167 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hawk-3.1.3.tgz</b></p></summary>
<p>HTTP Hawk Authentication Scheme</p>
<p>Library home page: <a href="https://registry.npmjs.org/hawk/-/hawk-3.1.3.tgz">https://registry.npmjs.org/hawk/-/hawk-3.1.3.tgz</a></p>
<p>Path to dependency file: /ecommerce-react-redux-saga-demo/redux-saga-cart/package.json</p>
<p>Path to vulnerable library: /redux-saga-cart/node_modules/hawk/package.json</p>
<p>
Dependency Hierarchy:
- webpack-1.15.0.tgz (Root Library)
- watchpack-0.2.9.tgz
- chokidar-1.7.0.tgz
- fsevents-1.1.1.tgz
- node-pre-gyp-0.6.33.tgz
- request-2.79.0.tgz
- :x: **hawk-3.1.3.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Hawk is an HTTP authentication scheme providing mechanisms for making authenticated HTTP requests with partial cryptographic verification of the request and response, covering the HTTP method, request URI, host, and optionally the request payload. Hawk used a regular expression to parse `Host` HTTP header (`Hawk.utils.parseHost()`), which was subject to regular expression DoS attack - meaning each added character in the attacker's input increases the computation time exponentially. `parseHost()` was patched in `9.0.1` to use built-in `URL` class to parse hostname instead. `Hawk.authenticate()` accepts `options` argument. If that contains `host` and `port`, those would be used instead of a call to `utils.parseHost()`.
<p>Publish Date: 2022-05-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29167>CVE-2022-29167</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/mozilla/hawk/security/advisories/GHSA-44pw-h2cw-w3vq">https://github.com/mozilla/hawk/security/advisories/GHSA-44pw-h2cw-w3vq</a></p>
<p>Release Date: 2022-05-05</p>
<p>Fix Resolution: hawk - 9.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-29167 (High) detected in hawk-3.1.3.tgz - ## CVE-2022-29167 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hawk-3.1.3.tgz</b></p></summary>
<p>HTTP Hawk Authentication Scheme</p>
<p>Library home page: <a href="https://registry.npmjs.org/hawk/-/hawk-3.1.3.tgz">https://registry.npmjs.org/hawk/-/hawk-3.1.3.tgz</a></p>
<p>Path to dependency file: /ecommerce-react-redux-saga-demo/redux-saga-cart/package.json</p>
<p>Path to vulnerable library: /redux-saga-cart/node_modules/hawk/package.json</p>
<p>
Dependency Hierarchy:
- webpack-1.15.0.tgz (Root Library)
- watchpack-0.2.9.tgz
- chokidar-1.7.0.tgz
- fsevents-1.1.1.tgz
- node-pre-gyp-0.6.33.tgz
- request-2.79.0.tgz
- :x: **hawk-3.1.3.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Hawk is an HTTP authentication scheme providing mechanisms for making authenticated HTTP requests with partial cryptographic verification of the request and response, covering the HTTP method, request URI, host, and optionally the request payload. Hawk used a regular expression to parse `Host` HTTP header (`Hawk.utils.parseHost()`), which was subject to regular expression DoS attack - meaning each added character in the attacker's input increases the computation time exponentially. `parseHost()` was patched in `9.0.1` to use built-in `URL` class to parse hostname instead. `Hawk.authenticate()` accepts `options` argument. If that contains `host` and `port`, those would be used instead of a call to `utils.parseHost()`.
<p>Publish Date: 2022-05-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29167>CVE-2022-29167</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/mozilla/hawk/security/advisories/GHSA-44pw-h2cw-w3vq">https://github.com/mozilla/hawk/security/advisories/GHSA-44pw-h2cw-w3vq</a></p>
<p>Release Date: 2022-05-05</p>
<p>Fix Resolution: hawk - 9.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in hawk tgz cve high severity vulnerability vulnerable library hawk tgz http hawk authentication scheme library home page a href path to dependency file ecommerce react redux saga demo redux saga cart package json path to vulnerable library redux saga cart node modules hawk package json dependency hierarchy webpack tgz root library watchpack tgz chokidar tgz fsevents tgz node pre gyp tgz request tgz x hawk tgz vulnerable library vulnerability details hawk is an http authentication scheme providing mechanisms for making authenticated http requests with partial cryptographic verification of the request and response covering the http method request uri host and optionally the request payload hawk used a regular expression to parse host http header hawk utils parsehost which was subject to regular expression dos attack meaning each added character in the attacker s input increases the computation time exponentially parsehost was patched in to use built in url class to parse hostname instead hawk authenticate accepts options argument if that contains host and port those would be used instead of a call to utils parsehost publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hawk step up your open source security game with whitesource | 0 |
61,229 | 14,619,108,739 | IssuesEvent | 2020-12-22 17:17:17 | kenferrara/calcite | https://api.github.com/repos/kenferrara/calcite | opened | CVE-2012-5783 (Medium) detected in commons-httpclient-3.1.jar | security vulnerability | ## CVE-2012-5783 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-httpclient-3.1.jar</b></p></summary>
<p>The HttpClient component supports the client-side of RFC 1945 (HTTP/1.0) and RFC 2616 (HTTP/1.1) , several related specifications (RFC 2109 (Cookies) , RFC 2617 (HTTP Authentication) , etc.), and provides a framework by which new request types (methods) or HTTP extensions can be created easily.</p>
<p>Path to dependency file: calcite/pig/build.gradle.kts</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-httpclient/commons-httpclient/3.1/964cd74171f427720480efdec40a7c7f6e58426a/commons-httpclient-3.1.jar</p>
<p>
Dependency Hierarchy:
- pig-0.16.0.jar (Root Library)
- :x: **commons-httpclient-3.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/calcite/commit/5e2bb91de24b843d35df04b29fe7158fad414ee0">5e2bb91de24b843d35df04b29fe7158fad414ee0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons HttpClient 3.x, as used in Amazon Flexible Payments Service (FPS) merchant Java SDK and other products, does not verify that the server hostname matches a domain name in the subject's Common Name (CN) or subjectAltName field of the X.509 certificate, which allows man-in-the-middle attackers to spoof SSL servers via an arbitrary valid certificate.
<p>Publish Date: 2012-11-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-5783>CVE-2012-5783</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://xforce.iss.net/xforce/xfdb/79984">http://xforce.iss.net/xforce/xfdb/79984</a></p>
<p>Release Date: 2017-12-31</p>
<p>Fix Resolution: Apply the appropriate patch for your system. See References.</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-httpclient","packageName":"commons-httpclient","packageVersion":"3.1","isTransitiveDependency":true,"dependencyTree":"org.apache.pig:pig:0.16.0;commons-httpclient:commons-httpclient:3.1","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2012-5783","vulnerabilityDetails":"Apache Commons HttpClient 3.x, as used in Amazon Flexible Payments Service (FPS) merchant Java SDK and other products, does not verify that the server hostname matches a domain name in the subject\u0027s Common Name (CN) or subjectAltName field of the X.509 certificate, which allows man-in-the-middle attackers to spoof SSL servers via an arbitrary valid certificate.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-5783","cvss2Severity":"medium","cvss2Score":"5.8","extraData":{}}</REMEDIATE> --> | True | CVE-2012-5783 (Medium) detected in commons-httpclient-3.1.jar - ## CVE-2012-5783 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-httpclient-3.1.jar</b></p></summary>
<p>The HttpClient component supports the client-side of RFC 1945 (HTTP/1.0) and RFC 2616 (HTTP/1.1) , several related specifications (RFC 2109 (Cookies) , RFC 2617 (HTTP Authentication) , etc.), and provides a framework by which new request types (methods) or HTTP extensions can be created easily.</p>
<p>Path to dependency file: calcite/pig/build.gradle.kts</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-httpclient/commons-httpclient/3.1/964cd74171f427720480efdec40a7c7f6e58426a/commons-httpclient-3.1.jar</p>
<p>
Dependency Hierarchy:
- pig-0.16.0.jar (Root Library)
- :x: **commons-httpclient-3.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kenferrara/calcite/commit/5e2bb91de24b843d35df04b29fe7158fad414ee0">5e2bb91de24b843d35df04b29fe7158fad414ee0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons HttpClient 3.x, as used in Amazon Flexible Payments Service (FPS) merchant Java SDK and other products, does not verify that the server hostname matches a domain name in the subject's Common Name (CN) or subjectAltName field of the X.509 certificate, which allows man-in-the-middle attackers to spoof SSL servers via an arbitrary valid certificate.
<p>Publish Date: 2012-11-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-5783>CVE-2012-5783</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://xforce.iss.net/xforce/xfdb/79984">http://xforce.iss.net/xforce/xfdb/79984</a></p>
<p>Release Date: 2017-12-31</p>
<p>Fix Resolution: Apply the appropriate patch for your system. See References.</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-httpclient","packageName":"commons-httpclient","packageVersion":"3.1","isTransitiveDependency":true,"dependencyTree":"org.apache.pig:pig:0.16.0;commons-httpclient:commons-httpclient:3.1","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2012-5783","vulnerabilityDetails":"Apache Commons HttpClient 3.x, as used in Amazon Flexible Payments Service (FPS) merchant Java SDK and other products, does not verify that the server hostname matches a domain name in the subject\u0027s Common Name (CN) or subjectAltName field of the X.509 certificate, which allows man-in-the-middle attackers to spoof SSL servers via an arbitrary valid certificate.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-5783","cvss2Severity":"medium","cvss2Score":"5.8","extraData":{}}</REMEDIATE> --> | non_defect | cve medium detected in commons httpclient jar cve medium severity vulnerability vulnerable library commons httpclient jar the httpclient component supports the client side of rfc http and rfc http several related specifications rfc cookies rfc http authentication etc and provides a framework by which new request types methods or http extensions can be created easily path to dependency file calcite pig build gradle kts path to vulnerable library home wss scanner gradle caches modules files commons httpclient commons httpclient commons httpclient jar dependency hierarchy pig jar root library x commons httpclient jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons httpclient x as used in amazon flexible payments service fps merchant java sdk and other products does not verify that the server hostname matches a domain name in the subject s common name cn or subjectaltname field of the x certificate which allows man in the middle attackers to spoof ssl servers via an arbitrary valid certificate publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution apply the appropriate patch for your system see references isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails apache commons httpclient x as used in amazon flexible payments service fps merchant java sdk and other products does not verify that the server hostname matches a domain name in the subject common name cn or subjectaltname field of the x certificate which allows man in the middle attackers to spoof ssl servers via an arbitrary valid certificate vulnerabilityurl | 0 |
79,707 | 28,498,450,255 | IssuesEvent | 2023-04-18 15:36:31 | vector-im/element-call | https://api.github.com/repos/vector-im/element-call | opened | Join with muted video and muted audio | T-Defect | ### Steps to reproduce
1. Join a call with User A no muted audio and no muted video
2. Join same call with User B muted audio and muted video
x. I bug not happen, than switch and join with User A muted and User B not muted. Than Bug will happen.

### Outcome
#### What did you expect?
User B can hear and See User A and User A can not see and hear User B.
#### What happened instead?
User A and User B can not see each other.
### Operating system
macOS
### Browser information
Chrome lates version
### URL for webapp
call.element.io
### Will you send logs?
No | 1.0 | Join with muted video and muted audio - ### Steps to reproduce
1. Join a call with User A no muted audio and no muted video
2. Join same call with User B muted audio and muted video
x. I bug not happen, than switch and join with User A muted and User B not muted. Than Bug will happen.

### Outcome
#### What did you expect?
User B can hear and See User A and User A can not see and hear User B.
#### What happened instead?
User A and User B can not see each other.
### Operating system
macOS
### Browser information
Chrome lates version
### URL for webapp
call.element.io
### Will you send logs?
No | defect | join with muted video and muted audio steps to reproduce join a call with user a no muted audio and no muted video join same call with user b muted audio and muted video x i bug not happen than switch and join with user a muted and user b not muted than bug will happen outcome what did you expect user b can hear and see user a and user a can not see and hear user b what happened instead user a and user b can not see each other operating system macos browser information chrome lates version url for webapp call element io will you send logs no | 1 |
21,861 | 3,573,522,106 | IssuesEvent | 2016-01-27 07:02:03 | hunnee/jsontoken | https://api.github.com/repos/hunnee/jsontoken | closed | Cannot check out the project in eclipse | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Import project to Eclipse
2. Specify the url of repository: http://jsontoken.googlecode.com/svn/
3. Start check out
What is the expected output? What do you see instead?
the project is expected to be imported in Eclipse.Instead the check out process
get stuck in the file "FakeClock.java"
What version of the product are you using? On what operating system?
Windows 7, eclipse indigo service release 2
Please provide any additional information below.
```
Original issue reported on code.google.com by `panipsi...@gmail.com` on 13 Jun 2013 at 12:00 | 1.0 | Cannot check out the project in eclipse - ```
What steps will reproduce the problem?
1. Import project to Eclipse
2. Specify the url of repository: http://jsontoken.googlecode.com/svn/
3. Start check out
What is the expected output? What do you see instead?
the project is expected to be imported in Eclipse.Instead the check out process
get stuck in the file "FakeClock.java"
What version of the product are you using? On what operating system?
Windows 7, eclipse indigo service release 2
Please provide any additional information below.
```
Original issue reported on code.google.com by `panipsi...@gmail.com` on 13 Jun 2013 at 12:00 | defect | cannot check out the project in eclipse what steps will reproduce the problem import project to eclipse specify the url of repository start check out what is the expected output what do you see instead the project is expected to be imported in eclipse instead the check out process get stuck in the file fakeclock java what version of the product are you using on what operating system windows eclipse indigo service release please provide any additional information below original issue reported on code google com by panipsi gmail com on jun at | 1 |
277,843 | 8,633,367,834 | IssuesEvent | 2018-11-22 13:40:19 | geosolutions-it/pyfulcrum | https://api.github.com/repos/geosolutions-it/pyfulcrum | closed | PyBackup entrypoint | Priority: Medium Task review | Develop a specific API to send Fulcrum's Records to the PyBackup module. An asynchronous mechanims needs to be included at this stage in order to grant high availability of the endpoint service, and reduce (throttle) downstream requests toward Fulcrum APIs which are needed to access and validate webhooks payloads.
Needed functionalities:
- [x] Module API
~~- [ ] Asynchronous mechanism to manage the incoming records and grant high availability~~ (commented out by https://github.com/geosolutions-it/pyfulcrum/issues/3#issuecomment-431051535) | 1.0 | PyBackup entrypoint - Develop a specific API to send Fulcrum's Records to the PyBackup module. An asynchronous mechanims needs to be included at this stage in order to grant high availability of the endpoint service, and reduce (throttle) downstream requests toward Fulcrum APIs which are needed to access and validate webhooks payloads.
Needed functionalities:
- [x] Module API
~~- [ ] Asynchronous mechanism to manage the incoming records and grant high availability~~ (commented out by https://github.com/geosolutions-it/pyfulcrum/issues/3#issuecomment-431051535) | non_defect | pybackup entrypoint develop a specific api to send fulcrum s records to the pybackup module an asynchronous mechanims needs to be included at this stage in order to grant high availability of the endpoint service and reduce throttle downstream requests toward fulcrum apis which are needed to access and validate webhooks payloads needed functionalities module api asynchronous mechanism to manage the incoming records and grant high availability commented out by | 0 |
64,495 | 18,717,058,353 | IssuesEvent | 2021-11-03 07:05:36 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | closed | [🐛 Bug]: AttributeError: 'dict' object has no attribute 'send_keys' | I-defect needs-triaging | ### What happened?
I just start the browser and find en element using Python:
```python
driver = webdriver.Chrome()
driver.implicitly_wait(10)
driver.get('https://account.xiaomi.com/fe/service/login/password')
el = driver.find_element(By.NAME, 'account')
el.send_keys('demo')
```
el should be an webelement object or NoSuchElementException, but I got this:
`AttributeError: 'dict' object has no attribute 'send_keys'`

### How can we reproduce the issue?
```shell
above
```
### Relevant log output
```shell
FAILED [100%]
test_selenum.py:12 (test_selenium)
def test_selenium():
"""测试selenium是否可用"""
driver = webdriver.Chrome()
driver.implicitly_wait(10)
# 访问被测试网址
driver.get('https://account.xiaomi.com/fe/service/login/password')
el = driver.find_element(By.NAME, 'account')
> el.send_keys('aaa')
E AttributeError: 'dict' object has no attribute 'send_keys'
test_selenum.py:20: AttributeError
```
### Operating System
Windows10
### Selenium version
4.0.0
### What are the browser(s) and version(s) where you see this issue?
Chrome90
### What are the browser driver(s) and version(s) where you see this issue?
90
### Are you using Selenium Grid?
_No response_ | 1.0 | [🐛 Bug]: AttributeError: 'dict' object has no attribute 'send_keys' - ### What happened?
I just start the browser and find en element using Python:
```python
driver = webdriver.Chrome()
driver.implicitly_wait(10)
driver.get('https://account.xiaomi.com/fe/service/login/password')
el = driver.find_element(By.NAME, 'account')
el.send_keys('demo')
```
el should be an webelement object or NoSuchElementException, but I got this:
`AttributeError: 'dict' object has no attribute 'send_keys'`

### How can we reproduce the issue?
```shell
above
```
### Relevant log output
```shell
FAILED [100%]
test_selenum.py:12 (test_selenium)
def test_selenium():
"""测试selenium是否可用"""
driver = webdriver.Chrome()
driver.implicitly_wait(10)
# 访问被测试网址
driver.get('https://account.xiaomi.com/fe/service/login/password')
el = driver.find_element(By.NAME, 'account')
> el.send_keys('aaa')
E AttributeError: 'dict' object has no attribute 'send_keys'
test_selenum.py:20: AttributeError
```
### Operating System
Windows10
### Selenium version
4.0.0
### What are the browser(s) and version(s) where you see this issue?
Chrome90
### What are the browser driver(s) and version(s) where you see this issue?
90
### Are you using Selenium Grid?
_No response_ | defect | attributeerror dict object has no attribute send keys what happened i just start the browser and find en element using python python driver webdriver chrome driver implicitly wait driver get el driver find element by name account el send keys demo el should be an webelement object or nosuchelementexception, but i got this attributeerror dict object has no attribute send keys how can we reproduce the issue shell above relevant log output shell failed test selenum py test selenium def test selenium 测试selenium是否可用 driver webdriver chrome driver implicitly wait 访问被测试网址 driver get el driver find element by name account el send keys aaa e attributeerror dict object has no attribute send keys test selenum py attributeerror operating system selenium version what are the browser s and version s where you see this issue what are the browser driver s and version s where you see this issue are you using selenium grid no response | 1 |
97,403 | 11,012,717,866 | IssuesEvent | 2019-12-04 18:53:14 | SUSE/kubecf | https://api.github.com/repos/SUSE/kubecf | opened | Reusing an external database after reinstalling kubecf does not work | Priority: High Type: Bug Type: Documentation | **Describe the bug**
On an attempt to reuse the external database, a kubecf reinstall won't work.
**To Reproduce**
- Deploy kubecf using an external database.
- Run smoke-tests.
- Delete kubecf installation.
- Deploy kubecf again without changing any values.
- Run smoke-tests - it will fail.
**Expected behavior**
Given the proper values, the smoke-tests on the reinstallation should work.
| 1.0 | Reusing an external database after reinstalling kubecf does not work - **Describe the bug**
On an attempt to reuse the external database, a kubecf reinstall won't work.
**To Reproduce**
- Deploy kubecf using an external database.
- Run smoke-tests.
- Delete kubecf installation.
- Deploy kubecf again without changing any values.
- Run smoke-tests - it will fail.
**Expected behavior**
Given the proper values, the smoke-tests on the reinstallation should work.
| non_defect | reusing an external database after reinstalling kubecf does not work describe the bug on an attempt to reuse the external database a kubecf reinstall won t work to reproduce deploy kubecf using an external database run smoke tests delete kubecf installation deploy kubecf again without changing any values run smoke tests it will fail expected behavior given the proper values the smoke tests on the reinstallation should work | 0 |
4,882 | 2,610,159,596 | IssuesEvent | 2015-02-26 18:50:39 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | closed | Text | auto-migrated Priority-Medium Type-Defect | ```
Tech 1 text screen display is missing. Check tech levels for the GC and check
the code and text file
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 31 Jan 2011 at 1:34 | 1.0 | Text - ```
Tech 1 text screen display is missing. Check tech levels for the GC and check
the code and text file
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 31 Jan 2011 at 1:34 | defect | text tech text screen display is missing check tech levels for the gc and check the code and text file original issue reported on code google com by gmail com on jan at | 1 |
22,236 | 15,055,896,462 | IssuesEvent | 2021-02-03 19:25:45 | commitdev/zero | https://api.github.com/repos/commitdev/zero | opened | bug: user-auth namespace already exist | bug infrastructure | when terraform applies, maybe because of the secret creation the user-auth namespace is already created
terraform errors out because it also tries to create it, we need to either change the https://github.com/commitdev/zero-aws-eks-stack/blob/main/templates/kubernetes/terraform/modules/kubernetes/user_auth.tf#L21 to be a `data` instead of `resource` | 1.0 | bug: user-auth namespace already exist - when terraform applies, maybe because of the secret creation the user-auth namespace is already created
terraform errors out because it also tries to create it, we need to either change the https://github.com/commitdev/zero-aws-eks-stack/blob/main/templates/kubernetes/terraform/modules/kubernetes/user_auth.tf#L21 to be a `data` instead of `resource` | non_defect | bug user auth namespace already exist when terraform applies maybe because of the secret creation the user auth namespace is already created terraform errors out because it also tries to create it we need to either change the to be a data instead of resource | 0 |
228,841 | 25,262,809,724 | IssuesEvent | 2022-11-16 00:45:38 | AlexRogalskiy/roadmap | https://api.github.com/repos/AlexRogalskiy/roadmap | closed | CVE-2019-16943 (High) detected in jackson-databind-2.1.3.jar | security vulnerability no-issue-activity | ## CVE-2019-16943 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.1.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: /modules/roadmap-all/pom.xml</p>
<p>Path to vulnerable library: /tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar,/tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar,/tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar</p>
<p>
Dependency Hierarchy:
- logback-jackson-0.1.5.jar (Root Library)
- :x: **jackson-databind-2.1.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/roadmap/commit/fff78511c51c60993515e1063ec6b51834154cd1">fff78511c51c60993515e1063ec6b51834154cd1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the p6spy (3.8.6) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of com.p6spy.engine.spy.P6DataSource mishandling.
<p>Publish Date: 2019-10-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-16943>CVE-2019-16943</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943</a></p>
<p>Release Date: 2019-10-01</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-16943 (High) detected in jackson-databind-2.1.3.jar - ## CVE-2019-16943 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.1.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: /modules/roadmap-all/pom.xml</p>
<p>Path to vulnerable library: /tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar,/tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar,/tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar</p>
<p>
Dependency Hierarchy:
- logback-jackson-0.1.5.jar (Root Library)
- :x: **jackson-databind-2.1.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/roadmap/commit/fff78511c51c60993515e1063ec6b51834154cd1">fff78511c51c60993515e1063ec6b51834154cd1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the p6spy (3.8.6) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of com.p6spy.engine.spy.P6DataSource mishandling.
<p>Publish Date: 2019-10-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-16943>CVE-2019-16943</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943</a></p>
<p>Release Date: 2019-10-01</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to dependency file modules roadmap all pom xml path to vulnerable library tmp ws ua rxrglz downloadresource pjgkyn jackson databind jar tmp ws ua rxrglz downloadresource pjgkyn jackson databind jar tmp ws ua rxrglz downloadresource pjgkyn jackson databind jar dependency hierarchy logback jackson jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the jar in the classpath and an attacker can find an rmi service endpoint to access it is possible to make the service execute a malicious payload this issue exists because of com engine spy mishandling publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with mend | 0 |
337 | 2,532,788,858 | IssuesEvent | 2015-01-23 18:30:11 | google/error-prone | https://api.github.com/repos/google/error-prone | closed | IJ plugin compilation fails with GC overhead limit exceeded | migrated Priority-Medium Type-Defect | _[Original issue](https://code.google.com/p/error-prone/issues/detail?id=250) created by **alexeagle@google.com** on 2014-04-28 at 06:23 PM_
---
<b>What steps will reproduce the problem?</b>
1. Set up an IntelliJ IDEA project with a very large number of sources
2. Use error-prone compiler plugin
Information:Using error-prone compiler to compile java sources
Information:Module "java" was fully rebuilt due to project configuration/dependencies changes
Information:Compilation completed with 1 error and 70 warnings in 1 min 34 sec
Information:1 error
Information:70 warnings
Error:java: java.lang.OutOfMemoryError: GC overhead limit exceeded
| 1.0 | IJ plugin compilation fails with GC overhead limit exceeded - _[Original issue](https://code.google.com/p/error-prone/issues/detail?id=250) created by **alexeagle@google.com** on 2014-04-28 at 06:23 PM_
---
<b>What steps will reproduce the problem?</b>
1. Set up an IntelliJ IDEA project with a very large number of sources
2. Use error-prone compiler plugin
Information:Using error-prone compiler to compile java sources
Information:Module "java" was fully rebuilt due to project configuration/dependencies changes
Information:Compilation completed with 1 error and 70 warnings in 1 min 34 sec
Information:1 error
Information:70 warnings
Error:java: java.lang.OutOfMemoryError: GC overhead limit exceeded
| defect | ij plugin compilation fails with gc overhead limit exceeded created by alexeagle google com on at pm what steps will reproduce the problem set up an intellij idea project with a very large number of sources use error prone compiler plugin information using error prone compiler to compile java sources information module java was fully rebuilt due to project configuration dependencies changes information compilation completed with error and warnings in min sec information error information warnings error java java lang outofmemoryerror gc overhead limit exceeded | 1 |
40,865 | 10,198,144,353 | IssuesEvent | 2019-08-13 03:47:54 | Psionics-ToEE/PsionicFiles | https://api.github.com/repos/Psionics-ToEE/PsionicFiles | opened | Mind blade weapon feat double dipping and dual wield bonus apply incorrectly. | defect low weight | Due to the workaround involving using the sickle type as the proto weapon type for mind blades, you can "double dip" on weapon feats like weapon focus to take the sickle versions for extra bonuses to the mind blade, when only the mind blade versions should apply. Detecting those sickle versions and not applying the mind blade versions could be done, but most people wouldn't waste a feat on a measly +1 attack bonus or on sickle since there is no indication of this in game.
Additionally, due to sickle classification, two weapon fighting bonus for light weapon in offhand works correctly still if the offhand is mind blade, however if it happens to be the LS or BS version of the mind blade (shape mind blade), the bonus would still apply when it should not. Depending on implementation of shape mind blade, it may be impossible to get a LS or BS mind blade in the secondary hand anyway. | 1.0 | Mind blade weapon feat double dipping and dual wield bonus apply incorrectly. - Due to the workaround involving using the sickle type as the proto weapon type for mind blades, you can "double dip" on weapon feats like weapon focus to take the sickle versions for extra bonuses to the mind blade, when only the mind blade versions should apply. Detecting those sickle versions and not applying the mind blade versions could be done, but most people wouldn't waste a feat on a measly +1 attack bonus or on sickle since there is no indication of this in game.
Additionally, due to sickle classification, two weapon fighting bonus for light weapon in offhand works correctly still if the offhand is mind blade, however if it happens to be the LS or BS version of the mind blade (shape mind blade), the bonus would still apply when it should not. Depending on implementation of shape mind blade, it may be impossible to get a LS or BS mind blade in the secondary hand anyway. | defect | mind blade weapon feat double dipping and dual wield bonus apply incorrectly due to the workaround involving using the sickle type as the proto weapon type for mind blades you can double dip on weapon feats like weapon focus to take the sickle versions for extra bonuses to the mind blade when only the mind blade versions should apply detecting those sickle versions and not applying the mind blade versions could be done but most people wouldn t waste a feat on a measly attack bonus or on sickle since there is no indication of this in game additionally due to sickle classification two weapon fighting bonus for light weapon in offhand works correctly still if the offhand is mind blade however if it happens to be the ls or bs version of the mind blade shape mind blade the bonus would still apply when it should not depending on implementation of shape mind blade it may be impossible to get a ls or bs mind blade in the secondary hand anyway | 1 |
10,680 | 2,622,180,257 | IssuesEvent | 2015-03-04 00:18:30 | byzhang/leveldb | https://api.github.com/repos/byzhang/leveldb | closed | The C API lacks a binary safe leveldb_get() function. | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
My C program that is using LevelDB for storing binary blobs and I ran into the
following problem with the LevelDB C API:
1)Leveldb_put() allows the programmer to pass a char* and a length for both the
"key" and "val", which allows the programmer to store binary data, which may
contain null bytes.
2)leveldb_get() returns a char* without a size parameter... Digging into
./leveldb/c.cc you can see that it is using a "std:string" to return this
char*... Not only is this not binary safe, but std::string? Having an
efficient slice, like the one provided in the C++ api would be much preferred.
So in short; using the C API its impossible to put() items without ever being
able to get() them back...
The solution is to provide a leveldb_get_binary() which allows the programmer
to pass in a reference to char* and size param. The call convention should be
consistent with the leveldb_put() function.
What is the expected output? What do you see instead?
I expected to obtain an unmolested blob of binary, I got a null terminated
string.
What version of the product are you using? On what operating system?
libleveldb-dev:amd64 0+20120530.gitdd0d562-2
installed with apt, under Ubuntu.
```
Original issue reported on code.google.com by `firealwa...@gmail.com` on 26 Feb 2013 at 9:24 | 1.0 | The C API lacks a binary safe leveldb_get() function. - ```
What steps will reproduce the problem?
My C program that is using LevelDB for storing binary blobs and I ran into the
following problem with the LevelDB C API:
1)Leveldb_put() allows the programmer to pass a char* and a length for both the
"key" and "val", which allows the programmer to store binary data, which may
contain null bytes.
2)leveldb_get() returns a char* without a size parameter... Digging into
./leveldb/c.cc you can see that it is using a "std:string" to return this
char*... Not only is this not binary safe, but std::string? Having an
efficient slice, like the one provided in the C++ api would be much preferred.
So in short; using the C API its impossible to put() items without ever being
able to get() them back...
The solution is to provide a leveldb_get_binary() which allows the programmer
to pass in a reference to char* and size param. The call convention should be
consistent with the leveldb_put() function.
What is the expected output? What do you see instead?
I expected to obtain an unmolested blob of binary, I got a null terminated
string.
What version of the product are you using? On what operating system?
libleveldb-dev:amd64 0+20120530.gitdd0d562-2
installed with apt, under Ubuntu.
```
Original issue reported on code.google.com by `firealwa...@gmail.com` on 26 Feb 2013 at 9:24 | defect | the c api lacks a binary safe leveldb get function what steps will reproduce the problem my c program that is using leveldb for storing binary blobs and i ran into the following problem with the leveldb c api leveldb put allows the programmer to pass a char and a length for both the key and val which allows the programmer to store binary data which may contain null bytes leveldb get returns a char without a size parameter digging into leveldb c cc you can see that it is using a std string to return this char not only is this not binary safe but std string having an efficient slice like the one provided in the c api would be much preferred so in short using the c api its impossible to put items without ever being able to get them back the solution is to provide a leveldb get binary which allows the programmer to pass in a reference to char and size param the call convention should be consistent with the leveldb put function what is the expected output what do you see instead i expected to obtain an unmolested blob of binary i got a null terminated string what version of the product are you using on what operating system libleveldb dev installed with apt under ubuntu original issue reported on code google com by firealwa gmail com on feb at | 1 |
6,665 | 9,682,769,815 | IssuesEvent | 2019-05-23 09:53:38 | goharbor/harbor | https://api.github.com/repos/goharbor/harbor | closed | Allow Robot users to upload Helm Charts | candidate/1.9.0 kind/requirement | **Is your feature request related to a problem? Please describe.**
I've integrated Harbor into my CI workflow, and automatically push Images via the Robot user. I also create Helm charts in CI, but to upload those, I have to have a second user with another set of credentials, etc.
**Describe the solution you'd like**
An option to allow a Robot user to upload Helm Charts.
**Describe the main design/architecture of your solution**
A simple mockup:

| 1.0 | Allow Robot users to upload Helm Charts - **Is your feature request related to a problem? Please describe.**
I've integrated Harbor into my CI workflow, and automatically push Images via the Robot user. I also create Helm charts in CI, but to upload those, I have to have a second user with another set of credentials, etc.
**Describe the solution you'd like**
An option to allow a Robot user to upload Helm Charts.
**Describe the main design/architecture of your solution**
A simple mockup:

| non_defect | allow robot users to upload helm charts is your feature request related to a problem please describe i ve integrated harbor into my ci workflow and automatically push images via the robot user i also create helm charts in ci but to upload those i have to have a second user with another set of credentials etc describe the solution you d like an option to allow a robot user to upload helm charts describe the main design architecture of your solution a simple mockup | 0 |
685,252 | 23,449,562,884 | IssuesEvent | 2022-08-16 00:13:32 | NuGet/Home | https://api.github.com/repos/NuGet/Home | closed | [MacOS, Linux] NuGet writes the NuGet.Config in ~/.config folder, but dotnet restore reads from ~/.nuget | Priority:2 Type:DCR Product:NuGet.exe Platform:Xplat Area:Settings Pipeline:In Progress Category:Customer Sprint | When using NuGet to add a private feed, `NuGet Sources Add` writes in `~/.config/NuGet/NuGet.Config` file. But when using `dotnet restore` the `~/.nuget/NuGet/NuGet.Config` file is used to read the `packageSources` and `packageSourceCredentials`.
This issue forces users to manually copy and paste all necessary information instead of having that generated automatically by the NuGet CLI.
DotNetCore repo issue: https://github.com/dotnet/core/issues/453 | 1.0 | [MacOS, Linux] NuGet writes the NuGet.Config in ~/.config folder, but dotnet restore reads from ~/.nuget - When using NuGet to add a private feed, `NuGet Sources Add` writes in `~/.config/NuGet/NuGet.Config` file. But when using `dotnet restore` the `~/.nuget/NuGet/NuGet.Config` file is used to read the `packageSources` and `packageSourceCredentials`.
This issue forces users to manually copy and paste all necessary information instead of having that generated automatically by the NuGet CLI.
DotNetCore repo issue: https://github.com/dotnet/core/issues/453 | non_defect | nuget writes the nuget config in config folder but dotnet restore reads from nuget when using nuget to add a private feed nuget sources add writes in config nuget nuget config file but when using dotnet restore the nuget nuget nuget config file is used to read the packagesources and packagesourcecredentials this issue forces users to manually copy and paste all necessary information instead of having that generated automatically by the nuget cli dotnetcore repo issue | 0 |
28,352 | 13,646,320,128 | IssuesEvent | 2020-09-25 22:50:21 | microsoft/STL | https://api.github.com/repos/microsoft/STL | closed | Update OS minimum version and remove XP runtime detection | performance work in progress | Currently STL has to target Vista as minimum OS version for x86 and x64
Update `filesystem.cpp`
https://github.com/microsoft/STL/blob/87dc1d3383a052216822fab1a37c8fe2794b0cdb/stl/src/filesystem.cpp#L56-L99
Update awint.hpp / winapisupp.cpp
*Note*: the ABI compatibility for XP toolset should stay. The XP toolset should build; it just doesn't need to support Windows XP anymore. | True | Update OS minimum version and remove XP runtime detection - Currently STL has to target Vista as minimum OS version for x86 and x64
Update `filesystem.cpp`
https://github.com/microsoft/STL/blob/87dc1d3383a052216822fab1a37c8fe2794b0cdb/stl/src/filesystem.cpp#L56-L99
Update awint.hpp / winapisupp.cpp
*Note*: the ABI compatibility for XP toolset should stay. The XP toolset should build; it just doesn't need to support Windows XP anymore. | non_defect | update os minimum version and remove xp runtime detection currently stl has to target vista as minimum os version for and update filesystem cpp update awint hpp winapisupp cpp note the abi compatibility for xp toolset should stay the xp toolset should build it just doesn t need to support windows xp anymore | 0 |
6,004 | 2,610,219,278 | IssuesEvent | 2015-02-26 19:09:42 | chrsmith/somefinders | https://api.github.com/repos/chrsmith/somefinders | opened | озвучка avast | auto-migrated Priority-Medium Type-Defect | ```
'''Богдан Овчинников'''
Привет всем не подскажите где можно найти
.озвучка avast. как то выкладывали уже
'''Арам Котов'''
Качай тут http://bit.ly/18TFMsK
'''Алим Лобанов'''
Спасибо вроде то но просит телефон вводить
'''Баян Зыков'''
Не это не влияет на баланс
'''Арсений Родионов'''
Не это не влияет на баланс
Информация о файле: озвучка avast
Загружен: В этом месяце
Скачан раз: 932
Рейтинг: 1045
Средняя скорость скачивания: 1282
Похожих файлов: 39
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 12:37 | 1.0 | озвучка avast - ```
'''Богдан Овчинников'''
Привет всем не подскажите где можно найти
.озвучка avast. как то выкладывали уже
'''Арам Котов'''
Качай тут http://bit.ly/18TFMsK
'''Алим Лобанов'''
Спасибо вроде то но просит телефон вводить
'''Баян Зыков'''
Не это не влияет на баланс
'''Арсений Родионов'''
Не это не влияет на баланс
Информация о файле: озвучка avast
Загружен: В этом месяце
Скачан раз: 932
Рейтинг: 1045
Средняя скорость скачивания: 1282
Похожих файлов: 39
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 12:37 | defect | озвучка avast богдан овчинников привет всем не подскажите где можно найти озвучка avast как то выкладывали уже арам котов качай тут алим лобанов спасибо вроде то но просит телефон вводить баян зыков не это не влияет на баланс арсений родионов не это не влияет на баланс информация о файле озвучка avast загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at | 1 |
149,934 | 23,550,923,210 | IssuesEvent | 2022-08-21 20:22:51 | ParadoxGameConverters/Vic3ToHoI4 | https://api.github.com/repos/ParadoxGameConverters/Vic3ToHoI4 | opened | Convert state categories | enhancement coding design | Vic2 to HoI4 mostly just made them big enough for the existing industry. This should be re-examined, and should handle cases like islands. Maybe population based? | 1.0 | Convert state categories - Vic2 to HoI4 mostly just made them big enough for the existing industry. This should be re-examined, and should handle cases like islands. Maybe population based? | non_defect | convert state categories to mostly just made them big enough for the existing industry this should be re examined and should handle cases like islands maybe population based | 0 |
25,585 | 7,727,653,109 | IssuesEvent | 2018-05-25 04:03:45 | Microsoft/MixedRealityToolkit-Unity | https://api.github.com/repos/Microsoft/MixedRealityToolkit-Unity | reopened | Unable to Deploy from Build Window | Build Tools HoloLens | ## Overview
Unable to deploy to Hololens.
## Expected Behavior
After successfully building a barebones project and building the Appx, the program deploys correctly to the Hololens
## Actual Behavior.
Builds project correctly. Appears to connect to Hololens.
Does not deploy to the Hololens when pressing "Install"
## Steps to reproduce
The correct WiFi IP address is provided and when I press "Connect" under deploy options, I get the correct Hololens device name. When I press "Open Device Portal" - IE opens the Hololens Device Portal correctly without issue.
When I press "Install" - I get a Network Error: Request timeout.
Then "Failed to install [APP].appx on [DEVICE].
See screenshot.

## Unity Editor Version
2017.3.1f1
## Mixed Reality Toolkit Release Version
2017.2.1.2
## Hololens OS Version
10.0.14393.2068
## Windows OS Version (Development PC)
10.0.16299 Build 16299
| 1.0 | Unable to Deploy from Build Window - ## Overview
Unable to deploy to Hololens.
## Expected Behavior
After successfully building a barebones project and building the Appx, the program deploys correctly to the Hololens
## Actual Behavior.
Builds project correctly. Appears to connect to Hololens.
Does not deploy to the Hololens when pressing "Install"
## Steps to reproduce
The correct WiFi IP address is provided and when I press "Connect" under deploy options, I get the correct Hololens device name. When I press "Open Device Portal" - IE opens the Hololens Device Portal correctly without issue.
When I press "Install" - I get a Network Error: Request timeout.
Then "Failed to install [APP].appx on [DEVICE].
See screenshot.

## Unity Editor Version
2017.3.1f1
## Mixed Reality Toolkit Release Version
2017.2.1.2
## Hololens OS Version
10.0.14393.2068
## Windows OS Version (Development PC)
10.0.16299 Build 16299
| non_defect | unable to deploy from build window overview unable to deploy to hololens expected behavior after successfully building a barebones project and building the appx the program deploys correctly to the hololens actual behavior builds project correctly appears to connect to hololens does not deploy to the hololens when pressing install steps to reproduce the correct wifi ip address is provided and when i press connect under deploy options i get the correct hololens device name when i press open device portal ie opens the hololens device portal correctly without issue when i press install i get a network error request timeout then failed to install appx on see screenshot unity editor version mixed reality toolkit release version hololens os version windows os version development pc build | 0 |
57,962 | 16,194,482,020 | IssuesEvent | 2021-05-04 13:03:24 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | BigQuery cannot reference exprsessions from SELECT DISTINCT in ORDER BY | C: DB: BigQuery C: Functionality E: Enterprise Edition P: Medium T: Defect | This doesn't work in BigQuery:
```sql
select distinct lower(TITLE)
from TEST.T_BOOK
order by lower(TITLE);
```
The error is:
> SQL Error [100032] [HY000]: [Simba][BigQueryJDBCDriver](100032) Error executing query job. Message: ORDER BY clause expression references column TITLE which is not visible after SELECT DISTINCT at [3:16]
A workaround could be to replace all `ORDER BY` expressions by the column index:
```sql
select distinct lower(TITLE)
from TEST.T_BOOK
order by 1;
```
But that only works if the expressions match directly. It doesn't work for inputs like that:
```sql
select distinct lower(TITLE)
from TEST.T_BOOK
order by reverse(lower(TITLE));
``` | 1.0 | BigQuery cannot reference exprsessions from SELECT DISTINCT in ORDER BY - This doesn't work in BigQuery:
```sql
select distinct lower(TITLE)
from TEST.T_BOOK
order by lower(TITLE);
```
The error is:
> SQL Error [100032] [HY000]: [Simba][BigQueryJDBCDriver](100032) Error executing query job. Message: ORDER BY clause expression references column TITLE which is not visible after SELECT DISTINCT at [3:16]
A workaround could be to replace all `ORDER BY` expressions by the column index:
```sql
select distinct lower(TITLE)
from TEST.T_BOOK
order by 1;
```
But that only works if the expressions match directly. It doesn't work for inputs like that:
```sql
select distinct lower(TITLE)
from TEST.T_BOOK
order by reverse(lower(TITLE));
``` | defect | bigquery cannot reference exprsessions from select distinct in order by this doesn t work in bigquery sql select distinct lower title from test t book order by lower title the error is sql error error executing query job message order by clause expression references column title which is not visible after select distinct at a workaround could be to replace all order by expressions by the column index sql select distinct lower title from test t book order by but that only works if the expressions match directly it doesn t work for inputs like that sql select distinct lower title from test t book order by reverse lower title | 1 |
7,323 | 2,610,363,453 | IssuesEvent | 2015-02-26 19:57:28 | chrsmith/scribefire-chrome | https://api.github.com/repos/chrsmith/scribefire-chrome | closed | tumblr authentication failing? | auto-migrated Priority-Medium tumblr Type-Defect | ```
What's the problem?
In the past week or so, when I try to post to my tumblr blog via scribefire, I
get an 'authentication failure'
I haven't changed my username/password and verified that the correct
credentials are cached.
What browser are you using?
Chrome ; but repros on firefox as well
What version of ScribeFire are you running?
```
-----
Original issue reported on code.google.com by `vinod.va...@gmail.com` on 7 Sep 2012 at 4:47
* Merged into: #766 | 1.0 | tumblr authentication failing? - ```
What's the problem?
In the past week or so, when I try to post to my tumblr blog via scribefire, I
get an 'authentication failure'
I haven't changed my username/password and verified that the correct
credentials are cached.
What browser are you using?
Chrome ; but repros on firefox as well
What version of ScribeFire are you running?
```
-----
Original issue reported on code.google.com by `vinod.va...@gmail.com` on 7 Sep 2012 at 4:47
* Merged into: #766 | defect | tumblr authentication failing what s the problem in the past week or so when i try to post to my tumblr blog via scribefire i get an authentication failure i haven t changed my username password and verified that the correct credentials are cached what browser are you using chrome but repros on firefox as well what version of scribefire are you running original issue reported on code google com by vinod va gmail com on sep at merged into | 1 |
57,838 | 16,100,064,387 | IssuesEvent | 2021-04-27 08:10:55 | line/armeria | https://api.github.com/repos/line/armeria | closed | 'Failed to force-initialize 'io.netty.channel.unix.Socket' message while using the armeria webflux sample project. | defect | Hello, I am inquiring with an error while using the armeria webflux sample project.
[https://github.com/line/armeria-examples/tree/master/spring-boot-webflux](https://github.com/line/armeria-examples/tree/master/spring-boot-webflux)
When this sample project is executed without modification, 'Failed to force-initialize 'io.netty.channel.unix.Socket' message is displayed.

[armeria_log.txt](https://github.com/line/armeria/files/6314962/armeria_log.txt)
There seems to be no problem with the use, but is this okay?
I am trying to use armeria for the first time, but I am anxious because this message is printed from the beginning.
If there is a problem, I hope it will be fixed.
| 1.0 | 'Failed to force-initialize 'io.netty.channel.unix.Socket' message while using the armeria webflux sample project. - Hello, I am inquiring with an error while using the armeria webflux sample project.
[https://github.com/line/armeria-examples/tree/master/spring-boot-webflux](https://github.com/line/armeria-examples/tree/master/spring-boot-webflux)
When this sample project is executed without modification, 'Failed to force-initialize 'io.netty.channel.unix.Socket' message is displayed.

[armeria_log.txt](https://github.com/line/armeria/files/6314962/armeria_log.txt)
There seems to be no problem with the use, but is this okay?
I am trying to use armeria for the first time, but I am anxious because this message is printed from the beginning.
If there is a problem, I hope it will be fixed.
| defect | failed to force initialize io netty channel unix socket message while using the armeria webflux sample project hello i am inquiring with an error while using the armeria webflux sample project when this sample project is executed without modification failed to force initialize io netty channel unix socket message is displayed there seems to be no problem with the use but is this okay i am trying to use armeria for the first time but i am anxious because this message is printed from the beginning if there is a problem i hope it will be fixed | 1 |
26,560 | 4,759,670,767 | IssuesEvent | 2016-10-24 23:30:40 | sukona/Grapevine | https://api.github.com/repos/sukona/Grapevine | reopened | ContentType is not correctly detected | defect Hacktoberfest Server | Hello,
While testing this excellent framework, I ran into some issues with content type detection.
Upon some introspection, I found the culprit: contentType might come richer than expected (either the mime type followed by the encoding or mime type coming as doubled).
My fix was a one-liner in the ContentType.cs file. Maybe it helps others ...
`public static ContentType FromString(this ContentType ct, string contentType)`
`{`
` return string.IsNullOrWhiteSpace(contentType)`
` ? ContentType.DEFAULT`
`//: Enum.GetValues(typeof(ContentType)).Cast<ContentType>().FirstOrDefault(t => t.ToValue().Equals(contentType));`
` : Enum.GetValues(typeof(ContentType)).Cast<ContentType>().FirstOrDefault(t => contentType.Contains(t.ToValue()));`
`}`
Best regards,
Marian | 1.0 | ContentType is not correctly detected - Hello,
While testing this excellent framework, I ran into some issues with content type detection.
Upon some introspection, I found the culprit: contentType might come richer than expected (either the mime type followed by the encoding or mime type coming as doubled).
My fix was a one-liner in the ContentType.cs file. Maybe it helps others ...
`public static ContentType FromString(this ContentType ct, string contentType)`
`{`
` return string.IsNullOrWhiteSpace(contentType)`
` ? ContentType.DEFAULT`
`//: Enum.GetValues(typeof(ContentType)).Cast<ContentType>().FirstOrDefault(t => t.ToValue().Equals(contentType));`
` : Enum.GetValues(typeof(ContentType)).Cast<ContentType>().FirstOrDefault(t => contentType.Contains(t.ToValue()));`
`}`
Best regards,
Marian | defect | contenttype is not correctly detected hello while testing this excellent framework i ran into some issues with content type detection upon some introspection i found the culprit contenttype might come richer than expected either the mime type followed by the encoding or mime type coming as doubled my fix was a one liner in the contenttype cs file maybe it helps others public static contenttype fromstring this contenttype ct string contenttype return string isnullorwhitespace contenttype contenttype default enum getvalues typeof contenttype cast firstordefault t t tovalue equals contenttype enum getvalues typeof contenttype cast firstordefault t contenttype contains t tovalue best regards marian | 1 |
12,325 | 3,601,040,255 | IssuesEvent | 2016-02-03 09:35:10 | mayuso/Datuk | https://api.github.com/repos/mayuso/Datuk | reopened | Examples | documentation | Datuk needs examples on how to use the API.
Examples needed:
* Excel + Data. (Missing)
* Txt + Data. (Missing)
* Classification. (Missing)
* Clustering. (Missing)
* Charts. (BarChart and XY done, PieChart missing).
| 1.0 | Examples - Datuk needs examples on how to use the API.
Examples needed:
* Excel + Data. (Missing)
* Txt + Data. (Missing)
* Classification. (Missing)
* Clustering. (Missing)
* Charts. (BarChart and XY done, PieChart missing).
| non_defect | examples datuk needs examples on how to use the api examples needed excel data missing txt data missing classification missing clustering missing charts barchart and xy done piechart missing | 0 |
9,109 | 2,615,131,038 | IssuesEvent | 2015-03-01 06:01:01 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | closed | Error parsing a json response | auto-migrated Priority-Medium Type-Defect | ```
Version of google-api-java-client (e.g. 1.5.0-beta)?
google-oauth-client-java7 -> 1.13.1-beta
google-http-client-gson -> 1.13.1-beta
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Java 7u10
Describe the problem.
i'm playing around with Dropbox Api, and i'm getting a json response like this
one
{
"referral_link": "https://www.dropbox.com/referrals/r1a2n3d4m5s6t7",
"display_name": "John P. User",
"uid": 12345678,
"country": "US",
"quota_info": {
"shared": 253738410565,
"quota": 107374182400000,
"normal": 680031877871
}
}
no problem sending/singing the req or getting the input stream from the
response, is when i try to parse the response using something like this
DropboxAccountInfo accountInfo = response.parseAs(DropboxAccountInfo.class);
it actually does return a new instance of the DropboxAccountInfo class but all
the field inside of it are null.
this is my DropboxAccountInfo class:
public class DropboxAccountInfo {
private String uid;
private String display_name;
private String country;
private String referral_link;
private DropboxQuota quota_info;
//Getters and setters not who do not necessary match with the variables names
private class DropboxQuota {
private long quota;
private long normal;
private long shared;
//Getters and setters not who do not necessary match with the variables names
}
now, i'm using a GsonObjectParser and it's working fine, the problem seems to
be on the ClassInfo.class when the constructor is running, the
nameToFieldInfoMap map is never filled up and i think to that's the root of the
entire problem.
this map is not filled up because on the line 174 FieldInfo fieldInfo =
FieldInfo.of(field); is always getting null, and this is because inside of the
"of" method the variable field is not an enum and then check for the @Key
annotation line (93), and it doesn't have that one cuz... just don't really
need to, anyways, i don't see why if it doesn't contain the @key annotation
this method should return null.
i hope have explained me well enough
How would you expect it to be fixed?
The parseAs() method should return a new instance of the class but with all the
values populated according to the json
IN ADVANCE THANKS SO MUCH AND SORRY FOR MY BAD ENGLISH, if you need more
information (or want to pair to track this down or whatever) please do not no
hesitate on contact me (lcc.julio@gmail.com)
```
Original issue reported on code.google.com by `lcc.ju...@gmail.com` on 7 Jan 2013 at 5:21 | 1.0 | Error parsing a json response - ```
Version of google-api-java-client (e.g. 1.5.0-beta)?
google-oauth-client-java7 -> 1.13.1-beta
google-http-client-gson -> 1.13.1-beta
Java environment (e.g. Java 6, Android 2.3, App Engine)?
Java 7u10
Describe the problem.
i'm playing around with Dropbox Api, and i'm getting a json response like this
one
{
"referral_link": "https://www.dropbox.com/referrals/r1a2n3d4m5s6t7",
"display_name": "John P. User",
"uid": 12345678,
"country": "US",
"quota_info": {
"shared": 253738410565,
"quota": 107374182400000,
"normal": 680031877871
}
}
no problem sending/singing the req or getting the input stream from the
response, is when i try to parse the response using something like this
DropboxAccountInfo accountInfo = response.parseAs(DropboxAccountInfo.class);
it actually does return a new instance of the DropboxAccountInfo class but all
the field inside of it are null.
this is my DropboxAccountInfo class:
public class DropboxAccountInfo {
private String uid;
private String display_name;
private String country;
private String referral_link;
private DropboxQuota quota_info;
//Getters and setters not who do not necessary match with the variables names
private class DropboxQuota {
private long quota;
private long normal;
private long shared;
//Getters and setters not who do not necessary match with the variables names
}
now, i'm using a GsonObjectParser and it's working fine, the problem seems to
be on the ClassInfo.class when the constructor is running, the
nameToFieldInfoMap map is never filled up and i think to that's the root of the
entire problem.
this map is not filled up because on the line 174 FieldInfo fieldInfo =
FieldInfo.of(field); is always getting null, and this is because inside of the
"of" method the variable field is not an enum and then check for the @Key
annotation line (93), and it doesn't have that one cuz... just don't really
need to, anyways, i don't see why if it doesn't contain the @key annotation
this method should return null.
i hope have explained me well enough
How would you expect it to be fixed?
The parseAs() method should return a new instance of the class but with all the
values populated according to the json
IN ADVANCE THANKS SO MUCH AND SORRY FOR MY BAD ENGLISH, if you need more
information (or want to pair to track this down or whatever) please do not no
hesitate on contact me (lcc.julio@gmail.com)
```
Original issue reported on code.google.com by `lcc.ju...@gmail.com` on 7 Jan 2013 at 5:21 | defect | error parsing a json response version of google api java client e g beta google oauth client beta google http client gson beta java environment e g java android app engine java describe the problem i m playing around with dropbox api and i m getting a json response like this one referral link display name john p user uid country us quota info shared quota normal no problem sending singing the req or getting the input stream from the response is when i try to parse the response using something like this dropboxaccountinfo accountinfo response parseas dropboxaccountinfo class it actually does return a new instance of the dropboxaccountinfo class but all the field inside of it are null this is my dropboxaccountinfo class public class dropboxaccountinfo private string uid private string display name private string country private string referral link private dropboxquota quota info getters and setters not who do not necessary match with the variables names private class dropboxquota private long quota private long normal private long shared getters and setters not who do not necessary match with the variables names now i m using a gsonobjectparser and it s working fine the problem seems to be on the classinfo class when the constructor is running the nametofieldinfomap map is never filled up and i think to that s the root of the entire problem this map is not filled up because on the line fieldinfo fieldinfo fieldinfo of field is always getting null and this is because inside of the of method the variable field is not an enum and then check for the key annotation line and it doesn t have that one cuz just don t really need to anyways i don t see why if it doesn t contain the key annotation this method should return null i hope have explained me well enough how would you expect it to be fixed the parseas method should return a new instance of the class but with all the values populated according to the json in advance thanks so much and sorry for my bad english if you need more information or want to pair to track this down or whatever please do not no hesitate on contact me lcc julio gmail com original issue reported on code google com by lcc ju gmail com on jan at | 1 |
79,713 | 28,500,180,645 | IssuesEvent | 2023-04-18 16:44:13 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | SelectCheckboxMenu: emptyLabel not considered as overlay-target and opens behind dialog | :lady_beetle: defect | ### Describe the bug
Dialog `bindEvents` checks for whether `mousedown` event target was `primefaces-overlay-target` and if not moves to top. This causes issues with clicks on `selectCheckboxMenu` `emptyLabel` which opens the menu/dropdown but dialog moves on top of it. When clicking on dropdown arrow or actual data label it works fine.
`SelectCheckboxMenu` `emptyLabel` fails this check on `line 379`:
https://github.com/primefaces/primefaces/blob/4e4430b40a2d1573137a975071edad1b3f953b66/primefaces/src/main/resources/META-INF/resources/primefaces/dialog/dialog.js#L377-L382
### Reproducer
Place selectCheckboxMenu inside popup and click on emptyLabel instead of dropdown arrow.
Menu opens but container dialog moves on top right after
### Expected behavior
Menu should open on top of the dialog
### PrimeFaces edition
Elite
### PrimeFaces version
12.0.3
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3
### Java version
8
### Browser(s)
_No response_ | 1.0 | SelectCheckboxMenu: emptyLabel not considered as overlay-target and opens behind dialog - ### Describe the bug
Dialog `bindEvents` checks for whether `mousedown` event target was `primefaces-overlay-target` and if not moves to top. This causes issues with clicks on `selectCheckboxMenu` `emptyLabel` which opens the menu/dropdown but dialog moves on top of it. When clicking on dropdown arrow or actual data label it works fine.
`SelectCheckboxMenu` `emptyLabel` fails this check on `line 379`:
https://github.com/primefaces/primefaces/blob/4e4430b40a2d1573137a975071edad1b3f953b66/primefaces/src/main/resources/META-INF/resources/primefaces/dialog/dialog.js#L377-L382
### Reproducer
Place selectCheckboxMenu inside popup and click on emptyLabel instead of dropdown arrow.
Menu opens but container dialog moves on top right after
### Expected behavior
Menu should open on top of the dialog
### PrimeFaces edition
Elite
### PrimeFaces version
12.0.3
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.3
### Java version
8
### Browser(s)
_No response_ | defect | selectcheckboxmenu emptylabel not considered as overlay target and opens behind dialog describe the bug dialog bindevents checks for whether mousedown event target was primefaces overlay target and if not moves to top this causes issues with clicks on selectcheckboxmenu emptylabel which opens the menu dropdown but dialog moves on top of it when clicking on dropdown arrow or actual data label it works fine selectcheckboxmenu emptylabel fails this check on line reproducer place selectcheckboxmenu inside popup and click on emptylabel instead of dropdown arrow menu opens but container dialog moves on top right after expected behavior menu should open on top of the dialog primefaces edition elite primefaces version theme no response jsf implementation mojarra jsf version java version browser s no response | 1 |
75,173 | 25,568,397,449 | IssuesEvent | 2022-11-30 15:51:01 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | closed | CMS sitewide alerts not visible to editors | Critical defect ⭐️ Sitewide CMS | ## Describe the defect
Admins are able to see sitewide alerts.
However, our shields and deploy warnings are not visible to everyone else.
Slack conversation https://dsva.slack.com/archives/CDHBKAL9W/p1669675816587709
This is incorrectly permed
It should be set for anonymous AND authenticated

## To Reproduce
Steps to reproduce the behavior:
1. As an admin set a sitewide alert to display
2. Logout and login as a non-admin
3. See that you see no sitewide alert
## AC / Expected behavior
Sitewide alerts must be visible to ALL users!
## Additional context
Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days).
## Desktop (please complete the following information if relevant, or delete)
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
## Labels
(You can delete this section once it's complete)
- [x] Issue type (red) (defaults to "Defect")
- [ ] CMS subsystem (green)
- [ ] CMS practice area (blue)
- [x] CMS workstream (orange) (not needed for bug tickets)
- [ ] CMS-supported product (black)
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [x] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
| 1.0 | CMS sitewide alerts not visible to editors - ## Describe the defect
Admins are able to see sitewide alerts.
However, our shields and deploy warnings are not visible to everyone else.
Slack conversation https://dsva.slack.com/archives/CDHBKAL9W/p1669675816587709
This is incorrectly permed
It should be set for anonymous AND authenticated

## To Reproduce
Steps to reproduce the behavior:
1. As an admin set a sitewide alert to display
2. Logout and login as a non-admin
3. See that you see no sitewide alert
## AC / Expected behavior
Sitewide alerts must be visible to ALL users!
## Additional context
Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days).
## Desktop (please complete the following information if relevant, or delete)
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
## Labels
(You can delete this section once it's complete)
- [x] Issue type (red) (defaults to "Defect")
- [ ] CMS subsystem (green)
- [ ] CMS practice area (blue)
- [x] CMS workstream (orange) (not needed for bug tickets)
- [ ] CMS-supported product (black)
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [x] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
| defect | cms sitewide alerts not visible to editors describe the defect admins are able to see sitewide alerts however our shields and deploy warnings are not visible to everyone else slack conversation this is incorrectly permed it should be set for anonymous and authenticated to reproduce steps to reproduce the behavior as an admin set a sitewide alert to display logout and login as a non admin see that you see no sitewide alert ac expected behavior sitewide alerts must be visible to all users additional context add any other context about the problem here reach out to the product managers to determine if it should be escalated as critical prevents users from accomplishing their work with no known workaround and needs to be addressed within business days desktop please complete the following information if relevant or delete os browser version labels you can delete this section once it s complete issue type red defaults to defect cms subsystem green cms practice area blue cms workstream orange not needed for bug tickets cms supported product black cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support | 1 |
34,047 | 7,330,874,230 | IssuesEvent | 2018-03-05 11:27:07 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | Atribute visible from MenuItem is not work in SplitButton | defect |
```
[ x ] bug report
```
**What is the motivation / use case for changing the behavior?**
Atribute visible from MenuItem is not work in SplitButton
**Please tell us about your environment:**
Windows 10
IDE: Webstorm
* **Angular version:** 5.2.1
* **PrimeNG version:** 5.2.0
* **Browser:**
Versão 63.0.3239.132 (Versão oficial) 64 bits
* **Language:**
"typescript": "~2.6.1"
"target": "es5",
* **Node (for AoT issues):**
v9.2.1
| 1.0 | Atribute visible from MenuItem is not work in SplitButton -
```
[ x ] bug report
```
**What is the motivation / use case for changing the behavior?**
Atribute visible from MenuItem is not work in SplitButton
**Please tell us about your environment:**
Windows 10
IDE: Webstorm
* **Angular version:** 5.2.1
* **PrimeNG version:** 5.2.0
* **Browser:**
Versão 63.0.3239.132 (Versão oficial) 64 bits
* **Language:**
"typescript": "~2.6.1"
"target": "es5",
* **Node (for AoT issues):**
v9.2.1
| defect | atribute visible from menuitem is not work in splitbutton bug report what is the motivation use case for changing the behavior atribute visible from menuitem is not work in splitbutton please tell us about your environment windows ide webstorm angular version primeng version browser versão versão oficial bits language typescript target node for aot issues | 1 |
78,978 | 27,849,345,489 | IssuesEvent | 2023-03-20 17:32:42 | idaholab/moose | https://api.github.com/repos/idaholab/moose | opened | Tagged residual being accumulated twice | C: Framework T: defect P: normal | ## Bug Description
MOOSE has the capability to save residual contributions from specific kernels using the [Tagging System](https://mooseframework.inl.gov/framework_development/tagging.html). You can accumulate the contribution for a specific tag by calling `NonlinearSystemBase::computeResidualTags(const std::set<TagID> & tags)`. This allows only the kernels associated with the given tags to be computed. However, it seems that tags that aren't requested are being accumulated by the `computeResidualTags` call. This is causing some unexpected behavior.
## Steps to Reproduce
Let's say we have two calls to computeResidual with different requested tags:
```C++
_fe_problem.computeResidualTags({residual_tag_1});
_fe_problem.computeResidualTags({residual_tag_2});
```
You will see that the vectors associated with `residual_tag_1` and `residual_tag_2` are accumulated in both of these calls. This is especially troublesome since we only zero and close the requested tags before and after the residual evaluation.
## Impact
Preventing Griffin from converting `save_in` functionality to vector tagging. | 1.0 | Tagged residual being accumulated twice - ## Bug Description
MOOSE has the capability to save residual contributions from specific kernels using the [Tagging System](https://mooseframework.inl.gov/framework_development/tagging.html). You can accumulate the contribution for a specific tag by calling `NonlinearSystemBase::computeResidualTags(const std::set<TagID> & tags)`. This allows only the kernels associated with the given tags to be computed. However, it seems that tags that aren't requested are being accumulated by the `computeResidualTags` call. This is causing some unexpected behavior.
## Steps to Reproduce
Let's say we have two calls to computeResidual with different requested tags:
```C++
_fe_problem.computeResidualTags({residual_tag_1});
_fe_problem.computeResidualTags({residual_tag_2});
```
You will see that the vectors associated with `residual_tag_1` and `residual_tag_2` are accumulated in both of these calls. This is especially troublesome since we only zero and close the requested tags before and after the residual evaluation.
## Impact
Preventing Griffin from converting `save_in` functionality to vector tagging. | defect | tagged residual being accumulated twice bug description moose has the capability to save residual contributions from specific kernels using the you can accumulate the contribution for a specific tag by calling nonlinearsystembase computeresidualtags const std set tags this allows only the kernels associated with the given tags to be computed however it seems that tags that aren t requested are being accumulated by the computeresidualtags call this is causing some unexpected behavior steps to reproduce let s say we have two calls to computeresidual with different requested tags c fe problem computeresidualtags residual tag fe problem computeresidualtags residual tag you will see that the vectors associated with residual tag and residual tag are accumulated in both of these calls this is especially troublesome since we only zero and close the requested tags before and after the residual evaluation impact preventing griffin from converting save in functionality to vector tagging | 1 |
60,055 | 17,023,322,800 | IssuesEvent | 2021-07-03 01:25:45 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Potlatch incorrectly handle GPX timestamps with fractions of second | Component: potlatch (flash editor) Priority: minor Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 10.26pm, Saturday, 15th November 2008]**
I have uploaded a GPX file with track point timestamps going up to fractions of second. The file was imported with no errors, however when I press edit on the web site, Potlatch shows it to me as if I have the first point in the GPX being somewhere far away (sometimes north-west, sometimes north-east, depending on track but constant for a track). I checked manually, the first point is where it should be (no such big error). The GPX view (clicking on the file name in the list of GPX tracks) shows the track correctly, just the Potlatch does it wrong (though I haven't tested with JOSM or similar).
For a sample look e.g. at this track: http://www.openstreetmap.org/user/dtbow/traces/250801
I have fixed in the meantime my application to not provide fractions of second, but in general, the GPX standard doesn't forbid doing it. | 1.0 | Potlatch incorrectly handle GPX timestamps with fractions of second - **[Submitted to the original trac issue database at 10.26pm, Saturday, 15th November 2008]**
I have uploaded a GPX file with track point timestamps going up to fractions of second. The file was imported with no errors, however when I press edit on the web site, Potlatch shows it to me as if I have the first point in the GPX being somewhere far away (sometimes north-west, sometimes north-east, depending on track but constant for a track). I checked manually, the first point is where it should be (no such big error). The GPX view (clicking on the file name in the list of GPX tracks) shows the track correctly, just the Potlatch does it wrong (though I haven't tested with JOSM or similar).
For a sample look e.g. at this track: http://www.openstreetmap.org/user/dtbow/traces/250801
I have fixed in the meantime my application to not provide fractions of second, but in general, the GPX standard doesn't forbid doing it. | defect | potlatch incorrectly handle gpx timestamps with fractions of second i have uploaded a gpx file with track point timestamps going up to fractions of second the file was imported with no errors however when i press edit on the web site potlatch shows it to me as if i have the first point in the gpx being somewhere far away sometimes north west sometimes north east depending on track but constant for a track i checked manually the first point is where it should be no such big error the gpx view clicking on the file name in the list of gpx tracks shows the track correctly just the potlatch does it wrong though i haven t tested with josm or similar for a sample look e g at this track i have fixed in the meantime my application to not provide fractions of second but in general the gpx standard doesn t forbid doing it | 1 |
17,427 | 3,004,869,415 | IssuesEvent | 2015-07-26 11:47:53 | emukidid/swiss-gc | https://api.github.com/repos/emukidid/swiss-gc | closed | Star fox Adventures no audio and wrong subtitles | auto-migrated Priority-Medium Type-Defect | ```
What step(s) will reproduce the problem?
1.Play the game on Swiss.
2.Select Story.
3.Watch the cutscenes.
What console are you running Swiss on (Wii/GC)?
Wii
What version of Swiss are you using?
Swiss r298
What disc was in the drive if any?
Wii Sports
What device was the data being loaded from (SD/HDD/DVD/DVD-R)?
SD
Please provide any additional information below.
When I play the story mode the voices of the character is mute and the
subtitles are wrong.
```
Original issue reported on code.google.com by `diego.so...@gmail.com` on 28 Feb 2015 at 12:26 | 1.0 | Star fox Adventures no audio and wrong subtitles - ```
What step(s) will reproduce the problem?
1.Play the game on Swiss.
2.Select Story.
3.Watch the cutscenes.
What console are you running Swiss on (Wii/GC)?
Wii
What version of Swiss are you using?
Swiss r298
What disc was in the drive if any?
Wii Sports
What device was the data being loaded from (SD/HDD/DVD/DVD-R)?
SD
Please provide any additional information below.
When I play the story mode the voices of the character is mute and the
subtitles are wrong.
```
Original issue reported on code.google.com by `diego.so...@gmail.com` on 28 Feb 2015 at 12:26 | defect | star fox adventures no audio and wrong subtitles what step s will reproduce the problem play the game on swiss select story watch the cutscenes what console are you running swiss on wii gc wii what version of swiss are you using swiss what disc was in the drive if any wii sports what device was the data being loaded from sd hdd dvd dvd r sd please provide any additional information below when i play the story mode the voices of the character is mute and the subtitles are wrong original issue reported on code google com by diego so gmail com on feb at | 1 |
31,676 | 13,613,139,636 | IssuesEvent | 2020-09-23 11:24:44 | dessant/buster | https://api.github.com/repos/dessant/buster | closed | hCaptcha image is not being solved | wontfix-service | The extension doesn't solve hCaptcha's. I've tried it multiple times on the page that required it and it just doesn't work. All times I had to manually do it myself.
EDIT: please add hCaptcha to the extension. | 1.0 | hCaptcha image is not being solved - The extension doesn't solve hCaptcha's. I've tried it multiple times on the page that required it and it just doesn't work. All times I had to manually do it myself.
EDIT: please add hCaptcha to the extension. | non_defect | hcaptcha image is not being solved the extension doesn t solve hcaptcha s i ve tried it multiple times on the page that required it and it just doesn t work all times i had to manually do it myself edit please add hcaptcha to the extension | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.