Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
33,376
7,106,978,963
IssuesEvent
2018-01-16 18:21:13
netty/netty
https://api.github.com/repos/netty/netty
closed
Wrapped direct ByteBufs do not respect the position of the underlying buffer
defect
### Expected behavior If I create a MappedByteBuffer then change it's position and limit then wrap that buffer, the ByteBuf should respect that position/offset. ### Actual behavior `ReadOnlyUnsafeDirectByteBuf` does not respect the position of the underlying MappedByteBuffer and will start at the base address rather than the position. ### Steps to reproduce ```scala val f = java.nio.channels.FileChannel.open(java.nio.file.FileSystems.getDefault.getPath("somefile")) val m = f.map(java.nio.channels.FileChannel.MapMode.READ_ONLY, 0, f.size) val b1 = io.netty.buffer.Unpooled.wrappedBuffer(m) val dup = m.duplicate dup.position(2) dup.limit(4) val b2 = io.netty.buffer.Unpooled.wrappedBuffer(dup) b1.slice(2, 2) == b2 // this should be true, but is not because b2 will start at the wrapped buffer's 0 index ``` ### Minimal yet complete reproducer code (or URL to code) See above ### Netty version 4.1.19.Final ### JVM version (e.g. `java -version`) java version "1.8.0_121" Java(TM) SE Runtime Environment (build 1.8.0_121-b13) Java HotSpot(TM) 64-Bit Server VM (build 25.121-b13, mixed mode) ### OS version (e.g. `uname -a`) Linux sprsquish 4.10.0-42-generic #46-Ubuntu SMP Mon Dec 4 14:38:01 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux
1.0
Wrapped direct ByteBufs do not respect the position of the underlying buffer - ### Expected behavior If I create a MappedByteBuffer then change it's position and limit then wrap that buffer, the ByteBuf should respect that position/offset. ### Actual behavior `ReadOnlyUnsafeDirectByteBuf` does not respect the position of the underlying MappedByteBuffer and will start at the base address rather than the position. ### Steps to reproduce ```scala val f = java.nio.channels.FileChannel.open(java.nio.file.FileSystems.getDefault.getPath("somefile")) val m = f.map(java.nio.channels.FileChannel.MapMode.READ_ONLY, 0, f.size) val b1 = io.netty.buffer.Unpooled.wrappedBuffer(m) val dup = m.duplicate dup.position(2) dup.limit(4) val b2 = io.netty.buffer.Unpooled.wrappedBuffer(dup) b1.slice(2, 2) == b2 // this should be true, but is not because b2 will start at the wrapped buffer's 0 index ``` ### Minimal yet complete reproducer code (or URL to code) See above ### Netty version 4.1.19.Final ### JVM version (e.g. `java -version`) java version "1.8.0_121" Java(TM) SE Runtime Environment (build 1.8.0_121-b13) Java HotSpot(TM) 64-Bit Server VM (build 25.121-b13, mixed mode) ### OS version (e.g. `uname -a`) Linux sprsquish 4.10.0-42-generic #46-Ubuntu SMP Mon Dec 4 14:38:01 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux
defect
wrapped direct bytebufs do not respect the position of the underlying buffer expected behavior if i create a mappedbytebuffer then change it s position and limit then wrap that buffer the bytebuf should respect that position offset actual behavior readonlyunsafedirectbytebuf does not respect the position of the underlying mappedbytebuffer and will start at the base address rather than the position steps to reproduce scala val f java nio channels filechannel open java nio file filesystems getdefault getpath somefile val m f map java nio channels filechannel mapmode read only f size val io netty buffer unpooled wrappedbuffer m val dup m duplicate dup position dup limit val io netty buffer unpooled wrappedbuffer dup slice this should be true but is not because will start at the wrapped buffer s index minimal yet complete reproducer code or url to code see above netty version final jvm version e g java version java version java tm se runtime environment build java hotspot tm bit server vm build mixed mode os version e g uname a linux sprsquish generic ubuntu smp mon dec utc gnu linux
1
43,855
11,865,628,145
IssuesEvent
2020-03-26 01:02:20
richgel999/jpeg-compressor
https://api.github.com/repos/richgel999/jpeg-compressor
closed
Shortened RGBA->YCC code
Priority-Medium Type-Defect auto-migrated
``` Attached patch ``` Original issue reported on code.google.com by `pornel...@gmail.com` on 29 Sep 2013 at 5:38 Attachments: - [0001-Use-standard-header.patch](https://storage.googleapis.com/google-code-attachments/jpeg-compressor/issue-3/comment-0/0001-Use-standard-header.patch) - [0002-Removed-duplicate-RGB-RGBA-conversions.patch](https://storage.googleapis.com/google-code-attachments/jpeg-compressor/issue-3/comment-0/0002-Removed-duplicate-RGB-RGBA-conversions.patch)
1.0
Shortened RGBA->YCC code - ``` Attached patch ``` Original issue reported on code.google.com by `pornel...@gmail.com` on 29 Sep 2013 at 5:38 Attachments: - [0001-Use-standard-header.patch](https://storage.googleapis.com/google-code-attachments/jpeg-compressor/issue-3/comment-0/0001-Use-standard-header.patch) - [0002-Removed-duplicate-RGB-RGBA-conversions.patch](https://storage.googleapis.com/google-code-attachments/jpeg-compressor/issue-3/comment-0/0002-Removed-duplicate-RGB-RGBA-conversions.patch)
defect
shortened rgba ycc code attached patch original issue reported on code google com by pornel gmail com on sep at attachments
1
517,700
15,018,657,154
IssuesEvent
2021-02-01 12:32:00
eventespresso/barista
https://api.github.com/repos/eventespresso/barista
closed
Move New Default Tickets UI Behind Feature Flag
C: UI/UX 🚽 D: EDTR ✏️ P2: HIGH priority 😮 S:1 new 👶🏻 T: task 🧹
Title says it all. We need to move ALL aspects of the new Default Tickets UI behind a Feature Flag (FF) so that it is not accessible to users unless they possess the appropriate permission (let's call it "use_default_ticket_manager")
1.0
Move New Default Tickets UI Behind Feature Flag - Title says it all. We need to move ALL aspects of the new Default Tickets UI behind a Feature Flag (FF) so that it is not accessible to users unless they possess the appropriate permission (let's call it "use_default_ticket_manager")
non_defect
move new default tickets ui behind feature flag title says it all we need to move all aspects of the new default tickets ui behind a feature flag ff so that it is not accessible to users unless they possess the appropriate permission let s call it use default ticket manager
0
47,465
13,056,197,234
IssuesEvent
2020-07-30 03:57:40
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
qt ports should check for opengl (Trac #596)
Migrated from Trac cmake defect
need to verify that qt isn't getting silently built w/o opengl support. People are getting the ports installed and then failing on OPENGL_FOUND when configuring the build of the glshovel... this should be impossible. Migrated from https://code.icecube.wisc.edu/ticket/596 ```json { "status": "closed", "changetime": "2010-02-15T21:42:56", "description": "need to verify that qt isn't getting silently built w/o opengl support. People are getting the ports installed and then failing on OPENGL_FOUND when configuring the build of the glshovel... this should be impossible.", "reporter": "troy", "cc": "", "resolution": "fixed", "_ts": "1266270176000000", "component": "cmake", "summary": "qt ports should check for opengl", "priority": "normal", "keywords": "", "time": "2010-02-09T15:34:42", "milestone": "", "owner": "troy", "type": "defect" } ```
1.0
qt ports should check for opengl (Trac #596) - need to verify that qt isn't getting silently built w/o opengl support. People are getting the ports installed and then failing on OPENGL_FOUND when configuring the build of the glshovel... this should be impossible. Migrated from https://code.icecube.wisc.edu/ticket/596 ```json { "status": "closed", "changetime": "2010-02-15T21:42:56", "description": "need to verify that qt isn't getting silently built w/o opengl support. People are getting the ports installed and then failing on OPENGL_FOUND when configuring the build of the glshovel... this should be impossible.", "reporter": "troy", "cc": "", "resolution": "fixed", "_ts": "1266270176000000", "component": "cmake", "summary": "qt ports should check for opengl", "priority": "normal", "keywords": "", "time": "2010-02-09T15:34:42", "milestone": "", "owner": "troy", "type": "defect" } ```
defect
qt ports should check for opengl trac need to verify that qt isn t getting silently built w o opengl support people are getting the ports installed and then failing on opengl found when configuring the build of the glshovel this should be impossible migrated from json status closed changetime description need to verify that qt isn t getting silently built w o opengl support people are getting the ports installed and then failing on opengl found when configuring the build of the glshovel this should be impossible reporter troy cc resolution fixed ts component cmake summary qt ports should check for opengl priority normal keywords time milestone owner troy type defect
1
98,786
20,798,233,514
IssuesEvent
2022-03-17 11:24:36
Onelinerhub/onelinerhub
https://api.github.com/repos/Onelinerhub/onelinerhub
opened
Short solution needed: "Custom log format" (nginx)
help wanted good first issue code nginx
Please help us write most modern and shortest code solution for this issue: **Custom log format** (technology: [nginx](https://onelinerhub.com/nginx)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to use comments to make solution explained. 3. Link to this issue in comments of pull request.
1.0
Short solution needed: "Custom log format" (nginx) - Please help us write most modern and shortest code solution for this issue: **Custom log format** (technology: [nginx](https://onelinerhub.com/nginx)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create pull request with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to use comments to make solution explained. 3. Link to this issue in comments of pull request.
non_defect
short solution needed custom log format nginx please help us write most modern and shortest code solution for this issue custom log format technology fast way just write the code solution in the comments prefered way create pull request with a new code file inside don t forget to use comments to make solution explained link to this issue in comments of pull request
0
12,037
3,250,687,281
IssuesEvent
2015-10-19 03:19:47
kumulsoft/Fixed-Assets
https://api.github.com/repos/kumulsoft/Fixed-Assets
closed
SETUP >> Manage Staff. Small Adjustments to the entry screen
bug enhancement Fixed Ready for testing UI
1. Rename section 'Contact Information' to 'Staff Information' 2. Rename label 'Employee Name' to 'Staff Name' 3. Contact Type must be defaulted to 'Staff' and Read Only 4. Arrange Centre and Location to be Side by Side (like Division and Section) 5. Remove/Hide the Address section 6. Position Field too is missing, put it back ![image](https://cloud.githubusercontent.com/assets/10192106/9671692/26943ee0-52d8-11e5-9bb9-c9fdea64d23e.png)
1.0
SETUP >> Manage Staff. Small Adjustments to the entry screen - 1. Rename section 'Contact Information' to 'Staff Information' 2. Rename label 'Employee Name' to 'Staff Name' 3. Contact Type must be defaulted to 'Staff' and Read Only 4. Arrange Centre and Location to be Side by Side (like Division and Section) 5. Remove/Hide the Address section 6. Position Field too is missing, put it back ![image](https://cloud.githubusercontent.com/assets/10192106/9671692/26943ee0-52d8-11e5-9bb9-c9fdea64d23e.png)
non_defect
setup manage staff small adjustments to the entry screen rename section contact information to staff information rename label employee name to staff name contact type must be defaulted to staff and read only arrange centre and location to be side by side like division and section remove hide the address section position field too is missing put it back
0
93,971
27,084,849,079
IssuesEvent
2023-02-14 16:18:20
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Maven downloads non-camunda artifacts from camunda-nexus instead of maven central
area/build component/build-pipeline
Since https://github.com/camunda/zeebe/pull/11498 for merged we observed maven to not just download camunda artifacts from the camunda nexus repo but also non camunda artifacts, e.g. https://github.com/camunda/zeebe/actions/runs/4173174037/jobs/7229130526 ``` [INFO] --- maven-install-plugin:2.4:install (default-install) @ zeebe-journal --- [INFO] Installing /runner/_work/zeebe/zeebe/journal/target/zeebe-journal-8.2.0-SNAPSHOT.jar to /home/runner/.m2/repository/io/camunda/zeebe-journal/8.2.0-SNAPSHOT/zeebe-journal-8.2.0-SNAPSHOT.jar [INFO] Installing /runner/_work/zeebe/zeebe/journal/.flattened-pom.xml to /home/runner/.m2/repository/io/camunda/zeebe-journal/8.2.0-SNAPSHOT/zeebe-journal-8.2.0-SNAPSHOT.pom [INFO] Downloaded from camunda-nexus: https://repository.nexus.camunda.cloud/content/groups/internal/io/netty/netty-handler/4.1.86.Final/netty-handler-4.1.86.Final.pom (4.5 kB at 14 kB/s) [INFO] Downloading from camunda-nexus: https://repository.nexus.camunda.cloud/content/groups/internal/io/netty/netty-transport-native-unix-common/4.1.86.Final/netty-transport-native-unix-common-4.1.86.Final.pom ``` On Github hosted runners we observed the download from this maven repo to be slow sometimes, ideally only camunda artifacts should get downloaded from this mirror.
2.0
Maven downloads non-camunda artifacts from camunda-nexus instead of maven central - Since https://github.com/camunda/zeebe/pull/11498 for merged we observed maven to not just download camunda artifacts from the camunda nexus repo but also non camunda artifacts, e.g. https://github.com/camunda/zeebe/actions/runs/4173174037/jobs/7229130526 ``` [INFO] --- maven-install-plugin:2.4:install (default-install) @ zeebe-journal --- [INFO] Installing /runner/_work/zeebe/zeebe/journal/target/zeebe-journal-8.2.0-SNAPSHOT.jar to /home/runner/.m2/repository/io/camunda/zeebe-journal/8.2.0-SNAPSHOT/zeebe-journal-8.2.0-SNAPSHOT.jar [INFO] Installing /runner/_work/zeebe/zeebe/journal/.flattened-pom.xml to /home/runner/.m2/repository/io/camunda/zeebe-journal/8.2.0-SNAPSHOT/zeebe-journal-8.2.0-SNAPSHOT.pom [INFO] Downloaded from camunda-nexus: https://repository.nexus.camunda.cloud/content/groups/internal/io/netty/netty-handler/4.1.86.Final/netty-handler-4.1.86.Final.pom (4.5 kB at 14 kB/s) [INFO] Downloading from camunda-nexus: https://repository.nexus.camunda.cloud/content/groups/internal/io/netty/netty-transport-native-unix-common/4.1.86.Final/netty-transport-native-unix-common-4.1.86.Final.pom ``` On Github hosted runners we observed the download from this maven repo to be slow sometimes, ideally only camunda artifacts should get downloaded from this mirror.
non_defect
maven downloads non camunda artifacts from camunda nexus instead of maven central since for merged we observed maven to not just download camunda artifacts from the camunda nexus repo but also non camunda artifacts e g maven install plugin install default install zeebe journal installing runner work zeebe zeebe journal target zeebe journal snapshot jar to home runner repository io camunda zeebe journal snapshot zeebe journal snapshot jar installing runner work zeebe zeebe journal flattened pom xml to home runner repository io camunda zeebe journal snapshot zeebe journal snapshot pom downloaded from camunda nexus kb at kb s downloading from camunda nexus on github hosted runners we observed the download from this maven repo to be slow sometimes ideally only camunda artifacts should get downloaded from this mirror
0
76,678
26,553,671,760
IssuesEvent
2023-01-20 10:07:17
decentraland/unity-renderer
https://api.github.com/repos/decentraland/unity-renderer
opened
[QA] Cannot enter worlds on desktop when using a deeplink with new decoupled loading screen on
medium defect
Description: With the new loading screen decoupled from kernel it is not possible to enter any of the Worlds directly on desktop by using a deep link - loading gets stuck at 0%. With the decoupled loading screen feature flag off the loading proceeds normally without issues. Issue occurs only on desktop clients and not on web. Reproduction rate: 5/5 Steps to reproduce: 1. Launch the desktop client on the release/release-20230119 branch with `ENABLE_DECOUPLED_LOADING_SCREEN&realm=menduz.dcl.eth` parameters. 2. Log in either via a guest or a wallet account. 3. Observe the loading screen get stuck at 0%. Results: With the decoupled loading screen feature flag on loading gets stuck at 0% and refuses to proceed on worlds. Expected results: With the decoupled loading screen feature flag on loading proceeds normally on worlds. Notes: - Platforms: Windows and MacOS desktop, wallet and guest login. - Issue occurs for the [release/release-20230119 branch](https://github.com/decentraland/explorer-desktop/pull/464). - Please see the attached screenshot and video for reference. ![image.png](https://images.zenhubusercontent.com/62e91e01ed547029fd717bd4/1be0b218-916e-4776-bab3-a7de0175afb7) https://images.zenhubusercontent.com/337227404/15ace7b0-103b-4ea1-9d95-e54668102dc5/2023_01_20_10_57_44.mp4
1.0
[QA] Cannot enter worlds on desktop when using a deeplink with new decoupled loading screen on - Description: With the new loading screen decoupled from kernel it is not possible to enter any of the Worlds directly on desktop by using a deep link - loading gets stuck at 0%. With the decoupled loading screen feature flag off the loading proceeds normally without issues. Issue occurs only on desktop clients and not on web. Reproduction rate: 5/5 Steps to reproduce: 1. Launch the desktop client on the release/release-20230119 branch with `ENABLE_DECOUPLED_LOADING_SCREEN&realm=menduz.dcl.eth` parameters. 2. Log in either via a guest or a wallet account. 3. Observe the loading screen get stuck at 0%. Results: With the decoupled loading screen feature flag on loading gets stuck at 0% and refuses to proceed on worlds. Expected results: With the decoupled loading screen feature flag on loading proceeds normally on worlds. Notes: - Platforms: Windows and MacOS desktop, wallet and guest login. - Issue occurs for the [release/release-20230119 branch](https://github.com/decentraland/explorer-desktop/pull/464). - Please see the attached screenshot and video for reference. ![image.png](https://images.zenhubusercontent.com/62e91e01ed547029fd717bd4/1be0b218-916e-4776-bab3-a7de0175afb7) https://images.zenhubusercontent.com/337227404/15ace7b0-103b-4ea1-9d95-e54668102dc5/2023_01_20_10_57_44.mp4
defect
cannot enter worlds on desktop when using a deeplink with new decoupled loading screen on description with the new loading screen decoupled from kernel it is not possible to enter any of the worlds directly on desktop by using a deep link loading gets stuck at with the decoupled loading screen feature flag off the loading proceeds normally without issues issue occurs only on desktop clients and not on web reproduction rate steps to reproduce launch the desktop client on the release release branch with enable decoupled loading screen realm menduz dcl eth parameters log in either via a guest or a wallet account observe the loading screen get stuck at results with the decoupled loading screen feature flag on loading gets stuck at and refuses to proceed on worlds expected results with the decoupled loading screen feature flag on loading proceeds normally on worlds notes platforms windows and macos desktop wallet and guest login issue occurs for the please see the attached screenshot and video for reference
1
40,919
10,225,606,945
IssuesEvent
2019-08-16 15:33:11
google/auto
https://api.github.com/repos/google/auto
closed
Autofactory compilation fails on Java 8
Component: factory P3 type=defect
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile (default-testCompile) on project auto-factory: Compilation failure: Compilation failure: [ERROR] /home/zubzub/git/auto/factory/src/test/java/com/google/auto/factory/processor/AutoFactoryDeclarationTest.java:[18,71] package com.google.auto.factory.processor.AutoFactoryDeclaration does not exist [ERROR] /home/zubzub/git/auto/factory/src/test/java/com/google/auto/factory/processor/AutoFactoryDeclarationTest.java:[18,1] static import only from classes and interfaces
1.0
Autofactory compilation fails on Java 8 - [ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.1:testCompile (default-testCompile) on project auto-factory: Compilation failure: Compilation failure: [ERROR] /home/zubzub/git/auto/factory/src/test/java/com/google/auto/factory/processor/AutoFactoryDeclarationTest.java:[18,71] package com.google.auto.factory.processor.AutoFactoryDeclaration does not exist [ERROR] /home/zubzub/git/auto/factory/src/test/java/com/google/auto/factory/processor/AutoFactoryDeclarationTest.java:[18,1] static import only from classes and interfaces
defect
autofactory compilation fails on java failed to execute goal org apache maven plugins maven compiler plugin testcompile default testcompile on project auto factory compilation failure compilation failure home zubzub git auto factory src test java com google auto factory processor autofactorydeclarationtest java package com google auto factory processor autofactorydeclaration does not exist home zubzub git auto factory src test java com google auto factory processor autofactorydeclarationtest java static import only from classes and interfaces
1
68,064
21,468,818,139
IssuesEvent
2022-04-26 07:38:01
klubcoin/lcn-mobile
https://api.github.com/repos/klubcoin/lcn-mobile
opened
Klubcoin app should navigate to Dashboard while press back handler button om Purchase Methods screen
Defect Should Have Task Major
### **Description:** Klubcoin app should navigate to Dashboard while press back handler button om Purchase Methods screen. **Build Environment:** Staging Candidate Environment **Affects Version:** **Device Platform:** Android **Device OS:** Android 12 **Test Device:** Samsung S21 ### **Pre-condition:** 1. User successfully installed Klubcoin App 2. User has an existing Klubcoin Wallet Account 4. User is currently at Klubcoin Dashboard ### **Steps to Reproduce:** 1. Tap Hamburger Button 2. Tap Buy 3. Tap Back Handler button ### **Expected Result:** Navigate to Dashboard. ### **Actual Result:** Navigate to another Purchase Methods screen.
1.0
Klubcoin app should navigate to Dashboard while press back handler button om Purchase Methods screen - ### **Description:** Klubcoin app should navigate to Dashboard while press back handler button om Purchase Methods screen. **Build Environment:** Staging Candidate Environment **Affects Version:** **Device Platform:** Android **Device OS:** Android 12 **Test Device:** Samsung S21 ### **Pre-condition:** 1. User successfully installed Klubcoin App 2. User has an existing Klubcoin Wallet Account 4. User is currently at Klubcoin Dashboard ### **Steps to Reproduce:** 1. Tap Hamburger Button 2. Tap Buy 3. Tap Back Handler button ### **Expected Result:** Navigate to Dashboard. ### **Actual Result:** Navigate to another Purchase Methods screen.
defect
klubcoin app should navigate to dashboard while press back handler button om purchase methods screen description klubcoin app should navigate to dashboard while press back handler button om purchase methods screen build environment staging candidate environment affects version device platform android device os android test device samsung pre condition user successfully installed klubcoin app user has an existing klubcoin wallet account user is currently at klubcoin dashboard steps to reproduce tap hamburger button tap buy tap back handler button expected result navigate to dashboard actual result navigate to another purchase methods screen
1
12,972
2,732,346,269
IssuesEvent
2015-04-17 04:48:04
rasmus/fast-member
https://api.github.com/repos/rasmus/fast-member
closed
TypeAccessor.WriteSetter throws exception, if property setter is not public
auto-migrated Priority-Medium Type-Defect
``` The method TypeAccessor.WriteSetter throws an ArgumentNullException, if the target type has a property with a non-public setter. (Line 141: http://code.google.com/p/fast-member/source/browse/FastMember/TypeAccessor.cs#14 1) Example: class TestTarget { public int Id { get; private set; } } I am not sure why `prop.GetSetMethod()` returns `null`, although `prop.CanWrite` obviously returned `true` in line 129. ``` Original issue reported on code.google.com by `schuster...@gmail.com` on 27 Jan 2012 at 8:22
1.0
TypeAccessor.WriteSetter throws exception, if property setter is not public - ``` The method TypeAccessor.WriteSetter throws an ArgumentNullException, if the target type has a property with a non-public setter. (Line 141: http://code.google.com/p/fast-member/source/browse/FastMember/TypeAccessor.cs#14 1) Example: class TestTarget { public int Id { get; private set; } } I am not sure why `prop.GetSetMethod()` returns `null`, although `prop.CanWrite` obviously returned `true` in line 129. ``` Original issue reported on code.google.com by `schuster...@gmail.com` on 27 Jan 2012 at 8:22
defect
typeaccessor writesetter throws exception if property setter is not public the method typeaccessor writesetter throws an argumentnullexception if the target type has a property with a non public setter line example class testtarget public int id get private set i am not sure why prop getsetmethod returns null although prop canwrite obviously returned true in line original issue reported on code google com by schuster gmail com on jan at
1
21,497
3,734,176,998
IssuesEvent
2016-03-08 04:54:40
department-of-veterans-affairs/roadrunner
https://api.github.com/repos/department-of-veterans-affairs/roadrunner
opened
best steez for links on landing page?
design
Wondering how best to style these links - buttons or just add link styling? ![screen shot 2016-03-07 at 8 53 33 pm](https://cloud.githubusercontent.com/assets/1307774/13592426/c39cfd5c-e4a6-11e5-8fdc-0af4fbe03123.png) If you don't have time I can take another look but didn't want to delay getting the rest of this pr out :)
1.0
best steez for links on landing page? - Wondering how best to style these links - buttons or just add link styling? ![screen shot 2016-03-07 at 8 53 33 pm](https://cloud.githubusercontent.com/assets/1307774/13592426/c39cfd5c-e4a6-11e5-8fdc-0af4fbe03123.png) If you don't have time I can take another look but didn't want to delay getting the rest of this pr out :)
non_defect
best steez for links on landing page wondering how best to style these links buttons or just add link styling if you don t have time i can take another look but didn t want to delay getting the rest of this pr out
0
11,207
3,193,179,500
IssuesEvent
2015-09-30 02:29:05
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
TestProcWithExceededActionQueueDepth is flaky
area/platform/mesos kind/flake priority/P0 team/test-infra
@jdef @karlkfi @davidopp Can we get this fixed ASAP? Thanks! ``` proc_test.go:288: starting test case nested at 2015-09-25 00:21:13.875176228 +0000 UTC proc_test.go:304: delegate chain invoked for nested at 2015-09-25 00:21:13.930546473 +0000 UTC proc_test.go:323: executing deferred action: nested at 2015-09-25 00:21:13.933936289 +0000 UTC proc_test.go:335: runDelegationTest received executed signal at 2015-09-25 00:21:13.949041708 +0000 UTC proc_test.go:290: runDelegationTest finished at 2015-09-25 00:21:14.035916116 +0000 UTC proc_test.go:390: unexpected error: cannot execute action because process has terminated ``` https://app.shippable.com/builds/56048e527291610b002dea79
1.0
TestProcWithExceededActionQueueDepth is flaky - @jdef @karlkfi @davidopp Can we get this fixed ASAP? Thanks! ``` proc_test.go:288: starting test case nested at 2015-09-25 00:21:13.875176228 +0000 UTC proc_test.go:304: delegate chain invoked for nested at 2015-09-25 00:21:13.930546473 +0000 UTC proc_test.go:323: executing deferred action: nested at 2015-09-25 00:21:13.933936289 +0000 UTC proc_test.go:335: runDelegationTest received executed signal at 2015-09-25 00:21:13.949041708 +0000 UTC proc_test.go:290: runDelegationTest finished at 2015-09-25 00:21:14.035916116 +0000 UTC proc_test.go:390: unexpected error: cannot execute action because process has terminated ``` https://app.shippable.com/builds/56048e527291610b002dea79
non_defect
testprocwithexceededactionqueuedepth is flaky jdef karlkfi davidopp can we get this fixed asap thanks proc test go starting test case nested at utc proc test go delegate chain invoked for nested at utc proc test go executing deferred action nested at utc proc test go rundelegationtest received executed signal at utc proc test go rundelegationtest finished at utc proc test go unexpected error cannot execute action because process has terminated
0
282,572
24,485,657,290
IssuesEvent
2022-10-09 11:56:41
Tencent/bk-ci
https://api.github.com/repos/Tencent/bk-ci
closed
bug:质量红线导致的流水线失败缺少错误信息
kind/bug for gray for test done area/ci/backend tested streams/tested streams/for test streams/for gray
问题:质量红线导致的流水线失败缺少错误信息,会导致统计等功能异常。 措施:质量红线导致的流水线失败应记录错误信息,记录的错误信息要注意区分errorType的类型
4.0
bug:质量红线导致的流水线失败缺少错误信息 - 问题:质量红线导致的流水线失败缺少错误信息,会导致统计等功能异常。 措施:质量红线导致的流水线失败应记录错误信息,记录的错误信息要注意区分errorType的类型
non_defect
bug 质量红线导致的流水线失败缺少错误信息 问题:质量红线导致的流水线失败缺少错误信息,会导致统计等功能异常。 措施:质量红线导致的流水线失败应记录错误信息,记录的错误信息要注意区分errortype的类型
0
53,085
13,260,877,196
IssuesEvent
2020-08-20 18:54:58
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
SLALIB/C needs a real makefile (Trac #677)
Migrated from Trac defect tools/ports
makefile is currently a PoS. needs to be re-done with proper make constructs, and variables. (ie: CC, CFLAGS) <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/677">https://code.icecube.wisc.edu/projects/icecube/ticket/677</a>, reported by negaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2012-05-29T21:39:28", "_ts": "1338327568000000", "description": "makefile is currently a PoS. needs to be re-done with proper make constructs, and variables. (ie: CC, CFLAGS)", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2012-05-29T19:00:40", "component": "tools/ports", "summary": "SLALIB/C needs a real makefile", "priority": "normal", "keywords": "slalib", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
SLALIB/C needs a real makefile (Trac #677) - makefile is currently a PoS. needs to be re-done with proper make constructs, and variables. (ie: CC, CFLAGS) <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/677">https://code.icecube.wisc.edu/projects/icecube/ticket/677</a>, reported by negaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2012-05-29T21:39:28", "_ts": "1338327568000000", "description": "makefile is currently a PoS. needs to be re-done with proper make constructs, and variables. (ie: CC, CFLAGS)", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2012-05-29T19:00:40", "component": "tools/ports", "summary": "SLALIB/C needs a real makefile", "priority": "normal", "keywords": "slalib", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
slalib c needs a real makefile trac makefile is currently a pos needs to be re done with proper make constructs and variables ie cc cflags migrated from json status closed changetime ts description makefile is currently a pos needs to be re done with proper make constructs and variables ie cc cflags reporter nega cc resolution fixed time component tools ports summary slalib c needs a real makefile priority normal keywords slalib milestone owner nega type defect
1
8,368
2,982,464,693
IssuesEvent
2015-07-17 11:26:06
gheber/kenzo
https://api.github.com/repos/gheber/kenzo
opened
ECL (15.2.21) test failure
bug test
-------------------------------- CHECK-HAT []: Unexpected Error: #<a SIMPLE-TYPE-ERROR> The assertion (EQ :EQUAL (CAT:A-CMPR3 CAT:CMPR CAT::DEL-I-DEL-J CAT::DEL-J-DEL-I+1)) failed.. --------------------------------
1.0
ECL (15.2.21) test failure - -------------------------------- CHECK-HAT []: Unexpected Error: #<a SIMPLE-TYPE-ERROR> The assertion (EQ :EQUAL (CAT:A-CMPR3 CAT:CMPR CAT::DEL-I-DEL-J CAT::DEL-J-DEL-I+1)) failed.. --------------------------------
non_defect
ecl test failure check hat unexpected error the assertion eq equal cat a cat cmpr cat del i del j cat del j del i failed
0
159,287
13,759,755,441
IssuesEvent
2020-10-07 03:59:25
dankamongmen/notcurses
https://api.github.com/repos/dankamongmen/notcurses
opened
reels ought support gapless mode
documentation enhancement
Currently, `ncreel` assumes (and enforces) exactly one line of gap between tablets. Users might want to run gapless, or even with multiline gaps. We ought support the former (0 lines of gap), and if we're gonna do that, it's really no trouble to support N lines of gap (except that 0 lines can be specified with a flag, while supporting arbitrarily many rows requires a new field in `ncreel_options`). Yeah, I think simply supporting gapless mode is sufficient. Extra gap lines can, after all, be effected by simply leaving empty lines in a tablet draw call. Yep, do the flag.
1.0
reels ought support gapless mode - Currently, `ncreel` assumes (and enforces) exactly one line of gap between tablets. Users might want to run gapless, or even with multiline gaps. We ought support the former (0 lines of gap), and if we're gonna do that, it's really no trouble to support N lines of gap (except that 0 lines can be specified with a flag, while supporting arbitrarily many rows requires a new field in `ncreel_options`). Yeah, I think simply supporting gapless mode is sufficient. Extra gap lines can, after all, be effected by simply leaving empty lines in a tablet draw call. Yep, do the flag.
non_defect
reels ought support gapless mode currently ncreel assumes and enforces exactly one line of gap between tablets users might want to run gapless or even with multiline gaps we ought support the former lines of gap and if we re gonna do that it s really no trouble to support n lines of gap except that lines can be specified with a flag while supporting arbitrarily many rows requires a new field in ncreel options yeah i think simply supporting gapless mode is sufficient extra gap lines can after all be effected by simply leaving empty lines in a tablet draw call yep do the flag
0
87,611
17,346,590,394
IssuesEvent
2021-07-29 00:17:11
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
superpmi: parallel mode doesn't handle crash of superpmi
area-CodeGen-coreclr
When investigating #56156, I found that if superpmi crashes in one of multiple superpmi parallel replay processes, that crash isn't caught and the superpmi replay "succeeds". This was on Linux/arm32. Perhaps we don't catch SIGBUS "Bus error" exceptions, compared to others? We do have try/catch in place. The crash is: ``` Thread 1 "superpmi" received signal SIGBUS, Bus error. Compiler::impImportStaticReadOnlyField (this=0xd17844, fldAddr=0xd665d9, lclTyp=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:7569 7569 lval = *((__int64*)fldAddr); (gdb) bt #0 Compiler::impImportStaticReadOnlyField (this=0xd17844, fldAddr=0xd665d9, lclTyp=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:7569 #1 0xf5f6e1ae in Compiler::impImportBlockCode (this=<optimized out>, block=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:14882 #2 0xf5f7266e in Compiler::impImportBlock(BasicBlock*)::$_0::operator()(Compiler::impImportBlock(BasicBlock*)::FilterVerificationExceptionsParam*) const (this=<optimized out>, pParam=0xfffed640) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:17695 #3 Compiler::impImportBlock (this=0xd17844, block=0xd25178) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:17705 #4 0xf5f746fc in Compiler::impImport (this=0xd17844) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:18778 #5 0xf5f131be in Compiler::fgImport (this=0xd17844) at /home/bruce/gh/runtime2/src/coreclr/jit/flowgraph.cpp:625 #6 0xf6018f36 in Phase::Run (this=0xfffed7b4) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.cpp:61 #7 0xf5ed3b20 in DoPhase (_compiler=0xd17844, _phase=PHASE_IMPORTATION, _action=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.h:136 #8 Compiler::compCompile (this=0xd17844, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:4511 #9 0xf5ed6aa2 in Compiler::compCompileHelper (this=0xd17844, classPtr=<optimized out>, compHnd=<optimized out>, methodInfo=<optimized out>, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:6407 #10 0xf5ed58d4 in Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::$_11::operator()(Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::__JITParam*) const ( this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5684 #11 Compiler::compCompile (this=0xd17844, classPtr=0xefe9f000, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5703 #12 0xf5ed745e in jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::{lambda(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*)#1}::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7053 #13 jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7078 #14 jitNativeCode (methodHnd=0xe7767664, classPtr=0xefe9f000, compHnd=0x5a9c58, methodInfo=0xd239f0, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0, inlineInfoPtr=0xfffede80) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7080 #15 0xf5f030cc in Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::$_0::operator()(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) const (this=<optimized out>, pParam=0xfffede6c) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:973 #16 Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::$_0::__invoke(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) (pParam=0xfffede6c) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:921 #17 0x004197c0 in RunWithErrorTrap(void (*)(void*), void*)::$_0::operator()(RunWithErrorTrap(void (*)(void*), void*)::TrapParam*) const (this=<optimized out>, pTrapParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi-shared/errorhandling.cpp:138 #18 RunWithErrorTrap (function=<optimized out>, param=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi-shared/errorhandling.cpp:140 #19 0xf5f0174a in Compiler::eeRunWithErrorTrap<Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param>(void (*)(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*), Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) (this=0xd147ac, function=0xd665d9, param=0xfffede6c) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.h:7957 #20 Compiler::fgInvokeInlineeCompiler (this=0xd147ac, call=0xd23860, inlineResult=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:920 #21 0xf5fee9f0 in Compiler::fgMorphCallInlineHelper (this=0xd147ac, call=0xd23860, result=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/morph.cpp:6721 #22 0xf5fee816 in Compiler::fgMorphCallInline (this=0xd147ac, call=0xd23860, inlineResult=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/morph.cpp:6603 #23 0xf5f0020a in Compiler::fgInline (this=0xd147ac) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:153 #24 0xf6018f36 in Phase::Run (this=0xfffee63c) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.cpp:61 #25 0xf5ed3cb6 in DoPhase (_compiler=0xd147ac, _phase=PHASE_MORPH_INLINE, _action=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.h:136 #26 Compiler::compCompile (this=0xd147ac, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:4684 #27 0xf5ed6aa2 in Compiler::compCompileHelper (this=0xd147ac, classPtr=<optimized out>, compHnd=<optimized out>, methodInfo=<optimized out>, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:6407 #28 0xf5ed58d4 in Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::$_11::operator()(Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::__JITParam*) const ( this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5684 #29 Compiler::compCompile (this=0xd147ac, classPtr=0xe77f360c, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5703 #30 0xf5ed745e in jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::{lambda(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*)#1}::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7053 #31 jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7078 #32 jitNativeCode (methodHnd=0xe77f9c54, classPtr=0xe77f360c, compHnd=0x5a9c58, methodInfo=0xfffeeda4, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0, inlineInfoPtr=0x0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7080 #33 0xf5edf25e in CILJit::compileMethod (this=<optimized out>, compHnd=0x5a9c58, methodInfo=0xfffeeda4, flags=<optimized out>, entryAddress=0xfffeed90, nativeSizeOfCode=0xfffeed8c) at /home/bruce/gh/runtime2/src/coreclr/jit/ee_il_dll.cpp:276 #34 0x0040d474 in JitInstance::CompileMethod(MethodContext*, int, bool)::$_0::operator()(JitInstance::CompileMethod(MethodContext*, int, bool)::Param*) const (this=<optimized out>, pParam=0xfffeed94) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/jitinstance.cpp:314 #35 JitInstance::CompileMethod (this=0x59d608, MethodToCompile=<optimized out>, mcIndex=0x2a03d, collectThroughput=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/jitinstance.cpp:373 #36 0x00410d90 in main (argc=<optimized out>, argv=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/superpmi.cpp:361 (gdb) ? fldAddr Undefined command: "". Try "help". (gdb) print fldAddr $1 = (void *) 0xd665d9 ``` The "successful" run looks like: ``` $ /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/superpmi -p /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/libclrjit.so /home/brucefo/bugs/spmicollect/small.mch Using child (/home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/superpmi) with args ( /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/libclrjit.so /home/brucefo/bugs/spmicollect/small.mch) workerCount=4, skipCleanup=0. Loaded 3 Jitted 3 FailedCompile 0 Excluded 0 Missing 0 ``` But there are 5 method contexts in this example: the first process that was supposed to handle 2 of them crashed.
1.0
superpmi: parallel mode doesn't handle crash of superpmi - When investigating #56156, I found that if superpmi crashes in one of multiple superpmi parallel replay processes, that crash isn't caught and the superpmi replay "succeeds". This was on Linux/arm32. Perhaps we don't catch SIGBUS "Bus error" exceptions, compared to others? We do have try/catch in place. The crash is: ``` Thread 1 "superpmi" received signal SIGBUS, Bus error. Compiler::impImportStaticReadOnlyField (this=0xd17844, fldAddr=0xd665d9, lclTyp=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:7569 7569 lval = *((__int64*)fldAddr); (gdb) bt #0 Compiler::impImportStaticReadOnlyField (this=0xd17844, fldAddr=0xd665d9, lclTyp=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:7569 #1 0xf5f6e1ae in Compiler::impImportBlockCode (this=<optimized out>, block=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:14882 #2 0xf5f7266e in Compiler::impImportBlock(BasicBlock*)::$_0::operator()(Compiler::impImportBlock(BasicBlock*)::FilterVerificationExceptionsParam*) const (this=<optimized out>, pParam=0xfffed640) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:17695 #3 Compiler::impImportBlock (this=0xd17844, block=0xd25178) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:17705 #4 0xf5f746fc in Compiler::impImport (this=0xd17844) at /home/bruce/gh/runtime2/src/coreclr/jit/importer.cpp:18778 #5 0xf5f131be in Compiler::fgImport (this=0xd17844) at /home/bruce/gh/runtime2/src/coreclr/jit/flowgraph.cpp:625 #6 0xf6018f36 in Phase::Run (this=0xfffed7b4) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.cpp:61 #7 0xf5ed3b20 in DoPhase (_compiler=0xd17844, _phase=PHASE_IMPORTATION, _action=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.h:136 #8 Compiler::compCompile (this=0xd17844, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:4511 #9 0xf5ed6aa2 in Compiler::compCompileHelper (this=0xd17844, classPtr=<optimized out>, compHnd=<optimized out>, methodInfo=<optimized out>, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:6407 #10 0xf5ed58d4 in Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::$_11::operator()(Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::__JITParam*) const ( this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5684 #11 Compiler::compCompile (this=0xd17844, classPtr=0xefe9f000, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5703 #12 0xf5ed745e in jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::{lambda(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*)#1}::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7053 #13 jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7078 #14 jitNativeCode (methodHnd=0xe7767664, classPtr=0xefe9f000, compHnd=0x5a9c58, methodInfo=0xd239f0, methodCodePtr=0xfffede80, methodCodeSize=0x0, compileFlags=0xfffeddf0, inlineInfoPtr=0xfffede80) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7080 #15 0xf5f030cc in Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::$_0::operator()(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) const (this=<optimized out>, pParam=0xfffede6c) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:973 #16 Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::$_0::__invoke(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) (pParam=0xfffede6c) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:921 #17 0x004197c0 in RunWithErrorTrap(void (*)(void*), void*)::$_0::operator()(RunWithErrorTrap(void (*)(void*), void*)::TrapParam*) const (this=<optimized out>, pTrapParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi-shared/errorhandling.cpp:138 #18 RunWithErrorTrap (function=<optimized out>, param=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi-shared/errorhandling.cpp:140 #19 0xf5f0174a in Compiler::eeRunWithErrorTrap<Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param>(void (*)(Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*), Compiler::fgInvokeInlineeCompiler(GenTreeCall*, InlineResult*)::Param*) (this=0xd147ac, function=0xd665d9, param=0xfffede6c) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.h:7957 #20 Compiler::fgInvokeInlineeCompiler (this=0xd147ac, call=0xd23860, inlineResult=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:920 #21 0xf5fee9f0 in Compiler::fgMorphCallInlineHelper (this=0xd147ac, call=0xd23860, result=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/morph.cpp:6721 #22 0xf5fee816 in Compiler::fgMorphCallInline (this=0xd147ac, call=0xd23860, inlineResult=0xfffee4f8) at /home/bruce/gh/runtime2/src/coreclr/jit/morph.cpp:6603 #23 0xf5f0020a in Compiler::fgInline (this=0xd147ac) at /home/bruce/gh/runtime2/src/coreclr/jit/fginline.cpp:153 #24 0xf6018f36 in Phase::Run (this=0xfffee63c) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.cpp:61 #25 0xf5ed3cb6 in DoPhase (_compiler=0xd147ac, _phase=PHASE_MORPH_INLINE, _action=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/phase.h:136 #26 Compiler::compCompile (this=0xd147ac, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:4684 #27 0xf5ed6aa2 in Compiler::compCompileHelper (this=0xd147ac, classPtr=<optimized out>, compHnd=<optimized out>, methodInfo=<optimized out>, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:6407 #28 0xf5ed58d4 in Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::$_11::operator()(Compiler::compCompile(CORINFO_MODULE_STRUCT_*, void**, unsigned int*, JitFlags*)::__JITParam*) const ( this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5684 #29 Compiler::compCompile (this=0xd147ac, classPtr=0xe77f360c, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:5703 #30 0xf5ed745e in jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::{lambda(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*)#1}::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7053 #31 jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::$_13::operator()(jitNativeCode(CORINFO_METHOD_STRUCT_*, CORINFO_MODULE_STRUCT_*, ICorJitInfo*, CORINFO_METHOD_INFO*, void**, unsigned int*, JitFlags*, void*)::__JITParam*) const (this=<optimized out>, __JITpParam=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7078 #32 jitNativeCode (methodHnd=0xe77f9c54, classPtr=0xe77f360c, compHnd=0x5a9c58, methodInfo=0xfffeeda4, methodCodePtr=0xfffeec8c, methodCodeSize=0xfffeed8c, compileFlags=0xfffeeca0, inlineInfoPtr=0x0) at /home/bruce/gh/runtime2/src/coreclr/jit/compiler.cpp:7080 #33 0xf5edf25e in CILJit::compileMethod (this=<optimized out>, compHnd=0x5a9c58, methodInfo=0xfffeeda4, flags=<optimized out>, entryAddress=0xfffeed90, nativeSizeOfCode=0xfffeed8c) at /home/bruce/gh/runtime2/src/coreclr/jit/ee_il_dll.cpp:276 #34 0x0040d474 in JitInstance::CompileMethod(MethodContext*, int, bool)::$_0::operator()(JitInstance::CompileMethod(MethodContext*, int, bool)::Param*) const (this=<optimized out>, pParam=0xfffeed94) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/jitinstance.cpp:314 #35 JitInstance::CompileMethod (this=0x59d608, MethodToCompile=<optimized out>, mcIndex=0x2a03d, collectThroughput=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/jitinstance.cpp:373 #36 0x00410d90 in main (argc=<optimized out>, argv=<optimized out>) at /home/bruce/gh/runtime2/src/coreclr/ToolBox/superpmi/superpmi/superpmi.cpp:361 (gdb) ? fldAddr Undefined command: "". Try "help". (gdb) print fldAddr $1 = (void *) 0xd665d9 ``` The "successful" run looks like: ``` $ /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/superpmi -p /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/libclrjit.so /home/brucefo/bugs/spmicollect/small.mch Using child (/home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/superpmi) with args ( /home/brucefo/build/runtime/artifacts/tests/coreclr/Linux.arm.Checked/Tests/Core_Root/libclrjit.so /home/brucefo/bugs/spmicollect/small.mch) workerCount=4, skipCleanup=0. Loaded 3 Jitted 3 FailedCompile 0 Excluded 0 Missing 0 ``` But there are 5 method contexts in this example: the first process that was supposed to handle 2 of them crashed.
non_defect
superpmi parallel mode doesn t handle crash of superpmi when investigating i found that if superpmi crashes in one of multiple superpmi parallel replay processes that crash isn t caught and the superpmi replay succeeds this was on linux perhaps we don t catch sigbus bus error exceptions compared to others we do have try catch in place the crash is thread superpmi received signal sigbus bus error compiler impimportstaticreadonlyfield this fldaddr lcltyp at home bruce gh src coreclr jit importer cpp lval fldaddr gdb bt compiler impimportstaticreadonlyfield this fldaddr lcltyp at home bruce gh src coreclr jit importer cpp in compiler impimportblockcode this block at home bruce gh src coreclr jit importer cpp in compiler impimportblock basicblock operator compiler impimportblock basicblock filterverificationexceptionsparam const this pparam at home bruce gh src coreclr jit importer cpp compiler impimportblock this block at home bruce gh src coreclr jit importer cpp in compiler impimport this at home bruce gh src coreclr jit importer cpp in compiler fgimport this at home bruce gh src coreclr jit flowgraph cpp in phase run this at home bruce gh src coreclr jit phase cpp in dophase compiler phase phase importation action at home bruce gh src coreclr jit phase h compiler compcompile this methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in compiler compcompilehelper this classptr comphnd methodinfo methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in compiler compcompile corinfo module struct void unsigned int jitflags operator compiler compcompile corinfo module struct void unsigned int jitflags jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp compiler compcompile this classptr methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const lambda jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const jitparam operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp jitnativecode methodhnd classptr comphnd methodinfo methodcodeptr methodcodesize compileflags inlineinfoptr at home bruce gh src coreclr jit compiler cpp in compiler fginvokeinlineecompiler gentreecall inlineresult operator compiler fginvokeinlineecompiler gentreecall inlineresult param const this pparam at home bruce gh src coreclr jit fginline cpp compiler fginvokeinlineecompiler gentreecall inlineresult invoke compiler fginvokeinlineecompiler gentreecall inlineresult param pparam at home bruce gh src coreclr jit fginline cpp in runwitherrortrap void void void operator runwitherrortrap void void void trapparam const this ptrapparam at home bruce gh src coreclr toolbox superpmi superpmi shared errorhandling cpp runwitherrortrap function param at home bruce gh src coreclr toolbox superpmi superpmi shared errorhandling cpp in compiler eerunwitherrortrap void compiler fginvokeinlineecompiler gentreecall inlineresult param compiler fginvokeinlineecompiler gentreecall inlineresult param this function param at home bruce gh src coreclr jit compiler h compiler fginvokeinlineecompiler this call inlineresult at home bruce gh src coreclr jit fginline cpp in compiler fgmorphcallinlinehelper this call result at home bruce gh src coreclr jit morph cpp in compiler fgmorphcallinline this call inlineresult at home bruce gh src coreclr jit morph cpp in compiler fginline this at home bruce gh src coreclr jit fginline cpp in phase run this at home bruce gh src coreclr jit phase cpp in dophase compiler phase phase morph inline action at home bruce gh src coreclr jit phase h compiler compcompile this methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in compiler compcompilehelper this classptr comphnd methodinfo methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in compiler compcompile corinfo module struct void unsigned int jitflags operator compiler compcompile corinfo module struct void unsigned int jitflags jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp compiler compcompile this classptr methodcodeptr methodcodesize compileflags at home bruce gh src coreclr jit compiler cpp in jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const lambda jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const jitparam operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void operator jitnativecode corinfo method struct corinfo module struct icorjitinfo corinfo method info void unsigned int jitflags void jitparam const this jitpparam at home bruce gh src coreclr jit compiler cpp jitnativecode methodhnd classptr comphnd methodinfo methodcodeptr methodcodesize compileflags inlineinfoptr at home bruce gh src coreclr jit compiler cpp in ciljit compilemethod this comphnd methodinfo flags entryaddress nativesizeofcode at home bruce gh src coreclr jit ee il dll cpp in jitinstance compilemethod methodcontext int bool operator jitinstance compilemethod methodcontext int bool param const this pparam at home bruce gh src coreclr toolbox superpmi superpmi jitinstance cpp jitinstance compilemethod this methodtocompile mcindex collectthroughput at home bruce gh src coreclr toolbox superpmi superpmi jitinstance cpp in main argc argv at home bruce gh src coreclr toolbox superpmi superpmi superpmi cpp gdb fldaddr undefined command try help gdb print fldaddr void the successful run looks like home brucefo build runtime artifacts tests coreclr linux arm checked tests core root superpmi p home brucefo build runtime artifacts tests coreclr linux arm checked tests core root libclrjit so home brucefo bugs spmicollect small mch using child home brucefo build runtime artifacts tests coreclr linux arm checked tests core root superpmi with args home brucefo build runtime artifacts tests coreclr linux arm checked tests core root libclrjit so home brucefo bugs spmicollect small mch workercount skipcleanup loaded jitted failedcompile excluded missing but there are method contexts in this example the first process that was supposed to handle of them crashed
0
35,520
7,759,210,765
IssuesEvent
2018-05-31 22:21:58
NREL/EnergyPlus
https://api.github.com/repos/NREL/EnergyPlus
closed
Multiple GenericContam controllers are not set correctly
Defect
Issue overview -------------- Issue #6565 fixed a bug to set multiple CO2 controllers correctly. The same issue occurs to handle multiple generic contam controllers. ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
1.0
Multiple GenericContam controllers are not set correctly - Issue overview -------------- Issue #6565 fixed a bug to set multiple CO2 controllers correctly. The same issue occurs to handle multiple generic contam controllers. ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
defect
multiple genericcontam controllers are not set correctly issue overview issue fixed a bug to set multiple controllers correctly the same issue occurs to handle multiple generic contam controllers details some additional details for this issue if relevant platform operating system version version of energyplus if using an intermediate build include sha unmethours link or helpdesk ticket number checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
1
96,997
3,980,280,968
IssuesEvent
2016-05-06 06:29:53
xcat2/xcat-core
https://api.github.com/repos/xcat2/xcat-core
closed
[FVT] [21 Apr 2016] Need verify Ubuntu 16.04 final release support against the latest xCAT master build
component:os_provision priority:high type:feature
Ubuntu 16.04 will be released on April 21st, 2016. Please refer the Ubuntu 16.04 release schedule [1]. [1] https://wiki.ubuntu.com/XenialXerus/ReleaseSchedule
1.0
[FVT] [21 Apr 2016] Need verify Ubuntu 16.04 final release support against the latest xCAT master build - Ubuntu 16.04 will be released on April 21st, 2016. Please refer the Ubuntu 16.04 release schedule [1]. [1] https://wiki.ubuntu.com/XenialXerus/ReleaseSchedule
non_defect
need verify ubuntu final release support against the latest xcat master build ubuntu will be released on april please refer the ubuntu release schedule
0
73,164
24,480,751,445
IssuesEvent
2022-10-08 20:00:54
FreeRADIUS/freeradius-server
https://api.github.com/repos/FreeRADIUS/freeradius-server
opened
sqlcounter[103]: Counter attribute Session-Timeout MUST be integer64
defect
### What type of defect/bug is this? Unexpected behaviour (obvious or verified by project member) ### How can the issue be reproduced? @alandekok how to fix this? ### Log output from the FreeRADIUS daemon ```shell BusyBox v1.35.0 (2022-10-05 19:50:01 UTC) built-in shell (ash) _______ ________ __ | |.-----.-----.-----.| | | |.----.| |_ | - || _ | -__| || | | || _|| _| |_______|| __|_____|__|__||________||__| |____| |__| W I R E L E S S F R E E D O M ----------------------------------------------------- OpenWrt 22.03.0, r19685-512e76967f ----------------------------------------------------- === WARNING! ===================================== There is no root password defined on this device! Use the "passwd" command to set up a new password in order to prevent unauthorized SSH logins. -------------------------------------------------- root@OpenWrt:~# radiusd -X FreeRADIUS Version 3.0.21 Copyright (C) 1999-2019 The FreeRADIUS server project and contributors There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE You may redistribute copies of FreeRADIUS under the terms of the GNU General Public License For more information about these matters, see the file named COPYRIGHT Starting - reading configuration files ... including dictionary file /usr/share/freeradius3/dictionary including dictionary file /etc/freeradius3/dictionary including configuration file /etc/freeradius3/radiusd.conf including configuration file /etc/freeradius3/clients.conf including files in directory /etc/freeradius3/mods-enabled/ including configuration file /etc/freeradius3/mods-enabled/expiration including configuration file /etc/freeradius3/mods-enabled/expr including configuration file /etc/freeradius3/mods-available/sql including configuration file /etc/freeradius3/mods-config/sql/main/mysql/queries.conf including configuration file /etc/freeradius3/mods-available/sqlcounter including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/dailycounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/monthlycounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/expire_on_login.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/noresetcounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/chillispot_max_bytes.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/max_all_mb.conf including files in directory /etc/freeradius3/sites-enabled/ including configuration file /etc/freeradius3/sites-enabled/default main { security { allow_core_dumps = no } name = "radiusd" prefix = "/usr" localstatedir = "/var" logdir = "/var/log" run_dir = "/var/run/radiusd" } main { name = "radiusd" prefix = "/usr" localstatedir = "/var" sbindir = "/usr/sbin" logdir = "/var/log" run_dir = "/var/run/radiusd" libdir = "/usr/lib/freeradius3" radacctdir = "/var/db/radacct" hostname_lookups = no max_request_time = 30 cleanup_delay = 5 max_requests = 16384 pidfile = "/var/run/radiusd/radiusd.pid" checkrad = "/usr/sbin/checkrad" debug_level = 0 proxy_requests = yes log { stripped_names = no auth = no auth_badpass = no auth_goodpass = no colourise = yes msg_denied = "You are already logged in - access denied" } resources { } security { max_attributes = 200 reject_delay = 1.000000 status_server = yes } } radiusd: #### Loading Realms and Home Servers #### radiusd: #### Loading Clients #### client 192.168.0.1 { ipaddr = 192.168.0.1 require_message_authenticator = no secret = <<< secret >>> nas_type = "other" proto = "*" limit { max_connections = 16 lifetime = 0 idle_timeout = 30 } } Debugger not attached # Creating Auth-Type = mschap # Creating Auth-Type = digest radiusd: #### Instantiating modules #### modules { # Loaded module rlm_expiration # Loading module "expiration" from file /etc/freeradius3/mods-enabled/expiration # Loaded module rlm_expr # Loading module "expr" from file /etc/freeradius3/mods-enabled/expr expr { safe_characters = "@abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789.-_: /äéöüàâæçèéêëîïôœùûüaÿÄÉÖÜßÀÂÆÇÈÉÊËÎÏÔŒÙÛÜŸ" } # Loaded module rlm_sql # Loading module "sql" from file /etc/freeradius3/mods-available/sql sql { driver = "rlm_sql_mysql" server = "localhost" port = 0 login = "easyhotspot" password = <<< secret >>> radius_db = "hotspot" read_groups = yes read_profiles = yes read_clients = no delete_stale_sessions = yes sql_user_name = "%{User-Name}" default_user_profile = "" client_query = "SELECT id, nasname, shortname, type, secret, server FROM nas" authorize_check_query = "SELECT id, username, attribute, value, op FROM radcheck WHERE username = '%{SQL-User-Name}' ORDER BY id" authorize_reply_query = "SELECT id, username, attribute, value, op FROM radreply WHERE username = '%{SQL-User-Name}' ORDER BY id" authorize_group_check_query = "SELECT id, groupname, attribute, Value, op FROM radgroupcheck WHERE groupname = '%{SQL-Group}' ORDER BY id" authorize_group_reply_query = "SELECT id, groupname, attribute, value, op FROM radgroupreply WHERE groupname = '%{SQL-Group}' ORDER BY id" group_membership_query = "SELECT groupname FROM radusergroup WHERE username = '%{SQL-User-Name}' ORDER BY priority" simul_count_query = "SELECT COUNT(*) FROM radacct WHERE username = '%{SQL-User-Name}' AND acctstoptime IS NULL" simul_verify_query = "SELECT radacctid, acctsessionid, username, nasipaddress, nasportid, framedipaddress, callingstationid, framedprotocol FROM radacct WHERE username = '%{SQL-User-Name}' AND acctstoptime IS NULL" safe_characters = "@abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789.-_: /" auto_escape = no accounting { reference = "%{tolower:type.%{%{Acct-Status-Type}:-%{Request-Processing-Stage}}.query}" type { accounting-on { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = '%{%{integer:Event-Timestamp}:-%l}' - UNIX_TIMESTAMP(acctstarttime), acctterminatecause = '%{%{Acct-Terminate-Cause}:-NAS-Reboot}' WHERE acctstoptime IS NULL AND nasipaddress = '%{NAS-IP-Address}' AND acctstarttime <= FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l})" } accounting-off { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = '%{%{integer:Event-Timestamp}:-%l}' - UNIX_TIMESTAMP(acctstarttime), acctterminatecause = '%{%{Acct-Terminate-Cause}:-NAS-Reboot}' WHERE acctstoptime IS NULL AND nasipaddress = '%{NAS-IP-Address}' AND acctstarttime <= FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l})" } start { query = "INSERT INTO radacct (acctsessionid, acctuniqueid, username, realm, nasipaddress, nasportid, nasporttype, acctstarttime, acctupdatetime, acctstoptime, acctsessiontime, acctauthentic, connectinfo_start, connectinfo_stop, acctinputoctets, acctoutputoctets, calledstationid, callingstationid, acctterminatecause, servicetype, framedprotocol, framedipaddress, framedipv6address, framedipv6prefix, framedinterfaceid, delegatedipv6prefix) VALUES ('%{Acct-Session-Id}', '%{Acct-Unique-Session-Id}', '%{SQL-User-Name}', '%{Realm}', '%{NAS-IP-Address}', '%{%{NAS-Port-ID}:-%{NAS-Port}}', '%{NAS-Port-Type}', FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), NULL, '0', '%{Acct-Authentic}', '%{Connect-Info}', '', '0', '0', '%{Called-Station-Id}', '%{Calling-Station-Id}', '', '%{Service-Type}', '%{Framed-Protocol}', '%{Framed-IP-Address}', '%{Framed-IPv6-Address}', '%{Framed-IPv6-Prefix}', '%{Framed-Interface-Id}', '%{Delegated-IPv6-Prefix}')" } interim-update { query = "UPDATE radacct SET acctupdatetime = (@acctupdatetime_old:=acctupdatetime), acctupdatetime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctinterval = %{%{integer:Event-Timestamp}:-%l} - UNIX_TIMESTAMP(@acctupdatetime_old), framedipaddress = '%{Framed-IP-Address}', framedipv6address = '%{Framed-IPv6-Address}', framedipv6prefix = '%{Framed-IPv6-Prefix}', framedinterfaceid = '%{Framed-Interface-Id}', delegatedipv6prefix = '%{Delegated-IPv6-Prefix}', acctsessiontime = %{%{Acct-Session-Time}:-NULL}, acctinputoctets = '%{%{Acct-Input-Gigawords}:-0}' << 32 | '%{%{Acct-Input-Octets}:-0}', acctoutputoctets = '%{%{Acct-Output-Gigawords}:-0}' << 32 | '%{%{Acct-Output-Octets}:-0}' WHERE AcctUniqueId = '%{Acct-Unique-Session-Id}'" } stop { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = %{%{Acct-Session-Time}:-NULL}, acctinputoctets = '%{%{Acct-Input-Gigawords}:-0}' << 32 | '%{%{Acct-Input-Octets}:-0}', acctoutputoctets = '%{%{Acct-Output-Gigawords}:-0}' << 32 | '%{%{Acct-Output-Octets}:-0}', acctterminatecause = '%{Acct-Terminate-Cause}', connectinfo_stop = '%{Connect-Info}' WHERE AcctUniqueId = '%{Acct-Unique-Session-Id}'" } } } post-auth { reference = ".query" query = "INSERT INTO radpostauth (username, pass, reply, authdate) VALUES ( '%{SQL-User-Name}', '%{%{User-Password}:-%{Chap-Password}}', '%{reply:Packet-Type}', '%S.%M')" } } rlm_sql (sql): Driver rlm_sql_mysql (module rlm_sql_mysql) loaded and linked Creating attribute SQL-Group # Loaded module rlm_sqlcounter # Loading module "dailycounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter dailycounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(acctsessiontime - GREATEST((%%b - UNIX_TIMESTAMP(acctstarttime)), 0)) FROM radacct WHERE username = '%{User-Name}' AND UNIX_TIMESTAMP(acctstarttime) + acctsessiontime > '%%b'" reset = "daily" counter_name = "Daily-Session-Time" check_name = "Max-Daily-Session" reply_name = "Session-Timeout" } # Loading module "monthlycounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter monthlycounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(acctsessiontime - GREATEST((%%b - UNIX_TIMESTAMP(acctstarttime)), 0)) FROM radacct WHERE username='%{User-Name}' AND UNIX_TIMESTAMP(acctstarttime) + acctsessiontime > '%%b'" reset = "monthly" counter_name = "Monthly-Session-Time" check_name = "Max-Monthly-Session" reply_name = "Session-Timeout" } # Loading module "expire_on_login" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter expire_on_login { sql_module_instance = "sql" key = "User-Name" query = "SELECT IFNULL( MAX(TIME_TO_SEC(TIMEDIFF(NOW(), acctstarttime))),0) FROM radacct WHERE UserName='%{User-Name}' ORDER BY acctstarttime LIMIT 1;" reset = "never" counter_name = "Expire-After-Initial-Login" check_name = "Expire-After" reply_name = "Session-Timeout" } # Loading module "noresetcounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter noresetcounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(Acctsessiontime) FROM radacct WHERE UserName='%{User-Name}'" reset = "never" counter_name = "Session-Timeout" check_name = "Session-Timeout" reply_name = "Session-Timeout" } /etc/freeradius3/mods-available/sqlcounter[103]: Counter attribute Session-Timeout MUST be integer64 /etc/freeradius3/mods-available/sqlcounter[103]: Instantiation failed for module "noresetcounter" root@OpenWrt:~#``` ``` ### Relevant log output from client utilities _No response_ ### Backtrace from LLDB or GDB _No response_
1.0
sqlcounter[103]: Counter attribute Session-Timeout MUST be integer64 - ### What type of defect/bug is this? Unexpected behaviour (obvious or verified by project member) ### How can the issue be reproduced? @alandekok how to fix this? ### Log output from the FreeRADIUS daemon ```shell BusyBox v1.35.0 (2022-10-05 19:50:01 UTC) built-in shell (ash) _______ ________ __ | |.-----.-----.-----.| | | |.----.| |_ | - || _ | -__| || | | || _|| _| |_______|| __|_____|__|__||________||__| |____| |__| W I R E L E S S F R E E D O M ----------------------------------------------------- OpenWrt 22.03.0, r19685-512e76967f ----------------------------------------------------- === WARNING! ===================================== There is no root password defined on this device! Use the "passwd" command to set up a new password in order to prevent unauthorized SSH logins. -------------------------------------------------- root@OpenWrt:~# radiusd -X FreeRADIUS Version 3.0.21 Copyright (C) 1999-2019 The FreeRADIUS server project and contributors There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE You may redistribute copies of FreeRADIUS under the terms of the GNU General Public License For more information about these matters, see the file named COPYRIGHT Starting - reading configuration files ... including dictionary file /usr/share/freeradius3/dictionary including dictionary file /etc/freeradius3/dictionary including configuration file /etc/freeradius3/radiusd.conf including configuration file /etc/freeradius3/clients.conf including files in directory /etc/freeradius3/mods-enabled/ including configuration file /etc/freeradius3/mods-enabled/expiration including configuration file /etc/freeradius3/mods-enabled/expr including configuration file /etc/freeradius3/mods-available/sql including configuration file /etc/freeradius3/mods-config/sql/main/mysql/queries.conf including configuration file /etc/freeradius3/mods-available/sqlcounter including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/dailycounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/monthlycounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/expire_on_login.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/noresetcounter.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/chillispot_max_bytes.conf including configuration file /etc/freeradius3/mods-config/sql/counter/mysql/max_all_mb.conf including files in directory /etc/freeradius3/sites-enabled/ including configuration file /etc/freeradius3/sites-enabled/default main { security { allow_core_dumps = no } name = "radiusd" prefix = "/usr" localstatedir = "/var" logdir = "/var/log" run_dir = "/var/run/radiusd" } main { name = "radiusd" prefix = "/usr" localstatedir = "/var" sbindir = "/usr/sbin" logdir = "/var/log" run_dir = "/var/run/radiusd" libdir = "/usr/lib/freeradius3" radacctdir = "/var/db/radacct" hostname_lookups = no max_request_time = 30 cleanup_delay = 5 max_requests = 16384 pidfile = "/var/run/radiusd/radiusd.pid" checkrad = "/usr/sbin/checkrad" debug_level = 0 proxy_requests = yes log { stripped_names = no auth = no auth_badpass = no auth_goodpass = no colourise = yes msg_denied = "You are already logged in - access denied" } resources { } security { max_attributes = 200 reject_delay = 1.000000 status_server = yes } } radiusd: #### Loading Realms and Home Servers #### radiusd: #### Loading Clients #### client 192.168.0.1 { ipaddr = 192.168.0.1 require_message_authenticator = no secret = <<< secret >>> nas_type = "other" proto = "*" limit { max_connections = 16 lifetime = 0 idle_timeout = 30 } } Debugger not attached # Creating Auth-Type = mschap # Creating Auth-Type = digest radiusd: #### Instantiating modules #### modules { # Loaded module rlm_expiration # Loading module "expiration" from file /etc/freeradius3/mods-enabled/expiration # Loaded module rlm_expr # Loading module "expr" from file /etc/freeradius3/mods-enabled/expr expr { safe_characters = "@abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789.-_: /äéöüàâæçèéêëîïôœùûüaÿÄÉÖÜßÀÂÆÇÈÉÊËÎÏÔŒÙÛÜŸ" } # Loaded module rlm_sql # Loading module "sql" from file /etc/freeradius3/mods-available/sql sql { driver = "rlm_sql_mysql" server = "localhost" port = 0 login = "easyhotspot" password = <<< secret >>> radius_db = "hotspot" read_groups = yes read_profiles = yes read_clients = no delete_stale_sessions = yes sql_user_name = "%{User-Name}" default_user_profile = "" client_query = "SELECT id, nasname, shortname, type, secret, server FROM nas" authorize_check_query = "SELECT id, username, attribute, value, op FROM radcheck WHERE username = '%{SQL-User-Name}' ORDER BY id" authorize_reply_query = "SELECT id, username, attribute, value, op FROM radreply WHERE username = '%{SQL-User-Name}' ORDER BY id" authorize_group_check_query = "SELECT id, groupname, attribute, Value, op FROM radgroupcheck WHERE groupname = '%{SQL-Group}' ORDER BY id" authorize_group_reply_query = "SELECT id, groupname, attribute, value, op FROM radgroupreply WHERE groupname = '%{SQL-Group}' ORDER BY id" group_membership_query = "SELECT groupname FROM radusergroup WHERE username = '%{SQL-User-Name}' ORDER BY priority" simul_count_query = "SELECT COUNT(*) FROM radacct WHERE username = '%{SQL-User-Name}' AND acctstoptime IS NULL" simul_verify_query = "SELECT radacctid, acctsessionid, username, nasipaddress, nasportid, framedipaddress, callingstationid, framedprotocol FROM radacct WHERE username = '%{SQL-User-Name}' AND acctstoptime IS NULL" safe_characters = "@abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789.-_: /" auto_escape = no accounting { reference = "%{tolower:type.%{%{Acct-Status-Type}:-%{Request-Processing-Stage}}.query}" type { accounting-on { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = '%{%{integer:Event-Timestamp}:-%l}' - UNIX_TIMESTAMP(acctstarttime), acctterminatecause = '%{%{Acct-Terminate-Cause}:-NAS-Reboot}' WHERE acctstoptime IS NULL AND nasipaddress = '%{NAS-IP-Address}' AND acctstarttime <= FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l})" } accounting-off { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = '%{%{integer:Event-Timestamp}:-%l}' - UNIX_TIMESTAMP(acctstarttime), acctterminatecause = '%{%{Acct-Terminate-Cause}:-NAS-Reboot}' WHERE acctstoptime IS NULL AND nasipaddress = '%{NAS-IP-Address}' AND acctstarttime <= FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l})" } start { query = "INSERT INTO radacct (acctsessionid, acctuniqueid, username, realm, nasipaddress, nasportid, nasporttype, acctstarttime, acctupdatetime, acctstoptime, acctsessiontime, acctauthentic, connectinfo_start, connectinfo_stop, acctinputoctets, acctoutputoctets, calledstationid, callingstationid, acctterminatecause, servicetype, framedprotocol, framedipaddress, framedipv6address, framedipv6prefix, framedinterfaceid, delegatedipv6prefix) VALUES ('%{Acct-Session-Id}', '%{Acct-Unique-Session-Id}', '%{SQL-User-Name}', '%{Realm}', '%{NAS-IP-Address}', '%{%{NAS-Port-ID}:-%{NAS-Port}}', '%{NAS-Port-Type}', FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), NULL, '0', '%{Acct-Authentic}', '%{Connect-Info}', '', '0', '0', '%{Called-Station-Id}', '%{Calling-Station-Id}', '', '%{Service-Type}', '%{Framed-Protocol}', '%{Framed-IP-Address}', '%{Framed-IPv6-Address}', '%{Framed-IPv6-Prefix}', '%{Framed-Interface-Id}', '%{Delegated-IPv6-Prefix}')" } interim-update { query = "UPDATE radacct SET acctupdatetime = (@acctupdatetime_old:=acctupdatetime), acctupdatetime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctinterval = %{%{integer:Event-Timestamp}:-%l} - UNIX_TIMESTAMP(@acctupdatetime_old), framedipaddress = '%{Framed-IP-Address}', framedipv6address = '%{Framed-IPv6-Address}', framedipv6prefix = '%{Framed-IPv6-Prefix}', framedinterfaceid = '%{Framed-Interface-Id}', delegatedipv6prefix = '%{Delegated-IPv6-Prefix}', acctsessiontime = %{%{Acct-Session-Time}:-NULL}, acctinputoctets = '%{%{Acct-Input-Gigawords}:-0}' << 32 | '%{%{Acct-Input-Octets}:-0}', acctoutputoctets = '%{%{Acct-Output-Gigawords}:-0}' << 32 | '%{%{Acct-Output-Octets}:-0}' WHERE AcctUniqueId = '%{Acct-Unique-Session-Id}'" } stop { query = "UPDATE radacct SET acctstoptime = FROM_UNIXTIME(%{%{integer:Event-Timestamp}:-%l}), acctsessiontime = %{%{Acct-Session-Time}:-NULL}, acctinputoctets = '%{%{Acct-Input-Gigawords}:-0}' << 32 | '%{%{Acct-Input-Octets}:-0}', acctoutputoctets = '%{%{Acct-Output-Gigawords}:-0}' << 32 | '%{%{Acct-Output-Octets}:-0}', acctterminatecause = '%{Acct-Terminate-Cause}', connectinfo_stop = '%{Connect-Info}' WHERE AcctUniqueId = '%{Acct-Unique-Session-Id}'" } } } post-auth { reference = ".query" query = "INSERT INTO radpostauth (username, pass, reply, authdate) VALUES ( '%{SQL-User-Name}', '%{%{User-Password}:-%{Chap-Password}}', '%{reply:Packet-Type}', '%S.%M')" } } rlm_sql (sql): Driver rlm_sql_mysql (module rlm_sql_mysql) loaded and linked Creating attribute SQL-Group # Loaded module rlm_sqlcounter # Loading module "dailycounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter dailycounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(acctsessiontime - GREATEST((%%b - UNIX_TIMESTAMP(acctstarttime)), 0)) FROM radacct WHERE username = '%{User-Name}' AND UNIX_TIMESTAMP(acctstarttime) + acctsessiontime > '%%b'" reset = "daily" counter_name = "Daily-Session-Time" check_name = "Max-Daily-Session" reply_name = "Session-Timeout" } # Loading module "monthlycounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter monthlycounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(acctsessiontime - GREATEST((%%b - UNIX_TIMESTAMP(acctstarttime)), 0)) FROM radacct WHERE username='%{User-Name}' AND UNIX_TIMESTAMP(acctstarttime) + acctsessiontime > '%%b'" reset = "monthly" counter_name = "Monthly-Session-Time" check_name = "Max-Monthly-Session" reply_name = "Session-Timeout" } # Loading module "expire_on_login" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter expire_on_login { sql_module_instance = "sql" key = "User-Name" query = "SELECT IFNULL( MAX(TIME_TO_SEC(TIMEDIFF(NOW(), acctstarttime))),0) FROM radacct WHERE UserName='%{User-Name}' ORDER BY acctstarttime LIMIT 1;" reset = "never" counter_name = "Expire-After-Initial-Login" check_name = "Expire-After" reply_name = "Session-Timeout" } # Loading module "noresetcounter" from file /etc/freeradius3/mods-available/sqlcounter sqlcounter noresetcounter { sql_module_instance = "sql" key = "User-Name" query = "SELECT SUM(Acctsessiontime) FROM radacct WHERE UserName='%{User-Name}'" reset = "never" counter_name = "Session-Timeout" check_name = "Session-Timeout" reply_name = "Session-Timeout" } /etc/freeradius3/mods-available/sqlcounter[103]: Counter attribute Session-Timeout MUST be integer64 /etc/freeradius3/mods-available/sqlcounter[103]: Instantiation failed for module "noresetcounter" root@OpenWrt:~#``` ``` ### Relevant log output from client utilities _No response_ ### Backtrace from LLDB or GDB _No response_
defect
sqlcounter counter attribute session timeout must be what type of defect bug is this unexpected behaviour obvious or verified by project member how can the issue be reproduced alandekok how to fix this log output from the freeradius daemon shell busybox utc built in shell ash w i r e l e s s f r e e d o m openwrt warning there is no root password defined on this device use the passwd command to set up a new password in order to prevent unauthorized ssh logins root openwrt radiusd x freeradius version copyright c the freeradius server project and contributors there is no warranty not even for merchantability or fitness for a particular purpose you may redistribute copies of freeradius under the terms of the gnu general public license for more information about these matters see the file named copyright starting reading configuration files including dictionary file usr share dictionary including dictionary file etc dictionary including configuration file etc radiusd conf including configuration file etc clients conf including files in directory etc mods enabled including configuration file etc mods enabled expiration including configuration file etc mods enabled expr including configuration file etc mods available sql including configuration file etc mods config sql main mysql queries conf including configuration file etc mods available sqlcounter including configuration file etc mods config sql counter mysql dailycounter conf including configuration file etc mods config sql counter mysql monthlycounter conf including configuration file etc mods config sql counter mysql expire on login conf including configuration file etc mods config sql counter mysql noresetcounter conf including configuration file etc mods config sql counter mysql chillispot max bytes conf including configuration file etc mods config sql counter mysql max all mb conf including files in directory etc sites enabled including configuration file etc sites enabled default main security allow core dumps no name radiusd prefix usr localstatedir var logdir var log run dir var run radiusd main name radiusd prefix usr localstatedir var sbindir usr sbin logdir var log run dir var run radiusd libdir usr lib radacctdir var db radacct hostname lookups no max request time cleanup delay max requests pidfile var run radiusd radiusd pid checkrad usr sbin checkrad debug level proxy requests yes log stripped names no auth no auth badpass no auth goodpass no colourise yes msg denied you are already logged in access denied resources security max attributes reject delay status server yes radiusd loading realms and home servers radiusd loading clients client ipaddr require message authenticator no secret nas type other proto limit max connections lifetime idle timeout debugger not attached creating auth type mschap creating auth type digest radiusd instantiating modules modules loaded module rlm expiration loading module expiration from file etc mods enabled expiration loaded module rlm expr loading module expr from file etc mods enabled expr expr safe characters äéöüàâæçèéêëîïôœùûüaÿäéöüßàâæçèéêëîïôœùûüÿ loaded module rlm sql loading module sql from file etc mods available sql sql driver rlm sql mysql server localhost port login easyhotspot password radius db hotspot read groups yes read profiles yes read clients no delete stale sessions yes sql user name user name default user profile client query select id nasname shortname type secret server from nas authorize check query select id username attribute value op from radcheck where username sql user name order by id authorize reply query select id username attribute value op from radreply where username sql user name order by id authorize group check query select id groupname attribute value op from radgroupcheck where groupname sql group order by id authorize group reply query select id groupname attribute value op from radgroupreply where groupname sql group order by id group membership query select groupname from radusergroup where username sql user name order by priority simul count query select count from radacct where username sql user name and acctstoptime is null simul verify query select radacctid acctsessionid username nasipaddress nasportid framedipaddress callingstationid framedprotocol from radacct where username sql user name and acctstoptime is null safe characters auto escape no accounting reference tolower type acct status type request processing stage query type accounting on query update radacct set acctstoptime from unixtime integer event timestamp l acctsessiontime integer event timestamp l unix timestamp acctstarttime acctterminatecause acct terminate cause nas reboot where acctstoptime is null and nasipaddress nas ip address and acctstarttime b reset daily counter name daily session time check name max daily session reply name session timeout loading module monthlycounter from file etc mods available sqlcounter sqlcounter monthlycounter sql module instance sql key user name query select sum acctsessiontime greatest b unix timestamp acctstarttime from radacct where username user name and unix timestamp acctstarttime acctsessiontime b reset monthly counter name monthly session time check name max monthly session reply name session timeout loading module expire on login from file etc mods available sqlcounter sqlcounter expire on login sql module instance sql key user name query select ifnull max time to sec timediff now acctstarttime from radacct where username user name order by acctstarttime limit reset never counter name expire after initial login check name expire after reply name session timeout loading module noresetcounter from file etc mods available sqlcounter sqlcounter noresetcounter sql module instance sql key user name query select sum acctsessiontime from radacct where username user name reset never counter name session timeout check name session timeout reply name session timeout etc mods available sqlcounter counter attribute session timeout must be etc mods available sqlcounter instantiation failed for module noresetcounter root openwrt relevant log output from client utilities no response backtrace from lldb or gdb no response
1
75,931
3,478,697,915
IssuesEvent
2015-12-28 14:46:23
nim-lang/nimble
https://api.github.com/repos/nim-lang/nimble
opened
Removing a newer version of a bin pkg does not switch bin symlinks
Bug Low Priority
* Install two versions (e.g. 0.6.2 and 0.7.0) of Nimble * The ``~/.nimble/bin/nimble`` will point to ``~/.nimble/pkgs/nimble-0.7.0/nimble`` * ``nimble remove nimble@0.7`` * The symlink in ``~/.nimble/bin/nimble`` still points to ``~/.nimble/pkgs/nimble-0.7.0/nimble``
1.0
Removing a newer version of a bin pkg does not switch bin symlinks - * Install two versions (e.g. 0.6.2 and 0.7.0) of Nimble * The ``~/.nimble/bin/nimble`` will point to ``~/.nimble/pkgs/nimble-0.7.0/nimble`` * ``nimble remove nimble@0.7`` * The symlink in ``~/.nimble/bin/nimble`` still points to ``~/.nimble/pkgs/nimble-0.7.0/nimble``
non_defect
removing a newer version of a bin pkg does not switch bin symlinks install two versions e g and of nimble the nimble bin nimble will point to nimble pkgs nimble nimble nimble remove nimble the symlink in nimble bin nimble still points to nimble pkgs nimble nimble
0
174,984
21,300,626,348
IssuesEvent
2022-04-15 02:16:56
farooqmir/React-Redux-Demonstration-with-api
https://api.github.com/repos/farooqmir/React-Redux-Demonstration-with-api
opened
CVE-2021-43138 (High) detected in multiple libraries
security vulnerability
## CVE-2021-43138 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>async-1.0.0.tgz</b>, <b>async-0.9.2.tgz</b>, <b>async-1.5.2.tgz</b></p></summary> <p> <details><summary><b>async-1.0.0.tgz</b></p></summary> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://registry.npmjs.org/async/-/async-1.0.0.tgz">https://registry.npmjs.org/async/-/async-1.0.0.tgz</a></p> <p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p> <p>Path to vulnerable library: /node_modules/winston/node_modules/async/package.json</p> <p> Dependency Hierarchy: - prompt-1.0.0.tgz (Root Library) - winston-2.1.1.tgz - :x: **async-1.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>async-0.9.2.tgz</b></p></summary> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://registry.npmjs.org/async/-/async-0.9.2.tgz">https://registry.npmjs.org/async/-/async-0.9.2.tgz</a></p> <p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p> <p>Path to vulnerable library: /node_modules/utile/node_modules/async/package.json</p> <p> Dependency Hierarchy: - prompt-1.0.0.tgz (Root Library) - utile-0.3.0.tgz - :x: **async-0.9.2.tgz** (Vulnerable Library) </details> <details><summary><b>async-1.5.2.tgz</b></p></summary> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://registry.npmjs.org/async/-/async-1.5.2.tgz">https://registry.npmjs.org/async/-/async-1.5.2.tgz</a></p> <p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p> <p>Path to vulnerable library: /node_modules/async/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.5.tgz (Root Library) - portscanner-2.1.1.tgz - :x: **async-1.5.2.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability exists in Async through 3.2.1 (fixed in 3.2.2) , which could let a malicious user obtain privileges via the mapValues() method. <p>Publish Date: 2022-04-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43138>CVE-2021-43138</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p> <p>Release Date: 2022-04-06</p> <p>Fix Resolution: async - v3.2.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-43138 (High) detected in multiple libraries - ## CVE-2021-43138 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>async-1.0.0.tgz</b>, <b>async-0.9.2.tgz</b>, <b>async-1.5.2.tgz</b></p></summary> <p> <details><summary><b>async-1.0.0.tgz</b></p></summary> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://registry.npmjs.org/async/-/async-1.0.0.tgz">https://registry.npmjs.org/async/-/async-1.0.0.tgz</a></p> <p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p> <p>Path to vulnerable library: /node_modules/winston/node_modules/async/package.json</p> <p> Dependency Hierarchy: - prompt-1.0.0.tgz (Root Library) - winston-2.1.1.tgz - :x: **async-1.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>async-0.9.2.tgz</b></p></summary> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://registry.npmjs.org/async/-/async-0.9.2.tgz">https://registry.npmjs.org/async/-/async-0.9.2.tgz</a></p> <p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p> <p>Path to vulnerable library: /node_modules/utile/node_modules/async/package.json</p> <p> Dependency Hierarchy: - prompt-1.0.0.tgz (Root Library) - utile-0.3.0.tgz - :x: **async-0.9.2.tgz** (Vulnerable Library) </details> <details><summary><b>async-1.5.2.tgz</b></p></summary> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://registry.npmjs.org/async/-/async-1.5.2.tgz">https://registry.npmjs.org/async/-/async-1.5.2.tgz</a></p> <p>Path to dependency file: /React-Redux-Demonstration-with-api/package.json</p> <p>Path to vulnerable library: /node_modules/async/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.5.tgz (Root Library) - portscanner-2.1.1.tgz - :x: **async-1.5.2.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability exists in Async through 3.2.1 (fixed in 3.2.2) , which could let a malicious user obtain privileges via the mapValues() method. <p>Publish Date: 2022-04-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-43138>CVE-2021-43138</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p> <p>Release Date: 2022-04-06</p> <p>Fix Resolution: async - v3.2.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries async tgz async tgz async tgz async tgz higher order functions and common patterns for asynchronous code library home page a href path to dependency file react redux demonstration with api package json path to vulnerable library node modules winston node modules async package json dependency hierarchy prompt tgz root library winston tgz x async tgz vulnerable library async tgz higher order functions and common patterns for asynchronous code library home page a href path to dependency file react redux demonstration with api package json path to vulnerable library node modules utile node modules async package json dependency hierarchy prompt tgz root library utile tgz x async tgz vulnerable library async tgz higher order functions and common patterns for asynchronous code library home page a href path to dependency file react redux demonstration with api package json path to vulnerable library node modules async package json dependency hierarchy browser sync tgz root library portscanner tgz x async tgz vulnerable library vulnerability details a vulnerability exists in async through fixed in which could let a malicious user obtain privileges via the mapvalues method publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution async step up your open source security game with whitesource
0
6,421
9,333,326,449
IssuesEvent
2019-03-28 14:12:23
gnosis/dex-contracts
https://api.github.com/repos/gnosis/dex-contracts
closed
Order Collection
requirement
One should be able to submit a *limit order* of the following form - **Buy** at most `X` of token `i` for token `j` if price `p_{i, j} < P` - **Sell** at most `Y` of token `i` for token `j` if price `p_{i, j} >= P` @twalth3r could you please confirm the order types that should be collected by the contract? Expressing a `Sell` Order in terms of integers should be as follows: - `buyToken: uint8` - `sellToken: uint8` - `buyAmount: uint` - `sellAmount: uint` - `priceBuyToken: uint` - `priceSellToken: uint`
1.0
Order Collection - One should be able to submit a *limit order* of the following form - **Buy** at most `X` of token `i` for token `j` if price `p_{i, j} < P` - **Sell** at most `Y` of token `i` for token `j` if price `p_{i, j} >= P` @twalth3r could you please confirm the order types that should be collected by the contract? Expressing a `Sell` Order in terms of integers should be as follows: - `buyToken: uint8` - `sellToken: uint8` - `buyAmount: uint` - `sellAmount: uint` - `priceBuyToken: uint` - `priceSellToken: uint`
non_defect
order collection one should be able to submit a limit order of the following form buy at most x of token i for token j if price p i j p sell at most y of token i for token j if price p i j p could you please confirm the order types that should be collected by the contract expressing a sell order in terms of integers should be as follows buytoken selltoken buyamount uint sellamount uint pricebuytoken uint priceselltoken uint
0
24,088
2,666,017,262
IssuesEvent
2015-03-21 03:23:36
scanmem/scanmem
https://api.github.com/repos/scanmem/scanmem
closed
Non-blocking attaching
auto-migrated Priority-Low Type-Enhancement
``` To attach the target process without blocking it ``` Original issue reported on code.google.com by `coolwan...@gmail.com` on 25 Apr 2013 at 2:00
1.0
Non-blocking attaching - ``` To attach the target process without blocking it ``` Original issue reported on code.google.com by `coolwan...@gmail.com` on 25 Apr 2013 at 2:00
non_defect
non blocking attaching to attach the target process without blocking it original issue reported on code google com by coolwan gmail com on apr at
0
14,705
2,831,388,625
IssuesEvent
2015-05-24 15:53:31
nobodyguy/dslrdashboard
https://api.github.com/repos/nobodyguy/dslrdashboard
closed
LRTimelapse Holy Grail Mode - Shutter Exceeding Set Limit
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. LRtimelapse - Holy Grail Mode (Sunset) 2. Starting exposure 1/60s, ISO 50 3. set Limit longest shutter to 2s, highest ISO: 800, 1/2 stop increment for both shutter and ISO. What is the expected output? What do you see instead? Expected DD to limit the shutter speed at 2s and proceed to increase ISO. Instead the DD keep on increasing the shutter speed past 2s (until 8s, at that point I stopped the timelapse recording) and the ISO was still kept at 50. What version of the product are you using? On what operating system? latest version (updated 21 Feb 2014) on Galaxy Note 2; Nikon d600 connected using wifi Please provide any additional information below. Previous version's Auto Holy Grail works flawlessly. Will try again tomorrow for sunrise to see if similar problem will recur Many thanks for the awesome apps! regards Kamrul ``` Original issue reported on code.google.com by `kamrul.a...@gmail.com` on 22 Feb 2014 at 1:33
1.0
LRTimelapse Holy Grail Mode - Shutter Exceeding Set Limit - ``` What steps will reproduce the problem? 1. LRtimelapse - Holy Grail Mode (Sunset) 2. Starting exposure 1/60s, ISO 50 3. set Limit longest shutter to 2s, highest ISO: 800, 1/2 stop increment for both shutter and ISO. What is the expected output? What do you see instead? Expected DD to limit the shutter speed at 2s and proceed to increase ISO. Instead the DD keep on increasing the shutter speed past 2s (until 8s, at that point I stopped the timelapse recording) and the ISO was still kept at 50. What version of the product are you using? On what operating system? latest version (updated 21 Feb 2014) on Galaxy Note 2; Nikon d600 connected using wifi Please provide any additional information below. Previous version's Auto Holy Grail works flawlessly. Will try again tomorrow for sunrise to see if similar problem will recur Many thanks for the awesome apps! regards Kamrul ``` Original issue reported on code.google.com by `kamrul.a...@gmail.com` on 22 Feb 2014 at 1:33
defect
lrtimelapse holy grail mode shutter exceeding set limit what steps will reproduce the problem lrtimelapse holy grail mode sunset starting exposure iso set limit longest shutter to highest iso stop increment for both shutter and iso what is the expected output what do you see instead expected dd to limit the shutter speed at and proceed to increase iso instead the dd keep on increasing the shutter speed past until at that point i stopped the timelapse recording and the iso was still kept at what version of the product are you using on what operating system latest version updated feb on galaxy note nikon connected using wifi please provide any additional information below previous version s auto holy grail works flawlessly will try again tomorrow for sunrise to see if similar problem will recur many thanks for the awesome apps regards kamrul original issue reported on code google com by kamrul a gmail com on feb at
1
127,531
5,031,946,942
IssuesEvent
2016-12-16 09:28:23
itsyouonline/identityserver
https://api.github.com/repos/itsyouonline/identityserver
closed
Possibility not to send invitation emails
priority_major state_verification type_feature
Some services would like to send the invitation email themselves so the possibility for not sending out invitation emails should be added when an organization invites someone. Proposed solution: add an `invitemail` parameter to the invitation creation api with the possible values of `default` and `none`, the default value is `default`. When `none` is supplied, no invitation email should be send.
1.0
Possibility not to send invitation emails - Some services would like to send the invitation email themselves so the possibility for not sending out invitation emails should be added when an organization invites someone. Proposed solution: add an `invitemail` parameter to the invitation creation api with the possible values of `default` and `none`, the default value is `default`. When `none` is supplied, no invitation email should be send.
non_defect
possibility not to send invitation emails some services would like to send the invitation email themselves so the possibility for not sending out invitation emails should be added when an organization invites someone proposed solution add an invitemail parameter to the invitation creation api with the possible values of default and none the default value is default when none is supplied no invitation email should be send
0
141,379
11,414,337,583
IssuesEvent
2020-02-02 02:01:27
istio/istio
https://api.github.com/repos/istio/istio
closed
New Test, TestAuthorizationPolicies_ListNamespacesOfServiceRoles, flaky due to non deterministic sorting
area/security kind/test failure lifecycle/needs-triage
https://prow.istio.io/view/gcs/istio-prow/pr-logs/pull/istio_istio/17755/unit-tests_istio/1981
1.0
New Test, TestAuthorizationPolicies_ListNamespacesOfServiceRoles, flaky due to non deterministic sorting - https://prow.istio.io/view/gcs/istio-prow/pr-logs/pull/istio_istio/17755/unit-tests_istio/1981
non_defect
new test testauthorizationpolicies listnamespacesofserviceroles flaky due to non deterministic sorting
0
83,942
15,720,761,497
IssuesEvent
2021-03-29 01:07:21
andygonzalez2010/store
https://api.github.com/repos/andygonzalez2010/store
opened
WS-2017-3767 (Medium) detected in spring-security-web-5.1.5.RELEASE.jar
security vulnerability
## WS-2017-3767 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-5.1.5.RELEASE.jar</b></p></summary> <p>spring-security-web</p> <p>Library home page: <a href="https://spring.io/spring-security">https://spring.io/spring-security</a></p> <p>Path to dependency file: store/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/5.1.5.RELEASE/spring-security-web-5.1.5.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-security-2.1.4.RELEASE.jar (Root Library) - :x: **spring-security-web-5.1.5.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-Site Request Forgery (CSRF) vulnerability was found in spring-security before 4.2.15, 5.0.15, 5.1.9, 5.2.3, and 5.3.1. SwitchUserFilter responds to all HTTP methods, making it vulnerable to CSRF attacks. <p>Publish Date: 2017-01-03 <p>URL: <a href=https://github.com/spring-projects/spring-security/commit/eed71243cb86833e7edf230e5e43ad89b01142f9>WS-2017-3767</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/spring-projects/spring-security/releases/tag/5.3.1.RELEASE">https://github.com/spring-projects/spring-security/releases/tag/5.3.1.RELEASE</a></p> <p>Release Date: 2017-01-03</p> <p>Fix Resolution: org.springframework.security:spring-security-web:4.2.15,5.0.15,5.1.9,5.2.3,5.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2017-3767 (Medium) detected in spring-security-web-5.1.5.RELEASE.jar - ## WS-2017-3767 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-5.1.5.RELEASE.jar</b></p></summary> <p>spring-security-web</p> <p>Library home page: <a href="https://spring.io/spring-security">https://spring.io/spring-security</a></p> <p>Path to dependency file: store/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/5.1.5.RELEASE/spring-security-web-5.1.5.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-security-2.1.4.RELEASE.jar (Root Library) - :x: **spring-security-web-5.1.5.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-Site Request Forgery (CSRF) vulnerability was found in spring-security before 4.2.15, 5.0.15, 5.1.9, 5.2.3, and 5.3.1. SwitchUserFilter responds to all HTTP methods, making it vulnerable to CSRF attacks. <p>Publish Date: 2017-01-03 <p>URL: <a href=https://github.com/spring-projects/spring-security/commit/eed71243cb86833e7edf230e5e43ad89b01142f9>WS-2017-3767</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/spring-projects/spring-security/releases/tag/5.3.1.RELEASE">https://github.com/spring-projects/spring-security/releases/tag/5.3.1.RELEASE</a></p> <p>Release Date: 2017-01-03</p> <p>Fix Resolution: org.springframework.security:spring-security-web:4.2.15,5.0.15,5.1.9,5.2.3,5.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
ws medium detected in spring security web release jar ws medium severity vulnerability vulnerable library spring security web release jar spring security web library home page a href path to dependency file store pom xml path to vulnerable library home wss scanner repository org springframework security spring security web release spring security web release jar dependency hierarchy spring boot starter security release jar root library x spring security web release jar vulnerable library found in base branch master vulnerability details cross site request forgery csrf vulnerability was found in spring security before and switchuserfilter responds to all http methods making it vulnerable to csrf attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework security spring security web step up your open source security game with whitesource
0
53,871
13,262,408,600
IssuesEvent
2020-08-20 21:43:56
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
better error reporting from boost::python (Trac #2221)
Migrated from Trac combo core defect
I keep getting crashes from steamshovel wit the error message ```text >>> libc++abi.dylib: terminating with uncaught exception of type boost::python::error_already_set Abort trap: 6 ``` Which is as far as I can tell boost::python's super helpful response to the python interpreter encountering an error. We need to find a way to print out the actual error in the python interpreter. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2221">https://code.icecube.wisc.edu/projects/icecube/ticket/2221</a>, reported by kjmeagher</summary> <p> ```json { "status": "closed", "changetime": "2019-07-01T13:45:37", "_ts": "1561988737587870", "description": "I keep getting crashes from steamshovel wit the error message\n{{{\n>>> libc++abi.dylib: terminating with uncaught exception of type boost::python::error_already_set\nAbort trap: 6\n}}}\nWhich is as far as I can tell boost::python's super helpful response to the python interpreter encountering an error. We need to find a way to print out the actual error in the python interpreter.", "reporter": "kjmeagher", "cc": "", "resolution": "wontfix", "time": "2018-12-06T16:36:20", "component": "combo core", "summary": "better error reporting from boost::python", "priority": "normal", "keywords": "", "milestone": "Autumnal Equinox 2019", "owner": "", "type": "defect" } ``` </p> </details>
1.0
better error reporting from boost::python (Trac #2221) - I keep getting crashes from steamshovel wit the error message ```text >>> libc++abi.dylib: terminating with uncaught exception of type boost::python::error_already_set Abort trap: 6 ``` Which is as far as I can tell boost::python's super helpful response to the python interpreter encountering an error. We need to find a way to print out the actual error in the python interpreter. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2221">https://code.icecube.wisc.edu/projects/icecube/ticket/2221</a>, reported by kjmeagher</summary> <p> ```json { "status": "closed", "changetime": "2019-07-01T13:45:37", "_ts": "1561988737587870", "description": "I keep getting crashes from steamshovel wit the error message\n{{{\n>>> libc++abi.dylib: terminating with uncaught exception of type boost::python::error_already_set\nAbort trap: 6\n}}}\nWhich is as far as I can tell boost::python's super helpful response to the python interpreter encountering an error. We need to find a way to print out the actual error in the python interpreter.", "reporter": "kjmeagher", "cc": "", "resolution": "wontfix", "time": "2018-12-06T16:36:20", "component": "combo core", "summary": "better error reporting from boost::python", "priority": "normal", "keywords": "", "milestone": "Autumnal Equinox 2019", "owner": "", "type": "defect" } ``` </p> </details>
defect
better error reporting from boost python trac i keep getting crashes from steamshovel wit the error message text libc abi dylib terminating with uncaught exception of type boost python error already set abort trap which is as far as i can tell boost python s super helpful response to the python interpreter encountering an error we need to find a way to print out the actual error in the python interpreter migrated from json status closed changetime ts description i keep getting crashes from steamshovel wit the error message n n libc abi dylib terminating with uncaught exception of type boost python error already set nabort trap n nwhich is as far as i can tell boost python s super helpful response to the python interpreter encountering an error we need to find a way to print out the actual error in the python interpreter reporter kjmeagher cc resolution wontfix time component combo core summary better error reporting from boost python priority normal keywords milestone autumnal equinox owner type defect
1
774,905
27,214,999,896
IssuesEvent
2023-02-20 20:36:43
ascheid/itsg33-pbmm-issue-gen
https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen
closed
SA-18 TAMPER RESISTANCE AND DETECTION
Priority: None defined
(A) The organization implements a tamper protection program for the information system, system component, or information system service.
1.0
SA-18 TAMPER RESISTANCE AND DETECTION - (A) The organization implements a tamper protection program for the information system, system component, or information system service.
non_defect
sa tamper resistance and detection a the organization implements a tamper protection program for the information system system component or information system service
0
116,084
17,358,340,407
IssuesEvent
2021-07-29 16:58:11
Tim-sandbox/webgoat-trng
https://api.github.com/repos/Tim-sandbox/webgoat-trng
opened
CVE-2021-22112 (High) detected in spring-security-web-5.2.1.RELEASE.jar
security vulnerability
## CVE-2021-22112 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-5.2.1.RELEASE.jar</b></p></summary> <p>spring-security-web</p> <p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p> <p>Path to dependency file: webgoat-trng/webgoat-integration-tests/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/5.2.1.RELEASE/spring-security-web-5.2.1.RELEASE.jar</p> <p> Dependency Hierarchy: - webwolf-v8.1.0.jar (Root Library) - spring-boot-starter-security-2.2.2.RELEASE.jar - :x: **spring-security-web-5.2.1.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/webgoat-trng/commit/35c20cdfd5acad60536ca6b0402f619e18097690">35c20cdfd5acad60536ca6b0402f619e18097690</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Security 5.4.x prior to 5.4.4, 5.3.x prior to 5.3.8.RELEASE, 5.2.x prior to 5.2.9.RELEASE, and older unsupported versions can fail to save the SecurityContext if it is changed more than once in a single request.A malicious user cannot cause the bug to happen (it must be programmed in). However, if the application's intent is to only allow the user to run with elevated privileges in a small portion of the application, the bug can be leveraged to extend those privileges to the rest of the application. <p>Publish Date: 2021-02-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22112>CVE-2021-22112</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22112">https://tanzu.vmware.com/security/cve-2021-22112</a></p> <p>Release Date: 2021-02-23</p> <p>Fix Resolution: org.springframework.security:spring-security-web:5.2.9,5.3.8,5.4.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework.security","packageName":"spring-security-web","packageVersion":"5.2.1.RELEASE","packageFilePaths":["/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.owasp.webgoat:webwolf:v8.1.0;org.springframework.boot:spring-boot-starter-security:2.2.2.RELEASE;org.springframework.security:spring-security-web:5.2.1.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework.security:spring-security-web:5.2.9,5.3.8,5.4.4"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-22112","vulnerabilityDetails":"Spring Security 5.4.x prior to 5.4.4, 5.3.x prior to 5.3.8.RELEASE, 5.2.x prior to 5.2.9.RELEASE, and older unsupported versions can fail to save the SecurityContext if it is changed more than once in a single request.A malicious user cannot cause the bug to happen (it must be programmed in). However, if the application\u0027s intent is to only allow the user to run with elevated privileges in a small portion of the application, the bug can be leveraged to extend those privileges to the rest of the application.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22112","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-22112 (High) detected in spring-security-web-5.2.1.RELEASE.jar - ## CVE-2021-22112 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-5.2.1.RELEASE.jar</b></p></summary> <p>spring-security-web</p> <p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p> <p>Path to dependency file: webgoat-trng/webgoat-integration-tests/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/5.2.1.RELEASE/spring-security-web-5.2.1.RELEASE.jar</p> <p> Dependency Hierarchy: - webwolf-v8.1.0.jar (Root Library) - spring-boot-starter-security-2.2.2.RELEASE.jar - :x: **spring-security-web-5.2.1.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/webgoat-trng/commit/35c20cdfd5acad60536ca6b0402f619e18097690">35c20cdfd5acad60536ca6b0402f619e18097690</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Security 5.4.x prior to 5.4.4, 5.3.x prior to 5.3.8.RELEASE, 5.2.x prior to 5.2.9.RELEASE, and older unsupported versions can fail to save the SecurityContext if it is changed more than once in a single request.A malicious user cannot cause the bug to happen (it must be programmed in). However, if the application's intent is to only allow the user to run with elevated privileges in a small portion of the application, the bug can be leveraged to extend those privileges to the rest of the application. <p>Publish Date: 2021-02-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22112>CVE-2021-22112</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22112">https://tanzu.vmware.com/security/cve-2021-22112</a></p> <p>Release Date: 2021-02-23</p> <p>Fix Resolution: org.springframework.security:spring-security-web:5.2.9,5.3.8,5.4.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework.security","packageName":"spring-security-web","packageVersion":"5.2.1.RELEASE","packageFilePaths":["/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.owasp.webgoat:webwolf:v8.1.0;org.springframework.boot:spring-boot-starter-security:2.2.2.RELEASE;org.springframework.security:spring-security-web:5.2.1.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework.security:spring-security-web:5.2.9,5.3.8,5.4.4"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-22112","vulnerabilityDetails":"Spring Security 5.4.x prior to 5.4.4, 5.3.x prior to 5.3.8.RELEASE, 5.2.x prior to 5.2.9.RELEASE, and older unsupported versions can fail to save the SecurityContext if it is changed more than once in a single request.A malicious user cannot cause the bug to happen (it must be programmed in). However, if the application\u0027s intent is to only allow the user to run with elevated privileges in a small portion of the application, the bug can be leveraged to extend those privileges to the rest of the application.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22112","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in spring security web release jar cve high severity vulnerability vulnerable library spring security web release jar spring security web library home page a href path to dependency file webgoat trng webgoat integration tests pom xml path to vulnerable library home wss scanner repository org springframework security spring security web release spring security web release jar dependency hierarchy webwolf jar root library spring boot starter security release jar x spring security web release jar vulnerable library found in head commit a href found in base branch main vulnerability details spring security x prior to x prior to release x prior to release and older unsupported versions can fail to save the securitycontext if it is changed more than once in a single request a malicious user cannot cause the bug to happen it must be programmed in however if the application s intent is to only allow the user to run with elevated privileges in a small portion of the application the bug can be leveraged to extend those privileges to the rest of the application publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework security spring security web isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org owasp webgoat webwolf org springframework boot spring boot starter security release org springframework security spring security web release isminimumfixversionavailable true minimumfixversion org springframework security spring security web basebranches vulnerabilityidentifier cve vulnerabilitydetails spring security x prior to x prior to release x prior to release and older unsupported versions can fail to save the securitycontext if it is changed more than once in a single request a malicious user cannot cause the bug to happen it must be programmed in however if the application intent is to only allow the user to run with elevated privileges in a small portion of the application the bug can be leveraged to extend those privileges to the rest of the application vulnerabilityurl
0
47,142
13,206,083,391
IssuesEvent
2020-08-14 19:22:47
kenferrara/layer5
https://api.github.com/repos/kenferrara/layer5
opened
CVE-2020-8161 (High) detected in rack-1.6.13.gem
security vulnerability
## CVE-2020-8161 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rack-1.6.13.gem</b></p></summary> <p>Rack provides a minimal, modular and adaptable interface for developing web applications in Ruby. By wrapping HTTP requests and responses in the simplest way possible, it unifies and distills the API for web servers, web frameworks, and software in between (the so-called middleware) into a single method call. Also see http://rack.github.io/. </p> <p>Library home page: <a href="https://rubygems.org/gems/rack-1.6.13.gem">https://rubygems.org/gems/rack-1.6.13.gem</a></p> <p> Dependency Hierarchy: - rack-jekyll-0.5.0.gem (Root Library) - :x: **rack-1.6.13.gem** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kenferrara/layer5/commit/eda0b835176f91fd68458c4f840b249240947718">eda0b835176f91fd68458c4f840b249240947718</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A directory traversal vulnerability exists in rack < 2.2.0 that allows an attacker perform directory traversal vulnerability in the Rack::Directory app that is bundled with Rack which could result in information disclosure. <p>Publish Date: 2020-07-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8161>CVE-2020-8161</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/rack/rack/tree/2.2.0">https://github.com/rack/rack/tree/2.2.0</a></p> <p>Release Date: 2020-06-01</p> <p>Fix Resolution: 2.2.0,2.1.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Ruby","packageName":"rack","packageVersion":"1.6.13","isTransitiveDependency":true,"dependencyTree":"rack-jekyll:0.5.0;rack:1.6.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.2.0,2.1.3"}],"vulnerabilityIdentifier":"CVE-2020-8161","vulnerabilityDetails":"A directory traversal vulnerability exists in rack \u003c 2.2.0 that allows an attacker perform directory traversal vulnerability in the Rack::Directory app that is bundled with Rack which could result in information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8161","cvss3Severity":"high","cvss3Score":"8.6","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-8161 (High) detected in rack-1.6.13.gem - ## CVE-2020-8161 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rack-1.6.13.gem</b></p></summary> <p>Rack provides a minimal, modular and adaptable interface for developing web applications in Ruby. By wrapping HTTP requests and responses in the simplest way possible, it unifies and distills the API for web servers, web frameworks, and software in between (the so-called middleware) into a single method call. Also see http://rack.github.io/. </p> <p>Library home page: <a href="https://rubygems.org/gems/rack-1.6.13.gem">https://rubygems.org/gems/rack-1.6.13.gem</a></p> <p> Dependency Hierarchy: - rack-jekyll-0.5.0.gem (Root Library) - :x: **rack-1.6.13.gem** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kenferrara/layer5/commit/eda0b835176f91fd68458c4f840b249240947718">eda0b835176f91fd68458c4f840b249240947718</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A directory traversal vulnerability exists in rack < 2.2.0 that allows an attacker perform directory traversal vulnerability in the Rack::Directory app that is bundled with Rack which could result in information disclosure. <p>Publish Date: 2020-07-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8161>CVE-2020-8161</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/rack/rack/tree/2.2.0">https://github.com/rack/rack/tree/2.2.0</a></p> <p>Release Date: 2020-06-01</p> <p>Fix Resolution: 2.2.0,2.1.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Ruby","packageName":"rack","packageVersion":"1.6.13","isTransitiveDependency":true,"dependencyTree":"rack-jekyll:0.5.0;rack:1.6.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.2.0,2.1.3"}],"vulnerabilityIdentifier":"CVE-2020-8161","vulnerabilityDetails":"A directory traversal vulnerability exists in rack \u003c 2.2.0 that allows an attacker perform directory traversal vulnerability in the Rack::Directory app that is bundled with Rack which could result in information disclosure.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8161","cvss3Severity":"high","cvss3Score":"8.6","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in rack gem cve high severity vulnerability vulnerable library rack gem rack provides a minimal modular and adaptable interface for developing web applications in ruby by wrapping http requests and responses in the simplest way possible it unifies and distills the api for web servers web frameworks and software in between the so called middleware into a single method call also see library home page a href dependency hierarchy rack jekyll gem root library x rack gem vulnerable library found in head commit a href vulnerability details a directory traversal vulnerability exists in rack that allows an attacker perform directory traversal vulnerability in the rack directory app that is bundled with rack which could result in information disclosure publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a directory traversal vulnerability exists in rack that allows an attacker perform directory traversal vulnerability in the rack directory app that is bundled with rack which could result in information disclosure vulnerabilityurl
0
42,013
5,412,272,179
IssuesEvent
2017-03-01 14:09:32
wellcometrust/wellcomecollection.org
https://api.github.com/repos/wellcometrust/wellcomecollection.org
closed
Explore - Article promo
design
The format text (example: Article) should be black (hex: 1d1d1d) and is clickable. The user can click image or title to go into article, but they can also click on format and go to the index page for that format. Hover for this text is the similar as regular body link, with teal underline appearing on hover ![screen shot 2017-02-01 at 15 34 20](https://cloud.githubusercontent.com/assets/23660297/22513619/accc0e5e-e894-11e6-9a91-24ced800d53d.png)
1.0
Explore - Article promo - The format text (example: Article) should be black (hex: 1d1d1d) and is clickable. The user can click image or title to go into article, but they can also click on format and go to the index page for that format. Hover for this text is the similar as regular body link, with teal underline appearing on hover ![screen shot 2017-02-01 at 15 34 20](https://cloud.githubusercontent.com/assets/23660297/22513619/accc0e5e-e894-11e6-9a91-24ced800d53d.png)
non_defect
explore article promo the format text example article should be black hex and is clickable the user can click image or title to go into article but they can also click on format and go to the index page for that format hover for this text is the similar as regular body link with teal underline appearing on hover
0
66,349
20,157,930,774
IssuesEvent
2022-02-09 18:15:05
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
ZFS 2.1.2 + NFSv4 ARC stuck pruning/evicting, flatlines throughput
Type: Defect
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | `Ubuntu` Distribution Version | `20.04.3 LTS` Kernel Version | `5.4.0-96-generic` Architecture | `x86_64` OpenZFS Version | `zfs-2.1.2-1` <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing ZFS is flatlined on throughput with an `arc_evict` and `arc_prune` process spinning at 100%. The workload is kernel NFS server (all NFSv4 clients) with ZFS 2.1.2 built from source. Characterized by high CPU `iowait` and throughput flatlining. #### RAM Graph <img width="1350" alt="Screen Shot 2022-02-09 at 9 06 55 AM" src="https://user-images.githubusercontent.com/16908622/153252531-9ee24c8e-3ecb-4242-b862-b860353f5625.png"> It's holding on to RAM pretty hard. This is where I begin to lose debugging expertise, having discovered slabs yesterday. :-) #### top ``` top - 16:55:29 up 21 days, 14:56, 1 user, load average: 195.23, 124.60, 134.24 Tasks: 993 total, 2 running, 991 sleeping, 0 stopped, 0 zombie %Cpu(s): 0.0 us, 7.3 sy, 0.0 ni, 5.0 id, 87.3 wa, 0.0 hi, 0.4 si, 0.0 st MiB Mem : 128622.1 total, 2501.0 free, 121637.5 used, 4483.5 buff/cache MiB Swap: 1907.0 total, 1042.7 free, 864.2 used. 6078.4 avail Mem PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 245944 root 20 0 0 0 0 R 100.0 0.0 549:36.86 arc_evict 245943 root 20 0 0 0 0 S 93.1 0.0 459:13.30 arc_prune 246555 root 20 0 953064 20876 2732 S 3.6 0.0 29:32.85 zed 3174406 root 20 0 12784 4872 3248 R 1.3 0.0 0:00.17 top 14154 root rt 0 356604 31088 8300 S 0.7 0.0 133:03.60 multipathd 3159749 root 20 0 0 0 0 D 0.7 0.0 0:00.11 nfsd ``` #### slabtop ``` # slabtop Active / Total Objects (% used) : 215263329 / 222780310 (96.6%) Active / Total Slabs (% used) : 6020764 / 6020764 (100.0%) Active / Total Caches (% used) : 141 / 199 (70.9%) Active / Total Size (% used) : 98093294.62K / 99549181.36K (98.5%) Minimum / Average / Maximum Object : 0.01K / 0.45K / 16.75K OBJS ACTIVE USE OBJ SIZE SLABS OBJ/SLAB CACHE SIZE NAME 27437256 27437256 100% 0.97K 831432 33 26605824K dnode_t 25866414 25866282 99% 0.38K 615867 42 9853872K dmu_buf_impl_t 24953184 24952001 99% 0.50K 779787 32 12476592K kmalloc-512 20654336 18708128 90% 0.06K 322724 64 1290896K kmalloc-64 14428134 14148450 98% 0.09K 343527 42 1374108K arc_buf_hdr_t_l2only 10290924 10290229 99% 0.09K 245022 42 980088K kmalloc-96 10017152 9881010 98% 0.03K 78259 128 313036K kmalloc-32 7602798 7575100 99% 0.19K 181019 42 1448152K dentry 7255278 7139757 98% 1.09K 250182 29 8005824K zfs_znode_cache 7225482 7140416 98% 0.24K 218954 33 1751632K sa_cache 5172144 5172144 100% 0.16K 107753 48 862024K nfsd4_stateids 5079424 5077737 99% 0.25K 158732 32 1269856K filp 4988088 4985288 99% 0.19K 118764 42 950112K cred_jar 4952610 4952610 100% 0.02K 29133 170 116532K lsm_file_cache 4951520 4951520 100% 0.28K 176840 28 1414720K nfsd4_files 4949376 4949376 100% 0.03K 38667 128 154668K fsnotify_mark_connector 4947816 4947816 100% 0.08K 97016 51 388064K Acpi-State 4947372 4947372 100% 0.11K 137427 36 549708K khugepaged_mm_slot 4862528 4858723 99% 0.06K 75977 64 303908K kmalloc-rcl-64 3813933 1054784 27% 0.31K 74783 51 1196528K arc_buf_hdr_t_full 3601017 3601017 100% 0.05K 49329 73 197316K nsproxy 3022149 1059649 35% 0.10K 77491 39 309964K abd_t 2598288 2591411 99% 0.57K 92796 28 1484736K radix_tree_node 1523068 1523068 100% 0.42K 41164 37 658624K nfsd4_openowners 1248128 1247029 99% 1.00K 39004 32 1248128K kmalloc-1k 1128681 1052509 93% 0.08K 22131 51 88524K arc_buf_t 960386 960375 99% 16.00K 480193 2 15366176K zio_buf_comb_16384 945488 942004 99% 8.00K 236372 4 7563904K kmalloc-8k 290745 268913 92% 0.10K 7455 39 29820K buffer_head 288000 287664 99% 0.13K 9600 30 38400K kernfs_node_cache 228160 226423 99% 0.06K 3565 64 14260K anon_vma_chain 211894 204295 96% 0.59K 3998 53 127936K inode_cache 209664 209664 100% 0.02K 819 256 3276K kmalloc-16 187432 187432 100% 0.07K 3347 56 13388K Acpi-Operand 169611 168672 99% 0.20K 4349 39 34792K vm_area_struct 121856 121856 100% 0.01K 238 512 952K kmalloc-8 116978 116702 99% 0.09K 2543 46 10172K anon_vma 107856 99165 91% 0.09K 2568 42 10272K kmalloc-rcl-96 ``` #### zpool ``` # zpool status pod-10 pool: pod-10 state: ONLINE config: NAME STATE READ WRITE CKSUM pod-10 ONLINE 0 0 0 raidz3-0 ONLINE 0 0 0 35000c500ae29a4bb ONLINE 0 0 0 35000c500ae95def3 ONLINE 0 0 0 35000c500ae96d1d7 ONLINE 0 0 0 35000c500ae9729af ONLINE 0 0 0 35000c500ae97296f ONLINE 0 0 0 35000c500ae96daeb ONLINE 0 0 0 35000c500ae968243 ONLINE 0 0 0 35000c500ae97269b ONLINE 0 0 0 35000c500ae970cd7 ONLINE 0 0 0 35000c500ae975a7f ONLINE 0 0 0 35000c500ae957c3b ONLINE 0 0 0 raidz3-1 ONLINE 0 0 0 35000c500ae96870f ONLINE 0 0 0 35000c500ae2bc057 ONLINE 0 0 0 35000c500ae9733b7 ONLINE 0 0 0 35000c500ae2bc4db ONLINE 0 0 0 35000c500ae96b4bf ONLINE 0 0 0 35000c500ae970e3b ONLINE 0 0 0 35000c500ae957bab ONLINE 0 0 0 35000c500ae96aa6f ONLINE 0 0 0 35000c500ae96833f ONLINE 0 0 0 35000c500ae96a4a3 ONLINE 0 0 0 35000c500ae2a6e9b ONLINE 0 0 0 raidz3-2 ONLINE 0 0 0 35000c500ae96c3f7 ONLINE 0 0 0 35000c500ae972ddf ONLINE 0 0 0 35000c500ae96bb4f ONLINE 0 0 0 35000c500ae95d66f ONLINE 0 0 0 35000c500ae96777f ONLINE 0 0 0 35000c500ae60fec3 ONLINE 0 0 0 35000c500ae96fcef ONLINE 0 0 0 35000c500ae966b0b ONLINE 0 0 0 35000c500ae96c823 ONLINE 0 0 0 35000c500ae95e363 ONLINE 0 0 0 35000c500ae96fab3 ONLINE 0 0 0 raidz3-3 ONLINE 0 0 0 35000c500ae34ceeb ONLINE 0 0 0 35000c500ae34d4c7 ONLINE 0 0 0 35000c500ae970af7 ONLINE 0 0 0 35000c500ae9597cb ONLINE 0 0 0 35000c500ae62c433 ONLINE 0 0 0 35000c500ae968857 ONLINE 0 0 0 35000c500ae970167 ONLINE 0 0 0 35000c500ae63517f ONLINE 0 0 0 35000c500ae961313 ONLINE 0 0 0 35000c500ae95d53b ONLINE 0 0 0 35000c500ae95cc5b ONLINE 0 0 0 raidz3-4 ONLINE 0 0 0 35000c500ae9737c3 ONLINE 0 0 0 35000c500ae970feb ONLINE 0 0 0 35000c500ae9686f3 ONLINE 0 0 0 35000c500ae97387b ONLINE 0 0 0 35000c500ae97403f ONLINE 0 0 0 35000c500ae95711f ONLINE 0 0 0 35000c500ae96cb23 ONLINE 0 0 0 35000c500ae2a6db7 ONLINE 0 0 0 35000c500ae9681a3 ONLINE 0 0 0 35000c500ae9688b7 ONLINE 0 0 0 35000c500ae97404b ONLINE 0 0 0 raidz3-5 ONLINE 0 0 0 35000c500ae956b87 ONLINE 0 0 0 35000c500ae974bf3 ONLINE 0 0 0 35000c500ae9744fb ONLINE 0 0 0 35000c500ae29e993 ONLINE 0 0 0 35000c500ae96ef57 ONLINE 0 0 0 35000c500ae974d2b ONLINE 0 0 0 35000c500ae970a0f ONLINE 0 0 0 35000c500ae39455f ONLINE 0 0 0 35000c500ae29d97f ONLINE 0 0 0 35000c500ae95712b ONLINE 0 0 0 35000c500ae9742c3 ONLINE 0 0 0 raidz3-6 ONLINE 0 0 0 35000c500ae96fc4b ONLINE 0 0 0 35000c500ae955eef ONLINE 0 0 0 35000c500ae95c243 ONLINE 0 0 0 35000c500ae974057 ONLINE 0 0 0 35000c500ae95e4cb ONLINE 0 0 0 35000c500ae96eab3 ONLINE 0 0 0 35000c500ae96c9bb ONLINE 0 0 0 35000c500ae959fd7 ONLINE 0 0 0 35000c500ae2a709f ONLINE 0 0 0 35000c500ae96d793 ONLINE 0 0 0 35000c500ae9728ab ONLINE 0 0 0 raidz3-7 ONLINE 0 0 0 35000c500ae962117 ONLINE 0 0 0 35000c500ae4044bf ONLINE 0 0 0 35000c500ae393b9b ONLINE 0 0 0 35000c500ae974223 ONLINE 0 0 0 35000c500ae95f193 ONLINE 0 0 0 35000c500ae957abb ONLINE 0 0 0 35000c500ae9563a3 ONLINE 0 0 0 35000c500ae9683ab ONLINE 0 0 0 35000c500ae96df8f ONLINE 0 0 0 35000c500ae96d543 ONLINE 0 0 0 35000c500ae972b2f ONLINE 0 0 0 raidz3-8 ONLINE 0 0 0 35000c500ae960517 ONLINE 0 0 0 35000c500ae95555b ONLINE 0 0 0 35000c500ae974637 ONLINE 0 0 0 35000c500ae972dd7 ONLINE 0 0 0 35000c500ae4323c7 ONLINE 0 0 0 35000c500ae96d617 ONLINE 0 0 0 35000c500ae9638b7 ONLINE 0 0 0 35000c500ae96ea0f ONLINE 0 0 0 35000c500ae96e3eb ONLINE 0 0 0 35000c500ae29a417 ONLINE 0 0 0 35000c500ae96fda3 ONLINE 0 0 0 logs mirror-9 ONLINE 0 0 0 wwn-0x55cd2e4152220c93-part1 ONLINE 0 0 0 wwn-0x55cd2e41519b023b-part1 ONLINE 0 0 0 cache wwn-0x55cd2e4152220c93-part2 ONLINE 0 0 0 wwn-0x55cd2e41519b023b-part2 ONLINE 0 0 0 wwn-0x55cd2e415221c8c7 ONLINE 0 0 0 spares 35000c500ae97206f AVAIL 35000c500ae965dbb AVAIL 35000c500ae2b926f AVAIL 35000c500ae976187 AVAIL 35000c500ae96dcf7 AVAIL 35000c500ae958777 AVAIL 35000c500ae41875f AVAIL errors: No known data errors ``` #### zfs fs ``` # zfs list NAME USED AVAIL REFER MOUNTPOINT pod-10 120T 926T 279K /pod-10 pod-10/pod-10 120T 926T 120T /srv/pod-10 ``` ``` # zfs get primarycache,secondarycache pod-10/pod-10 NAME PROPERTY VALUE SOURCE pod-10/pod-10 primarycache metadata local pod-10/pod-10 secondarycache metadata local ``` Intention is for ARC/L2ARC to be entirely metadata. #### initial arcstats arcstats reported metadata usage above the limit. ``` ARC size (current): 102.8 % 64.6 GiB Target size (adaptive): 100.0 % 62.8 GiB Min size (hard limit): 6.2 % 3.9 GiB Max size (high water): 16:1 62.8 GiB Most Frequently Used (MFU) cache size: 13.0 % 2.6 GiB Most Recently Used (MRU) cache size: 87.0 % 17.7 GiB Metadata cache size (hard limit): 75.0 % 47.1 GiB Metadata cache size (current): 135.8 % 64.0 GiB Dnode cache size (hard limit): 10.0 % 4.7 GiB Dnode cache size (current): 537.3 % 25.3 GiB ``` #### initial adjustment Attempted to remediate by increasing ARC size by half of remaining RAM and increasing the metadata allocation in the ARC. ``` echo 86762369024 > /sys/module/zfs/parameters/zfs_arc_max echo 90 > /sys/module/zfs/parameters/zfs_arc_meta_limit_percent echo 50 > /sys/module/zfs/parameters/zfs_arc_dnode_limit_percent ``` #### current issue This remediated the issue temporarily, and the prune processes stopped, but we are back! :) Stopped NFS server and it seems to be free-ing memory albeit very slowly, many `dp_sync_taskq` processes... But removing the NFS server is very un-ideal. Given they are both living in the kernel, it's difficult for me personally to determine who is eating up the memory. ``` top - 18:05:52 up 21 days, 16:07, 2 users, load average: 2.52, 7.31, 35.56 Tasks: 807 total, 1 running, 806 sleeping, 0 stopped, 0 zombie %Cpu(s): 0.0 us, 1.5 sy, 0.0 ni, 95.9 id, 2.5 wa, 0.0 hi, 0.1 si, 0.0 st MiB Mem : 128622.1 total, 17620.3 free, 108471.5 used, 2530.3 buff/cache MiB Swap: 1907.0 total, 1764.7 free, 142.2 used. 19277.1 avail Mem PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 248929 root 20 0 0 0 0 S 2.3 0.0 349:34.92 txg_sync 248864 root 39 19 0 0 0 S 1.7 0.0 33:52.18 dp_sync_taskq 1767630 root 20 0 0 0 0 D 1.7 0.0 3:35.82 kworker/23:2+events 248854 root 39 19 0 0 0 S 1.3 0.0 33:48.50 dp_sync_taskq 248855 root 39 19 0 0 0 S 1.3 0.0 33:51.96 dp_sync_taskq 248856 root 39 19 0 0 0 S 1.3 0.0 33:49.72 dp_sync_taskq 248857 root 39 19 0 0 0 S 1.3 0.0 33:49.15 dp_sync_taskq 248859 root 39 19 0 0 0 S 1.3 0.0 33:54.62 dp_sync_taskq 248860 root 39 19 0 0 0 S 1.3 0.0 33:50.09 dp_sync_taskq 248862 root 39 19 0 0 0 S 1.3 0.0 33:49.23 dp_sync_taskq 248865 root 39 19 0 0 0 S 1.3 0.0 33:50.26 dp_sync_taskq 248867 root 39 19 0 0 0 S 1.3 0.0 33:48.57 dp_sync_taskq 248870 root 39 19 0 0 0 S 1.3 0.0 33:51.20 dp_sync_taskq 248871 root 39 19 0 0 0 S 1.3 0.0 33:48.09 dp_sync_taskq 248872 root 39 19 0 0 0 S 1.3 0.0 33:51.06 dp_sync_taskq 248873 root 39 19 0 0 0 S 1.3 0.0 33:46.16 dp_sync_taskq ``` ``` # arc_summary ------------------------------------------------------------------------ ZFS Subsystem Report Wed Feb 09 16:23:14 2022 Linux 5.4.0-96-generic 2.1.2-1 Machine: r8-n9 (x86_64) 2.1.2-1 ARC status: HEALTHY Memory throttle count: 0 ARC size (current): 90.1 % 72.8 GiB Target size (adaptive): 4.9 % 3.9 GiB Min size (hard limit): 4.9 % 3.9 GiB Max size (high water): 20:1 80.8 GiB Most Frequently Used (MFU) cache size: 13.8 % 3.0 GiB Most Recently Used (MRU) cache size: 86.2 % 18.7 GiB Metadata cache size (hard limit): 90.0 % 72.7 GiB Metadata cache size (current): 100.2 % 72.8 GiB Dnode cache size (hard limit): 50.0 % 36.4 GiB Dnode cache size (current): 80.9 % 29.4 GiB ARC hash breakdown: Elements max: 15.7M Elements current: 97.2 % 15.2M Collisions: 1.1G Chain max: 10 Chains: 3.9M ARC misc: Deleted: 28.5M Mutex misses: 6.6G Eviction skips: 93.2G Eviction skips due to L2 writes: 10.3k L2 cached evictions: 1.4 TiB L2 eligible evictions: 73.1 GiB L2 eligible MFU evictions: 4.8 % 3.5 GiB L2 eligible MRU evictions: 95.2 % 69.6 GiB L2 ineligible evictions: 707.3 GiB ARC total accesses (hits + misses): 11.3G Cache hit ratio: 86.4 % 9.8G Cache miss ratio: 13.6 % 1.5G Actual hit ratio (MFU + MRU hits): 86.3 % 9.8G Data demand efficiency: 1.6 % 1.6G Data prefetch efficiency: n/a 0 Cache hits by cache type: Most frequently used (MFU): 96.3 % 9.4G Most recently used (MRU): 3.6 % 356.4M Most frequently used (MFU) ghost: < 0.1 % 1.3M Most recently used (MRU) ghost: < 0.1 % 1.0M Cache hits by data type: Demand data: 0.3 % 25.4M Demand prefetch data: 0.0 % 0 Demand metadata: 99.7 % 9.8G Demand prefetch metadata: < 0.1 % 2.4M Cache misses by data type: Demand data: 99.3 % 1.5G Demand prefetch data: 0.0 % 0 Demand metadata: 0.4 % 5.9M Demand prefetch metadata: 0.3 % 4.8M DMU prefetch efficiency: 2.7G Hit ratio: 5.7 % 152.9M Miss ratio: 94.3 % 2.5G L2ARC status: HEALTHY Low memory aborts: 24.5k Free on write: 281 R/W clashes: 0 Bad checksums: 0 I/O errors: 0 L2ARC size (adaptive): 1.3 TiB Compressed: 5.6 % 74.6 GiB Header size: 0.1 % 1.3 GiB MFU allocated size: 19.4 % 14.5 GiB MRU allocated size: 85.0 % 63.4 GiB Prefetch allocated size: 0.1 % 67.6 MiB Data (buffer content) allocated size: 0.0 % 0 Bytes Metadata (buffer content) allocated size: 104.5 % 77.9 GiB L2ARC breakdown: 1.5G Hit ratio: 0.2 % 3.8M Miss ratio: 99.8 % 1.5G Feeds: 1.8M L2ARC writes: Writes sent: 100 % 1.4M L2ARC evicts: Lock retries: 0 Upon reading: 0 Solaris Porting Layer (SPL): spl_hostid 0 spl_hostid_path /etc/hostid spl_kmem_alloc_max 1048576 spl_kmem_alloc_warn 65536 spl_kmem_cache_kmem_threads 4 spl_kmem_cache_magazine_size 0 spl_kmem_cache_max_size 32 spl_kmem_cache_obj_per_slab 8 spl_kmem_cache_reclaim 0 spl_kmem_cache_slab_limit 16384 spl_max_show_tasks 512 spl_panic_halt 0 spl_schedule_hrtimeout_slack_us 0 spl_taskq_kick 0 spl_taskq_thread_bind 0 spl_taskq_thread_dynamic 1 spl_taskq_thread_priority 1 spl_taskq_thread_sequential 4 Tunables: dbuf_cache_hiwater_pct 10 dbuf_cache_lowater_pct 10 dbuf_cache_max_bytes 18446744073709551615 dbuf_cache_shift 5 dbuf_metadata_cache_max_bytes 18446744073709551615 dbuf_metadata_cache_shift 6 dmu_object_alloc_chunk_shift 7 dmu_prefetch_max 134217728 ignore_hole_birth 1 l2arc_feed_again 1 l2arc_feed_min_ms 200 l2arc_feed_secs 1 l2arc_headroom 2 l2arc_headroom_boost 200 l2arc_meta_percent 33 l2arc_mfuonly 0 l2arc_noprefetch 1 l2arc_norw 0 l2arc_rebuild_blocks_min_l2size 1073741824 l2arc_rebuild_enabled 1 l2arc_trim_ahead 0 l2arc_write_boost 8388608 l2arc_write_max 8388608 metaslab_aliquot 524288 metaslab_bias_enabled 1 metaslab_debug_load 0 metaslab_debug_unload 0 metaslab_df_max_search 16777216 metaslab_df_use_largest_segment 0 metaslab_force_ganging 16777217 metaslab_fragmentation_factor_enabled 1 metaslab_lba_weighting_enabled 1 metaslab_preload_enabled 1 metaslab_unload_delay 32 metaslab_unload_delay_ms 600000 send_holes_without_birth_time 1 spa_asize_inflation 24 spa_config_path /etc/zfs/zpool.cache spa_load_print_vdev_tree 0 spa_load_verify_data 1 spa_load_verify_metadata 1 spa_load_verify_shift 4 spa_slop_shift 5 vdev_file_logical_ashift 9 vdev_file_physical_ashift 9 vdev_removal_max_span 32768 vdev_validate_skip 0 zap_iterate_prefetch 1 zfetch_array_rd_sz 1048576 zfetch_max_distance 8388608 zfetch_max_idistance 67108864 zfetch_max_streams 8 zfetch_min_sec_reap 2 zfs_abd_scatter_enabled 1 zfs_abd_scatter_max_order 10 zfs_abd_scatter_min_size 1536 zfs_admin_snapshot 0 zfs_allow_redacted_dataset_mount 0 zfs_arc_average_blocksize 8192 zfs_arc_dnode_limit 0 zfs_arc_dnode_limit_percent 50 zfs_arc_dnode_reduce_percent 10 zfs_arc_evict_batch_limit 10 zfs_arc_eviction_pct 200 zfs_arc_grow_retry 0 zfs_arc_lotsfree_percent 10 zfs_arc_max 86762369024 zfs_arc_meta_adjust_restarts 4096 zfs_arc_meta_limit 0 zfs_arc_meta_limit_percent 90 zfs_arc_meta_min 0 zfs_arc_meta_prune 10000 zfs_arc_meta_strategy 1 zfs_arc_min 0 zfs_arc_min_prefetch_ms 0 zfs_arc_min_prescient_prefetch_ms 0 zfs_arc_p_dampener_disable 1 zfs_arc_p_min_shift 0 zfs_arc_pc_percent 0 zfs_arc_shrink_shift 0 zfs_arc_shrinker_limit 10000 zfs_arc_sys_free 0 zfs_async_block_max_blocks 18446744073709551615 zfs_autoimport_disable 1 zfs_checksum_events_per_second 20 zfs_commit_timeout_pct 5 zfs_compressed_arc_enabled 1 zfs_condense_indirect_commit_entry_delay_ms 0 zfs_condense_indirect_obsolete_pct 25 zfs_condense_indirect_vdevs_enable 1 zfs_condense_max_obsolete_bytes 1073741824 zfs_condense_min_mapping_bytes 131072 zfs_dbgmsg_enable 1 zfs_dbgmsg_maxsize 4194304 zfs_dbuf_state_index 0 zfs_ddt_data_is_special 1 zfs_deadman_checktime_ms 60000 zfs_deadman_enabled 1 zfs_deadman_failmode wait zfs_deadman_synctime_ms 600000 zfs_deadman_ziotime_ms 300000 zfs_dedup_prefetch 0 zfs_delay_min_dirty_percent 60 zfs_delay_scale 500000 zfs_delete_blocks 20480 zfs_dirty_data_max 4294967296 zfs_dirty_data_max_max 4294967296 zfs_dirty_data_max_max_percent 25 zfs_dirty_data_max_percent 10 zfs_dirty_data_sync_percent 20 zfs_disable_ivset_guid_check 0 zfs_dmu_offset_next_sync 0 zfs_embedded_slog_min_ms 64 zfs_expire_snapshot 300 zfs_fallocate_reserve_percent 110 zfs_flags 0 zfs_free_bpobj_enabled 1 zfs_free_leak_on_eio 0 zfs_free_min_time_ms 1000 zfs_history_output_max 1048576 zfs_immediate_write_sz 32768 zfs_initialize_chunk_size 1048576 zfs_initialize_value 16045690984833335022 zfs_keep_log_spacemaps_at_export 0 zfs_key_max_salt_uses 400000000 zfs_livelist_condense_new_alloc 0 zfs_livelist_condense_sync_cancel 0 zfs_livelist_condense_sync_pause 0 zfs_livelist_condense_zthr_cancel 0 zfs_livelist_condense_zthr_pause 0 zfs_livelist_max_entries 500000 zfs_livelist_min_percent_shared 75 zfs_lua_max_instrlimit 100000000 zfs_lua_max_memlimit 104857600 zfs_max_async_dedup_frees 100000 zfs_max_log_walking 5 zfs_max_logsm_summary_length 10 zfs_max_missing_tvds 0 zfs_max_nvlist_src_size 0 zfs_max_recordsize 1048576 zfs_metaslab_find_max_tries 100 zfs_metaslab_fragmentation_threshold 70 zfs_metaslab_max_size_cache_sec 3600 zfs_metaslab_mem_limit 25 zfs_metaslab_segment_weight_enabled 1 zfs_metaslab_switch_threshold 2 zfs_metaslab_try_hard_before_gang 0 zfs_mg_fragmentation_threshold 95 zfs_mg_noalloc_threshold 0 zfs_min_metaslabs_to_flush 1 zfs_multihost_fail_intervals 10 zfs_multihost_history 0 zfs_multihost_import_intervals 20 zfs_multihost_interval 1000 zfs_multilist_num_sublists 0 zfs_no_scrub_io 0 zfs_no_scrub_prefetch 0 zfs_nocacheflush 0 zfs_nopwrite_enabled 1 zfs_object_mutex_size 64 zfs_obsolete_min_time_ms 500 zfs_override_estimate_recordsize 0 zfs_pd_bytes_max 52428800 zfs_per_txg_dirty_frees_percent 5 zfs_prefetch_disable 0 zfs_read_history 0 zfs_read_history_hits 0 zfs_rebuild_max_segment 1048576 zfs_rebuild_scrub_enabled 1 zfs_rebuild_vdev_limit 33554432 zfs_reconstruct_indirect_combinations_max 4096 zfs_recover 0 zfs_recv_queue_ff 20 zfs_recv_queue_length 16777216 zfs_recv_write_batch_size 1048576 zfs_removal_ignore_errors 0 zfs_removal_suspend_progress 0 zfs_remove_max_segment 16777216 zfs_resilver_disable_defer 0 zfs_resilver_min_time_ms 3000 zfs_scan_checkpoint_intval 7200 zfs_scan_fill_weight 3 zfs_scan_ignore_errors 0 zfs_scan_issue_strategy 0 zfs_scan_legacy 0 zfs_scan_max_ext_gap 2097152 zfs_scan_mem_lim_fact 20 zfs_scan_mem_lim_soft_fact 20 zfs_scan_strict_mem_lim 0 zfs_scan_suspend_progress 0 zfs_scan_vdev_limit 4194304 zfs_scrub_min_time_ms 1000 zfs_send_corrupt_data 0 zfs_send_no_prefetch_queue_ff 20 zfs_send_no_prefetch_queue_length 1048576 zfs_send_queue_ff 20 zfs_send_queue_length 16777216 zfs_send_unmodified_spill_blocks 1 zfs_slow_io_events_per_second 20 zfs_spa_discard_memory_limit 16777216 zfs_special_class_metadata_reserve_pct 25 zfs_sync_pass_deferred_free 2 zfs_sync_pass_dont_compress 8 zfs_sync_pass_rewrite 2 zfs_sync_taskq_batch_pct 75 zfs_traverse_indirect_prefetch_limit 32 zfs_trim_extent_bytes_max 134217728 zfs_trim_extent_bytes_min 32768 zfs_trim_metaslab_skip 0 zfs_trim_queue_limit 10 zfs_trim_txg_batch 32 zfs_txg_history 100 zfs_txg_timeout 5 zfs_unflushed_log_block_max 262144 zfs_unflushed_log_block_min 1000 zfs_unflushed_log_block_pct 400 zfs_unflushed_max_mem_amt 1073741824 zfs_unflushed_max_mem_ppm 1000 zfs_unlink_suspend_progress 0 zfs_user_indirect_is_special 1 zfs_vdev_aggregate_trim 0 zfs_vdev_aggregation_limit 1048576 zfs_vdev_aggregation_limit_non_rotating 131072 zfs_vdev_async_read_max_active 3 zfs_vdev_async_read_min_active 1 zfs_vdev_async_write_active_max_dirty_percent 60 zfs_vdev_async_write_active_min_dirty_percent 30 zfs_vdev_async_write_max_active 10 zfs_vdev_async_write_min_active 2 zfs_vdev_cache_bshift 16 zfs_vdev_cache_max 16384 zfs_vdev_cache_size 0 zfs_vdev_default_ms_count 200 zfs_vdev_default_ms_shift 29 zfs_vdev_initializing_max_active 1 zfs_vdev_initializing_min_active 1 zfs_vdev_max_active 1000 zfs_vdev_max_auto_ashift 16 zfs_vdev_min_auto_ashift 9 zfs_vdev_min_ms_count 16 zfs_vdev_mirror_non_rotating_inc 0 zfs_vdev_mirror_non_rotating_seek_inc 1 zfs_vdev_mirror_rotating_inc 0 zfs_vdev_mirror_rotating_seek_inc 5 zfs_vdev_mirror_rotating_seek_offset 1048576 zfs_vdev_ms_count_limit 131072 zfs_vdev_nia_credit 5 zfs_vdev_nia_delay 5 zfs_vdev_queue_depth_pct 1000 zfs_vdev_raidz_impl cycle [fastest] original scalar sse2 ssse3 avx2 avx512f avx512bw zfs_vdev_read_gap_limit 32768 zfs_vdev_rebuild_max_active 3 zfs_vdev_rebuild_min_active 1 zfs_vdev_removal_max_active 2 zfs_vdev_removal_min_active 1 zfs_vdev_scheduler unused zfs_vdev_scrub_max_active 3 zfs_vdev_scrub_min_active 1 zfs_vdev_sync_read_max_active 10 zfs_vdev_sync_read_min_active 10 zfs_vdev_sync_write_max_active 10 zfs_vdev_sync_write_min_active 10 zfs_vdev_trim_max_active 2 zfs_vdev_trim_min_active 1 zfs_vdev_write_gap_limit 4096 zfs_vnops_read_chunk_size 1048576 zfs_zevent_len_max 512 zfs_zevent_retain_expire_secs 900 zfs_zevent_retain_max 2000 zfs_zil_clean_taskq_maxalloc 1048576 zfs_zil_clean_taskq_minalloc 1024 zfs_zil_clean_taskq_nthr_pct 100 zil_maxblocksize 131072 zil_nocacheflush 0 zil_replay_disable 0 zil_slog_bulk 786432 zio_deadman_log_all 0 zio_dva_throttle_enabled 1 zio_requeue_io_start_cut_in_line 1 zio_slow_io_ms 30000 zio_taskq_batch_pct 80 zio_taskq_batch_tpq 0 zvol_inhibit_dev 0 zvol_major 230 zvol_max_discard_blocks 16384 zvol_prefetch_bytes 131072 zvol_request_sync 0 zvol_threads 32 zvol_volmode 1 VDEV cache disabled, skipping section ZIL committed transactions: 3.2G Commit requests: 404.3M Flushes to stable storage: 369.8M Transactions to SLOG storage pool: 167.6 TiB 1.6G Transactions to non-SLOG storage pool: 0 Bytes 0 ``` #### zed logs ``` -- Logs begin at Wed 2022-01-12 17:47:33 UTC, end at Wed 2022-02-09 17:58:06 UTC. -- Jan 19 02:15:35 r8-n9 systemd[1]: Started ZFS Event Daemon (zed). Jan 19 02:15:35 r8-n9 zed[246555]: ZFS Event Daemon 2.1.2-1 (PID 246555) Jan 19 02:15:35 r8-n9 zed[246555]: Processing events since eid=0 Jan 19 02:16:16 r8-n9 zed[249150]: eid=38 class=config_sync pool='pod-10' Jan 19 02:25:41 r8-n9 zed[256421]: eid=45 class=vdev_add pool='pod-10' Feb 09 14:46:20 r8-n9 zed[1838695]: eid=58 class=delay pool='pod-10' vdev=35000c500ae95cc5b size=4096 offset=4740717096960 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1036122 Feb 09 14:46:20 r8-n9 zed[1838693]: eid=57 class=delay pool='pod-10' vdev=35000c500ae970e3b size=20480 offset=4785921613824 priority=0 err=0 flags=0x40080c80 delay=30174ms Feb 09 14:46:20 r8-n9 zed[1838700]: eid=60 class=delay pool='pod-10' vdev=35000c500ae34d4c7 size=4096 offset=4731204771840 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1059557 Feb 09 14:46:20 r8-n9 zed[1838703]: eid=62 class=delay pool='pod-10' vdev=35000c500ae96a4a3 size=4096 offset=4785771388928 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1033656 Feb 09 14:46:21 r8-n9 zed[1838803]: eid=70 class=delay pool='pod-10' vdev=35000c500ae95d66f size=4096 offset=4774821298176 priority=0 err=0 flags=0x180980 delay=30547ms bookmark=269:0:0:1041464 Feb 09 14:46:21 r8-n9 zed[1838814]: eid=76 class=delay pool='pod-10' vdev=35000c500ae62c433 size=4096 offset=4721981607936 priority=0 err=0 flags=0x180980 delay=30159ms bookmark=269:0:0:1044753 Feb 09 14:46:22 r8-n9 zed[1838836]: eid=77 class=delay pool='pod-10' vdev=35000c500ae29d97f size=4096 offset=4781092179968 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1053826 Feb 09 14:46:22 r8-n9 zed[1838842]: eid=78 class=delay pool='pod-10' vdev=35000c500ae29a4bb size=4096 offset=4776777699328 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1031840 Feb 09 14:46:22 r8-n9 zed[1838846]: eid=82 class=delay pool='pod-10' vdev=35000c500ae96aa6f size=4096 offset=4786110545920 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1033105 Feb 09 14:46:22 r8-n9 zed[1838875]: eid=84 class=delay pool='pod-10' vdev=35000c500ae95555b size=53248 offset=4774958784512 priority=0 err=0 flags=0x40080c80 delay=30460ms Feb 09 14:46:22 r8-n9 zed[1838878]: eid=83 class=delay pool='pod-10' vdev=35000c500ae957c3b size=4096 offset=4743807332352 priority=0 err=0 flags=0x180980 delay=30668ms bookmark=269:0:0:1054047 Feb 09 14:46:22 r8-n9 zed[1838884]: eid=87 class=delay pool='pod-10' vdev=35000c500ae96d793 size=4096 offset=4768750641152 priority=0 err=0 flags=0x180980 delay=30444ms bookmark=269:0:0:1042517 Feb 09 14:46:22 r8-n9 zed[1838935]: eid=89 class=delay pool='pod-10' vdev=35000c500ae972b2f size=4096 offset=4769971810304 priority=0 err=0 flags=0x180980 delay=30623ms bookmark=269:0:0:1033391 Feb 09 14:46:22 r8-n9 zed[1838942]: eid=91 class=delay pool='pod-10' vdev=35000c500ae9737c3 size=4096 offset=4728066199552 priority=0 err=0 flags=0x180980 delay=30623ms bookmark=269:0:0:1057873 Feb 09 14:46:22 r8-n9 zed[1838963]: eid=94 class=delay pool='pod-10' vdev=35000c500ae95d53b size=4096 offset=4712411389952 priority=0 err=0 flags=0x180980 delay=30831ms bookmark=269:0:0:1049636 Feb 09 14:46:22 r8-n9 zed[1838967]: eid=95 class=delay pool='pod-10' vdev=35000c500ae970cd7 size=4096 offset=4776225902592 priority=0 err=0 flags=0x180980 delay=30626ms bookmark=269:0:0:1046985 Feb 09 14:46:22 r8-n9 zed[1838972]: eid=97 class=delay pool='pod-10' vdev=35000c500ae2a6e9b size=4096 offset=4782662496256 priority=0 err=0 flags=0x180980 delay=30831ms bookmark=269:0:0:1048773 Feb 09 14:46:22 r8-n9 zed[1838969]: eid=96 class=delay pool='pod-10' vdev=35000c500ae955eef size=20480 offset=4774534471680 priority=0 err=0 flags=0x40080c80 delay=30626ms ``` #### vmstat ``` # cat /proc/vmstat nr_free_pages 497162 nr_zone_inactive_anon 77677 nr_zone_active_anon 81958 nr_zone_inactive_file 16392 nr_zone_active_file 10007 nr_zone_unevictable 7960 nr_zone_write_pending 844 nr_mlock 7960 nr_page_table_pages 1963 nr_kernel_stack 19296 nr_bounce 0 nr_zspages 0 nr_free_cma 0 numa_hit 151121274140 numa_miss 652564522 numa_foreign 652564522 numa_interleave 96982 numa_local 133392770941 numa_other 18381067721 nr_inactive_anon 77677 nr_active_anon 81958 nr_inactive_file 16392 nr_active_file 10007 nr_unevictable 7960 nr_slab_reclaimable 1104575 nr_slab_unreclaimable 28741700 nr_isolated_anon 0 nr_isolated_file 0 workingset_nodes 1973 workingset_refault 78664 workingset_activate 26788 workingset_restore 14173 workingset_nodereclaim 816 nr_anon_pages 162434 nr_mapped 20118 nr_file_pages 32246 nr_dirty 844 nr_writeback 0 nr_writeback_temp 0 nr_shmem 369 nr_shmem_hugepages 0 nr_shmem_pmdmapped 0 nr_file_hugepages 0 nr_file_pmdmapped 0 nr_anon_transparent_hugepages 0 nr_unstable 0 nr_vmscan_write 68263 nr_vmscan_immediate_reclaim 641 nr_dirtied 3379059 nr_written 3113482 nr_kernel_misc_reclaimable 0 nr_dirty_threshold 73743 nr_dirty_background_threshold 36826 pgpgin 194613865828 pgpgout 594833359918 pswpin 4218 pswpout 68258 pgalloc_dma 0 pgalloc_dma32 324399540 pgalloc_normal 245349772461 pgalloc_movable 0 allocstall_dma 0 allocstall_dma32 0 allocstall_normal 1 allocstall_movable 2 pgskip_dma 0 pgskip_dma32 0 pgskip_normal 0 pgskip_movable 0 pgfree 245674776191 pgactivate 730915 pgdeactivate 358591 pglazyfree 6389 pgfault 243384077 pgmajfault 15967 pglazyfreed 0 pgrefill 382033 pgsteal_kswapd 292238 pgsteal_direct 1375 pgscan_kswapd 477828 pgscan_direct 1399 pgscan_direct_throttle 0 zone_reclaim_failed 0 pginodesteal 74 slabs_scanned 459729476 kswapd_inodesteal 26293 kswapd_low_wmark_hit_quickly 1 kswapd_high_wmark_hit_quickly 14 pageoutrun 277 pgrotated 69682 drop_pagecache 1 drop_slab 1 oom_kill 0 numa_pte_updates 4332995 numa_huge_pte_updates 12 numa_hint_faults 3863526 numa_hint_faults_local 3281256 numa_pages_migrated 327088 pgmigrate_success 411548 pgmigrate_fail 3419 compact_migrate_scanned 10896483 compact_free_scanned 460891 compact_isolated 170144 compact_stall 4 compact_fail 4 compact_success 0 compact_daemon_wake 255 compact_daemon_migrate_scanned 1507268 compact_daemon_free_scanned 317747 htlb_buddy_alloc_success 0 htlb_buddy_alloc_fail 0 unevictable_pgs_culled 81639 unevictable_pgs_scanned 0 unevictable_pgs_rescued 14620 unevictable_pgs_mlocked 26080 unevictable_pgs_munlocked 16730 unevictable_pgs_cleared 1390 unevictable_pgs_stranded 1390 thp_fault_alloc 10 thp_fault_fallback 0 thp_collapse_alloc 8 thp_collapse_alloc_failed 2 thp_file_alloc 0 thp_file_mapped 0 thp_split_page 0 thp_split_page_failed 0 thp_deferred_split_page 18 thp_split_pmd 9 thp_split_pud 0 thp_zero_page_alloc 0 thp_zero_page_alloc_failed 0 thp_swpout 0 thp_swpout_fallback 0 balloon_inflate 0 balloon_deflate 0 balloon_migrate 0 swap_ra 1519 swap_ra_hit 833 ``` #### buddyinfo ``` Node: 0 Zone: DMA Free KiB in zone: 15876.00 Fragment size Free fragments Total available KiB 4096 1 4.0 8192 0 0.0 16384 0 0.0 32768 0 0.0 65536 2 128.0 131072 1 128.0 262144 1 256.0 524288 0 0.0 1048576 1 1024.0 2097152 1 2048.0 4194304 3 12288.0 Zone: DMA32 Free KiB in zone: 252500.00 Fragment size Free fragments Total available KiB 4096 1923 7692.0 8192 5143 41144.0 16384 2581 41296.0 32768 142 4544.0 65536 26 1664.0 131072 16 2048.0 262144 78 19968.0 524288 70 35840.0 1048576 36 36864.0 2097152 22 45056.0 4194304 4 16384.0 Zone: Normal Free KiB in zone: 107160.00 Fragment size Free fragments Total available KiB 4096 1586 6344.0 8192 8328 66624.0 16384 1987 31792.0 32768 27 864.0 65536 24 1536.0 131072 0 0.0 262144 0 0.0 524288 0 0.0 1048576 0 0.0 2097152 0 0.0 4194304 0 0.0 Node: 1 Zone: Normal Free KiB in zone: 2152516.00 Fragment size Free fragments Total available KiB 4096 149157 596628.0 8192 107244 857952.0 16384 3545 56720.0 32768 4110 131520.0 65536 3156 201984.0 131072 1610 206080.0 262144 297 76032.0 524288 48 24576.0 1048576 1 1024.0 2097152 0 0.0 4194304 0 0.0 ``` ### Describe how to reproduce the problem Uncertain but it has occurred on 2 separate servers so it is likely to happen again. ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
1.0
ZFS 2.1.2 + NFSv4 ARC stuck pruning/evicting, flatlines throughput - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | `Ubuntu` Distribution Version | `20.04.3 LTS` Kernel Version | `5.4.0-96-generic` Architecture | `x86_64` OpenZFS Version | `zfs-2.1.2-1` <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing ZFS is flatlined on throughput with an `arc_evict` and `arc_prune` process spinning at 100%. The workload is kernel NFS server (all NFSv4 clients) with ZFS 2.1.2 built from source. Characterized by high CPU `iowait` and throughput flatlining. #### RAM Graph <img width="1350" alt="Screen Shot 2022-02-09 at 9 06 55 AM" src="https://user-images.githubusercontent.com/16908622/153252531-9ee24c8e-3ecb-4242-b862-b860353f5625.png"> It's holding on to RAM pretty hard. This is where I begin to lose debugging expertise, having discovered slabs yesterday. :-) #### top ``` top - 16:55:29 up 21 days, 14:56, 1 user, load average: 195.23, 124.60, 134.24 Tasks: 993 total, 2 running, 991 sleeping, 0 stopped, 0 zombie %Cpu(s): 0.0 us, 7.3 sy, 0.0 ni, 5.0 id, 87.3 wa, 0.0 hi, 0.4 si, 0.0 st MiB Mem : 128622.1 total, 2501.0 free, 121637.5 used, 4483.5 buff/cache MiB Swap: 1907.0 total, 1042.7 free, 864.2 used. 6078.4 avail Mem PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 245944 root 20 0 0 0 0 R 100.0 0.0 549:36.86 arc_evict 245943 root 20 0 0 0 0 S 93.1 0.0 459:13.30 arc_prune 246555 root 20 0 953064 20876 2732 S 3.6 0.0 29:32.85 zed 3174406 root 20 0 12784 4872 3248 R 1.3 0.0 0:00.17 top 14154 root rt 0 356604 31088 8300 S 0.7 0.0 133:03.60 multipathd 3159749 root 20 0 0 0 0 D 0.7 0.0 0:00.11 nfsd ``` #### slabtop ``` # slabtop Active / Total Objects (% used) : 215263329 / 222780310 (96.6%) Active / Total Slabs (% used) : 6020764 / 6020764 (100.0%) Active / Total Caches (% used) : 141 / 199 (70.9%) Active / Total Size (% used) : 98093294.62K / 99549181.36K (98.5%) Minimum / Average / Maximum Object : 0.01K / 0.45K / 16.75K OBJS ACTIVE USE OBJ SIZE SLABS OBJ/SLAB CACHE SIZE NAME 27437256 27437256 100% 0.97K 831432 33 26605824K dnode_t 25866414 25866282 99% 0.38K 615867 42 9853872K dmu_buf_impl_t 24953184 24952001 99% 0.50K 779787 32 12476592K kmalloc-512 20654336 18708128 90% 0.06K 322724 64 1290896K kmalloc-64 14428134 14148450 98% 0.09K 343527 42 1374108K arc_buf_hdr_t_l2only 10290924 10290229 99% 0.09K 245022 42 980088K kmalloc-96 10017152 9881010 98% 0.03K 78259 128 313036K kmalloc-32 7602798 7575100 99% 0.19K 181019 42 1448152K dentry 7255278 7139757 98% 1.09K 250182 29 8005824K zfs_znode_cache 7225482 7140416 98% 0.24K 218954 33 1751632K sa_cache 5172144 5172144 100% 0.16K 107753 48 862024K nfsd4_stateids 5079424 5077737 99% 0.25K 158732 32 1269856K filp 4988088 4985288 99% 0.19K 118764 42 950112K cred_jar 4952610 4952610 100% 0.02K 29133 170 116532K lsm_file_cache 4951520 4951520 100% 0.28K 176840 28 1414720K nfsd4_files 4949376 4949376 100% 0.03K 38667 128 154668K fsnotify_mark_connector 4947816 4947816 100% 0.08K 97016 51 388064K Acpi-State 4947372 4947372 100% 0.11K 137427 36 549708K khugepaged_mm_slot 4862528 4858723 99% 0.06K 75977 64 303908K kmalloc-rcl-64 3813933 1054784 27% 0.31K 74783 51 1196528K arc_buf_hdr_t_full 3601017 3601017 100% 0.05K 49329 73 197316K nsproxy 3022149 1059649 35% 0.10K 77491 39 309964K abd_t 2598288 2591411 99% 0.57K 92796 28 1484736K radix_tree_node 1523068 1523068 100% 0.42K 41164 37 658624K nfsd4_openowners 1248128 1247029 99% 1.00K 39004 32 1248128K kmalloc-1k 1128681 1052509 93% 0.08K 22131 51 88524K arc_buf_t 960386 960375 99% 16.00K 480193 2 15366176K zio_buf_comb_16384 945488 942004 99% 8.00K 236372 4 7563904K kmalloc-8k 290745 268913 92% 0.10K 7455 39 29820K buffer_head 288000 287664 99% 0.13K 9600 30 38400K kernfs_node_cache 228160 226423 99% 0.06K 3565 64 14260K anon_vma_chain 211894 204295 96% 0.59K 3998 53 127936K inode_cache 209664 209664 100% 0.02K 819 256 3276K kmalloc-16 187432 187432 100% 0.07K 3347 56 13388K Acpi-Operand 169611 168672 99% 0.20K 4349 39 34792K vm_area_struct 121856 121856 100% 0.01K 238 512 952K kmalloc-8 116978 116702 99% 0.09K 2543 46 10172K anon_vma 107856 99165 91% 0.09K 2568 42 10272K kmalloc-rcl-96 ``` #### zpool ``` # zpool status pod-10 pool: pod-10 state: ONLINE config: NAME STATE READ WRITE CKSUM pod-10 ONLINE 0 0 0 raidz3-0 ONLINE 0 0 0 35000c500ae29a4bb ONLINE 0 0 0 35000c500ae95def3 ONLINE 0 0 0 35000c500ae96d1d7 ONLINE 0 0 0 35000c500ae9729af ONLINE 0 0 0 35000c500ae97296f ONLINE 0 0 0 35000c500ae96daeb ONLINE 0 0 0 35000c500ae968243 ONLINE 0 0 0 35000c500ae97269b ONLINE 0 0 0 35000c500ae970cd7 ONLINE 0 0 0 35000c500ae975a7f ONLINE 0 0 0 35000c500ae957c3b ONLINE 0 0 0 raidz3-1 ONLINE 0 0 0 35000c500ae96870f ONLINE 0 0 0 35000c500ae2bc057 ONLINE 0 0 0 35000c500ae9733b7 ONLINE 0 0 0 35000c500ae2bc4db ONLINE 0 0 0 35000c500ae96b4bf ONLINE 0 0 0 35000c500ae970e3b ONLINE 0 0 0 35000c500ae957bab ONLINE 0 0 0 35000c500ae96aa6f ONLINE 0 0 0 35000c500ae96833f ONLINE 0 0 0 35000c500ae96a4a3 ONLINE 0 0 0 35000c500ae2a6e9b ONLINE 0 0 0 raidz3-2 ONLINE 0 0 0 35000c500ae96c3f7 ONLINE 0 0 0 35000c500ae972ddf ONLINE 0 0 0 35000c500ae96bb4f ONLINE 0 0 0 35000c500ae95d66f ONLINE 0 0 0 35000c500ae96777f ONLINE 0 0 0 35000c500ae60fec3 ONLINE 0 0 0 35000c500ae96fcef ONLINE 0 0 0 35000c500ae966b0b ONLINE 0 0 0 35000c500ae96c823 ONLINE 0 0 0 35000c500ae95e363 ONLINE 0 0 0 35000c500ae96fab3 ONLINE 0 0 0 raidz3-3 ONLINE 0 0 0 35000c500ae34ceeb ONLINE 0 0 0 35000c500ae34d4c7 ONLINE 0 0 0 35000c500ae970af7 ONLINE 0 0 0 35000c500ae9597cb ONLINE 0 0 0 35000c500ae62c433 ONLINE 0 0 0 35000c500ae968857 ONLINE 0 0 0 35000c500ae970167 ONLINE 0 0 0 35000c500ae63517f ONLINE 0 0 0 35000c500ae961313 ONLINE 0 0 0 35000c500ae95d53b ONLINE 0 0 0 35000c500ae95cc5b ONLINE 0 0 0 raidz3-4 ONLINE 0 0 0 35000c500ae9737c3 ONLINE 0 0 0 35000c500ae970feb ONLINE 0 0 0 35000c500ae9686f3 ONLINE 0 0 0 35000c500ae97387b ONLINE 0 0 0 35000c500ae97403f ONLINE 0 0 0 35000c500ae95711f ONLINE 0 0 0 35000c500ae96cb23 ONLINE 0 0 0 35000c500ae2a6db7 ONLINE 0 0 0 35000c500ae9681a3 ONLINE 0 0 0 35000c500ae9688b7 ONLINE 0 0 0 35000c500ae97404b ONLINE 0 0 0 raidz3-5 ONLINE 0 0 0 35000c500ae956b87 ONLINE 0 0 0 35000c500ae974bf3 ONLINE 0 0 0 35000c500ae9744fb ONLINE 0 0 0 35000c500ae29e993 ONLINE 0 0 0 35000c500ae96ef57 ONLINE 0 0 0 35000c500ae974d2b ONLINE 0 0 0 35000c500ae970a0f ONLINE 0 0 0 35000c500ae39455f ONLINE 0 0 0 35000c500ae29d97f ONLINE 0 0 0 35000c500ae95712b ONLINE 0 0 0 35000c500ae9742c3 ONLINE 0 0 0 raidz3-6 ONLINE 0 0 0 35000c500ae96fc4b ONLINE 0 0 0 35000c500ae955eef ONLINE 0 0 0 35000c500ae95c243 ONLINE 0 0 0 35000c500ae974057 ONLINE 0 0 0 35000c500ae95e4cb ONLINE 0 0 0 35000c500ae96eab3 ONLINE 0 0 0 35000c500ae96c9bb ONLINE 0 0 0 35000c500ae959fd7 ONLINE 0 0 0 35000c500ae2a709f ONLINE 0 0 0 35000c500ae96d793 ONLINE 0 0 0 35000c500ae9728ab ONLINE 0 0 0 raidz3-7 ONLINE 0 0 0 35000c500ae962117 ONLINE 0 0 0 35000c500ae4044bf ONLINE 0 0 0 35000c500ae393b9b ONLINE 0 0 0 35000c500ae974223 ONLINE 0 0 0 35000c500ae95f193 ONLINE 0 0 0 35000c500ae957abb ONLINE 0 0 0 35000c500ae9563a3 ONLINE 0 0 0 35000c500ae9683ab ONLINE 0 0 0 35000c500ae96df8f ONLINE 0 0 0 35000c500ae96d543 ONLINE 0 0 0 35000c500ae972b2f ONLINE 0 0 0 raidz3-8 ONLINE 0 0 0 35000c500ae960517 ONLINE 0 0 0 35000c500ae95555b ONLINE 0 0 0 35000c500ae974637 ONLINE 0 0 0 35000c500ae972dd7 ONLINE 0 0 0 35000c500ae4323c7 ONLINE 0 0 0 35000c500ae96d617 ONLINE 0 0 0 35000c500ae9638b7 ONLINE 0 0 0 35000c500ae96ea0f ONLINE 0 0 0 35000c500ae96e3eb ONLINE 0 0 0 35000c500ae29a417 ONLINE 0 0 0 35000c500ae96fda3 ONLINE 0 0 0 logs mirror-9 ONLINE 0 0 0 wwn-0x55cd2e4152220c93-part1 ONLINE 0 0 0 wwn-0x55cd2e41519b023b-part1 ONLINE 0 0 0 cache wwn-0x55cd2e4152220c93-part2 ONLINE 0 0 0 wwn-0x55cd2e41519b023b-part2 ONLINE 0 0 0 wwn-0x55cd2e415221c8c7 ONLINE 0 0 0 spares 35000c500ae97206f AVAIL 35000c500ae965dbb AVAIL 35000c500ae2b926f AVAIL 35000c500ae976187 AVAIL 35000c500ae96dcf7 AVAIL 35000c500ae958777 AVAIL 35000c500ae41875f AVAIL errors: No known data errors ``` #### zfs fs ``` # zfs list NAME USED AVAIL REFER MOUNTPOINT pod-10 120T 926T 279K /pod-10 pod-10/pod-10 120T 926T 120T /srv/pod-10 ``` ``` # zfs get primarycache,secondarycache pod-10/pod-10 NAME PROPERTY VALUE SOURCE pod-10/pod-10 primarycache metadata local pod-10/pod-10 secondarycache metadata local ``` Intention is for ARC/L2ARC to be entirely metadata. #### initial arcstats arcstats reported metadata usage above the limit. ``` ARC size (current): 102.8 % 64.6 GiB Target size (adaptive): 100.0 % 62.8 GiB Min size (hard limit): 6.2 % 3.9 GiB Max size (high water): 16:1 62.8 GiB Most Frequently Used (MFU) cache size: 13.0 % 2.6 GiB Most Recently Used (MRU) cache size: 87.0 % 17.7 GiB Metadata cache size (hard limit): 75.0 % 47.1 GiB Metadata cache size (current): 135.8 % 64.0 GiB Dnode cache size (hard limit): 10.0 % 4.7 GiB Dnode cache size (current): 537.3 % 25.3 GiB ``` #### initial adjustment Attempted to remediate by increasing ARC size by half of remaining RAM and increasing the metadata allocation in the ARC. ``` echo 86762369024 > /sys/module/zfs/parameters/zfs_arc_max echo 90 > /sys/module/zfs/parameters/zfs_arc_meta_limit_percent echo 50 > /sys/module/zfs/parameters/zfs_arc_dnode_limit_percent ``` #### current issue This remediated the issue temporarily, and the prune processes stopped, but we are back! :) Stopped NFS server and it seems to be free-ing memory albeit very slowly, many `dp_sync_taskq` processes... But removing the NFS server is very un-ideal. Given they are both living in the kernel, it's difficult for me personally to determine who is eating up the memory. ``` top - 18:05:52 up 21 days, 16:07, 2 users, load average: 2.52, 7.31, 35.56 Tasks: 807 total, 1 running, 806 sleeping, 0 stopped, 0 zombie %Cpu(s): 0.0 us, 1.5 sy, 0.0 ni, 95.9 id, 2.5 wa, 0.0 hi, 0.1 si, 0.0 st MiB Mem : 128622.1 total, 17620.3 free, 108471.5 used, 2530.3 buff/cache MiB Swap: 1907.0 total, 1764.7 free, 142.2 used. 19277.1 avail Mem PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 248929 root 20 0 0 0 0 S 2.3 0.0 349:34.92 txg_sync 248864 root 39 19 0 0 0 S 1.7 0.0 33:52.18 dp_sync_taskq 1767630 root 20 0 0 0 0 D 1.7 0.0 3:35.82 kworker/23:2+events 248854 root 39 19 0 0 0 S 1.3 0.0 33:48.50 dp_sync_taskq 248855 root 39 19 0 0 0 S 1.3 0.0 33:51.96 dp_sync_taskq 248856 root 39 19 0 0 0 S 1.3 0.0 33:49.72 dp_sync_taskq 248857 root 39 19 0 0 0 S 1.3 0.0 33:49.15 dp_sync_taskq 248859 root 39 19 0 0 0 S 1.3 0.0 33:54.62 dp_sync_taskq 248860 root 39 19 0 0 0 S 1.3 0.0 33:50.09 dp_sync_taskq 248862 root 39 19 0 0 0 S 1.3 0.0 33:49.23 dp_sync_taskq 248865 root 39 19 0 0 0 S 1.3 0.0 33:50.26 dp_sync_taskq 248867 root 39 19 0 0 0 S 1.3 0.0 33:48.57 dp_sync_taskq 248870 root 39 19 0 0 0 S 1.3 0.0 33:51.20 dp_sync_taskq 248871 root 39 19 0 0 0 S 1.3 0.0 33:48.09 dp_sync_taskq 248872 root 39 19 0 0 0 S 1.3 0.0 33:51.06 dp_sync_taskq 248873 root 39 19 0 0 0 S 1.3 0.0 33:46.16 dp_sync_taskq ``` ``` # arc_summary ------------------------------------------------------------------------ ZFS Subsystem Report Wed Feb 09 16:23:14 2022 Linux 5.4.0-96-generic 2.1.2-1 Machine: r8-n9 (x86_64) 2.1.2-1 ARC status: HEALTHY Memory throttle count: 0 ARC size (current): 90.1 % 72.8 GiB Target size (adaptive): 4.9 % 3.9 GiB Min size (hard limit): 4.9 % 3.9 GiB Max size (high water): 20:1 80.8 GiB Most Frequently Used (MFU) cache size: 13.8 % 3.0 GiB Most Recently Used (MRU) cache size: 86.2 % 18.7 GiB Metadata cache size (hard limit): 90.0 % 72.7 GiB Metadata cache size (current): 100.2 % 72.8 GiB Dnode cache size (hard limit): 50.0 % 36.4 GiB Dnode cache size (current): 80.9 % 29.4 GiB ARC hash breakdown: Elements max: 15.7M Elements current: 97.2 % 15.2M Collisions: 1.1G Chain max: 10 Chains: 3.9M ARC misc: Deleted: 28.5M Mutex misses: 6.6G Eviction skips: 93.2G Eviction skips due to L2 writes: 10.3k L2 cached evictions: 1.4 TiB L2 eligible evictions: 73.1 GiB L2 eligible MFU evictions: 4.8 % 3.5 GiB L2 eligible MRU evictions: 95.2 % 69.6 GiB L2 ineligible evictions: 707.3 GiB ARC total accesses (hits + misses): 11.3G Cache hit ratio: 86.4 % 9.8G Cache miss ratio: 13.6 % 1.5G Actual hit ratio (MFU + MRU hits): 86.3 % 9.8G Data demand efficiency: 1.6 % 1.6G Data prefetch efficiency: n/a 0 Cache hits by cache type: Most frequently used (MFU): 96.3 % 9.4G Most recently used (MRU): 3.6 % 356.4M Most frequently used (MFU) ghost: < 0.1 % 1.3M Most recently used (MRU) ghost: < 0.1 % 1.0M Cache hits by data type: Demand data: 0.3 % 25.4M Demand prefetch data: 0.0 % 0 Demand metadata: 99.7 % 9.8G Demand prefetch metadata: < 0.1 % 2.4M Cache misses by data type: Demand data: 99.3 % 1.5G Demand prefetch data: 0.0 % 0 Demand metadata: 0.4 % 5.9M Demand prefetch metadata: 0.3 % 4.8M DMU prefetch efficiency: 2.7G Hit ratio: 5.7 % 152.9M Miss ratio: 94.3 % 2.5G L2ARC status: HEALTHY Low memory aborts: 24.5k Free on write: 281 R/W clashes: 0 Bad checksums: 0 I/O errors: 0 L2ARC size (adaptive): 1.3 TiB Compressed: 5.6 % 74.6 GiB Header size: 0.1 % 1.3 GiB MFU allocated size: 19.4 % 14.5 GiB MRU allocated size: 85.0 % 63.4 GiB Prefetch allocated size: 0.1 % 67.6 MiB Data (buffer content) allocated size: 0.0 % 0 Bytes Metadata (buffer content) allocated size: 104.5 % 77.9 GiB L2ARC breakdown: 1.5G Hit ratio: 0.2 % 3.8M Miss ratio: 99.8 % 1.5G Feeds: 1.8M L2ARC writes: Writes sent: 100 % 1.4M L2ARC evicts: Lock retries: 0 Upon reading: 0 Solaris Porting Layer (SPL): spl_hostid 0 spl_hostid_path /etc/hostid spl_kmem_alloc_max 1048576 spl_kmem_alloc_warn 65536 spl_kmem_cache_kmem_threads 4 spl_kmem_cache_magazine_size 0 spl_kmem_cache_max_size 32 spl_kmem_cache_obj_per_slab 8 spl_kmem_cache_reclaim 0 spl_kmem_cache_slab_limit 16384 spl_max_show_tasks 512 spl_panic_halt 0 spl_schedule_hrtimeout_slack_us 0 spl_taskq_kick 0 spl_taskq_thread_bind 0 spl_taskq_thread_dynamic 1 spl_taskq_thread_priority 1 spl_taskq_thread_sequential 4 Tunables: dbuf_cache_hiwater_pct 10 dbuf_cache_lowater_pct 10 dbuf_cache_max_bytes 18446744073709551615 dbuf_cache_shift 5 dbuf_metadata_cache_max_bytes 18446744073709551615 dbuf_metadata_cache_shift 6 dmu_object_alloc_chunk_shift 7 dmu_prefetch_max 134217728 ignore_hole_birth 1 l2arc_feed_again 1 l2arc_feed_min_ms 200 l2arc_feed_secs 1 l2arc_headroom 2 l2arc_headroom_boost 200 l2arc_meta_percent 33 l2arc_mfuonly 0 l2arc_noprefetch 1 l2arc_norw 0 l2arc_rebuild_blocks_min_l2size 1073741824 l2arc_rebuild_enabled 1 l2arc_trim_ahead 0 l2arc_write_boost 8388608 l2arc_write_max 8388608 metaslab_aliquot 524288 metaslab_bias_enabled 1 metaslab_debug_load 0 metaslab_debug_unload 0 metaslab_df_max_search 16777216 metaslab_df_use_largest_segment 0 metaslab_force_ganging 16777217 metaslab_fragmentation_factor_enabled 1 metaslab_lba_weighting_enabled 1 metaslab_preload_enabled 1 metaslab_unload_delay 32 metaslab_unload_delay_ms 600000 send_holes_without_birth_time 1 spa_asize_inflation 24 spa_config_path /etc/zfs/zpool.cache spa_load_print_vdev_tree 0 spa_load_verify_data 1 spa_load_verify_metadata 1 spa_load_verify_shift 4 spa_slop_shift 5 vdev_file_logical_ashift 9 vdev_file_physical_ashift 9 vdev_removal_max_span 32768 vdev_validate_skip 0 zap_iterate_prefetch 1 zfetch_array_rd_sz 1048576 zfetch_max_distance 8388608 zfetch_max_idistance 67108864 zfetch_max_streams 8 zfetch_min_sec_reap 2 zfs_abd_scatter_enabled 1 zfs_abd_scatter_max_order 10 zfs_abd_scatter_min_size 1536 zfs_admin_snapshot 0 zfs_allow_redacted_dataset_mount 0 zfs_arc_average_blocksize 8192 zfs_arc_dnode_limit 0 zfs_arc_dnode_limit_percent 50 zfs_arc_dnode_reduce_percent 10 zfs_arc_evict_batch_limit 10 zfs_arc_eviction_pct 200 zfs_arc_grow_retry 0 zfs_arc_lotsfree_percent 10 zfs_arc_max 86762369024 zfs_arc_meta_adjust_restarts 4096 zfs_arc_meta_limit 0 zfs_arc_meta_limit_percent 90 zfs_arc_meta_min 0 zfs_arc_meta_prune 10000 zfs_arc_meta_strategy 1 zfs_arc_min 0 zfs_arc_min_prefetch_ms 0 zfs_arc_min_prescient_prefetch_ms 0 zfs_arc_p_dampener_disable 1 zfs_arc_p_min_shift 0 zfs_arc_pc_percent 0 zfs_arc_shrink_shift 0 zfs_arc_shrinker_limit 10000 zfs_arc_sys_free 0 zfs_async_block_max_blocks 18446744073709551615 zfs_autoimport_disable 1 zfs_checksum_events_per_second 20 zfs_commit_timeout_pct 5 zfs_compressed_arc_enabled 1 zfs_condense_indirect_commit_entry_delay_ms 0 zfs_condense_indirect_obsolete_pct 25 zfs_condense_indirect_vdevs_enable 1 zfs_condense_max_obsolete_bytes 1073741824 zfs_condense_min_mapping_bytes 131072 zfs_dbgmsg_enable 1 zfs_dbgmsg_maxsize 4194304 zfs_dbuf_state_index 0 zfs_ddt_data_is_special 1 zfs_deadman_checktime_ms 60000 zfs_deadman_enabled 1 zfs_deadman_failmode wait zfs_deadman_synctime_ms 600000 zfs_deadman_ziotime_ms 300000 zfs_dedup_prefetch 0 zfs_delay_min_dirty_percent 60 zfs_delay_scale 500000 zfs_delete_blocks 20480 zfs_dirty_data_max 4294967296 zfs_dirty_data_max_max 4294967296 zfs_dirty_data_max_max_percent 25 zfs_dirty_data_max_percent 10 zfs_dirty_data_sync_percent 20 zfs_disable_ivset_guid_check 0 zfs_dmu_offset_next_sync 0 zfs_embedded_slog_min_ms 64 zfs_expire_snapshot 300 zfs_fallocate_reserve_percent 110 zfs_flags 0 zfs_free_bpobj_enabled 1 zfs_free_leak_on_eio 0 zfs_free_min_time_ms 1000 zfs_history_output_max 1048576 zfs_immediate_write_sz 32768 zfs_initialize_chunk_size 1048576 zfs_initialize_value 16045690984833335022 zfs_keep_log_spacemaps_at_export 0 zfs_key_max_salt_uses 400000000 zfs_livelist_condense_new_alloc 0 zfs_livelist_condense_sync_cancel 0 zfs_livelist_condense_sync_pause 0 zfs_livelist_condense_zthr_cancel 0 zfs_livelist_condense_zthr_pause 0 zfs_livelist_max_entries 500000 zfs_livelist_min_percent_shared 75 zfs_lua_max_instrlimit 100000000 zfs_lua_max_memlimit 104857600 zfs_max_async_dedup_frees 100000 zfs_max_log_walking 5 zfs_max_logsm_summary_length 10 zfs_max_missing_tvds 0 zfs_max_nvlist_src_size 0 zfs_max_recordsize 1048576 zfs_metaslab_find_max_tries 100 zfs_metaslab_fragmentation_threshold 70 zfs_metaslab_max_size_cache_sec 3600 zfs_metaslab_mem_limit 25 zfs_metaslab_segment_weight_enabled 1 zfs_metaslab_switch_threshold 2 zfs_metaslab_try_hard_before_gang 0 zfs_mg_fragmentation_threshold 95 zfs_mg_noalloc_threshold 0 zfs_min_metaslabs_to_flush 1 zfs_multihost_fail_intervals 10 zfs_multihost_history 0 zfs_multihost_import_intervals 20 zfs_multihost_interval 1000 zfs_multilist_num_sublists 0 zfs_no_scrub_io 0 zfs_no_scrub_prefetch 0 zfs_nocacheflush 0 zfs_nopwrite_enabled 1 zfs_object_mutex_size 64 zfs_obsolete_min_time_ms 500 zfs_override_estimate_recordsize 0 zfs_pd_bytes_max 52428800 zfs_per_txg_dirty_frees_percent 5 zfs_prefetch_disable 0 zfs_read_history 0 zfs_read_history_hits 0 zfs_rebuild_max_segment 1048576 zfs_rebuild_scrub_enabled 1 zfs_rebuild_vdev_limit 33554432 zfs_reconstruct_indirect_combinations_max 4096 zfs_recover 0 zfs_recv_queue_ff 20 zfs_recv_queue_length 16777216 zfs_recv_write_batch_size 1048576 zfs_removal_ignore_errors 0 zfs_removal_suspend_progress 0 zfs_remove_max_segment 16777216 zfs_resilver_disable_defer 0 zfs_resilver_min_time_ms 3000 zfs_scan_checkpoint_intval 7200 zfs_scan_fill_weight 3 zfs_scan_ignore_errors 0 zfs_scan_issue_strategy 0 zfs_scan_legacy 0 zfs_scan_max_ext_gap 2097152 zfs_scan_mem_lim_fact 20 zfs_scan_mem_lim_soft_fact 20 zfs_scan_strict_mem_lim 0 zfs_scan_suspend_progress 0 zfs_scan_vdev_limit 4194304 zfs_scrub_min_time_ms 1000 zfs_send_corrupt_data 0 zfs_send_no_prefetch_queue_ff 20 zfs_send_no_prefetch_queue_length 1048576 zfs_send_queue_ff 20 zfs_send_queue_length 16777216 zfs_send_unmodified_spill_blocks 1 zfs_slow_io_events_per_second 20 zfs_spa_discard_memory_limit 16777216 zfs_special_class_metadata_reserve_pct 25 zfs_sync_pass_deferred_free 2 zfs_sync_pass_dont_compress 8 zfs_sync_pass_rewrite 2 zfs_sync_taskq_batch_pct 75 zfs_traverse_indirect_prefetch_limit 32 zfs_trim_extent_bytes_max 134217728 zfs_trim_extent_bytes_min 32768 zfs_trim_metaslab_skip 0 zfs_trim_queue_limit 10 zfs_trim_txg_batch 32 zfs_txg_history 100 zfs_txg_timeout 5 zfs_unflushed_log_block_max 262144 zfs_unflushed_log_block_min 1000 zfs_unflushed_log_block_pct 400 zfs_unflushed_max_mem_amt 1073741824 zfs_unflushed_max_mem_ppm 1000 zfs_unlink_suspend_progress 0 zfs_user_indirect_is_special 1 zfs_vdev_aggregate_trim 0 zfs_vdev_aggregation_limit 1048576 zfs_vdev_aggregation_limit_non_rotating 131072 zfs_vdev_async_read_max_active 3 zfs_vdev_async_read_min_active 1 zfs_vdev_async_write_active_max_dirty_percent 60 zfs_vdev_async_write_active_min_dirty_percent 30 zfs_vdev_async_write_max_active 10 zfs_vdev_async_write_min_active 2 zfs_vdev_cache_bshift 16 zfs_vdev_cache_max 16384 zfs_vdev_cache_size 0 zfs_vdev_default_ms_count 200 zfs_vdev_default_ms_shift 29 zfs_vdev_initializing_max_active 1 zfs_vdev_initializing_min_active 1 zfs_vdev_max_active 1000 zfs_vdev_max_auto_ashift 16 zfs_vdev_min_auto_ashift 9 zfs_vdev_min_ms_count 16 zfs_vdev_mirror_non_rotating_inc 0 zfs_vdev_mirror_non_rotating_seek_inc 1 zfs_vdev_mirror_rotating_inc 0 zfs_vdev_mirror_rotating_seek_inc 5 zfs_vdev_mirror_rotating_seek_offset 1048576 zfs_vdev_ms_count_limit 131072 zfs_vdev_nia_credit 5 zfs_vdev_nia_delay 5 zfs_vdev_queue_depth_pct 1000 zfs_vdev_raidz_impl cycle [fastest] original scalar sse2 ssse3 avx2 avx512f avx512bw zfs_vdev_read_gap_limit 32768 zfs_vdev_rebuild_max_active 3 zfs_vdev_rebuild_min_active 1 zfs_vdev_removal_max_active 2 zfs_vdev_removal_min_active 1 zfs_vdev_scheduler unused zfs_vdev_scrub_max_active 3 zfs_vdev_scrub_min_active 1 zfs_vdev_sync_read_max_active 10 zfs_vdev_sync_read_min_active 10 zfs_vdev_sync_write_max_active 10 zfs_vdev_sync_write_min_active 10 zfs_vdev_trim_max_active 2 zfs_vdev_trim_min_active 1 zfs_vdev_write_gap_limit 4096 zfs_vnops_read_chunk_size 1048576 zfs_zevent_len_max 512 zfs_zevent_retain_expire_secs 900 zfs_zevent_retain_max 2000 zfs_zil_clean_taskq_maxalloc 1048576 zfs_zil_clean_taskq_minalloc 1024 zfs_zil_clean_taskq_nthr_pct 100 zil_maxblocksize 131072 zil_nocacheflush 0 zil_replay_disable 0 zil_slog_bulk 786432 zio_deadman_log_all 0 zio_dva_throttle_enabled 1 zio_requeue_io_start_cut_in_line 1 zio_slow_io_ms 30000 zio_taskq_batch_pct 80 zio_taskq_batch_tpq 0 zvol_inhibit_dev 0 zvol_major 230 zvol_max_discard_blocks 16384 zvol_prefetch_bytes 131072 zvol_request_sync 0 zvol_threads 32 zvol_volmode 1 VDEV cache disabled, skipping section ZIL committed transactions: 3.2G Commit requests: 404.3M Flushes to stable storage: 369.8M Transactions to SLOG storage pool: 167.6 TiB 1.6G Transactions to non-SLOG storage pool: 0 Bytes 0 ``` #### zed logs ``` -- Logs begin at Wed 2022-01-12 17:47:33 UTC, end at Wed 2022-02-09 17:58:06 UTC. -- Jan 19 02:15:35 r8-n9 systemd[1]: Started ZFS Event Daemon (zed). Jan 19 02:15:35 r8-n9 zed[246555]: ZFS Event Daemon 2.1.2-1 (PID 246555) Jan 19 02:15:35 r8-n9 zed[246555]: Processing events since eid=0 Jan 19 02:16:16 r8-n9 zed[249150]: eid=38 class=config_sync pool='pod-10' Jan 19 02:25:41 r8-n9 zed[256421]: eid=45 class=vdev_add pool='pod-10' Feb 09 14:46:20 r8-n9 zed[1838695]: eid=58 class=delay pool='pod-10' vdev=35000c500ae95cc5b size=4096 offset=4740717096960 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1036122 Feb 09 14:46:20 r8-n9 zed[1838693]: eid=57 class=delay pool='pod-10' vdev=35000c500ae970e3b size=20480 offset=4785921613824 priority=0 err=0 flags=0x40080c80 delay=30174ms Feb 09 14:46:20 r8-n9 zed[1838700]: eid=60 class=delay pool='pod-10' vdev=35000c500ae34d4c7 size=4096 offset=4731204771840 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1059557 Feb 09 14:46:20 r8-n9 zed[1838703]: eid=62 class=delay pool='pod-10' vdev=35000c500ae96a4a3 size=4096 offset=4785771388928 priority=0 err=0 flags=0x180980 delay=30049ms bookmark=269:0:0:1033656 Feb 09 14:46:21 r8-n9 zed[1838803]: eid=70 class=delay pool='pod-10' vdev=35000c500ae95d66f size=4096 offset=4774821298176 priority=0 err=0 flags=0x180980 delay=30547ms bookmark=269:0:0:1041464 Feb 09 14:46:21 r8-n9 zed[1838814]: eid=76 class=delay pool='pod-10' vdev=35000c500ae62c433 size=4096 offset=4721981607936 priority=0 err=0 flags=0x180980 delay=30159ms bookmark=269:0:0:1044753 Feb 09 14:46:22 r8-n9 zed[1838836]: eid=77 class=delay pool='pod-10' vdev=35000c500ae29d97f size=4096 offset=4781092179968 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1053826 Feb 09 14:46:22 r8-n9 zed[1838842]: eid=78 class=delay pool='pod-10' vdev=35000c500ae29a4bb size=4096 offset=4776777699328 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1031840 Feb 09 14:46:22 r8-n9 zed[1838846]: eid=82 class=delay pool='pod-10' vdev=35000c500ae96aa6f size=4096 offset=4786110545920 priority=0 err=0 flags=0x180980 delay=30374ms bookmark=269:0:0:1033105 Feb 09 14:46:22 r8-n9 zed[1838875]: eid=84 class=delay pool='pod-10' vdev=35000c500ae95555b size=53248 offset=4774958784512 priority=0 err=0 flags=0x40080c80 delay=30460ms Feb 09 14:46:22 r8-n9 zed[1838878]: eid=83 class=delay pool='pod-10' vdev=35000c500ae957c3b size=4096 offset=4743807332352 priority=0 err=0 flags=0x180980 delay=30668ms bookmark=269:0:0:1054047 Feb 09 14:46:22 r8-n9 zed[1838884]: eid=87 class=delay pool='pod-10' vdev=35000c500ae96d793 size=4096 offset=4768750641152 priority=0 err=0 flags=0x180980 delay=30444ms bookmark=269:0:0:1042517 Feb 09 14:46:22 r8-n9 zed[1838935]: eid=89 class=delay pool='pod-10' vdev=35000c500ae972b2f size=4096 offset=4769971810304 priority=0 err=0 flags=0x180980 delay=30623ms bookmark=269:0:0:1033391 Feb 09 14:46:22 r8-n9 zed[1838942]: eid=91 class=delay pool='pod-10' vdev=35000c500ae9737c3 size=4096 offset=4728066199552 priority=0 err=0 flags=0x180980 delay=30623ms bookmark=269:0:0:1057873 Feb 09 14:46:22 r8-n9 zed[1838963]: eid=94 class=delay pool='pod-10' vdev=35000c500ae95d53b size=4096 offset=4712411389952 priority=0 err=0 flags=0x180980 delay=30831ms bookmark=269:0:0:1049636 Feb 09 14:46:22 r8-n9 zed[1838967]: eid=95 class=delay pool='pod-10' vdev=35000c500ae970cd7 size=4096 offset=4776225902592 priority=0 err=0 flags=0x180980 delay=30626ms bookmark=269:0:0:1046985 Feb 09 14:46:22 r8-n9 zed[1838972]: eid=97 class=delay pool='pod-10' vdev=35000c500ae2a6e9b size=4096 offset=4782662496256 priority=0 err=0 flags=0x180980 delay=30831ms bookmark=269:0:0:1048773 Feb 09 14:46:22 r8-n9 zed[1838969]: eid=96 class=delay pool='pod-10' vdev=35000c500ae955eef size=20480 offset=4774534471680 priority=0 err=0 flags=0x40080c80 delay=30626ms ``` #### vmstat ``` # cat /proc/vmstat nr_free_pages 497162 nr_zone_inactive_anon 77677 nr_zone_active_anon 81958 nr_zone_inactive_file 16392 nr_zone_active_file 10007 nr_zone_unevictable 7960 nr_zone_write_pending 844 nr_mlock 7960 nr_page_table_pages 1963 nr_kernel_stack 19296 nr_bounce 0 nr_zspages 0 nr_free_cma 0 numa_hit 151121274140 numa_miss 652564522 numa_foreign 652564522 numa_interleave 96982 numa_local 133392770941 numa_other 18381067721 nr_inactive_anon 77677 nr_active_anon 81958 nr_inactive_file 16392 nr_active_file 10007 nr_unevictable 7960 nr_slab_reclaimable 1104575 nr_slab_unreclaimable 28741700 nr_isolated_anon 0 nr_isolated_file 0 workingset_nodes 1973 workingset_refault 78664 workingset_activate 26788 workingset_restore 14173 workingset_nodereclaim 816 nr_anon_pages 162434 nr_mapped 20118 nr_file_pages 32246 nr_dirty 844 nr_writeback 0 nr_writeback_temp 0 nr_shmem 369 nr_shmem_hugepages 0 nr_shmem_pmdmapped 0 nr_file_hugepages 0 nr_file_pmdmapped 0 nr_anon_transparent_hugepages 0 nr_unstable 0 nr_vmscan_write 68263 nr_vmscan_immediate_reclaim 641 nr_dirtied 3379059 nr_written 3113482 nr_kernel_misc_reclaimable 0 nr_dirty_threshold 73743 nr_dirty_background_threshold 36826 pgpgin 194613865828 pgpgout 594833359918 pswpin 4218 pswpout 68258 pgalloc_dma 0 pgalloc_dma32 324399540 pgalloc_normal 245349772461 pgalloc_movable 0 allocstall_dma 0 allocstall_dma32 0 allocstall_normal 1 allocstall_movable 2 pgskip_dma 0 pgskip_dma32 0 pgskip_normal 0 pgskip_movable 0 pgfree 245674776191 pgactivate 730915 pgdeactivate 358591 pglazyfree 6389 pgfault 243384077 pgmajfault 15967 pglazyfreed 0 pgrefill 382033 pgsteal_kswapd 292238 pgsteal_direct 1375 pgscan_kswapd 477828 pgscan_direct 1399 pgscan_direct_throttle 0 zone_reclaim_failed 0 pginodesteal 74 slabs_scanned 459729476 kswapd_inodesteal 26293 kswapd_low_wmark_hit_quickly 1 kswapd_high_wmark_hit_quickly 14 pageoutrun 277 pgrotated 69682 drop_pagecache 1 drop_slab 1 oom_kill 0 numa_pte_updates 4332995 numa_huge_pte_updates 12 numa_hint_faults 3863526 numa_hint_faults_local 3281256 numa_pages_migrated 327088 pgmigrate_success 411548 pgmigrate_fail 3419 compact_migrate_scanned 10896483 compact_free_scanned 460891 compact_isolated 170144 compact_stall 4 compact_fail 4 compact_success 0 compact_daemon_wake 255 compact_daemon_migrate_scanned 1507268 compact_daemon_free_scanned 317747 htlb_buddy_alloc_success 0 htlb_buddy_alloc_fail 0 unevictable_pgs_culled 81639 unevictable_pgs_scanned 0 unevictable_pgs_rescued 14620 unevictable_pgs_mlocked 26080 unevictable_pgs_munlocked 16730 unevictable_pgs_cleared 1390 unevictable_pgs_stranded 1390 thp_fault_alloc 10 thp_fault_fallback 0 thp_collapse_alloc 8 thp_collapse_alloc_failed 2 thp_file_alloc 0 thp_file_mapped 0 thp_split_page 0 thp_split_page_failed 0 thp_deferred_split_page 18 thp_split_pmd 9 thp_split_pud 0 thp_zero_page_alloc 0 thp_zero_page_alloc_failed 0 thp_swpout 0 thp_swpout_fallback 0 balloon_inflate 0 balloon_deflate 0 balloon_migrate 0 swap_ra 1519 swap_ra_hit 833 ``` #### buddyinfo ``` Node: 0 Zone: DMA Free KiB in zone: 15876.00 Fragment size Free fragments Total available KiB 4096 1 4.0 8192 0 0.0 16384 0 0.0 32768 0 0.0 65536 2 128.0 131072 1 128.0 262144 1 256.0 524288 0 0.0 1048576 1 1024.0 2097152 1 2048.0 4194304 3 12288.0 Zone: DMA32 Free KiB in zone: 252500.00 Fragment size Free fragments Total available KiB 4096 1923 7692.0 8192 5143 41144.0 16384 2581 41296.0 32768 142 4544.0 65536 26 1664.0 131072 16 2048.0 262144 78 19968.0 524288 70 35840.0 1048576 36 36864.0 2097152 22 45056.0 4194304 4 16384.0 Zone: Normal Free KiB in zone: 107160.00 Fragment size Free fragments Total available KiB 4096 1586 6344.0 8192 8328 66624.0 16384 1987 31792.0 32768 27 864.0 65536 24 1536.0 131072 0 0.0 262144 0 0.0 524288 0 0.0 1048576 0 0.0 2097152 0 0.0 4194304 0 0.0 Node: 1 Zone: Normal Free KiB in zone: 2152516.00 Fragment size Free fragments Total available KiB 4096 149157 596628.0 8192 107244 857952.0 16384 3545 56720.0 32768 4110 131520.0 65536 3156 201984.0 131072 1610 206080.0 262144 297 76032.0 524288 48 24576.0 1048576 1 1024.0 2097152 0 0.0 4194304 0 0.0 ``` ### Describe how to reproduce the problem Uncertain but it has occurred on 2 separate servers so it is likely to happen again. ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
defect
zfs arc stuck pruning evicting flatlines throughput thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name ubuntu distribution version lts kernel version generic architecture openzfs version zfs command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing zfs is flatlined on throughput with an arc evict and arc prune process spinning at the workload is kernel nfs server all clients with zfs built from source characterized by high cpu iowait and throughput flatlining ram graph img width alt screen shot at am src it s holding on to ram pretty hard this is where i begin to lose debugging expertise having discovered slabs yesterday top top up days user load average tasks total running sleeping stopped zombie cpu s us sy ni id wa hi si st mib mem total free used buff cache mib swap total free used avail mem pid user pr ni virt res shr s cpu mem time command root r arc evict root s arc prune root s zed root r top root rt s multipathd root d nfsd slabtop slabtop active total objects used active total slabs used active total caches used active total size used minimum average maximum object objs active use obj size slabs obj slab cache size name dnode t dmu buf impl t kmalloc kmalloc arc buf hdr t kmalloc kmalloc dentry zfs znode cache sa cache stateids filp cred jar lsm file cache files fsnotify mark connector acpi state khugepaged mm slot kmalloc rcl arc buf hdr t full nsproxy abd t radix tree node openowners kmalloc arc buf t zio buf comb kmalloc buffer head kernfs node cache anon vma chain inode cache kmalloc acpi operand vm area struct kmalloc anon vma kmalloc rcl zpool zpool status pod pool pod state online config name state read write cksum pod online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online online logs mirror online wwn online wwn online cache wwn online wwn online wwn online spares avail avail avail avail avail avail avail errors no known data errors zfs fs zfs list name used avail refer mountpoint pod pod pod pod srv pod zfs get primarycache secondarycache pod pod name property value source pod pod primarycache metadata local pod pod secondarycache metadata local intention is for arc to be entirely metadata initial arcstats arcstats reported metadata usage above the limit arc size current gib target size adaptive gib min size hard limit gib max size high water gib most frequently used mfu cache size gib most recently used mru cache size gib metadata cache size hard limit gib metadata cache size current gib dnode cache size hard limit gib dnode cache size current gib initial adjustment attempted to remediate by increasing arc size by half of remaining ram and increasing the metadata allocation in the arc echo sys module zfs parameters zfs arc max echo sys module zfs parameters zfs arc meta limit percent echo sys module zfs parameters zfs arc dnode limit percent current issue this remediated the issue temporarily and the prune processes stopped but we are back stopped nfs server and it seems to be free ing memory albeit very slowly many dp sync taskq processes but removing the nfs server is very un ideal given they are both living in the kernel it s difficult for me personally to determine who is eating up the memory top up days users load average tasks total running sleeping stopped zombie cpu s us sy ni id wa hi si st mib mem total free used buff cache mib swap total free used avail mem pid user pr ni virt res shr s cpu mem time command root s txg sync root s dp sync taskq root d kworker events root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq root s dp sync taskq arc summary zfs subsystem report wed feb linux generic machine arc status healthy memory throttle count arc size current gib target size adaptive gib min size hard limit gib max size high water gib most frequently used mfu cache size gib most recently used mru cache size gib metadata cache size hard limit gib metadata cache size current gib dnode cache size hard limit gib dnode cache size current gib arc hash breakdown elements max elements current collisions chain max chains arc misc deleted mutex misses eviction skips eviction skips due to writes cached evictions tib eligible evictions gib eligible mfu evictions gib eligible mru evictions gib ineligible evictions gib arc total accesses hits misses cache hit ratio cache miss ratio actual hit ratio mfu mru hits data demand efficiency data prefetch efficiency n a cache hits by cache type most frequently used mfu most recently used mru most frequently used mfu ghost most recently used mru ghost cache hits by data type demand data demand prefetch data demand metadata demand prefetch metadata cache misses by data type demand data demand prefetch data demand metadata demand prefetch metadata dmu prefetch efficiency hit ratio miss ratio status healthy low memory aborts free on write r w clashes bad checksums i o errors size adaptive tib compressed gib header size gib mfu allocated size gib mru allocated size gib prefetch allocated size mib data buffer content allocated size bytes metadata buffer content allocated size gib breakdown hit ratio miss ratio feeds writes writes sent evicts lock retries upon reading solaris porting layer spl spl hostid spl hostid path etc hostid spl kmem alloc max spl kmem alloc warn spl kmem cache kmem threads spl kmem cache magazine size spl kmem cache max size spl kmem cache obj per slab spl kmem cache reclaim spl kmem cache slab limit spl max show tasks spl panic halt spl schedule hrtimeout slack us spl taskq kick spl taskq thread bind spl taskq thread dynamic spl taskq thread priority spl taskq thread sequential tunables dbuf cache hiwater pct dbuf cache lowater pct dbuf cache max bytes dbuf cache shift dbuf metadata cache max bytes dbuf metadata cache shift dmu object alloc chunk shift dmu prefetch max ignore hole birth feed again feed min ms feed secs headroom headroom boost meta percent mfuonly noprefetch norw rebuild blocks min rebuild enabled trim ahead write boost write max metaslab aliquot metaslab bias enabled metaslab debug load metaslab debug unload metaslab df max search metaslab df use largest segment metaslab force ganging metaslab fragmentation factor enabled metaslab lba weighting enabled metaslab preload enabled metaslab unload delay metaslab unload delay ms send holes without birth time spa asize inflation spa config path etc zfs zpool cache spa load print vdev tree spa load verify data spa load verify metadata spa load verify shift spa slop shift vdev file logical ashift vdev file physical ashift vdev removal max span vdev validate skip zap iterate prefetch zfetch array rd sz zfetch max distance zfetch max idistance zfetch max streams zfetch min sec reap zfs abd scatter enabled zfs abd scatter max order zfs abd scatter min size zfs admin snapshot zfs allow redacted dataset mount zfs arc average blocksize zfs arc dnode limit zfs arc dnode limit percent zfs arc dnode reduce percent zfs arc evict batch limit zfs arc eviction pct zfs arc grow retry zfs arc lotsfree percent zfs arc max zfs arc meta adjust restarts zfs arc meta limit zfs arc meta limit percent zfs arc meta min zfs arc meta prune zfs arc meta strategy zfs arc min zfs arc min prefetch ms zfs arc min prescient prefetch ms zfs arc p dampener disable zfs arc p min shift zfs arc pc percent zfs arc shrink shift zfs arc shrinker limit zfs arc sys free zfs async block max blocks zfs autoimport disable zfs checksum events per second zfs commit timeout pct zfs compressed arc enabled zfs condense indirect commit entry delay ms zfs condense indirect obsolete pct zfs condense indirect vdevs enable zfs condense max obsolete bytes zfs condense min mapping bytes zfs dbgmsg enable zfs dbgmsg maxsize zfs dbuf state index zfs ddt data is special zfs deadman checktime ms zfs deadman enabled zfs deadman failmode wait zfs deadman synctime ms zfs deadman ziotime ms zfs dedup prefetch zfs delay min dirty percent zfs delay scale zfs delete blocks zfs dirty data max zfs dirty data max max zfs dirty data max max percent zfs dirty data max percent zfs dirty data sync percent zfs disable ivset guid check zfs dmu offset next sync zfs embedded slog min ms zfs expire snapshot zfs fallocate reserve percent zfs flags zfs free bpobj enabled zfs free leak on eio zfs free min time ms zfs history output max zfs immediate write sz zfs initialize chunk size zfs initialize value zfs keep log spacemaps at export zfs key max salt uses zfs livelist condense new alloc zfs livelist condense sync cancel zfs livelist condense sync pause zfs livelist condense zthr cancel zfs livelist condense zthr pause zfs livelist max entries zfs livelist min percent shared zfs lua max instrlimit zfs lua max memlimit zfs max async dedup frees zfs max log walking zfs max logsm summary length zfs max missing tvds zfs max nvlist src size zfs max recordsize zfs metaslab find max tries zfs metaslab fragmentation threshold zfs metaslab max size cache sec zfs metaslab mem limit zfs metaslab segment weight enabled zfs metaslab switch threshold zfs metaslab try hard before gang zfs mg fragmentation threshold zfs mg noalloc threshold zfs min metaslabs to flush zfs multihost fail intervals zfs multihost history zfs multihost import intervals zfs multihost interval zfs multilist num sublists zfs no scrub io zfs no scrub prefetch zfs nocacheflush zfs nopwrite enabled zfs object mutex size zfs obsolete min time ms zfs override estimate recordsize zfs pd bytes max zfs per txg dirty frees percent zfs prefetch disable zfs read history zfs read history hits zfs rebuild max segment zfs rebuild scrub enabled zfs rebuild vdev limit zfs reconstruct indirect combinations max zfs recover zfs recv queue ff zfs recv queue length zfs recv write batch size zfs removal ignore errors zfs removal suspend progress zfs remove max segment zfs resilver disable defer zfs resilver min time ms zfs scan checkpoint intval zfs scan fill weight zfs scan ignore errors zfs scan issue strategy zfs scan legacy zfs scan max ext gap zfs scan mem lim fact zfs scan mem lim soft fact zfs scan strict mem lim zfs scan suspend progress zfs scan vdev limit zfs scrub min time ms zfs send corrupt data zfs send no prefetch queue ff zfs send no prefetch queue length zfs send queue ff zfs send queue length zfs send unmodified spill blocks zfs slow io events per second zfs spa discard memory limit zfs special class metadata reserve pct zfs sync pass deferred free zfs sync pass dont compress zfs sync pass rewrite zfs sync taskq batch pct zfs traverse indirect prefetch limit zfs trim extent bytes max zfs trim extent bytes min zfs trim metaslab skip zfs trim queue limit zfs trim txg batch zfs txg history zfs txg timeout zfs unflushed log block max zfs unflushed log block min zfs unflushed log block pct zfs unflushed max mem amt zfs unflushed max mem ppm zfs unlink suspend progress zfs user indirect is special zfs vdev aggregate trim zfs vdev aggregation limit zfs vdev aggregation limit non rotating zfs vdev async read max active zfs vdev async read min active zfs vdev async write active max dirty percent zfs vdev async write active min dirty percent zfs vdev async write max active zfs vdev async write min active zfs vdev cache bshift zfs vdev cache max zfs vdev cache size zfs vdev default ms count zfs vdev default ms shift zfs vdev initializing max active zfs vdev initializing min active zfs vdev max active zfs vdev max auto ashift zfs vdev min auto ashift zfs vdev min ms count zfs vdev mirror non rotating inc zfs vdev mirror non rotating seek inc zfs vdev mirror rotating inc zfs vdev mirror rotating seek inc zfs vdev mirror rotating seek offset zfs vdev ms count limit zfs vdev nia credit zfs vdev nia delay zfs vdev queue depth pct zfs vdev raidz impl cycle original scalar zfs vdev read gap limit zfs vdev rebuild max active zfs vdev rebuild min active zfs vdev removal max active zfs vdev removal min active zfs vdev scheduler unused zfs vdev scrub max active zfs vdev scrub min active zfs vdev sync read max active zfs vdev sync read min active zfs vdev sync write max active zfs vdev sync write min active zfs vdev trim max active zfs vdev trim min active zfs vdev write gap limit zfs vnops read chunk size zfs zevent len max zfs zevent retain expire secs zfs zevent retain max zfs zil clean taskq maxalloc zfs zil clean taskq minalloc zfs zil clean taskq nthr pct zil maxblocksize zil nocacheflush zil replay disable zil slog bulk zio deadman log all zio dva throttle enabled zio requeue io start cut in line zio slow io ms zio taskq batch pct zio taskq batch tpq zvol inhibit dev zvol major zvol max discard blocks zvol prefetch bytes zvol request sync zvol threads zvol volmode vdev cache disabled skipping section zil committed transactions commit requests flushes to stable storage transactions to slog storage pool tib transactions to non slog storage pool bytes zed logs logs begin at wed utc end at wed utc jan systemd started zfs event daemon zed jan zed zfs event daemon pid jan zed processing events since eid jan zed eid class config sync pool pod jan zed eid class vdev add pool pod feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay bookmark feb zed eid class delay pool pod vdev size offset priority err flags delay vmstat cat proc vmstat nr free pages nr zone inactive anon nr zone active anon nr zone inactive file nr zone active file nr zone unevictable nr zone write pending nr mlock nr page table pages nr kernel stack nr bounce nr zspages nr free cma numa hit numa miss numa foreign numa interleave numa local numa other nr inactive anon nr active anon nr inactive file nr active file nr unevictable nr slab reclaimable nr slab unreclaimable nr isolated anon nr isolated file workingset nodes workingset refault workingset activate workingset restore workingset nodereclaim nr anon pages nr mapped nr file pages nr dirty nr writeback nr writeback temp nr shmem nr shmem hugepages nr shmem pmdmapped nr file hugepages nr file pmdmapped nr anon transparent hugepages nr unstable nr vmscan write nr vmscan immediate reclaim nr dirtied nr written nr kernel misc reclaimable nr dirty threshold nr dirty background threshold pgpgin pgpgout pswpin pswpout pgalloc dma pgalloc pgalloc normal pgalloc movable allocstall dma allocstall allocstall normal allocstall movable pgskip dma pgskip pgskip normal pgskip movable pgfree pgactivate pgdeactivate pglazyfree pgfault pgmajfault pglazyfreed pgrefill pgsteal kswapd pgsteal direct pgscan kswapd pgscan direct pgscan direct throttle zone reclaim failed pginodesteal slabs scanned kswapd inodesteal kswapd low wmark hit quickly kswapd high wmark hit quickly pageoutrun pgrotated drop pagecache drop slab oom kill numa pte updates numa huge pte updates numa hint faults numa hint faults local numa pages migrated pgmigrate success pgmigrate fail compact migrate scanned compact free scanned compact isolated compact stall compact fail compact success compact daemon wake compact daemon migrate scanned compact daemon free scanned htlb buddy alloc success htlb buddy alloc fail unevictable pgs culled unevictable pgs scanned unevictable pgs rescued unevictable pgs mlocked unevictable pgs munlocked unevictable pgs cleared unevictable pgs stranded thp fault alloc thp fault fallback thp collapse alloc thp collapse alloc failed thp file alloc thp file mapped thp split page thp split page failed thp deferred split page thp split pmd thp split pud thp zero page alloc thp zero page alloc failed thp swpout thp swpout fallback balloon inflate balloon deflate balloon migrate swap ra swap ra hit buddyinfo node zone dma free kib in zone fragment size free fragments total available kib zone free kib in zone fragment size free fragments total available kib zone normal free kib in zone fragment size free fragments total available kib node zone normal free kib in zone fragment size free fragments total available kib describe how to reproduce the problem uncertain but it has occurred on separate servers so it is likely to happen again include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with
1
44,633
12,301,484,555
IssuesEvent
2020-05-11 15:28:43
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Bad formatting when selecting NULL
C: Functionality E: All Editions P: Low R: Fixed T: Defect
Generating the formatted SQL for: ```java var n = inline(null, INTEGER); ctx.select(n, n); ``` Produces ill formatted SQL: ```sql select null, null ``` The first column is indented too much.
1.0
Bad formatting when selecting NULL - Generating the formatted SQL for: ```java var n = inline(null, INTEGER); ctx.select(n, n); ``` Produces ill formatted SQL: ```sql select null, null ``` The first column is indented too much.
defect
bad formatting when selecting null generating the formatted sql for java var n inline null integer ctx select n n produces ill formatted sql sql select null null the first column is indented too much
1
52,243
13,211,411,442
IssuesEvent
2020-08-15 22:57:21
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
[vemcal] missing mysql breaks the build (Trac #1909)
Incomplete Migration Migrated from Trac combo reconstruction defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1909">https://code.icecube.wisc.edu/projects/icecube/ticket/1909</a>, reported by david.schultzand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-09-18T05:49:06", "_ts": "1568785746891317", "description": "On a standard py2-v2 cvmfs environment:\n\n{{{\n-- + vemcal\n-- +-- python [symlinks] \nCMake Error at cmake/tools.cmake:71 (message):\n Attempt to use tool 'mysql' which wasn't found\nCall Stack (most recent call first):\n cmake/tools.cmake:112 (use_tool)\n cmake/project.cmake:252 (use_tools)\n vemcal/CMakeLists.txt:39 (i3_add_library)\n}}}\n\nThis should probably be optional.", "reporter": "david.schultz", "cc": "", "resolution": "worksforme", "time": "2016-11-15T16:55:59", "component": "combo reconstruction", "summary": "[vemcal] missing mysql breaks the build", "priority": "major", "keywords": "", "milestone": "Long-Term Future", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
[vemcal] missing mysql breaks the build (Trac #1909) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1909">https://code.icecube.wisc.edu/projects/icecube/ticket/1909</a>, reported by david.schultzand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-09-18T05:49:06", "_ts": "1568785746891317", "description": "On a standard py2-v2 cvmfs environment:\n\n{{{\n-- + vemcal\n-- +-- python [symlinks] \nCMake Error at cmake/tools.cmake:71 (message):\n Attempt to use tool 'mysql' which wasn't found\nCall Stack (most recent call first):\n cmake/tools.cmake:112 (use_tool)\n cmake/project.cmake:252 (use_tools)\n vemcal/CMakeLists.txt:39 (i3_add_library)\n}}}\n\nThis should probably be optional.", "reporter": "david.schultz", "cc": "", "resolution": "worksforme", "time": "2016-11-15T16:55:59", "component": "combo reconstruction", "summary": "[vemcal] missing mysql breaks the build", "priority": "major", "keywords": "", "milestone": "Long-Term Future", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
missing mysql breaks the build trac migrated from json status closed changetime ts description on a standard cvmfs environment n n n vemcal n python ncmake error at cmake tools cmake message n attempt to use tool mysql which wasn t found ncall stack most recent call first n cmake tools cmake use tool n cmake project cmake use tools n vemcal cmakelists txt add library n n nthis should probably be optional reporter david schultz cc resolution worksforme time component combo reconstruction summary missing mysql breaks the build priority major keywords milestone long term future owner nega type defect
1
291,796
21,940,546,270
IssuesEvent
2022-05-23 17:36:59
pharmaverse/admiral
https://api.github.com/repos/pharmaverse/admiral
closed
Documentation: Get Started has incorrect trt variables in mutate
documentation good first issue help wanted release 20-May-2022 contribution
### Please select a category the issue is focused on? User Guides ### Let us know where something needs a refresh or put your idea here! ![image](https://user-images.githubusercontent.com/10111024/157339179-c4b6565b-5da4-46d4-8de3-be9ac4cf9d74.png) Treatment variables are not correct. Should be TRT01P = ARM and TRT01A = ACTARM
1.0
Documentation: Get Started has incorrect trt variables in mutate - ### Please select a category the issue is focused on? User Guides ### Let us know where something needs a refresh or put your idea here! ![image](https://user-images.githubusercontent.com/10111024/157339179-c4b6565b-5da4-46d4-8de3-be9ac4cf9d74.png) Treatment variables are not correct. Should be TRT01P = ARM and TRT01A = ACTARM
non_defect
documentation get started has incorrect trt variables in mutate please select a category the issue is focused on user guides let us know where something needs a refresh or put your idea here treatment variables are not correct should be arm and actarm
0
25,976
4,539,518,146
IssuesEvent
2016-09-09 11:21:03
netty/netty
https://api.github.com/repos/netty/netty
closed
HTTP/2: SETTINGS_HEADER_TABLE_SIZE should be an unsigned int
defect
The HTTP/2 spec demands that the max value for `SETTINGS_HEADER_TABLE_SIZE ` should be an unsigned 32-bit integer. However, it seems that some [limitations in HPACK](https://github.com/netty/netty/blob/4.1/codec-http2/src/main/java/io/netty/handler/codec/http2/Http2CodecUtil.java#L90) prevent us from supporting that. h2spec test suite 6.5 fails due to this limitation. The exception is `Setting HEADER_TABLE_SIZE is invalid: 4294967295`. ``` × Sends a SETTINGS frame - The endpoint MUST sends a SETTINGS frame with ACK. Expected: SETTINGS frame (Flags: 1) Actual: GOAWAY frame (Length: 56, Flags: 0, ErrorCode: PROTOCOL_ERROR) ```
1.0
HTTP/2: SETTINGS_HEADER_TABLE_SIZE should be an unsigned int - The HTTP/2 spec demands that the max value for `SETTINGS_HEADER_TABLE_SIZE ` should be an unsigned 32-bit integer. However, it seems that some [limitations in HPACK](https://github.com/netty/netty/blob/4.1/codec-http2/src/main/java/io/netty/handler/codec/http2/Http2CodecUtil.java#L90) prevent us from supporting that. h2spec test suite 6.5 fails due to this limitation. The exception is `Setting HEADER_TABLE_SIZE is invalid: 4294967295`. ``` × Sends a SETTINGS frame - The endpoint MUST sends a SETTINGS frame with ACK. Expected: SETTINGS frame (Flags: 1) Actual: GOAWAY frame (Length: 56, Flags: 0, ErrorCode: PROTOCOL_ERROR) ```
defect
http settings header table size should be an unsigned int the http spec demands that the max value for settings header table size should be an unsigned bit integer however it seems that some prevent us from supporting that test suite fails due to this limitation the exception is setting header table size is invalid × sends a settings frame the endpoint must sends a settings frame with ack expected settings frame flags actual goaway frame length flags errorcode protocol error
1
13,422
5,358,982,495
IssuesEvent
2017-02-21 00:45:19
elegantthemes/Divi-Beta
https://api.github.com/repos/elegantthemes/Divi-Beta
closed
Builder Sync :: Number Counter :: updating number and percent sign on BB causes console error when switching to VB
BUILDER SYNC DESIGN SIGNOFF QUALITY ASSURED READY FOR REVIEW
### Problem: Updating number and percent sign on BB causes console error when switching to VB: ![number-counter-issue](https://cloud.githubusercontent.com/assets/916442/23092708/b3565960-f603-11e6-95c4-7ffd29c1181b.gif) ### Steps To Reproduce: 1. Open a page on BB and VB . Open their console 2. Update number and percent sign on BB 3. Switch to VB 4. You'll see an error thrown ## Attached PR - https://github.com/elegantthemes/submodule-builder/pull/1744
1.0
Builder Sync :: Number Counter :: updating number and percent sign on BB causes console error when switching to VB - ### Problem: Updating number and percent sign on BB causes console error when switching to VB: ![number-counter-issue](https://cloud.githubusercontent.com/assets/916442/23092708/b3565960-f603-11e6-95c4-7ffd29c1181b.gif) ### Steps To Reproduce: 1. Open a page on BB and VB . Open their console 2. Update number and percent sign on BB 3. Switch to VB 4. You'll see an error thrown ## Attached PR - https://github.com/elegantthemes/submodule-builder/pull/1744
non_defect
builder sync number counter updating number and percent sign on bb causes console error when switching to vb problem updating number and percent sign on bb causes console error when switching to vb steps to reproduce open a page on bb and vb open their console update number and percent sign on bb switch to vb you ll see an error thrown attached pr
0
46,681
13,055,958,985
IssuesEvent
2020-07-30 03:14:17
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
IceHive - crashes during icetray-inspect (Trac #1708)
Incomplete Migration Migrated from Trac combo simulation defect
Migrated from https://code.icecube.wisc.edu/ticket/1708 ```json { "status": "closed", "changetime": "2019-02-13T14:12:58", "description": "this is preventing the docs from being built\n\nmaybe `Limits::RingLimits::RingLimits()` is getting bum vector?\n\nroot-ified stack trace:\n\n{{{\n===========================================================\nThere was a crash.\nThis is the entire stack trace of all threads:\n===========================================================\n\nThread 5 (Thread 0x7ff746184700 (LWP 19476)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 4 (Thread 0x7ff745783700 (LWP 19477)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 3 (Thread 0x7ff744d82700 (LWP 19478)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 2 (Thread 0x7ff744381700 (LWP 19479)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 1 (Thread 0x7ff81190b700 (LWP 19361)):\n#0 0x00007ff81078d69d in waitpid () from /lib64/libc.so.6\n#1 0x00007ff81071f609 in do_system () from /lib64/libc.so.6\n#2 0x00007ff81071f940 in system () from /lib64/libc.so.6\n#3 0x00007ff7f2dfbad8 in TUnixSystem::StackTrace() () from /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/root-v5.34.18/lib/libCore.so\n#4 0x00007ff7f2dfa5d3 in TUnixSystem::DispatchSignals(ESignals) () from /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/root-v5.34.18/lib/libCore.so\n#5 <signal handler called>\n#6 0x00007ff742492f52 in std::_Construct<Limits::LimitPair, Limits::LimitPair> (__p=0x7ff73d7aa010, __value=...) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_construct.h:80\n#7 0x00007ff74249121c in std::__uninitialized_copy<false>::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#8 0x00007ff74248f3d5 in std::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#9 0x00007ff74248d1c8 in std::__uninitialized_copy_a<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*, Limits::LimitPair> (__first=Cannot access memory at address 0x1\n#10 0x00007ff74248c0bb in std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> >::vector (this=0x7fff1a3c9f00, __x=std::vector of length 4693611, capacity 0 = {...}) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_vector.h:243\n#11 0x00007ff742499af3 in Limits::RingLimits::RingLimits (this=0x7fff1a3c9f00) at ../../source/IceHive/public/IceHive/IceHiveHelpers.h:38\n#12 0x00007ff742092bd1 in I3HiveCluster<I3RecoPulse>::I3HiveCluster (this=0x483e450, context=...) at ../../source/IceHive/private/IceHive/I3HiveCluster.h:130\n#13 0x00007ff74209135c in StandardCreate<I3Module, I3HiveCluster<I3RecoPulse> >::Create (c=...) at ../../source/icetray/public/icetray/I3Factory.h:110\n#14 0x00007ff805c9d73b in boost::detail::function::function_invoker1<boost::shared_ptr<I3Module> (*)(I3Context const&), boost::shared_ptr<I3Module>, I3Context const&>::invoke (function_ptr=..., a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:94\n#15 0x00007ff805c0bc74 in boost::function1<boost::shared_ptr<I3Module>, I3Context const&>::operator() (this=0x7fff1a3ca2d0, a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:989\n#16 0x00007ff805dcd4ee in module_default_config (name=\"I3HiveCluster<I3RecoPulse>\") at ../../source/icetray/private/pybindings/I3ModuleFactory.cxx:36\n#17 0x00007ff805dd0ce9 in boost::python::detail::invoke<boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::result_converter, I3ConfigurationPtr (*)(const std::string&), boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::c_t0>(boost::python::detail::invoke_tag_<false, false>, const impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::result_converter &, I3ConfigurationPtr (*&)(const std::string &), impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::c_t0 &) (rc=..., f=\n0x1d22fb8, ac0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/invoke.hpp:75\n#18 0x00007ff805dd072a in impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::operator() (this=0x1d22fb8, args_=0x4037290) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/caller.hpp:223\n#19 0x00007ff805dd0051 in boost::python::objects::caller_py_function_impl<boost::python::detail::caller<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> > >::operator() (this=0x1d22fb0, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/object/py_function.hpp:38\n#20 0x00007ff804eb43c5 in boost::python::objects::py_function::operator() (this=0x1d23440, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/object/py_function.hpp:143\n#21 0x00007ff804eb1f7b in boost::python::objects::function::call (this=0x1d23430, args=0x4037290, keywords=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:226\n#22 0x00007ff804eb372f in boost::python::objects::(anonymous namespace)::bind_return::operator() (this=0x7fff1a3ca900) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:581\n#23 0x00007ff804eb40f7 in boost::detail::function::void_function_ref_invoker0<boost::python::objects::<unnamed>::bind_return, void>::invoke(boost::detail::function::function_buffer &) (function_obj_ptr=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:187\n#24 0x00007ff804eca39a in boost::function0<void>::operator() (this=0x7fff1a3ca8b0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#25 0x00007ff804ec9dfc in boost::python::detail::exception_handler::operator() (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:74\n#26 0x00007ff7fdc810d0 in boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>::operator() (this=0x2040c98, handler=..., f=..., translate=0x7ff7fdc59bec <translate(not_found_exception const&)>) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/translate_exception.hpp:46\n#27 0x00007ff7fdc7e3dd in boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> >::operator()<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list2<boost::python::detail::exception_handler const&, boost::function0<void> const&> > (this=0x2040ca0, f=..., a=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind.hpp:361\n#28 0x00007ff7fdc79dde in boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >::operator()<boost::python::detail::exception_handler, boost::function0<void> > (this=0x2040c98, a1=..., a2=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind/bind_template.hpp:102\n#29 0x00007ff7fdc7504c in boost::detail::function::function_obj_invoker2<boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >, bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::invoke (function_obj_ptr=..., a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:131\n#30 0x00007ff804eca315 in boost::function2<bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::operator() (this=0x2040c90, a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#31 0x00007ff804eca1a3 in boost::python::detail::exception_handler::handle (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/detail/exception_handler.hpp:41\n#32 0x00007ff804ec9b13 in boost::python::handle_exception_impl (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:24\n#33 0x00007ff804eb3dab in boost::python::handle_exception<boost::python::objects::<unnamed>::bind_return>(boost::python::objects::(anonymous namespace)::bind_return) (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/errors.hpp:29\n#34 0x00007ff804eb380f in boost::python::objects::function_call (func=0x1d23430, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:613\n#35 0x00007ff811363dd3 in PyObject_Call (func=0x1d23430, arg=<value optimized out>, kw=<value optimized out>) at Objects/abstract.c:2529\n#36 0x00007ff81141126d in do_call (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4239\n#37 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4044\n#38 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#39 0x00007ff81141281e in fast_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4107\n#40 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4042\n#41 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#42 0x00007ff81141391e in PyEval_EvalCodeEx (co=0x7ff8117a3b30, globals=<value optimized out>, locals=<value optimized out>, args=<value optimized out>, argcount=0, kws=0x0, kwcount=0, defs=0x0, defcount=0, closure=0x0) at Python/ceval.c:3253\n#43 0x00007ff811413a32 in PyEval_EvalCode (co=<value optimized out>, globals=<value optimized out>, locals=<value optimized out>) at Python/ceval.c:667\n#44 0x00007ff8114334d0 in run_mod (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1353\n#45 PyRun_FileExFlags (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1339\n#46 0x00007ff81143369c in PyRun_SimpleFileExFlags (fp=0x1a14a80, filename=0x7fff1a3cc2de \"/build/buildslave/kipple_cvmfs/docs/build/bin/icetray-inspect\", closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:943\n#47 0x00007ff8114490ec in Py_Main (argc=<value optimized out>, argv=<value optimized out>) at Modules/main.c:639\n#48 0x00007ff8106ffd5d in __libc_start_main () from /lib64/libc.so.6\n#49 0x0000000000400649 in _start ()\n===========================================================\n\n\nThe lines below might hint at the cause of the crash.\nIf they do not help you then please submit a bug report at\nhttp://root.cern.ch/bugs. Please post the ENTIRE stack trace\nfrom above as an attachment in addition to anything else\nthat might help us fixing this issue.\n===========================================================\n#6 0x00007ff742492f52 in std::_Construct<Limits::LimitPair, Limits::LimitPair> (__p=0x7ff73d7aa010, __value=...) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_construct.h:80\n#7 0x00007ff74249121c in std::__uninitialized_copy<false>::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#8 0x00007ff74248f3d5 in std::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#9 0x00007ff74248d1c8 in std::__uninitialized_copy_a<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*, Limits::LimitPair> (__first=Cannot access memory at address 0x1\n#10 0x00007ff74248c0bb in std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> >::vector (this=0x7fff1a3c9f00, __x=std::vector of length 4693611, capacity 0 = {...}) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_vector.h:243\n#11 0x00007ff742499af3 in Limits::RingLimits::RingLimits (this=0x7fff1a3c9f00) at ../../source/IceHive/public/IceHive/IceHiveHelpers.h:38\n#12 0x00007ff742092bd1 in I3HiveCluster<I3RecoPulse>::I3HiveCluster (this=0x483e450, context=...) at ../../source/IceHive/private/IceHive/I3HiveCluster.h:130\n#13 0x00007ff74209135c in StandardCreate<I3Module, I3HiveCluster<I3RecoPulse> >::Create (c=...) at ../../source/icetray/public/icetray/I3Factory.h:110\n#14 0x00007ff805c9d73b in boost::detail::function::function_invoker1<boost::shared_ptr<I3Module> (*)(I3Context const&), boost::shared_ptr<I3Module>, I3Context const&>::invoke (function_ptr=..., a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:94\n#15 0x00007ff805c0bc74 in boost::function1<boost::shared_ptr<I3Module>, I3Context const&>::operator() (this=0x7fff1a3ca2d0, a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:989\n#16 0x00007ff805dcd4ee in module_default_config (name=\"I3HiveCluster<I3RecoPulse>\") at ../../source/icetray/private/pybindings/I3ModuleFactory.cxx:36\n#17 0x00007ff805dd0ce9 in boost::python::detail::invoke<boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::result_converter, I3ConfigurationPtr (*)(const std::string&), boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::c_t0>(boost::python::detail::invoke_tag_<false, false>, const impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::result_converter &, I3ConfigurationPtr (*&)(const std::string &), impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::c_t0 &) (rc=..., f=\n0x1d22fb8, ac0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/invoke.hpp:75\n#18 0x00007ff805dd072a in impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::operator() (this=0x1d22fb8, args_=0x4037290) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/caller.hpp:223\n#19 0x00007ff805dd0051 in boost::python::objects::caller_py_function_impl<boost::python::detail::caller<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> > >::operator() (this=0x1d22fb0, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/object/py_function.hpp:38\n#20 0x00007ff804eb43c5 in boost::python::objects::py_function::operator() (this=0x1d23440, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/object/py_function.hpp:143\n#21 0x00007ff804eb1f7b in boost::python::objects::function::call (this=0x1d23430, args=0x4037290, keywords=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:226\n#22 0x00007ff804eb372f in boost::python::objects::(anonymous namespace)::bind_return::operator() (this=0x7fff1a3ca900) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:581\n#23 0x00007ff804eb40f7 in boost::detail::function::void_function_ref_invoker0<boost::python::objects::<unnamed>::bind_return, void>::invoke(boost::detail::function::function_buffer &) (function_obj_ptr=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:187\n#24 0x00007ff804eca39a in boost::function0<void>::operator() (this=0x7fff1a3ca8b0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#25 0x00007ff804ec9dfc in boost::python::detail::exception_handler::operator() (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:74\n#26 0x00007ff7fdc810d0 in boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>::operator() (this=0x2040c98, handler=..., f=..., translate=0x7ff7fdc59bec <translate(not_found_exception const&)>) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/translate_exception.hpp:46\n#27 0x00007ff7fdc7e3dd in boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> >::operator()<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list2<boost::python::detail::exception_handler const&, boost::function0<void> const&> > (this=0x2040ca0, f=..., a=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind.hpp:361\n#28 0x00007ff7fdc79dde in boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >::operator()<boost::python::detail::exception_handler, boost::function0<void> > (this=0x2040c98, a1=..., a2=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind/bind_template.hpp:102\n#29 0x00007ff7fdc7504c in boost::detail::function::function_obj_invoker2<boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >, bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::invoke (function_obj_ptr=..., a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:131\n#30 0x00007ff804eca315 in boost::function2<bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::operator() (this=0x2040c90, a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#31 0x00007ff804eca1a3 in boost::python::detail::exception_handler::handle (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/detail/exception_handler.hpp:41\n#32 0x00007ff804ec9b13 in boost::python::handle_exception_impl (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:24\n#33 0x00007ff804eb3dab in boost::python::handle_exception<boost::python::objects::<unnamed>::bind_return>(boost::python::objects::(anonymous namespace)::bind_return) (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/errors.hpp:29\n#34 0x00007ff804eb380f in boost::python::objects::function_call (func=0x1d23430, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:613\n#35 0x00007ff811363dd3 in PyObject_Call (func=0x1d23430, arg=<value optimized out>, kw=<value optimized out>) at Objects/abstract.c:2529\n#36 0x00007ff81141126d in do_call (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4239\n#37 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4044\n#38 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#39 0x00007ff81141281e in fast_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4107\n#40 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4042\n#41 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#42 0x00007ff81141391e in PyEval_EvalCodeEx (co=0x7ff8117a3b30, globals=<value optimized out>, locals=<value optimized out>, args=<value optimized out>, argcount=0, kws=0x0, kwcount=0, defs=0x0, defcount=0, closure=0x0) at Python/ceval.c:3253\n#43 0x00007ff811413a32 in PyEval_EvalCode (co=<value optimized out>, globals=<value optimized out>, locals=<value optimized out>) at Python/ceval.c:667\n#44 0x00007ff8114334d0 in run_mod (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1353\n#45 PyRun_FileExFlags (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1339\n#46 0x00007ff81143369c in PyRun_SimpleFileExFlags (fp=0x1a14a80, filename=0x7fff1a3cc2de \"/build/buildslave/kipple_cvmfs/docs/build/bin/icetray-inspect\", closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:943\n#47 0x00007ff8114490ec in Py_Main (argc=<value optimized out>, argv=<value optimized out>) at Modules/main.c:639\n#48 0x00007ff8106ffd5d in __libc_start_main () from /lib64/libc.so.6\n#49 0x0000000000400649 in _start ()\n===========================================================\n\n\nmake[3]: *** [CMakeFiles/inspect] Error 1\nmake[2]: *** [CMakeFiles/inspect.dir/all] Error 2\nmake[1]: *** [CMakeFiles/docs.dir/rule] Error 2\nmake: *** [docs] Error 2\n(virtenv) [buildslave@kipple build]$\n}}}", "reporter": "nega", "cc": "olivas", "resolution": "wontfix", "_ts": "1550067178841456", "component": "combo simulation", "summary": "IceHive - crashes during icetray-inspect", "priority": "major", "keywords": "icetray-inspect icehive documentation", "time": "2016-05-17T17:07:39", "milestone": "", "owner": "nega", "type": "defect" } ```
1.0
IceHive - crashes during icetray-inspect (Trac #1708) - Migrated from https://code.icecube.wisc.edu/ticket/1708 ```json { "status": "closed", "changetime": "2019-02-13T14:12:58", "description": "this is preventing the docs from being built\n\nmaybe `Limits::RingLimits::RingLimits()` is getting bum vector?\n\nroot-ified stack trace:\n\n{{{\n===========================================================\nThere was a crash.\nThis is the entire stack trace of all threads:\n===========================================================\n\nThread 5 (Thread 0x7ff746184700 (LWP 19476)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 4 (Thread 0x7ff745783700 (LWP 19477)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 3 (Thread 0x7ff744d82700 (LWP 19478)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 2 (Thread 0x7ff744381700 (LWP 19479)):\n#0 0x00007ff81110b68c in pthread_cond_wait\n\nGLIBC_2.3.2 () from /lib64/libpthread.so.0\n#1 0x00007ff7461b7e88 in th_worker (tidptr=<value optimized out>) at numexpr/module.cpp:57\n#2 0x00007ff811107aa1 in start_thread () from /lib64/libpthread.so.0\n#3 0x00007ff8107c993d in clone () from /lib64/libc.so.6\n\nThread 1 (Thread 0x7ff81190b700 (LWP 19361)):\n#0 0x00007ff81078d69d in waitpid () from /lib64/libc.so.6\n#1 0x00007ff81071f609 in do_system () from /lib64/libc.so.6\n#2 0x00007ff81071f940 in system () from /lib64/libc.so.6\n#3 0x00007ff7f2dfbad8 in TUnixSystem::StackTrace() () from /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/root-v5.34.18/lib/libCore.so\n#4 0x00007ff7f2dfa5d3 in TUnixSystem::DispatchSignals(ESignals) () from /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/root-v5.34.18/lib/libCore.so\n#5 <signal handler called>\n#6 0x00007ff742492f52 in std::_Construct<Limits::LimitPair, Limits::LimitPair> (__p=0x7ff73d7aa010, __value=...) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_construct.h:80\n#7 0x00007ff74249121c in std::__uninitialized_copy<false>::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#8 0x00007ff74248f3d5 in std::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#9 0x00007ff74248d1c8 in std::__uninitialized_copy_a<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*, Limits::LimitPair> (__first=Cannot access memory at address 0x1\n#10 0x00007ff74248c0bb in std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> >::vector (this=0x7fff1a3c9f00, __x=std::vector of length 4693611, capacity 0 = {...}) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_vector.h:243\n#11 0x00007ff742499af3 in Limits::RingLimits::RingLimits (this=0x7fff1a3c9f00) at ../../source/IceHive/public/IceHive/IceHiveHelpers.h:38\n#12 0x00007ff742092bd1 in I3HiveCluster<I3RecoPulse>::I3HiveCluster (this=0x483e450, context=...) at ../../source/IceHive/private/IceHive/I3HiveCluster.h:130\n#13 0x00007ff74209135c in StandardCreate<I3Module, I3HiveCluster<I3RecoPulse> >::Create (c=...) at ../../source/icetray/public/icetray/I3Factory.h:110\n#14 0x00007ff805c9d73b in boost::detail::function::function_invoker1<boost::shared_ptr<I3Module> (*)(I3Context const&), boost::shared_ptr<I3Module>, I3Context const&>::invoke (function_ptr=..., a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:94\n#15 0x00007ff805c0bc74 in boost::function1<boost::shared_ptr<I3Module>, I3Context const&>::operator() (this=0x7fff1a3ca2d0, a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:989\n#16 0x00007ff805dcd4ee in module_default_config (name=\"I3HiveCluster<I3RecoPulse>\") at ../../source/icetray/private/pybindings/I3ModuleFactory.cxx:36\n#17 0x00007ff805dd0ce9 in boost::python::detail::invoke<boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::result_converter, I3ConfigurationPtr (*)(const std::string&), boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::c_t0>(boost::python::detail::invoke_tag_<false, false>, const impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::result_converter &, I3ConfigurationPtr (*&)(const std::string &), impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::c_t0 &) (rc=..., f=\n0x1d22fb8, ac0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/invoke.hpp:75\n#18 0x00007ff805dd072a in impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::operator() (this=0x1d22fb8, args_=0x4037290) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/caller.hpp:223\n#19 0x00007ff805dd0051 in boost::python::objects::caller_py_function_impl<boost::python::detail::caller<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> > >::operator() (this=0x1d22fb0, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/object/py_function.hpp:38\n#20 0x00007ff804eb43c5 in boost::python::objects::py_function::operator() (this=0x1d23440, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/object/py_function.hpp:143\n#21 0x00007ff804eb1f7b in boost::python::objects::function::call (this=0x1d23430, args=0x4037290, keywords=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:226\n#22 0x00007ff804eb372f in boost::python::objects::(anonymous namespace)::bind_return::operator() (this=0x7fff1a3ca900) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:581\n#23 0x00007ff804eb40f7 in boost::detail::function::void_function_ref_invoker0<boost::python::objects::<unnamed>::bind_return, void>::invoke(boost::detail::function::function_buffer &) (function_obj_ptr=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:187\n#24 0x00007ff804eca39a in boost::function0<void>::operator() (this=0x7fff1a3ca8b0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#25 0x00007ff804ec9dfc in boost::python::detail::exception_handler::operator() (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:74\n#26 0x00007ff7fdc810d0 in boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>::operator() (this=0x2040c98, handler=..., f=..., translate=0x7ff7fdc59bec <translate(not_found_exception const&)>) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/translate_exception.hpp:46\n#27 0x00007ff7fdc7e3dd in boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> >::operator()<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list2<boost::python::detail::exception_handler const&, boost::function0<void> const&> > (this=0x2040ca0, f=..., a=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind.hpp:361\n#28 0x00007ff7fdc79dde in boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >::operator()<boost::python::detail::exception_handler, boost::function0<void> > (this=0x2040c98, a1=..., a2=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind/bind_template.hpp:102\n#29 0x00007ff7fdc7504c in boost::detail::function::function_obj_invoker2<boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >, bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::invoke (function_obj_ptr=..., a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:131\n#30 0x00007ff804eca315 in boost::function2<bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::operator() (this=0x2040c90, a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#31 0x00007ff804eca1a3 in boost::python::detail::exception_handler::handle (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/detail/exception_handler.hpp:41\n#32 0x00007ff804ec9b13 in boost::python::handle_exception_impl (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:24\n#33 0x00007ff804eb3dab in boost::python::handle_exception<boost::python::objects::<unnamed>::bind_return>(boost::python::objects::(anonymous namespace)::bind_return) (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/errors.hpp:29\n#34 0x00007ff804eb380f in boost::python::objects::function_call (func=0x1d23430, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:613\n#35 0x00007ff811363dd3 in PyObject_Call (func=0x1d23430, arg=<value optimized out>, kw=<value optimized out>) at Objects/abstract.c:2529\n#36 0x00007ff81141126d in do_call (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4239\n#37 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4044\n#38 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#39 0x00007ff81141281e in fast_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4107\n#40 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4042\n#41 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#42 0x00007ff81141391e in PyEval_EvalCodeEx (co=0x7ff8117a3b30, globals=<value optimized out>, locals=<value optimized out>, args=<value optimized out>, argcount=0, kws=0x0, kwcount=0, defs=0x0, defcount=0, closure=0x0) at Python/ceval.c:3253\n#43 0x00007ff811413a32 in PyEval_EvalCode (co=<value optimized out>, globals=<value optimized out>, locals=<value optimized out>) at Python/ceval.c:667\n#44 0x00007ff8114334d0 in run_mod (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1353\n#45 PyRun_FileExFlags (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1339\n#46 0x00007ff81143369c in PyRun_SimpleFileExFlags (fp=0x1a14a80, filename=0x7fff1a3cc2de \"/build/buildslave/kipple_cvmfs/docs/build/bin/icetray-inspect\", closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:943\n#47 0x00007ff8114490ec in Py_Main (argc=<value optimized out>, argv=<value optimized out>) at Modules/main.c:639\n#48 0x00007ff8106ffd5d in __libc_start_main () from /lib64/libc.so.6\n#49 0x0000000000400649 in _start ()\n===========================================================\n\n\nThe lines below might hint at the cause of the crash.\nIf they do not help you then please submit a bug report at\nhttp://root.cern.ch/bugs. Please post the ENTIRE stack trace\nfrom above as an attachment in addition to anything else\nthat might help us fixing this issue.\n===========================================================\n#6 0x00007ff742492f52 in std::_Construct<Limits::LimitPair, Limits::LimitPair> (__p=0x7ff73d7aa010, __value=...) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_construct.h:80\n#7 0x00007ff74249121c in std::__uninitialized_copy<false>::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#8 0x00007ff74248f3d5 in std::uninitialized_copy<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*> (__first=Cannot access memory at address 0x1\n#9 0x00007ff74248d1c8 in std::__uninitialized_copy_a<__gnu_cxx::__normal_iterator<Limits::LimitPair const*, std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> > >, Limits::LimitPair*, Limits::LimitPair> (__first=Cannot access memory at address 0x1\n#10 0x00007ff74248c0bb in std::vector<Limits::LimitPair, std::allocator<Limits::LimitPair> >::vector (this=0x7fff1a3c9f00, __x=std::vector of length 4693611, capacity 0 = {...}) at /usr/lib/gcc/x86_64-redhat-linux/4.4.7/../../../../include/c++/4.4.7/bits/stl_vector.h:243\n#11 0x00007ff742499af3 in Limits::RingLimits::RingLimits (this=0x7fff1a3c9f00) at ../../source/IceHive/public/IceHive/IceHiveHelpers.h:38\n#12 0x00007ff742092bd1 in I3HiveCluster<I3RecoPulse>::I3HiveCluster (this=0x483e450, context=...) at ../../source/IceHive/private/IceHive/I3HiveCluster.h:130\n#13 0x00007ff74209135c in StandardCreate<I3Module, I3HiveCluster<I3RecoPulse> >::Create (c=...) at ../../source/icetray/public/icetray/I3Factory.h:110\n#14 0x00007ff805c9d73b in boost::detail::function::function_invoker1<boost::shared_ptr<I3Module> (*)(I3Context const&), boost::shared_ptr<I3Module>, I3Context const&>::invoke (function_ptr=..., a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:94\n#15 0x00007ff805c0bc74 in boost::function1<boost::shared_ptr<I3Module>, I3Context const&>::operator() (this=0x7fff1a3ca2d0, a0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:989\n#16 0x00007ff805dcd4ee in module_default_config (name=\"I3HiveCluster<I3RecoPulse>\") at ../../source/icetray/private/pybindings/I3ModuleFactory.cxx:36\n#17 0x00007ff805dd0ce9 in boost::python::detail::invoke<boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::result_converter, I3ConfigurationPtr (*)(const std::string&), boost::python::detail::caller_arity<1u>::impl<F, Policies, Sig>::operator()(PyObject*, PyObject*) [with F = I3ConfigurationPtr (*)(const std::string&), Policies = boost::python::default_call_policies, Sig = boost::mpl::vector2<boost::shared_ptr<I3Configuration>, const std::string&>]::c_t0>(boost::python::detail::invoke_tag_<false, false>, const impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::result_converter &, I3ConfigurationPtr (*&)(const std::string &), impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::c_t0 &) (rc=..., f=\n0x1d22fb8, ac0=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/invoke.hpp:75\n#18 0x00007ff805dd072a in impl<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> >::operator() (this=0x1d22fb8, args_=0x4037290) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/caller.hpp:223\n#19 0x00007ff805dd0051 in boost::python::objects::caller_py_function_impl<boost::python::detail::caller<I3ConfigurationPtr (*)(std::string const&), boost::python::default_call_policies, boost::mpl::vector2<boost::shared_ptr<I3Configuration>, std::string const&> > >::operator() (this=0x1d22fb0, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/object/py_function.hpp:38\n#20 0x00007ff804eb43c5 in boost::python::objects::py_function::operator() (this=0x1d23440, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/object/py_function.hpp:143\n#21 0x00007ff804eb1f7b in boost::python::objects::function::call (this=0x1d23430, args=0x4037290, keywords=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:226\n#22 0x00007ff804eb372f in boost::python::objects::(anonymous namespace)::bind_return::operator() (this=0x7fff1a3ca900) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:581\n#23 0x00007ff804eb40f7 in boost::detail::function::void_function_ref_invoker0<boost::python::objects::<unnamed>::bind_return, void>::invoke(boost::detail::function::function_buffer &) (function_obj_ptr=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:187\n#24 0x00007ff804eca39a in boost::function0<void>::operator() (this=0x7fff1a3ca8b0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#25 0x00007ff804ec9dfc in boost::python::detail::exception_handler::operator() (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:74\n#26 0x00007ff7fdc810d0 in boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>::operator() (this=0x2040c98, handler=..., f=..., translate=0x7ff7fdc59bec <translate(not_found_exception const&)>) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/python/detail/translate_exception.hpp:46\n#27 0x00007ff7fdc7e3dd in boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> >::operator()<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list2<boost::python::detail::exception_handler const&, boost::function0<void> const&> > (this=0x2040ca0, f=..., a=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind.hpp:361\n#28 0x00007ff7fdc79dde in boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >::operator()<boost::python::detail::exception_handler, boost::function0<void> > (this=0x2040c98, a1=..., a2=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/bind/bind_template.hpp:102\n#29 0x00007ff7fdc7504c in boost::detail::function::function_obj_invoker2<boost::_bi::bind_t<bool, boost::python::detail::translate_exception<not_found_exception, void (*)(not_found_exception const&)>, boost::_bi::list3<boost::arg<1>, boost::arg<2>, boost::_bi::value<void (*)(not_found_exception const&)> > >, bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::invoke (function_obj_ptr=..., a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/include/boost-1.38.0/boost/function/function_template.hpp:131\n#30 0x00007ff804eca315 in boost::function2<bool, boost::python::detail::exception_handler const&, boost::function0<void> const&>::operator() (this=0x2040c90, a0=..., a1=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/function/include/boost/function/function_template.hpp:989\n#31 0x00007ff804eca1a3 in boost::python::detail::exception_handler::handle (this=0x2040c90, f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/detail/exception_handler.hpp:41\n#32 0x00007ff804ec9b13 in boost::python::handle_exception_impl (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/errors.cpp:24\n#33 0x00007ff804eb3dab in boost::python::handle_exception<boost::python::objects::<unnamed>::bind_return>(boost::python::objects::(anonymous namespace)::bind_return) (f=...) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/include/boost/python/errors.hpp:29\n#34 0x00007ff804eb380f in boost::python::objects::function_call (func=0x1d23430, args=0x4037290, kw=0x0) at /cvmfs/icecube.opensciencegrid.org/py2-v1/RHEL_6_x86_64/i3ports/var/db/dports/build/file._cvmfs_icecube.opensciencegrid.org_py2-v1_RHEL_6_x86_64_i3ports_var_db_dports_sources_rsync.code.icecube.wisc.edu_icecube-tools-ports_devel_boost_1.38.0/work/boost-1.38.0/libs/python/src/object/function.cpp:613\n#35 0x00007ff811363dd3 in PyObject_Call (func=0x1d23430, arg=<value optimized out>, kw=<value optimized out>) at Objects/abstract.c:2529\n#36 0x00007ff81141126d in do_call (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4239\n#37 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4044\n#38 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#39 0x00007ff81141281e in fast_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4107\n#40 call_function (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:4042\n#41 PyEval_EvalFrameEx (f=<value optimized out>, throwflag=<value optimized out>) at Python/ceval.c:2666\n#42 0x00007ff81141391e in PyEval_EvalCodeEx (co=0x7ff8117a3b30, globals=<value optimized out>, locals=<value optimized out>, args=<value optimized out>, argcount=0, kws=0x0, kwcount=0, defs=0x0, defcount=0, closure=0x0) at Python/ceval.c:3253\n#43 0x00007ff811413a32 in PyEval_EvalCode (co=<value optimized out>, globals=<value optimized out>, locals=<value optimized out>) at Python/ceval.c:667\n#44 0x00007ff8114334d0 in run_mod (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1353\n#45 PyRun_FileExFlags (fp=0x1a14a80, filename=<value optimized out>, start=<value optimized out>, globals=0x197a3a0, locals=0x197a3a0, closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:1339\n#46 0x00007ff81143369c in PyRun_SimpleFileExFlags (fp=0x1a14a80, filename=0x7fff1a3cc2de \"/build/buildslave/kipple_cvmfs/docs/build/bin/icetray-inspect\", closeit=1, flags=0x7fff1a3caf00) at Python/pythonrun.c:943\n#47 0x00007ff8114490ec in Py_Main (argc=<value optimized out>, argv=<value optimized out>) at Modules/main.c:639\n#48 0x00007ff8106ffd5d in __libc_start_main () from /lib64/libc.so.6\n#49 0x0000000000400649 in _start ()\n===========================================================\n\n\nmake[3]: *** [CMakeFiles/inspect] Error 1\nmake[2]: *** [CMakeFiles/inspect.dir/all] Error 2\nmake[1]: *** [CMakeFiles/docs.dir/rule] Error 2\nmake: *** [docs] Error 2\n(virtenv) [buildslave@kipple build]$\n}}}", "reporter": "nega", "cc": "olivas", "resolution": "wontfix", "_ts": "1550067178841456", "component": "combo simulation", "summary": "IceHive - crashes during icetray-inspect", "priority": "major", "keywords": "icetray-inspect icehive documentation", "time": "2016-05-17T17:07:39", "milestone": "", "owner": "nega", "type": "defect" } ```
defect
icehive crashes during icetray inspect trac migrated from json status closed changetime description this is preventing the docs from being built n nmaybe limits ringlimits ringlimits is getting bum vector n nroot ified stack trace n n n nthere was a crash nthis is the entire stack trace of all threads n n nthread thread lwp n in pthread cond wait n nglibc from libpthread so n in th worker tidptr at numexpr module cpp n in start thread from libpthread so n in clone from libc so n nthread thread lwp n in pthread cond wait n nglibc from libpthread so n in th worker tidptr at numexpr module cpp n in start thread from libpthread so n in clone from libc so n nthread thread lwp n in pthread cond wait n nglibc from libpthread so n in th worker tidptr at numexpr module cpp n in start thread from libpthread so n in clone from libc so n nthread thread lwp n in pthread cond wait n nglibc from libpthread so n in th worker tidptr at numexpr module cpp n in start thread from libpthread so n in clone from libc so n nthread thread lwp n in waitpid from libc so n in do system from libc so n in system from libc so n in tunixsystem stacktrace from cvmfs icecube opensciencegrid org rhel root lib libcore so n in tunixsystem dispatchsignals esignals from cvmfs icecube opensciencegrid org rhel root lib libcore so n n in std construct p value at usr lib gcc redhat linux include c bits stl construct h n in std uninitialized copy uninitialized copy limits limitpair first cannot access memory at address n in std uninitialized copy limits limitpair first cannot access memory at address n in std uninitialized copy a limits limitpair limits limitpair first cannot access memory at address n in std vector vector this x std vector of length capacity at usr lib gcc redhat linux include c bits stl vector h n in limits ringlimits ringlimits this at source icehive public icehive icehivehelpers h n in this context at source icehive private icehive h n in standardcreate create c at source icetray public icetray h n in boost detail function function const boost shared ptr const invoke function ptr at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in boost const operator this at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in module default config name at source icetray private pybindings cxx n in boost python detail invoke impl operator pyobject pyobject result converter const std string boost python detail caller arity impl operator pyobject pyobject c boost python detail invoke tag const impl std string const result converter const std string impl std string const c rc f at cvmfs icecube opensciencegrid org rhel include boost boost python detail invoke hpp n in impl std string const operator this args at cvmfs icecube opensciencegrid org rhel include boost boost python detail caller hpp n in boost python objects caller py function impl std string const operator this args kw at cvmfs icecube opensciencegrid org rhel include boost boost python object py function hpp n in boost python objects py function operator this args kw at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python object py function hpp n in boost python objects function call this args keywords at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in boost python objects anonymous namespace bind return operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in boost detail function void function ref bind return void invoke boost detail function function buffer function obj ptr at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost python detail exception handler operator this f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src errors cpp n in boost python detail translate exception operator this handler f translate at cvmfs icecube opensciencegrid org rhel include boost boost python detail translate exception hpp n in boost bi boost arg boost bi value operator boost bi const this f a at cvmfs icecube opensciencegrid org rhel include boost boost bind hpp n in boost bi bind t boost bi boost arg boost bi value operator this at cvmfs icecube opensciencegrid org rhel include boost boost bind bind template hpp n in boost detail function function obj boost bi boost arg boost bi value bool boost python detail exception handler const boost const invoke function obj ptr at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in boost const operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost python detail exception handler handle this f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python detail exception handler hpp n in boost python handle exception impl f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src errors cpp n in boost python handle exception bind return boost python objects anonymous namespace bind return f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python errors hpp n in boost python objects function call func args kw at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in pyobject call func arg kw at objects abstract c n in do call f throwflag at python ceval c n call function f throwflag at python ceval c n pyeval evalframeex f throwflag at python ceval c n in fast function f throwflag at python ceval c n call function f throwflag at python ceval c n pyeval evalframeex f throwflag at python ceval c n in pyeval evalcodeex co globals locals args argcount kws kwcount defs defcount closure at python ceval c n in pyeval evalcode co globals locals at python ceval c n in run mod fp filename start globals locals closeit flags at python pythonrun c n pyrun fileexflags fp filename start globals locals closeit flags at python pythonrun c n in pyrun simplefileexflags fp filename build buildslave kipple cvmfs docs build bin icetray inspect closeit flags at python pythonrun c n in py main argc argv at modules main c n in libc start main from libc so n in start n n n nthe lines below might hint at the cause of the crash nif they do not help you then please submit a bug report at n please post the entire stack trace nfrom above as an attachment in addition to anything else nthat might help us fixing this issue n n in std construct p value at usr lib gcc redhat linux include c bits stl construct h n in std uninitialized copy uninitialized copy limits limitpair first cannot access memory at address n in std uninitialized copy limits limitpair first cannot access memory at address n in std uninitialized copy a limits limitpair limits limitpair first cannot access memory at address n in std vector vector this x std vector of length capacity at usr lib gcc redhat linux include c bits stl vector h n in limits ringlimits ringlimits this at source icehive public icehive icehivehelpers h n in this context at source icehive private icehive h n in standardcreate create c at source icetray public icetray h n in boost detail function function const boost shared ptr const invoke function ptr at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in boost const operator this at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in module default config name at source icetray private pybindings cxx n in boost python detail invoke impl operator pyobject pyobject result converter const std string boost python detail caller arity impl operator pyobject pyobject c boost python detail invoke tag const impl std string const result converter const std string impl std string const c rc f at cvmfs icecube opensciencegrid org rhel include boost boost python detail invoke hpp n in impl std string const operator this args at cvmfs icecube opensciencegrid org rhel include boost boost python detail caller hpp n in boost python objects caller py function impl std string const operator this args kw at cvmfs icecube opensciencegrid org rhel include boost boost python object py function hpp n in boost python objects py function operator this args kw at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python object py function hpp n in boost python objects function call this args keywords at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in boost python objects anonymous namespace bind return operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in boost detail function void function ref bind return void invoke boost detail function function buffer function obj ptr at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost python detail exception handler operator this f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src errors cpp n in boost python detail translate exception operator this handler f translate at cvmfs icecube opensciencegrid org rhel include boost boost python detail translate exception hpp n in boost bi boost arg boost bi value operator boost bi const this f a at cvmfs icecube opensciencegrid org rhel include boost boost bind hpp n in boost bi bind t boost bi boost arg boost bi value operator this at cvmfs icecube opensciencegrid org rhel include boost boost bind bind template hpp n in boost detail function function obj boost bi boost arg boost bi value bool boost python detail exception handler const boost const invoke function obj ptr at cvmfs icecube opensciencegrid org rhel include boost boost function function template hpp n in boost const operator this at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs function include boost function function template hpp n in boost python detail exception handler handle this f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python detail exception handler hpp n in boost python handle exception impl f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src errors cpp n in boost python handle exception bind return boost python objects anonymous namespace bind return f at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python include boost python errors hpp n in boost python objects function call func args kw at cvmfs icecube opensciencegrid org rhel var db dports build file cvmfs icecube opensciencegrid org rhel var db dports sources rsync code icecube wisc edu icecube tools ports devel boost work boost libs python src object function cpp n in pyobject call func arg kw at objects abstract c n in do call f throwflag at python ceval c n call function f throwflag at python ceval c n pyeval evalframeex f throwflag at python ceval c n in fast function f throwflag at python ceval c n call function f throwflag at python ceval c n pyeval evalframeex f throwflag at python ceval c n in pyeval evalcodeex co globals locals args argcount kws kwcount defs defcount closure at python ceval c n in pyeval evalcode co globals locals at python ceval c n in run mod fp filename start globals locals closeit flags at python pythonrun c n pyrun fileexflags fp filename start globals locals closeit flags at python pythonrun c n in pyrun simplefileexflags fp filename build buildslave kipple cvmfs docs build bin icetray inspect closeit flags at python pythonrun c n in py main argc argv at modules main c n in libc start main from libc so n in start n n n nmake error nmake error nmake error nmake error n virtenv n reporter nega cc olivas resolution wontfix ts component combo simulation summary icehive crashes during icetray inspect priority major keywords icetray inspect icehive documentation time milestone owner nega type defect
1
80,421
30,282,624,431
IssuesEvent
2023-07-08 09:00:24
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
zdb -R shows incorrect data
Type: Defect
### System information Type | Version/Name --- | --- Distribution Name | ubuntu Distribution Version | 22.4 Kernel Version | ubuntu 5.15.0-1033-raspi Architecture | aarch64 aarch64 aarch64 GNU/Linux OpenZFS Version | zfs-2.2.99-1, zfs-kmod-2.1.5-1ubuntu6~22.04.1 ### Describe the problem you're observing ## Background: I have a 4disk raid z2 pool that had one disk failed. As I can't import the pool any longer ( crashes on import on macos, freebsd, spews out errors on linux ). Importing it with readonly works, but anything else but using zpool status will crash the machine (freebsd,macos) or hangs zfs (linux). So now I'm in the process of getting my data back with zdb when I encountered this bug. ``` root@ubuntu:/home/ubuntu# zpool status pool: ocean state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scan: resilver in progress since Sat Jul 1 10:01:24 2023 0B / 11.6T scanned, 0B / 11.6T issued 0B resilvered, 0.00% done, no estimated completion time config: NAME STATE READ WRITE CKSUM ocean DEGRADED 0 0 0 raidz2-0 DEGRADED 0 0 0 18276338438250889807 FAULTED 0 0 0 was /dev/sdd2 sdc1 ONLINE 0 0 0 sdb2 ONLINE 0 0 0 sdd2 ONLINE 0 0 0 errors: 3962 data errors, use '-v' for a list root@ubuntu:/home/ubuntu# zdb -u -e ocean Uberblock: magic = 0000000000bab10c version = 5000 txg = 18360509 guid_sum = 15765822683839940218 timestamp = 1688241822 UTC = Sat Jul 1 20:03:42 2023 mmp_magic = 00000000a11cea11 mmp_delay = 0 mmp_valid = 0 checkpoint_txg = 0 root@ubuntu:/home/ubuntu# zdb -dP -e ocean Dataset mos [META], ID 0, cr_txg 4, 1092430080, 6633 objects Dataset ocean/tmtinkerbell@2023-06-24T00:00:01 [ZPL], ID 5114, cr_txg 18302756, 1111802459136, 24930 objects ... Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 4065675030144, 9979 objects ... dmu_bonus_hold(os, object, dl, &dl->dl_dbuf) == 0 (0x34 == 0) ASSERT at module/zfs/dsl_deadlist.c:308:dsl_deadlist_open()Aborted (core dumped) root@ubuntu:/home/ubuntu# zdb -dd -e ocean/movies Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 3.70T, 9979 objects ZIL header: claim_txg 18360433, claim_blk_seq 1, claim_lr_seq 0 replay_seq 0, flags 0x2 Object lvl iblk dblk dsize dnsize lsize %full type 0 6 128K 16K 14.7M 512 96.9M 5.03 DMU dnode -1 1 128K 512 11.5K 512 512 100.00 ZFS user/group/project used -2 1 128K 512 11.5K 512 512 100.00 ZFS user/group/project used 1 1 128K 1K 11.5K 512 1K 100.00 ZFS master node 2 1 128K 512 11.5K 512 512 100.00 SA master node 3 1 128K 21.5K 11.5K 512 21.5K 100.00 ZFS delete queue 4 2 128K 16K 35K 512 32K 100.00 ZFS directory 5 1 128K 2K 11.5K 512 2K 100.00 SA attr registration ... ``` And looking at one file: ``` root@ubuntu:~# zdb -ddddd -K $pass -e ocean/movies 5268 Unlocked encryption root: ocean/movies Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 3.70T, 9979 objects, rootbp DVA[0]=<0:2400014a000:3000> DVA[1]=<0:f20081e4000:3000> [L0 DMU objset] fletcher4 uncompressed authenticated LE contiguous unique double size=800L/800P birth=18360433L/18360433P fill=9979 cksum=00000021c040a809:000025b93e7be9da:00190442ec3d9de4:0c1fdfcfea494199 Object lvl iblk dblk dsize dnsize lsize %full type 5268 1 128K 105K 40.5K 512 105K 100.00 ZFS plain file 168 bonus System attributes dnode flags: USED_BYTES USERUSED_ACCOUNTED dnode maxblkid: 0 path /Pictures/Testbilder/NecPattern.png uid 1000 gid 1000 atime Tue Dec 31 08:24:53 2019 mtime Thu Feb 11 09:05:10 2016 ctime Wed Jan 27 17:59:02 2021 crtime Tue Jan 26 20:40:27 2021 gen 4641518 mode 100644 size 107030 parent 5267 links 1 pflags 40800000004 Indirect blocks: 0 L0 0:10a1263d2000:15000 1a400L/9000P F=1 B=4641518/4641518 cksum=246684c5894cb36e:ecb92ef5e414d48a:13123f3d0f3e78df:9619bdcef80cd664 segment [0000000000000000, 000000000001a400) size 105K ``` ## The Issue Using **zdb -R** on an unencrypted pool works and gives me back the data, but on encrypted pools that does not work as shown here there is no PNG header as it should ( described in the how to replicate section) ``` root@ubuntu:/home/ubuntu# zdb -R -K $pass $disks -e ocean/movies 0:10a1263d2000:1a400 |head Found vdev type: raidz 0:10a1263d2000:1a400 0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef 000000: 9c042f5028cc72d6 1eb5378942ca849d .r.(P/.....B.7.. 000010: 39d46c5bb5ef339d 9aeb9cbf22549f9a .3..[l.9..T".... 000020: c4bcb91876bfb295 9b297424b1886cdd ...v.....l..$t). 000030: dab8c9b2e13c8f06 f9517b44e365816a ..<.....j.e.D{Q. 000040: f63da5f46b496aa5 d2a3744d20577b5a .jIk..=.Z{W Mt.. ... ``` ### Describe how to reproduce the problem Create a pool with an ecrypted pool inside and copy an image.png into it. ``` mkfile 100m sda sdb sdc sdd pass='testtest' disks='-p /Users/jolly/Desktop/zfsdebug2/sda -p /Users/jolly/Desktop/zfsdebug2/sdb -p /Users/jolly/Desktop/zfsdebug2/sdc -p /Users/jolly/Desktop/zfsdebug2/sdd' # zpool create test raidz2 $disks # create -o encryption=on -o keylocation=prompt -o keyformat=passphrase test/movies # cp Image.png /Volumes/test/movies/ ``` look for the image size in ``` # zdb -dd $disks -e test/movies physmem = 2097152 pages (32.00 GB) looking up movies in obj34 Dataset test/movies [ZPL], ID 388, cr_txg 32, 2.08M, 104 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 6 128K 16K 280K 512 288K 18.06 DMU dnode -1 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used -2 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used -3 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used 1 1 128K 1K 1.50K 512 1K 100.00 ZFS master node 2 2 128K 128K 1.66M 512 1.75M 100.00 ZFS plain file 3 1 128K 512 1.50K 512 512 100.00 ZFS directory 4 1 128K 512 512 512 512 100.00 ZFS plain file 5 1 128K 6.50K 4.50K 512 6.50K 100.00 ZFS directory 6 1 128K 512 0 512 512 0.00 ZFS plain file 7 1 128K 512 1.50K 512 512 100.00 ZFS directory 9 1 128K 105K 33K 512 105K 100.00 ZFS plain file 10 1 128K 512 1.50K 512 512 100.00 ZFS directory ... ``` I my case 105k Bytes - so it's object number 9 Then find out the blocks ``` # zdb -ddddd $disks -e test/movies 9 physmem = 2097152 pages (32.00 GB) looking up movies in obj34 Dataset test/movies [ZPL], ID 388, cr_txg 32, 2.08M, 104 objects, rootbp DVA[0]=<0:ce95800:2400> DVA[1]=<0:41f1400:2400> [L0 DMU objset] fletcher4 uncompressed authenticated LE contiguous unique double size=1000L/1000P birth=43L/43P fill=104 cksum=00000022e10e3971:000064b3a2f91165:009866b8a335c5af:9fae33cbc48176dc Object lvl iblk dblk dsize dnsize lsize %full type 9 1 128K 105K 33K 512 105K 100.00 ZFS plain file 176 bonus System attributes dnode flags: USED_BYTES USERUSED_ACCOUNTED USEROBJUSED_ACCOUNTED dnode maxblkid: 0 (bonus encrypted) (object encrypted) Indirect blocks: 0 L0 0:c969600:10800 1a400L/8200P F=1 B=40/40 cksum=3243e353334e32fb:6d3c14cd8a3ce5d2:31f1bac457bf51ec:413b385e911ad9e2 segment [0000000000000000, 000000000001a400) size 105K ``` and now finally try to get the image data at 0:c969600:1a400 ``` bash-3.2# zdb -R -K $pass $disks -e test/movies 0:c969600:1a400 |head Found vdev type: raidz physmem = 2097152 pages (32.00 GB) 0:c969600:1a400L 0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef 000000: 5727568bba70a4aa 13faac8aec9d2530 W'V..p........%0 000010: 4e83318bd6144700 c1711eed6c4bff06 N.1...G..q..lK.. 000020: 73bc77792344f8c6 071924efb202fa5a s.wy#D....$....Z 000030: 48e5991759d8d5e9 5b647ef25695d182 H...Y...[d~.V... 000040: f87618227293d53f 958eb1815f8df46b .v."r..?...._..k ``` those do not contain the PNG header as on an unencrypted pool: ``` bash-3.2# zdb -R -p /Users/jolly/Desktop/zfsdebug2/sda -p /Users/jolly/Desktop/zfsdebug2/sdb -p /Users/jolly/Desktop/zfsdebug2/sdc -e test 0:4604200:10800 |head Found vdev type: raidz physmem = 2097152 pages (32.00 GB) 0:4604200:10800 0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef 000000: 0000803ff0698950 4e470d0a1a0a0000 ...?.i.PNG...... 000010: 000d494844520000 0780000004380802 ..IHDR.......8.. 000020: 00000067b1561400 00002c7445587443 ...g.V....,tEXtC 000030: 72656174696f6e20 54696d6500536174 reation Time.Sat ```
1.0
zdb -R shows incorrect data - ### System information Type | Version/Name --- | --- Distribution Name | ubuntu Distribution Version | 22.4 Kernel Version | ubuntu 5.15.0-1033-raspi Architecture | aarch64 aarch64 aarch64 GNU/Linux OpenZFS Version | zfs-2.2.99-1, zfs-kmod-2.1.5-1ubuntu6~22.04.1 ### Describe the problem you're observing ## Background: I have a 4disk raid z2 pool that had one disk failed. As I can't import the pool any longer ( crashes on import on macos, freebsd, spews out errors on linux ). Importing it with readonly works, but anything else but using zpool status will crash the machine (freebsd,macos) or hangs zfs (linux). So now I'm in the process of getting my data back with zdb when I encountered this bug. ``` root@ubuntu:/home/ubuntu# zpool status pool: ocean state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scan: resilver in progress since Sat Jul 1 10:01:24 2023 0B / 11.6T scanned, 0B / 11.6T issued 0B resilvered, 0.00% done, no estimated completion time config: NAME STATE READ WRITE CKSUM ocean DEGRADED 0 0 0 raidz2-0 DEGRADED 0 0 0 18276338438250889807 FAULTED 0 0 0 was /dev/sdd2 sdc1 ONLINE 0 0 0 sdb2 ONLINE 0 0 0 sdd2 ONLINE 0 0 0 errors: 3962 data errors, use '-v' for a list root@ubuntu:/home/ubuntu# zdb -u -e ocean Uberblock: magic = 0000000000bab10c version = 5000 txg = 18360509 guid_sum = 15765822683839940218 timestamp = 1688241822 UTC = Sat Jul 1 20:03:42 2023 mmp_magic = 00000000a11cea11 mmp_delay = 0 mmp_valid = 0 checkpoint_txg = 0 root@ubuntu:/home/ubuntu# zdb -dP -e ocean Dataset mos [META], ID 0, cr_txg 4, 1092430080, 6633 objects Dataset ocean/tmtinkerbell@2023-06-24T00:00:01 [ZPL], ID 5114, cr_txg 18302756, 1111802459136, 24930 objects ... Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 4065675030144, 9979 objects ... dmu_bonus_hold(os, object, dl, &dl->dl_dbuf) == 0 (0x34 == 0) ASSERT at module/zfs/dsl_deadlist.c:308:dsl_deadlist_open()Aborted (core dumped) root@ubuntu:/home/ubuntu# zdb -dd -e ocean/movies Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 3.70T, 9979 objects ZIL header: claim_txg 18360433, claim_blk_seq 1, claim_lr_seq 0 replay_seq 0, flags 0x2 Object lvl iblk dblk dsize dnsize lsize %full type 0 6 128K 16K 14.7M 512 96.9M 5.03 DMU dnode -1 1 128K 512 11.5K 512 512 100.00 ZFS user/group/project used -2 1 128K 512 11.5K 512 512 100.00 ZFS user/group/project used 1 1 128K 1K 11.5K 512 1K 100.00 ZFS master node 2 1 128K 512 11.5K 512 512 100.00 SA master node 3 1 128K 21.5K 11.5K 512 21.5K 100.00 ZFS delete queue 4 2 128K 16K 35K 512 32K 100.00 ZFS directory 5 1 128K 2K 11.5K 512 2K 100.00 SA attr registration ... ``` And looking at one file: ``` root@ubuntu:~# zdb -ddddd -K $pass -e ocean/movies 5268 Unlocked encryption root: ocean/movies Dataset ocean/movies [ZPL], ID 772, cr_txg 1430, 3.70T, 9979 objects, rootbp DVA[0]=<0:2400014a000:3000> DVA[1]=<0:f20081e4000:3000> [L0 DMU objset] fletcher4 uncompressed authenticated LE contiguous unique double size=800L/800P birth=18360433L/18360433P fill=9979 cksum=00000021c040a809:000025b93e7be9da:00190442ec3d9de4:0c1fdfcfea494199 Object lvl iblk dblk dsize dnsize lsize %full type 5268 1 128K 105K 40.5K 512 105K 100.00 ZFS plain file 168 bonus System attributes dnode flags: USED_BYTES USERUSED_ACCOUNTED dnode maxblkid: 0 path /Pictures/Testbilder/NecPattern.png uid 1000 gid 1000 atime Tue Dec 31 08:24:53 2019 mtime Thu Feb 11 09:05:10 2016 ctime Wed Jan 27 17:59:02 2021 crtime Tue Jan 26 20:40:27 2021 gen 4641518 mode 100644 size 107030 parent 5267 links 1 pflags 40800000004 Indirect blocks: 0 L0 0:10a1263d2000:15000 1a400L/9000P F=1 B=4641518/4641518 cksum=246684c5894cb36e:ecb92ef5e414d48a:13123f3d0f3e78df:9619bdcef80cd664 segment [0000000000000000, 000000000001a400) size 105K ``` ## The Issue Using **zdb -R** on an unencrypted pool works and gives me back the data, but on encrypted pools that does not work as shown here there is no PNG header as it should ( described in the how to replicate section) ``` root@ubuntu:/home/ubuntu# zdb -R -K $pass $disks -e ocean/movies 0:10a1263d2000:1a400 |head Found vdev type: raidz 0:10a1263d2000:1a400 0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef 000000: 9c042f5028cc72d6 1eb5378942ca849d .r.(P/.....B.7.. 000010: 39d46c5bb5ef339d 9aeb9cbf22549f9a .3..[l.9..T".... 000020: c4bcb91876bfb295 9b297424b1886cdd ...v.....l..$t). 000030: dab8c9b2e13c8f06 f9517b44e365816a ..<.....j.e.D{Q. 000040: f63da5f46b496aa5 d2a3744d20577b5a .jIk..=.Z{W Mt.. ... ``` ### Describe how to reproduce the problem Create a pool with an ecrypted pool inside and copy an image.png into it. ``` mkfile 100m sda sdb sdc sdd pass='testtest' disks='-p /Users/jolly/Desktop/zfsdebug2/sda -p /Users/jolly/Desktop/zfsdebug2/sdb -p /Users/jolly/Desktop/zfsdebug2/sdc -p /Users/jolly/Desktop/zfsdebug2/sdd' # zpool create test raidz2 $disks # create -o encryption=on -o keylocation=prompt -o keyformat=passphrase test/movies # cp Image.png /Volumes/test/movies/ ``` look for the image size in ``` # zdb -dd $disks -e test/movies physmem = 2097152 pages (32.00 GB) looking up movies in obj34 Dataset test/movies [ZPL], ID 388, cr_txg 32, 2.08M, 104 objects ZIL header: claim_txg 0, claim_blk_seq 0, claim_lr_seq 0 replay_seq 0, flags 0x0 Object lvl iblk dblk dsize dnsize lsize %full type 0 6 128K 16K 280K 512 288K 18.06 DMU dnode -1 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used -2 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used -3 1 128K 512 1.50K 512 512 100.00 ZFS user/group/project used 1 1 128K 1K 1.50K 512 1K 100.00 ZFS master node 2 2 128K 128K 1.66M 512 1.75M 100.00 ZFS plain file 3 1 128K 512 1.50K 512 512 100.00 ZFS directory 4 1 128K 512 512 512 512 100.00 ZFS plain file 5 1 128K 6.50K 4.50K 512 6.50K 100.00 ZFS directory 6 1 128K 512 0 512 512 0.00 ZFS plain file 7 1 128K 512 1.50K 512 512 100.00 ZFS directory 9 1 128K 105K 33K 512 105K 100.00 ZFS plain file 10 1 128K 512 1.50K 512 512 100.00 ZFS directory ... ``` I my case 105k Bytes - so it's object number 9 Then find out the blocks ``` # zdb -ddddd $disks -e test/movies 9 physmem = 2097152 pages (32.00 GB) looking up movies in obj34 Dataset test/movies [ZPL], ID 388, cr_txg 32, 2.08M, 104 objects, rootbp DVA[0]=<0:ce95800:2400> DVA[1]=<0:41f1400:2400> [L0 DMU objset] fletcher4 uncompressed authenticated LE contiguous unique double size=1000L/1000P birth=43L/43P fill=104 cksum=00000022e10e3971:000064b3a2f91165:009866b8a335c5af:9fae33cbc48176dc Object lvl iblk dblk dsize dnsize lsize %full type 9 1 128K 105K 33K 512 105K 100.00 ZFS plain file 176 bonus System attributes dnode flags: USED_BYTES USERUSED_ACCOUNTED USEROBJUSED_ACCOUNTED dnode maxblkid: 0 (bonus encrypted) (object encrypted) Indirect blocks: 0 L0 0:c969600:10800 1a400L/8200P F=1 B=40/40 cksum=3243e353334e32fb:6d3c14cd8a3ce5d2:31f1bac457bf51ec:413b385e911ad9e2 segment [0000000000000000, 000000000001a400) size 105K ``` and now finally try to get the image data at 0:c969600:1a400 ``` bash-3.2# zdb -R -K $pass $disks -e test/movies 0:c969600:1a400 |head Found vdev type: raidz physmem = 2097152 pages (32.00 GB) 0:c969600:1a400L 0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef 000000: 5727568bba70a4aa 13faac8aec9d2530 W'V..p........%0 000010: 4e83318bd6144700 c1711eed6c4bff06 N.1...G..q..lK.. 000020: 73bc77792344f8c6 071924efb202fa5a s.wy#D....$....Z 000030: 48e5991759d8d5e9 5b647ef25695d182 H...Y...[d~.V... 000040: f87618227293d53f 958eb1815f8df46b .v."r..?...._..k ``` those do not contain the PNG header as on an unencrypted pool: ``` bash-3.2# zdb -R -p /Users/jolly/Desktop/zfsdebug2/sda -p /Users/jolly/Desktop/zfsdebug2/sdb -p /Users/jolly/Desktop/zfsdebug2/sdc -e test 0:4604200:10800 |head Found vdev type: raidz physmem = 2097152 pages (32.00 GB) 0:4604200:10800 0 1 2 3 4 5 6 7 8 9 a b c d e f 0123456789abcdef 000000: 0000803ff0698950 4e470d0a1a0a0000 ...?.i.PNG...... 000010: 000d494844520000 0780000004380802 ..IHDR.......8.. 000020: 00000067b1561400 00002c7445587443 ...g.V....,tEXtC 000030: 72656174696f6e20 54696d6500536174 reation Time.Sat ```
defect
zdb r shows incorrect data system information type version name distribution name ubuntu distribution version kernel version ubuntu raspi architecture gnu linux openzfs version zfs zfs kmod describe the problem you re observing background i have a raid pool that had one disk failed as i can t import the pool any longer crashes on import on macos freebsd spews out errors on linux importing it with readonly works but anything else but using zpool status will crash the machine freebsd macos or hangs zfs linux so now i m in the process of getting my data back with zdb when i encountered this bug root ubuntu home ubuntu zpool status pool ocean state degraded status one or more devices is currently being resilvered the pool will continue to function possibly in a degraded state action wait for the resilver to complete scan resilver in progress since sat jul scanned issued resilvered done no estimated completion time config name state read write cksum ocean degraded degraded faulted was dev online online online errors data errors use v for a list root ubuntu home ubuntu zdb u e ocean uberblock magic version txg guid sum timestamp utc sat jul mmp magic mmp delay mmp valid checkpoint txg root ubuntu home ubuntu zdb dp e ocean dataset mos id cr txg objects dataset ocean tmtinkerbell id cr txg objects dataset ocean movies id cr txg objects dmu bonus hold os object dl dl dl dbuf assert at module zfs dsl deadlist c dsl deadlist open aborted core dumped root ubuntu home ubuntu zdb dd e ocean movies dataset ocean movies id cr txg objects zil header claim txg claim blk seq claim lr seq replay seq flags object lvl iblk dblk dsize dnsize lsize full type dmu dnode zfs user group project used zfs user group project used zfs master node sa master node zfs delete queue zfs directory sa attr registration and looking at one file root ubuntu zdb ddddd k pass e ocean movies unlocked encryption root ocean movies dataset ocean movies id cr txg objects rootbp dva dva uncompressed authenticated le contiguous unique double size birth fill cksum object lvl iblk dblk dsize dnsize lsize full type zfs plain file bonus system attributes dnode flags used bytes userused accounted dnode maxblkid path pictures testbilder necpattern png uid gid atime tue dec mtime thu feb ctime wed jan crtime tue jan gen mode size parent links pflags indirect blocks f b cksum segment size the issue using zdb r on an unencrypted pool works and gives me back the data but on encrypted pools that does not work as shown here there is no png header as it should described in the how to replicate section root ubuntu home ubuntu zdb r k pass disks e ocean movies head found vdev type raidz a b c d e f r p b l t v l t j e d q jik z w mt describe how to reproduce the problem create a pool with an ecrypted pool inside and copy an image png into it mkfile sda sdb sdc sdd pass testtest disks p users jolly desktop sda p users jolly desktop sdb p users jolly desktop sdc p users jolly desktop sdd zpool create test disks create o encryption on o keylocation prompt o keyformat passphrase test movies cp image png volumes test movies look for the image size in zdb dd disks e test movies physmem pages gb looking up movies in dataset test movies id cr txg objects zil header claim txg claim blk seq claim lr seq replay seq flags object lvl iblk dblk dsize dnsize lsize full type dmu dnode zfs user group project used zfs user group project used zfs user group project used zfs master node zfs plain file zfs directory zfs plain file zfs directory zfs plain file zfs directory zfs plain file zfs directory i my case bytes so it s object number then find out the blocks zdb ddddd disks e test movies physmem pages gb looking up movies in dataset test movies id cr txg objects rootbp dva dva uncompressed authenticated le contiguous unique double size birth fill cksum object lvl iblk dblk dsize dnsize lsize full type zfs plain file bonus system attributes dnode flags used bytes userused accounted userobjused accounted dnode maxblkid bonus encrypted object encrypted indirect blocks f b cksum segment size and now finally try to get the image data at bash zdb r k pass disks e test movies head found vdev type raidz physmem pages gb a b c d e f w v p n g q lk s wy d z h y d v v r k those do not contain the png header as on an unencrypted pool bash zdb r p users jolly desktop sda p users jolly desktop sdb p users jolly desktop sdc e test head found vdev type raidz physmem pages gb a b c d e f i png ihdr g v textc reation time sat
1
487,915
14,061,255,102
IssuesEvent
2020-11-03 07:48:55
mobigen/IRIS-BigData-Platform
https://api.github.com/repos/mobigen/IRIS-BigData-Platform
closed
[LH빅데이터 플랫폼] 알티베이스 연동 I/F 추가
#Discovery Priority: P0 R: 20.10.30 Status: Backlog
## 기능 요청 ## LH빅데이터 플랫폼 과제 수행을 위해 **알티베이스 연동 I/F 추가 필요** ## 원하는 솔루션 설명 ## IRIS Discovery 메뉴의 "연결정보" 설정 기능에 알티베이스 추가 ![image](https://user-images.githubusercontent.com/4444717/94379681-f301a880-016c-11eb-9f3a-8ec9fa862b60.png) IRIS Studio에서 DBMS 연결유형에 알티베이스 추가 ![image](https://user-images.githubusercontent.com/4444717/94379757-47a52380-016d-11eb-9967-6c9d07f01eb5.png) ## 고려한 다른 대안 ## ## 기타 ##
1.0
[LH빅데이터 플랫폼] 알티베이스 연동 I/F 추가 - ## 기능 요청 ## LH빅데이터 플랫폼 과제 수행을 위해 **알티베이스 연동 I/F 추가 필요** ## 원하는 솔루션 설명 ## IRIS Discovery 메뉴의 "연결정보" 설정 기능에 알티베이스 추가 ![image](https://user-images.githubusercontent.com/4444717/94379681-f301a880-016c-11eb-9f3a-8ec9fa862b60.png) IRIS Studio에서 DBMS 연결유형에 알티베이스 추가 ![image](https://user-images.githubusercontent.com/4444717/94379757-47a52380-016d-11eb-9967-6c9d07f01eb5.png) ## 고려한 다른 대안 ## ## 기타 ##
non_defect
알티베이스 연동 i f 추가 기능 요청 lh빅데이터 플랫폼 과제 수행을 위해 알티베이스 연동 i f 추가 필요 원하는 솔루션 설명 iris discovery 메뉴의 연결정보 설정 기능에 알티베이스 추가 iris studio에서 dbms 연결유형에 알티베이스 추가 고려한 다른 대안 기타
0
39,051
9,186,855,205
IssuesEvent
2019-03-06 00:23:03
vmware/vic
https://api.github.com/repos/vmware/vic
closed
share templates across VCHs
area/ux component/install component/isos component/portlayer/execution kind/defect/performance severity/3-moderate
When using vmfork enabled VCHs as a tenancy mechanism in large numbers it's necessary to share the vmfork templates between those VCHs using the same bootstrap versions. This allows amortization of the template overhead across all containerVMs and not just those within a single VCH. This is a core scenario for service provider usage and for on-prem dev/test. This requires a mechanism for discovering templates and performing reference counting of some kind. As an initial implementation it's acceptable to have the location of templates specified by the admin, however we should still manage template versions within that location or the upgrade paths become problematic.
1.0
share templates across VCHs - When using vmfork enabled VCHs as a tenancy mechanism in large numbers it's necessary to share the vmfork templates between those VCHs using the same bootstrap versions. This allows amortization of the template overhead across all containerVMs and not just those within a single VCH. This is a core scenario for service provider usage and for on-prem dev/test. This requires a mechanism for discovering templates and performing reference counting of some kind. As an initial implementation it's acceptable to have the location of templates specified by the admin, however we should still manage template versions within that location or the upgrade paths become problematic.
defect
share templates across vchs when using vmfork enabled vchs as a tenancy mechanism in large numbers it s necessary to share the vmfork templates between those vchs using the same bootstrap versions this allows amortization of the template overhead across all containervms and not just those within a single vch this is a core scenario for service provider usage and for on prem dev test this requires a mechanism for discovering templates and performing reference counting of some kind as an initial implementation it s acceptable to have the location of templates specified by the admin however we should still manage template versions within that location or the upgrade paths become problematic
1
40,429
9,992,287,876
IssuesEvent
2019-07-11 13:09:35
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Review correctness of jOOQ's CURRENT_TIME, CURRENT_TIMESTAMP functions
C: Functionality E: All Editions P: Medium T: Defect T: Incompatible change
The SQL standard specifies that - `CURRENT_TIME` produces a `TIME WITH TIME ZONE` value (Java `OffsetTime`) - `CURRENT_TIMESTAMP` produces a `TIMESTAMP WITH TIME ZONE` value (Java `OffsetDateTime`) - `LOCALTIME` produces a `TIME WITHOUT TIME ZONE` value (Java `LocalTime` or `java.sql.Time`) - `LOCALTIMESTAMP` produces a `TIMESTAMP WITHOUT TIME ZONE` value (Java `LocalDateTime` or `java.sql.Timestamp`) This is not what the current `currentTimestamp()` and other functions suggest. Historically, as jOOQ hasn't always supported `WITH TIME ZONE` types, all of these functions map to the `WITHOUT TIME ZONE` counterparts. This may produce different, inconsistent results depending on the dialect. It is difficult to see a solution here that doesn't break backwards compatibility of behaviour.
1.0
Review correctness of jOOQ's CURRENT_TIME, CURRENT_TIMESTAMP functions - The SQL standard specifies that - `CURRENT_TIME` produces a `TIME WITH TIME ZONE` value (Java `OffsetTime`) - `CURRENT_TIMESTAMP` produces a `TIMESTAMP WITH TIME ZONE` value (Java `OffsetDateTime`) - `LOCALTIME` produces a `TIME WITHOUT TIME ZONE` value (Java `LocalTime` or `java.sql.Time`) - `LOCALTIMESTAMP` produces a `TIMESTAMP WITHOUT TIME ZONE` value (Java `LocalDateTime` or `java.sql.Timestamp`) This is not what the current `currentTimestamp()` and other functions suggest. Historically, as jOOQ hasn't always supported `WITH TIME ZONE` types, all of these functions map to the `WITHOUT TIME ZONE` counterparts. This may produce different, inconsistent results depending on the dialect. It is difficult to see a solution here that doesn't break backwards compatibility of behaviour.
defect
review correctness of jooq s current time current timestamp functions the sql standard specifies that current time produces a time with time zone value java offsettime current timestamp produces a timestamp with time zone value java offsetdatetime localtime produces a time without time zone value java localtime or java sql time localtimestamp produces a timestamp without time zone value java localdatetime or java sql timestamp this is not what the current currenttimestamp and other functions suggest historically as jooq hasn t always supported with time zone types all of these functions map to the without time zone counterparts this may produce different inconsistent results depending on the dialect it is difficult to see a solution here that doesn t break backwards compatibility of behaviour
1
28,190
5,217,162,160
IssuesEvent
2017-01-26 13:00:11
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
Offset when opening up menu
defect
**I'm submitting a ...** (check one with "x") ``` [x] bug report => Search github for a similar issue or PR before submitting [ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap [ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35 ``` **Plunkr Case (Bug Reports)** Please fork the plunkr below and create a case demonstrating your bug report. Issues without a plunkr have much less possibility to be reviewed. http://plnkr.co/edit/NtWWnN **Current behavior** if an anchor tag contains another element, the menu attaches to that element **Expected behavior** the menu should always attach to the anchor tag **Minimal reproduction of the problem with instructions** <!-- If the current behavior is a bug or you can illustrate your feature request better with an example, please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via https://plnkr.co or similar (you can use this template as a starting point: http://plnkr.co/edit/tpl:AvJOMERrnz94ekVua0u5). --> **What is the motivation / use case for changing the behavior?** <!-- Describe the motivation or the concrete use case --> **Please tell us about your environment:** <!-- Operating system, IDE, package manager, HTTP server, ... --> * **Angular version:** 2.0.X <!-- Check whether this is still an issue in the most recent Angular version --> * **PrimeNG version:** 2.0.X <!-- Check whether this is still an issue in the most recent Angular version --> * **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] <!-- All browsers where this could be reproduced --> * **Language:** [all | TypeScript X.X | ES6/7 | ES5] * **Node (for AoT issues):** `node --version` =
1.0
Offset when opening up menu - **I'm submitting a ...** (check one with "x") ``` [x] bug report => Search github for a similar issue or PR before submitting [ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap [ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35 ``` **Plunkr Case (Bug Reports)** Please fork the plunkr below and create a case demonstrating your bug report. Issues without a plunkr have much less possibility to be reviewed. http://plnkr.co/edit/NtWWnN **Current behavior** if an anchor tag contains another element, the menu attaches to that element **Expected behavior** the menu should always attach to the anchor tag **Minimal reproduction of the problem with instructions** <!-- If the current behavior is a bug or you can illustrate your feature request better with an example, please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via https://plnkr.co or similar (you can use this template as a starting point: http://plnkr.co/edit/tpl:AvJOMERrnz94ekVua0u5). --> **What is the motivation / use case for changing the behavior?** <!-- Describe the motivation or the concrete use case --> **Please tell us about your environment:** <!-- Operating system, IDE, package manager, HTTP server, ... --> * **Angular version:** 2.0.X <!-- Check whether this is still an issue in the most recent Angular version --> * **PrimeNG version:** 2.0.X <!-- Check whether this is still an issue in the most recent Angular version --> * **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] <!-- All browsers where this could be reproduced --> * **Language:** [all | TypeScript X.X | ES6/7 | ES5] * **Node (for AoT issues):** `node --version` =
defect
offset when opening up menu i m submitting a check one with x bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see plunkr case bug reports please fork the plunkr below and create a case demonstrating your bug report issues without a plunkr have much less possibility to be reviewed current behavior if an anchor tag contains another element the menu attaches to that element expected behavior the menu should always attach to the anchor tag minimal reproduction of the problem with instructions if the current behavior is a bug or you can illustrate your feature request better with an example please provide the steps to reproduce and if possible a minimal demo of the problem via or similar you can use this template as a starting point what is the motivation use case for changing the behavior please tell us about your environment angular version x primeng version x browser language node for aot issues node version
1
41,920
10,707,853,738
IssuesEvent
2019-10-24 18:23:02
STEllAR-GROUP/phylanx
https://api.github.com/repos/STEllAR-GROUP/phylanx
closed
`fold_left`, `fold_right` and `fmap` do not work with a lazy function
category: @Phylanx type: compatibility issue type: defect
Having the following code: ```py import numpy as np from phylanx import Phylanx, PhylanxSession, execution_tree PhylanxSession.init(1) def variable(value, dtype=None, name=None, constraint=None): if dtype is None: dtype = "float32" if constraint is not None: raise TypeError("Constraint is the projection function to be " "applied to the variable after an optimizer update") from phylanx.ast.physl import PhySL if isinstance(value, PhySL.eval_wrapper): return execution_tree.variable(value.code(), dtype) if isinstance(value, execution_tree.variable): return value return execution_tree.variable(value, dtype=dtype, name=name) def eval(func): return func.eval() @Phylanx def map_fn_eager(fn, elems, dtype): return fmap(fn, elems) def map_fn(fn, elems, name=None, dtype=None): return map_fn_eager.lazy(fn, elems, dtype) @Phylanx def sum_eager(x, axis=None, keepdims=False): return np.sum(x, axis, keepdims) def sum(x, axis=None, keepdims=False): return sum_eager.lazy(x, axis, keepdims) def test_map(): x = [1,2,3] vx = variable(x) kx = eval(map_fn(sum, vx)) return kx test_map() ``` results in: ```pytb Traceback (most recent call last): File "test51.py", line 63, in <module> print(test_map()) File "test51.py", line 47, in test_map kx = eval(foldl(sum, vx)) File "test51.py", line 41, in foldl return foldl_eager.lazy(fn, elems, initializer) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\transducer.py", line 123, in lazy return self.backend.lazy(map(self.map_decorated, args)) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 526, in lazy return self.eval_wrapper(self, tuple(map(self.map_wrapped, args))) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\transducer.py", line 111, in map_decorated fn_physl = PhySL(val, fn_ast, {}) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 387, in __init__ self.ir = self.apply_rule(tree.body[0]) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule return eval('self._%s' % node_name)(node) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 951, in _FunctionDef body = self.block(node.body) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 440, in block block = tuple(map(self.apply_rule, node)) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule return eval('self._%s' % node_name)(node) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 1144, in _Return value = self.apply_rule(node.value) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule return eval('self._%s' % node_name)(node) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 693, in _Call symbol = self.apply_rule(node.func) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule return eval('self._%s' % node_name)(node) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 654, in _Attribute 'Cannot transform: %s' % attr) NotImplementedError: Phylanx does not support non-NumPy member functions.Cannot transform: sum_eager.lazy ```
1.0
`fold_left`, `fold_right` and `fmap` do not work with a lazy function - Having the following code: ```py import numpy as np from phylanx import Phylanx, PhylanxSession, execution_tree PhylanxSession.init(1) def variable(value, dtype=None, name=None, constraint=None): if dtype is None: dtype = "float32" if constraint is not None: raise TypeError("Constraint is the projection function to be " "applied to the variable after an optimizer update") from phylanx.ast.physl import PhySL if isinstance(value, PhySL.eval_wrapper): return execution_tree.variable(value.code(), dtype) if isinstance(value, execution_tree.variable): return value return execution_tree.variable(value, dtype=dtype, name=name) def eval(func): return func.eval() @Phylanx def map_fn_eager(fn, elems, dtype): return fmap(fn, elems) def map_fn(fn, elems, name=None, dtype=None): return map_fn_eager.lazy(fn, elems, dtype) @Phylanx def sum_eager(x, axis=None, keepdims=False): return np.sum(x, axis, keepdims) def sum(x, axis=None, keepdims=False): return sum_eager.lazy(x, axis, keepdims) def test_map(): x = [1,2,3] vx = variable(x) kx = eval(map_fn(sum, vx)) return kx test_map() ``` results in: ```pytb Traceback (most recent call last): File "test51.py", line 63, in <module> print(test_map()) File "test51.py", line 47, in test_map kx = eval(foldl(sum, vx)) File "test51.py", line 41, in foldl return foldl_eager.lazy(fn, elems, initializer) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\transducer.py", line 123, in lazy return self.backend.lazy(map(self.map_decorated, args)) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 526, in lazy return self.eval_wrapper(self, tuple(map(self.map_wrapped, args))) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\transducer.py", line 111, in map_decorated fn_physl = PhySL(val, fn_ast, {}) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 387, in __init__ self.ir = self.apply_rule(tree.body[0]) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule return eval('self._%s' % node_name)(node) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 951, in _FunctionDef body = self.block(node.body) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 440, in block block = tuple(map(self.apply_rule, node)) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule return eval('self._%s' % node_name)(node) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 1144, in _Return value = self.apply_rule(node.value) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule return eval('self._%s' % node_name)(node) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 693, in _Call symbol = self.apply_rule(node.func) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 434, in apply_rule return eval('self._%s' % node_name)(node) File "C:\Repos\phylanx\cmake-build-debug\python\build\lib.win-amd64-3.6\phylanx\ast\physl.py", line 654, in _Attribute 'Cannot transform: %s' % attr) NotImplementedError: Phylanx does not support non-NumPy member functions.Cannot transform: sum_eager.lazy ```
defect
fold left fold right and fmap do not work with a lazy function having the following code py import numpy as np from phylanx import phylanx phylanxsession execution tree phylanxsession init def variable value dtype none name none constraint none if dtype is none dtype if constraint is not none raise typeerror constraint is the projection function to be applied to the variable after an optimizer update from phylanx ast physl import physl if isinstance value physl eval wrapper return execution tree variable value code dtype if isinstance value execution tree variable return value return execution tree variable value dtype dtype name name def eval func return func eval phylanx def map fn eager fn elems dtype return fmap fn elems def map fn fn elems name none dtype none return map fn eager lazy fn elems dtype phylanx def sum eager x axis none keepdims false return np sum x axis keepdims def sum x axis none keepdims false return sum eager lazy x axis keepdims def test map x vx variable x kx eval map fn sum vx return kx test map results in pytb traceback most recent call last file py line in print test map file py line in test map kx eval foldl sum vx file py line in foldl return foldl eager lazy fn elems initializer file c repos phylanx cmake build debug python build lib win phylanx ast transducer py line in lazy return self backend lazy map self map decorated args file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in lazy return self eval wrapper self tuple map self map wrapped args file c repos phylanx cmake build debug python build lib win phylanx ast transducer py line in map decorated fn physl physl val fn ast file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in init self ir self apply rule tree body file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in apply rule return eval self s node name node file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in functiondef body self block node body file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in block block tuple map self apply rule node file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in apply rule return eval self s node name node file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in return value self apply rule node value file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in apply rule return eval self s node name node file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in call symbol self apply rule node func file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in apply rule return eval self s node name node file c repos phylanx cmake build debug python build lib win phylanx ast physl py line in attribute cannot transform s attr notimplementederror phylanx does not support non numpy member functions cannot transform sum eager lazy
1
24,697
4,074,930,861
IssuesEvent
2016-05-28 20:25:14
Nuand/bladeRF
https://api.github.com/repos/Nuand/bladeRF
closed
[libbladeRF] Clarify required SMB clock port config prior to trigger API usage
documentation defect
[This documentation](https://nuand.com/libbladeRF-doc/v1.7.0/group___f_n___t_r_i_g.html) does not explicitly note the requirement to have the SMB clock configuration completed prior to using these trigger features. This is required and should be explicitly be noted as a prerequisite.
1.0
[libbladeRF] Clarify required SMB clock port config prior to trigger API usage - [This documentation](https://nuand.com/libbladeRF-doc/v1.7.0/group___f_n___t_r_i_g.html) does not explicitly note the requirement to have the SMB clock configuration completed prior to using these trigger features. This is required and should be explicitly be noted as a prerequisite.
defect
clarify required smb clock port config prior to trigger api usage does not explicitly note the requirement to have the smb clock configuration completed prior to using these trigger features this is required and should be explicitly be noted as a prerequisite
1
292,021
21,946,706,960
IssuesEvent
2022-05-24 01:56:29
AndyBoyee604/Taskinator
https://api.github.com/repos/AndyBoyee604/Taskinator
opened
Initial Setup
documentation
## Requirements * Create the task tracking HTML page that needs a: * Header * Main Content area for the task list * footer * Use the style sheet provided * Add functionality to the button to add tasks to the list
1.0
Initial Setup - ## Requirements * Create the task tracking HTML page that needs a: * Header * Main Content area for the task list * footer * Use the style sheet provided * Add functionality to the button to add tasks to the list
non_defect
initial setup requirements create the task tracking html page that needs a header main content area for the task list footer use the style sheet provided add functionality to the button to add tasks to the list
0
269,648
8,441,481,436
IssuesEvent
2018-10-18 10:20:47
fac-14/sleuth-coop
https://api.github.com/repos/fac-14/sleuth-coop
opened
Password validation consistency
T25m bug priority-1
The validation on the password for the sign up form and the login form is different. You can use a 3 character password on sign up but your password has to be min 4 characters on the log in!
1.0
Password validation consistency - The validation on the password for the sign up form and the login form is different. You can use a 3 character password on sign up but your password has to be min 4 characters on the log in!
non_defect
password validation consistency the validation on the password for the sign up form and the login form is different you can use a character password on sign up but your password has to be min characters on the log in
0
77,906
27,229,151,217
IssuesEvent
2023-02-21 11:56:28
scoutplan/scoutplan
https://api.github.com/repos/scoutplan/scoutplan
closed
[Scoutplan Production/production] Net::OpenTimeout: Timeout to open TCP connection to smtp.mailgun.org:587 (exceeds 5 seconds)
defect
## Backtrace [View full backtrace and more info at honeybadger.io](https://app.honeybadger.io/projects/97676/faults/93912390)
1.0
[Scoutplan Production/production] Net::OpenTimeout: Timeout to open TCP connection to smtp.mailgun.org:587 (exceeds 5 seconds) - ## Backtrace [View full backtrace and more info at honeybadger.io](https://app.honeybadger.io/projects/97676/faults/93912390)
defect
net opentimeout timeout to open tcp connection to smtp mailgun org exceeds seconds backtrace
1
57,385
15,761,689,092
IssuesEvent
2021-03-31 10:15:01
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
[COGNITION]: Actionable links should have visual cues
508-defect-2 508/Accessibility direct deposit vsa-authenticated-exp
## Description [COGNITION]: Actionable links should have visual cues <img width="788" alt="direct-deposit-edit-links-issue" src="https://user-images.githubusercontent.com/57469/68237941-2cba4580-ffd6-11e9-823c-5627a2f09ed2.png"> Actionable elements must be identified visually, by convention, and by information provided to assistive technologies. The Edit links have a color difference from the text, but the visual cue is not sufficiently clear that it is actionable. ## Point of Contact Jennifer ## Acceptance Criteria As a user of VA.gov, I want to open the Direct Deposit information update functionality. These action links should be clearly visible as actionable links. ## Environment * Operating System: all * Browser: all ## Steps to Recreate 1. Log into staging.va.gov with user 1 2. Click on the person menu in the top right corner 3. Click "Profile" from that menu. Your profile view will load. 4. Click the "Direct deposit information" link to jump to the direct deposit section 5. See any of the Edit links next to Bank name, Account number, or Account type to open the DD modal. They are currently coded as buttons, but styled as links, and need to follow the link styling from [https://design.va.gov/design/typography#links](https://design.va.gov/design/typography#links).
1.0
[COGNITION]: Actionable links should have visual cues - ## Description [COGNITION]: Actionable links should have visual cues <img width="788" alt="direct-deposit-edit-links-issue" src="https://user-images.githubusercontent.com/57469/68237941-2cba4580-ffd6-11e9-823c-5627a2f09ed2.png"> Actionable elements must be identified visually, by convention, and by information provided to assistive technologies. The Edit links have a color difference from the text, but the visual cue is not sufficiently clear that it is actionable. ## Point of Contact Jennifer ## Acceptance Criteria As a user of VA.gov, I want to open the Direct Deposit information update functionality. These action links should be clearly visible as actionable links. ## Environment * Operating System: all * Browser: all ## Steps to Recreate 1. Log into staging.va.gov with user 1 2. Click on the person menu in the top right corner 3. Click "Profile" from that menu. Your profile view will load. 4. Click the "Direct deposit information" link to jump to the direct deposit section 5. See any of the Edit links next to Bank name, Account number, or Account type to open the DD modal. They are currently coded as buttons, but styled as links, and need to follow the link styling from [https://design.va.gov/design/typography#links](https://design.va.gov/design/typography#links).
defect
actionable links should have visual cues description actionable links should have visual cues img width alt direct deposit edit links issue src actionable elements must be identified visually by convention and by information provided to assistive technologies the edit links have a color difference from the text but the visual cue is not sufficiently clear that it is actionable point of contact jennifer acceptance criteria as a user of va gov i want to open the direct deposit information update functionality these action links should be clearly visible as actionable links environment operating system all browser all steps to recreate log into staging va gov with user click on the person menu in the top right corner click profile from that menu your profile view will load click the direct deposit information link to jump to the direct deposit section see any of the edit links next to bank name account number or account type to open the dd modal they are currently coded as buttons but styled as links and need to follow the link styling from
1
63,039
17,355,452,752
IssuesEvent
2021-07-29 13:56:25
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
NPE from SQL
Type: Defect
Using 5.0-BETA-1, the following code throws a NullPointerException ``` public static void main(String[] args) throws Exception { Config config = new Config(); HazelcastInstance hazelcastInstance = Hazelcast.newHazelcastInstance(config); String sql = "CREATE OR REPLACE MAPPING \"neil\" " + "TYPE IMap " + " OPTIONS ( " + " 'keyFormat' = 'java'," + " 'keyJavaClass' = '" + Integer.class.getCanonicalName() + "'," + " 'valueFormat' = 'java'," + " 'valueJavaClass' = '" + Integer.class.getCanonicalName() + "'" + " )"; hazelcastInstance.getSql().execute(sql); hazelcastInstance.shutdown(); } ``` classpath is ``` [INFO] +- com.hazelcast:hazelcast:jar:5.0-BETA-1:compile [INFO] \- com.hazelcast:hazelcast-sql:jar:5.0-BETA-1:compile ``` With 5.0-BETA-1 I get ``` Exception in thread "main" com.hazelcast.sql.HazelcastSqlException at com.hazelcast.sql.impl.QueryUtils.toPublicException(QueryUtils.java:79) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:233) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:199) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:195) at com.hazelcast.sql.SqlService.execute(SqlService.java:146) at neil.Application.main(Application.java:20) Caused by: com.hazelcast.sql.impl.QueryException at com.hazelcast.sql.impl.QueryException.error(QueryException.java:60) at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:86) at com.hazelcast.sql.impl.calcite.OptimizerContext.parse(OptimizerContext.java:136) at com.hazelcast.sql.impl.calcite.CalciteSqlOptimizer.prepare(CalciteSqlOptimizer.java:127) at com.hazelcast.sql.impl.SqlServiceImpl.prepare(SqlServiceImpl.java:287) at com.hazelcast.sql.impl.SqlServiceImpl.query0(SqlServiceImpl.java:268) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:220) ... 4 more Caused by: java.lang.NullPointerException at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:91) at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:76) ... 9 more ``` 5.0-SNAPSHOT (`20210729 - 12d65bd`) gives a different error ``` Exception in thread "main" com.hazelcast.sql.HazelcastSqlException: Encountered "CREATE" at line 1, column 1. at com.hazelcast.sql.impl.QueryUtils.toPublicException(QueryUtils.java:79) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:238) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:200) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:196) at com.hazelcast.sql.SqlService.execute(SqlService.java:146) at neil.Application.main(Application.java:20) Caused by: com.hazelcast.sql.impl.QueryException: Encountered "CREATE" at line 1, column 1. at com.hazelcast.sql.impl.QueryException.error(QueryException.java:60) at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:95) at com.hazelcast.sql.impl.calcite.OptimizerContext.parse(OptimizerContext.java:140) at com.hazelcast.sql.impl.calcite.CalciteSqlOptimizer.prepare(CalciteSqlOptimizer.java:127) at com.hazelcast.sql.impl.SqlServiceImpl.prepare(SqlServiceImpl.java:292) at com.hazelcast.sql.impl.SqlServiceImpl.query0(SqlServiceImpl.java:273) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:225) ... 4 more Caused by: com.hazelcast.org.apache.calcite.sql.parser.SqlParseException: Encountered "CREATE" at line 1, column 1. Was expecting one of: "ABS" ... "ALTER" ... "ARRAY" ... "AVG" ... "CALL" ... "CARDINALITY" ... "CASE" ... "CAST" ... "CEIL" ... "CEILING" ... "CHAR_LENGTH" ... "CHARACTER_LENGTH" ... "CLASSIFIER" ... "COALESCE" ... "COLLECT" ... "CONVERT" ... "COUNT" ... "COVAR_POP" ... "COVAR_SAMP" ... "CUME_DIST" ... ``` `CREATE` isn't listed, so not sure if this is a true defect or due to beta changes
1.0
NPE from SQL - Using 5.0-BETA-1, the following code throws a NullPointerException ``` public static void main(String[] args) throws Exception { Config config = new Config(); HazelcastInstance hazelcastInstance = Hazelcast.newHazelcastInstance(config); String sql = "CREATE OR REPLACE MAPPING \"neil\" " + "TYPE IMap " + " OPTIONS ( " + " 'keyFormat' = 'java'," + " 'keyJavaClass' = '" + Integer.class.getCanonicalName() + "'," + " 'valueFormat' = 'java'," + " 'valueJavaClass' = '" + Integer.class.getCanonicalName() + "'" + " )"; hazelcastInstance.getSql().execute(sql); hazelcastInstance.shutdown(); } ``` classpath is ``` [INFO] +- com.hazelcast:hazelcast:jar:5.0-BETA-1:compile [INFO] \- com.hazelcast:hazelcast-sql:jar:5.0-BETA-1:compile ``` With 5.0-BETA-1 I get ``` Exception in thread "main" com.hazelcast.sql.HazelcastSqlException at com.hazelcast.sql.impl.QueryUtils.toPublicException(QueryUtils.java:79) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:233) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:199) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:195) at com.hazelcast.sql.SqlService.execute(SqlService.java:146) at neil.Application.main(Application.java:20) Caused by: com.hazelcast.sql.impl.QueryException at com.hazelcast.sql.impl.QueryException.error(QueryException.java:60) at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:86) at com.hazelcast.sql.impl.calcite.OptimizerContext.parse(OptimizerContext.java:136) at com.hazelcast.sql.impl.calcite.CalciteSqlOptimizer.prepare(CalciteSqlOptimizer.java:127) at com.hazelcast.sql.impl.SqlServiceImpl.prepare(SqlServiceImpl.java:287) at com.hazelcast.sql.impl.SqlServiceImpl.query0(SqlServiceImpl.java:268) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:220) ... 4 more Caused by: java.lang.NullPointerException at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:91) at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:76) ... 9 more ``` 5.0-SNAPSHOT (`20210729 - 12d65bd`) gives a different error ``` Exception in thread "main" com.hazelcast.sql.HazelcastSqlException: Encountered "CREATE" at line 1, column 1. at com.hazelcast.sql.impl.QueryUtils.toPublicException(QueryUtils.java:79) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:238) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:200) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:196) at com.hazelcast.sql.SqlService.execute(SqlService.java:146) at neil.Application.main(Application.java:20) Caused by: com.hazelcast.sql.impl.QueryException: Encountered "CREATE" at line 1, column 1. at com.hazelcast.sql.impl.QueryException.error(QueryException.java:60) at com.hazelcast.sql.impl.calcite.parse.QueryParser.parse(QueryParser.java:95) at com.hazelcast.sql.impl.calcite.OptimizerContext.parse(OptimizerContext.java:140) at com.hazelcast.sql.impl.calcite.CalciteSqlOptimizer.prepare(CalciteSqlOptimizer.java:127) at com.hazelcast.sql.impl.SqlServiceImpl.prepare(SqlServiceImpl.java:292) at com.hazelcast.sql.impl.SqlServiceImpl.query0(SqlServiceImpl.java:273) at com.hazelcast.sql.impl.SqlServiceImpl.execute(SqlServiceImpl.java:225) ... 4 more Caused by: com.hazelcast.org.apache.calcite.sql.parser.SqlParseException: Encountered "CREATE" at line 1, column 1. Was expecting one of: "ABS" ... "ALTER" ... "ARRAY" ... "AVG" ... "CALL" ... "CARDINALITY" ... "CASE" ... "CAST" ... "CEIL" ... "CEILING" ... "CHAR_LENGTH" ... "CHARACTER_LENGTH" ... "CLASSIFIER" ... "COALESCE" ... "COLLECT" ... "CONVERT" ... "COUNT" ... "COVAR_POP" ... "COVAR_SAMP" ... "CUME_DIST" ... ``` `CREATE` isn't listed, so not sure if this is a true defect or due to beta changes
defect
npe from sql using beta the following code throws a nullpointerexception public static void main string args throws exception config config new config hazelcastinstance hazelcastinstance hazelcast newhazelcastinstance config string sql create or replace mapping neil type imap options keyformat java keyjavaclass integer class getcanonicalname valueformat java valuejavaclass integer class getcanonicalname hazelcastinstance getsql execute sql hazelcastinstance shutdown classpath is com hazelcast hazelcast jar beta compile com hazelcast hazelcast sql jar beta compile with beta i get exception in thread main com hazelcast sql hazelcastsqlexception at com hazelcast sql impl queryutils topublicexception queryutils java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql sqlservice execute sqlservice java at neil application main application java caused by com hazelcast sql impl queryexception at com hazelcast sql impl queryexception error queryexception java at com hazelcast sql impl calcite parse queryparser parse queryparser java at com hazelcast sql impl calcite optimizercontext parse optimizercontext java at com hazelcast sql impl calcite calcitesqloptimizer prepare calcitesqloptimizer java at com hazelcast sql impl sqlserviceimpl prepare sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java more caused by java lang nullpointerexception at com hazelcast sql impl calcite parse queryparser parse queryparser java at com hazelcast sql impl calcite parse queryparser parse queryparser java more snapshot gives a different error exception in thread main com hazelcast sql hazelcastsqlexception encountered create at line column at com hazelcast sql impl queryutils topublicexception queryutils java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java at com hazelcast sql sqlservice execute sqlservice java at neil application main application java caused by com hazelcast sql impl queryexception encountered create at line column at com hazelcast sql impl queryexception error queryexception java at com hazelcast sql impl calcite parse queryparser parse queryparser java at com hazelcast sql impl calcite optimizercontext parse optimizercontext java at com hazelcast sql impl calcite calcitesqloptimizer prepare calcitesqloptimizer java at com hazelcast sql impl sqlserviceimpl prepare sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl sqlserviceimpl java at com hazelcast sql impl sqlserviceimpl execute sqlserviceimpl java more caused by com hazelcast org apache calcite sql parser sqlparseexception encountered create at line column was expecting one of abs alter array avg call cardinality case cast ceil ceiling char length character length classifier coalesce collect convert count covar pop covar samp cume dist create isn t listed so not sure if this is a true defect or due to beta changes
1
21,192
16,616,407,692
IssuesEvent
2021-06-02 17:15:46
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
opened
About access restricting to system database . It's kinda cumbersome.
usability
```sql create user foo identified by '123'; revoke all on *.* from foo; create role RO; CREATE ROW POLICY ro_query_log_filter ON system.query_log USING 1 AS RESTRICTIVE TO RO; CREATE ROW POLICY ro_part_log_filter ON system.part_log USING 1 AS RESTRICTIVE TO RO; CREATE ROW POLICY ro_trace_log_filter ON system.trace_log USING 1 AS RESTRICTIVE TO RO; CREATE ROW POLICY ro_processes_filter ON system.processes USING 1 AS RESTRICTIVE TO RO; grant RO to foo; ALTER USER foo DEFAULT ROLE RO SETTINGS NONE; clickhouse-client -u foo --password=123 select count() from system.trace_log; 0 rows in set. select count() from system.query_log; 0 rows in set. select count() from system.tables; 0 rows in set. select query from system.processes; 0 rows in set ``` Also this is a bug: ``` use system; DB::Exception: foo: Not enough privileges. ``` Granting access to own rows in query_log/trace_log is even more cumbersome: ``` CREATE POLICY ${user_name}_query_log_filter ON system.query_log FOR SELECT USING initial_user='${user_name}' TO ${user_name}; ```
True
About access restricting to system database . It's kinda cumbersome. - ```sql create user foo identified by '123'; revoke all on *.* from foo; create role RO; CREATE ROW POLICY ro_query_log_filter ON system.query_log USING 1 AS RESTRICTIVE TO RO; CREATE ROW POLICY ro_part_log_filter ON system.part_log USING 1 AS RESTRICTIVE TO RO; CREATE ROW POLICY ro_trace_log_filter ON system.trace_log USING 1 AS RESTRICTIVE TO RO; CREATE ROW POLICY ro_processes_filter ON system.processes USING 1 AS RESTRICTIVE TO RO; grant RO to foo; ALTER USER foo DEFAULT ROLE RO SETTINGS NONE; clickhouse-client -u foo --password=123 select count() from system.trace_log; 0 rows in set. select count() from system.query_log; 0 rows in set. select count() from system.tables; 0 rows in set. select query from system.processes; 0 rows in set ``` Also this is a bug: ``` use system; DB::Exception: foo: Not enough privileges. ``` Granting access to own rows in query_log/trace_log is even more cumbersome: ``` CREATE POLICY ${user_name}_query_log_filter ON system.query_log FOR SELECT USING initial_user='${user_name}' TO ${user_name}; ```
non_defect
about access restricting to system database it s kinda cumbersome sql create user foo identified by revoke all on from foo create role ro create row policy ro query log filter on system query log using as restrictive to ro create row policy ro part log filter on system part log using as restrictive to ro create row policy ro trace log filter on system trace log using as restrictive to ro create row policy ro processes filter on system processes using as restrictive to ro grant ro to foo alter user foo default role ro settings none clickhouse client u foo password select count from system trace log rows in set select count from system query log rows in set select count from system tables rows in set select query from system processes rows in set also this is a bug use system db exception foo not enough privileges granting access to own rows in query log trace log is even more cumbersome create policy user name query log filter on system query log for select using initial user user name to user name
0
232,613
17,788,731,710
IssuesEvent
2021-08-31 14:02:08
pythonarcade/arcade
https://api.github.com/repos/pythonarcade/arcade
closed
[DOC] Example for UI event register
enhancement documentation
## Documentation request: Hello! I'm a newcomer from Qt UI designing. In PyQt5 or PySide2, there's a "signal & slot" mechanism in all event handler, which is relative to `EventDispatcher` in Pyglet. In Qt: ```python # Slot function (an optional marked callable object) @Slot() def close(self): ... # Signal connect to callable exit_btn.clicked.connect(lambda: print("Exit!")) exit_btn.clicked.connect(self.close) ``` In `arcade.gui`: ```python exit_btn.push_handlers(on_click=lambda: print("Exit!")) exit_btn.push_handlers(on_click=close_window) ``` This is also a better design instead of override event function directly when creating a lot UI elements. Not yet explained in documentation after my searching. ```python # Directly assignment or create new inheritance # This looks like a monkey patch to me exit_btn.on_click = close_window ``` ### Where is it located? The API of `arcade.gui`. ### What is wrong with it? How can it be improved? There is still no example for `arcade.gui` package. Add a little example code for `EventDispatcher`. (like above) Or a bigger one, like a kind of extensible option menu before the game. ```python from arcade import set_background_color, start_render, run from arcade.application import Window, View from arcade.csscolor import CORNFLOWER_BLUE, GREENYELLOW, BLACK from arcade.text import draw_text from arcade.gui import UIManager, UIFlatButton class MainView(View): def __init__(self): super(MainView, self).__init__() self.title = draw_text('Game', self.window.width // 2, self.window.height // 2, BLACK, font_size=72, anchor_x='center') def on_draw(self): start_render() set_background_color(GREENYELLOW) self.title.draw() class OptionView(View): def __init__(self): super(OptionView, self).__init__() self.title = draw_text('Option', self.window.width // 2, self.window.height // 2, BLACK, font_size=72, anchor_x='center') def on_draw(self): start_render() set_background_color(CORNFLOWER_BLUE) self.title.draw() class Main(Window): def __init__(self): super(Main, self).__init__() self.ui = UIManager(self) def setup(self): self.show_main_ui() def show_main_ui(self): self.ui.purge_ui_elements() opt_btn = UIFlatButton('Option', self.width // 2, self.height // 2 - 50, 300, 40) opt_btn.push_handlers(on_click=self.show_opt_ui) self.ui.add_ui_element(opt_btn) self.show_view(MainView()) def show_opt_ui(self): self.ui.purge_ui_elements() back_btn = UIFlatButton('Back', self.width // 2, self.height // 2 - 50, 300, 40) back_btn.push_handlers(on_click=self.show_main_ui) self.ui.add_ui_element(back_btn) self.show_view(OptionView()) def on_draw(self): start_render() self.current_view.on_draw() self.ui.on_draw() def main(): w = Main() w.setup() run() if __name__ == '__main__': main() ```
1.0
[DOC] Example for UI event register - ## Documentation request: Hello! I'm a newcomer from Qt UI designing. In PyQt5 or PySide2, there's a "signal & slot" mechanism in all event handler, which is relative to `EventDispatcher` in Pyglet. In Qt: ```python # Slot function (an optional marked callable object) @Slot() def close(self): ... # Signal connect to callable exit_btn.clicked.connect(lambda: print("Exit!")) exit_btn.clicked.connect(self.close) ``` In `arcade.gui`: ```python exit_btn.push_handlers(on_click=lambda: print("Exit!")) exit_btn.push_handlers(on_click=close_window) ``` This is also a better design instead of override event function directly when creating a lot UI elements. Not yet explained in documentation after my searching. ```python # Directly assignment or create new inheritance # This looks like a monkey patch to me exit_btn.on_click = close_window ``` ### Where is it located? The API of `arcade.gui`. ### What is wrong with it? How can it be improved? There is still no example for `arcade.gui` package. Add a little example code for `EventDispatcher`. (like above) Or a bigger one, like a kind of extensible option menu before the game. ```python from arcade import set_background_color, start_render, run from arcade.application import Window, View from arcade.csscolor import CORNFLOWER_BLUE, GREENYELLOW, BLACK from arcade.text import draw_text from arcade.gui import UIManager, UIFlatButton class MainView(View): def __init__(self): super(MainView, self).__init__() self.title = draw_text('Game', self.window.width // 2, self.window.height // 2, BLACK, font_size=72, anchor_x='center') def on_draw(self): start_render() set_background_color(GREENYELLOW) self.title.draw() class OptionView(View): def __init__(self): super(OptionView, self).__init__() self.title = draw_text('Option', self.window.width // 2, self.window.height // 2, BLACK, font_size=72, anchor_x='center') def on_draw(self): start_render() set_background_color(CORNFLOWER_BLUE) self.title.draw() class Main(Window): def __init__(self): super(Main, self).__init__() self.ui = UIManager(self) def setup(self): self.show_main_ui() def show_main_ui(self): self.ui.purge_ui_elements() opt_btn = UIFlatButton('Option', self.width // 2, self.height // 2 - 50, 300, 40) opt_btn.push_handlers(on_click=self.show_opt_ui) self.ui.add_ui_element(opt_btn) self.show_view(MainView()) def show_opt_ui(self): self.ui.purge_ui_elements() back_btn = UIFlatButton('Back', self.width // 2, self.height // 2 - 50, 300, 40) back_btn.push_handlers(on_click=self.show_main_ui) self.ui.add_ui_element(back_btn) self.show_view(OptionView()) def on_draw(self): start_render() self.current_view.on_draw() self.ui.on_draw() def main(): w = Main() w.setup() run() if __name__ == '__main__': main() ```
non_defect
example for ui event register documentation request hello i m a newcomer from qt ui designing in or there s a signal slot mechanism in all event handler which is relative to eventdispatcher in pyglet in qt python slot function an optional marked callable object slot def close self signal connect to callable exit btn clicked connect lambda print exit exit btn clicked connect self close in arcade gui python exit btn push handlers on click lambda print exit exit btn push handlers on click close window this is also a better design instead of override event function directly when creating a lot ui elements not yet explained in documentation after my searching python directly assignment or create new inheritance this looks like a monkey patch to me exit btn on click close window where is it located the api of arcade gui what is wrong with it how can it be improved there is still no example for arcade gui package add a little example code for eventdispatcher like above or a bigger one like a kind of extensible option menu before the game python from arcade import set background color start render run from arcade application import window view from arcade csscolor import cornflower blue greenyellow black from arcade text import draw text from arcade gui import uimanager uiflatbutton class mainview view def init self super mainview self init self title draw text game self window width self window height black font size anchor x center def on draw self start render set background color greenyellow self title draw class optionview view def init self super optionview self init self title draw text option self window width self window height black font size anchor x center def on draw self start render set background color cornflower blue self title draw class main window def init self super main self init self ui uimanager self def setup self self show main ui def show main ui self self ui purge ui elements opt btn uiflatbutton option self width self height opt btn push handlers on click self show opt ui self ui add ui element opt btn self show view mainview def show opt ui self self ui purge ui elements back btn uiflatbutton back self width self height back btn push handlers on click self show main ui self ui add ui element back btn self show view optionview def on draw self start render self current view on draw self ui on draw def main w main w setup run if name main main
0
21,638
3,737,403,224
IssuesEvent
2016-03-08 19:09:38
ELENA-LANG/elena-lang
https://api.github.com/repos/ELENA-LANG/elena-lang
closed
shorthand property settings
Design Idea Discussion
In the following code #class MyClass { #field theField1. #method Field1 = theField1. #method set &Field1:val [ theField1 := val. ] } the following shorthand syntax is proposed myObj Field1 := 2. which is compiled as myObject set &Field1:2.
1.0
shorthand property settings - In the following code #class MyClass { #field theField1. #method Field1 = theField1. #method set &Field1:val [ theField1 := val. ] } the following shorthand syntax is proposed myObj Field1 := 2. which is compiled as myObject set &Field1:2.
non_defect
shorthand property settings in the following code class myclass field method method set val the following shorthand syntax is proposed myobj which is compiled as myobject set
0
11,401
9,344,868,968
IssuesEvent
2019-03-30 01:42:43
microsoftgraph/msgraph-sdk-dotnet
https://api.github.com/repos/microsoftgraph/msgraph-sdk-dotnet
closed
FR: Enable parallel/out-of-sequence large file uploads in session
enhancement (service)
Currently it appears that a [file] UploadSession only supports receiving content chunks in the exact order the file needs to be constructed in (exact sequence). This defeats a straightforward method to improve large file upload performance by orders of magnitude. If we can put content chunks in any sequence, we can leverage .Net parallel tasks and substantially improve customer experience. Since the UploadSession (and the library's helper-class ChunkedUploadProvider) already keep track of which byte-ranges are missing, it appears to be a reasonably simple update for the service to allow this (and simply collate all the chunks and append them when all ranges have been received). I have sample/prototype ParallelUpload code to offer the project if/when the service can support it (because I didn't read the fine-print before I wrote it, only to track down ServiceExceptions when testing). Thanks for considering it.
1.0
FR: Enable parallel/out-of-sequence large file uploads in session - Currently it appears that a [file] UploadSession only supports receiving content chunks in the exact order the file needs to be constructed in (exact sequence). This defeats a straightforward method to improve large file upload performance by orders of magnitude. If we can put content chunks in any sequence, we can leverage .Net parallel tasks and substantially improve customer experience. Since the UploadSession (and the library's helper-class ChunkedUploadProvider) already keep track of which byte-ranges are missing, it appears to be a reasonably simple update for the service to allow this (and simply collate all the chunks and append them when all ranges have been received). I have sample/prototype ParallelUpload code to offer the project if/when the service can support it (because I didn't read the fine-print before I wrote it, only to track down ServiceExceptions when testing). Thanks for considering it.
non_defect
fr enable parallel out of sequence large file uploads in session currently it appears that a uploadsession only supports receiving content chunks in the exact order the file needs to be constructed in exact sequence this defeats a straightforward method to improve large file upload performance by orders of magnitude if we can put content chunks in any sequence we can leverage net parallel tasks and substantially improve customer experience since the uploadsession and the library s helper class chunkeduploadprovider already keep track of which byte ranges are missing it appears to be a reasonably simple update for the service to allow this and simply collate all the chunks and append them when all ranges have been received i have sample prototype parallelupload code to offer the project if when the service can support it because i didn t read the fine print before i wrote it only to track down serviceexceptions when testing thanks for considering it
0
22,668
3,681,675,469
IssuesEvent
2016-02-24 05:03:10
sccn/labstreaminglayer
https://api.github.com/repos/sccn/labstreaminglayer
closed
Fix for vis_stream does not work with Matlab R2014b
auto-migrated Priority-Medium Type-Defect
``` It seems Matlab's new graphics system has broken vis_stream. The only problem is in supergui line 109, finputcheck requires fig to be numeric but it is now an object. I just emptied the type column (i.e. '') and it works again. ``` Original issue reported on code.google.com by `chadwick...@gmail.com` on 3 Dec 2014 at 6:47
1.0
Fix for vis_stream does not work with Matlab R2014b - ``` It seems Matlab's new graphics system has broken vis_stream. The only problem is in supergui line 109, finputcheck requires fig to be numeric but it is now an object. I just emptied the type column (i.e. '') and it works again. ``` Original issue reported on code.google.com by `chadwick...@gmail.com` on 3 Dec 2014 at 6:47
defect
fix for vis stream does not work with matlab it seems matlab s new graphics system has broken vis stream the only problem is in supergui line finputcheck requires fig to be numeric but it is now an object i just emptied the type column i e and it works again original issue reported on code google com by chadwick gmail com on dec at
1
71,879
23,839,014,148
IssuesEvent
2022-09-06 08:43:12
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
App Layout: Slow to load filter lists
T-Defect Team: Delight Z-AppLayout
### Steps to reproduce When the user switches between filters on the home page it takes several moments to load the list (even if it's a short list) ### Outcome #### What did you expect? #### What happened instead? ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store _No response_ ### Homeserver _No response_ ### Will you send logs? No ### Are you willing to provide a PR? No
1.0
App Layout: Slow to load filter lists - ### Steps to reproduce When the user switches between filters on the home page it takes several moments to load the list (even if it's a short list) ### Outcome #### What did you expect? #### What happened instead? ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store _No response_ ### Homeserver _No response_ ### Will you send logs? No ### Are you willing to provide a PR? No
defect
app layout slow to load filter lists steps to reproduce when the user switches between filters on the home page it takes several moments to load the list even if it s a short list outcome what did you expect what happened instead your phone model no response operating system version no response application version and app store no response homeserver no response will you send logs no are you willing to provide a pr no
1
68,803
21,910,850,580
IssuesEvent
2022-05-21 02:59:30
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
PANIC at zfs_znode.c zfs_znode_sa_init() - Regression in closed issue #10971
Type: Defect Component: Encryption
### System information Type | Version/Name --- | --- Distribution Name | Pop!_OS | Distribution Version | 20.04 & 22.04 | Kernel Version | 5.16.15 & 5.17.5 | Architecture | x86_64 | OpenZFS Version | zfs-2.1.4-1pop1~1650059719~20.04~9934f01 | ### Describe the problem you're observing Regression of bug in #10971 On 20.04 it used to work very well for many months, days ago I upgraded to 22.04 and this started popping up whenever pods in my k8s cluster access files in the raid array. There was a firmware update for my MOBO so I did that. To downgrade the OS I did a clean install, and `zpool import -f`. Would deleting the pool and recover from backup data help? It's many TBs so I would like to avoid this. ### Describe how to reproduce the problem Use Pop!_OS 22.04 on an encrypted pool ### Include any warning/errors/backtraces from the system logs ``` [70099.854321] CPU: 3 PID: 3404665 Comm: .NET ThreadPool Tainted: P OE 5.17.5-76051705-generic #202204271406~1651504840~22.04~63e51bd [70099.854323] Hardware name: To Be Filled By O.E.M. To Be Filled By O.E.M./Z590 Steel Legend WiFi 6E, BIOS P1.80 07/22/2021 [70099.854324] Call Trace: [70099.854325] <TASK> [70099.854327] show_stack+0x52/0x59 [70099.854330] dump_stack_lvl+0x4c/0x64 [70099.854333] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed [70099.854334] dump_stack+0x10/0x13 [70099.854335] PANIC at zfs_znode.c:339:zfs_znode_sa_init() [70099.854336] Showing stack for process 3417543 [70099.854335] spl_dumpstack+0x29/0x2c [spl] [70099.854341] spl_panic+0xd1/0xe9 [spl] [70099.854345] ? sa_cache_constructor+0x27/0x50 [zfs] [70099.854410] ? sa_build_index+0x2c1/0x310 [zfs] [70099.854451] ? sa_build_index+0x25d/0x310 [zfs] [70099.854491] ? dmu_buf_set_user_ie+0x62/0x80 [zfs] [70099.854518] zfs_znode_sa_init+0xdf/0xf0 [zfs] [70099.854569] zfs_znode_alloc+0x101/0x750 [zfs] [70099.854617] ? __kmalloc_node+0x1c4/0x3e0 [70099.854620] ? arc_untransform+0x23/0x80 [zfs] [70099.854644] ? aggsum_add+0x1a2/0x1c0 [zfs] [70099.854666] ? dmu_buf_unlock_parent+0x39/0x80 [zfs] [70099.854691] ? dbuf_read_impl.constprop.0+0x5a1/0x7f0 [zfs] [70099.854717] ? spl_kmem_cache_alloc+0x9d/0x780 [spl] [70099.854720] ? spl_kmem_cache_alloc+0x9d/0x780 [spl] [70099.854723] ? aggsum_add+0x1a2/0x1c0 [zfs] [70099.854745] ? aggsum_add+0x1a2/0x1c0 [zfs] [70099.854768] ? arc_space_consume+0x54/0x120 [zfs] [70099.854791] ? dbuf_create+0x5aa/0x5e0 [zfs] [70099.854816] ? dbuf_read+0x11b/0x620 [zfs] [70099.854843] ? dnode_rele_and_unlock+0x68/0xf0 [zfs] [70099.854875] ? dmu_object_info_from_dnode+0x82/0xb0 [zfs] [70099.854902] zfs_zget+0x249/0x290 [zfs] [70099.854949] zfs_dirent_lock+0x4a7/0x670 [zfs] [70099.854996] zfs_dirlook+0x91/0x2e0 [zfs] [70099.855044] ? zfs_zaccess+0x216/0x420 [zfs] [70099.855089] zfs_lookup+0x25b/0x410 [zfs] [70099.855132] zpl_lookup+0xcf/0x1e0 [zfs] [70099.855175] __lookup_slow+0x81/0x150 [70099.855178] walk_component+0x142/0x1c0 [70099.855179] link_path_walk.part.0.constprop.0+0x24b/0x3d0 [70099.855180] ? path_init+0x2c2/0x3f0 [70099.855181] path_lookupat+0x3e/0x1b0 [70099.855182] filename_lookup+0xcf/0x1d0 [70099.855183] ? __check_object_size+0x1a/0x20 [70099.855185] ? strncpy_from_user+0x44/0x140 [70099.855187] ? getname_flags.part.0+0x4c/0x1b0 [70099.855188] user_path_at_empty+0x3f/0x60 [70099.855189] vfs_statx+0x7a/0x120 [70099.855191] __do_sys_newlstat+0x3e/0x70 [70099.855192] ? restore_fpregs_from_fpstate+0x47/0xb0 [70099.855194] ? switch_fpu_return+0x4e/0xd0 [70099.855196] ? exit_to_user_mode_prepare+0x37/0xb0 [70099.855197] __x64_sys_newlstat+0x16/0x20 [70099.855199] do_syscall_64+0x59/0xc0 [70099.855200] ? exit_to_user_mode_prepare+0x37/0xb0 [70099.855201] ? syscall_exit_to_user_mode+0x27/0x50 [70099.855203] ? __do_sys_geteuid+0x28/0x30 [70099.855204] ? do_syscall_64+0x69/0xc0 [70099.855204] ? exit_to_user_mode_prepare+0x37/0xb0 [70099.855206] ? syscall_exit_to_user_mode+0x27/0x50 [70099.855207] ? do_syscall_64+0x69/0xc0 [70099.855208] ? do_syscall_64+0x69/0xc0 [70099.855208] entry_SYSCALL_64_after_hwframe+0x44/0xae [70099.855210] RIP: 0033:0x7f7ce4111b45 [70099.855211] Code: 51 13 0d 00 64 c7 00 16 00 00 00 b8 ff ff ff ff c3 0f 1f 40 00 48 89 f0 83 ff 01 77 30 48 89 c7 48 89 d6 b8 06 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 03 c3 66 90 48 8b 15 19 13 0d 00 f7 d8 64 89 [70099.855212] RSP: 002b:00007f7bebffdc68 EFLAGS: 00000246 ORIG_RAX: 0000000000000006 [70099.855214] RAX: ffffffffffffffda RBX: 00007f7c35f786c8 RCX: 00007f7ce4111b45 [70099.855215] RDX: 00007f7bebffdc70 RSI: 00007f7bebffdc70 RDI: 00007f7bebffddc0 [70099.855215] RBP: 00007f7bebffddb0 R08: 00007f7bebffdd10 R09: 0000000000000047 [70099.855216] R10: 00007f7c6ab09538 R11: 0000000000000246 R12: 0000000000000000 [70099.855217] R13: 0000000000000001 R14: 00007f7c35f786c8 R15: 00007f7be0001770 [70099.855218] </TASK> ``` After downgrading back to 20.04 ``` [28502.940773] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed [28502.940781] PANIC at zfs_znode.c:339:zfs_znode_sa_init() [28502.940784] Showing stack for process 2986375 [28502.940787] CPU: 10 PID: 2986375 Comm: .NET ThreadPool Tainted: P OE 5.16.15-76051615-generic #202203161444~1647964027~20.04~e706226-Ubuntu [28502.940789] Hardware name: To Be Filled By O.E.M. Z590 Steel Legend WiFi 6E/Z590 Steel Legend WiFi 6E, BIOS P2.00 10/21/2021 [28502.940791] Call Trace: [28502.940794] <TASK> [28502.940800] dump_stack_lvl+0x4c/0x63 [28502.940811] dump_stack+0x10/0x12 [28502.940816] spl_dumpstack+0x29/0x2b [spl] [28502.940828] spl_panic+0xd1/0xe9 [spl] [28502.940833] ? sa_cache_constructor+0x27/0x50 [zfs] [28502.940976] ? spl_kmem_cache_alloc+0xa3/0x7e0 [spl] [28502.940986] ? dmu_buf_replace_user+0x60/0x80 [zfs] [28502.941041] ? dmu_buf_set_user_ie+0x1a/0x20 [zfs] [28502.941094] zfs_znode_sa_init.isra.0+0xdd/0xf0 [zfs] [28502.941175] zfs_znode_alloc+0x102/0x720 [zfs] [28502.941273] ? __kmalloc_node+0x1db/0x540 [28502.941281] ? aggsum_add+0x1aa/0x1c0 [zfs] [28502.941351] ? dmu_buf_unlock_parent+0x38/0x80 [zfs] [28502.941412] ? dbuf_read_impl.constprop.0+0x654/0x760 [zfs] [28502.941464] ? spl_kmem_cache_alloc+0xa3/0x7e0 [spl] [28502.941471] ? arc_buf_access+0x266/0x280 [zfs] [28502.941520] ? aggsum_add+0x1aa/0x1c0 [zfs] [28502.941567] ? arc_space_consume+0x54/0x120 [zfs] [28502.941615] ? dbuf_read+0x1ed/0x5f0 [zfs] [28502.941664] ? dnode_rele_and_unlock+0x68/0xf0 [zfs] [28502.941718] ? dmu_object_info_from_dnode+0x82/0xb0 [zfs] [28502.941769] zfs_zget+0x24a/0x290 [zfs] [28502.941838] zfs_dirent_lock+0x347/0x680 [zfs] [28502.941906] zfs_dirlook+0x90/0x2e0 [zfs] [28502.941972] ? zfs_zaccess+0x224/0x410 [zfs] [28502.942039] zfs_lookup+0x25c/0x410 [zfs] [28502.942104] zpl_lookup+0xc9/0x1e0 [zfs] [28502.942173] ? __legitimize_path.isra.0+0x31/0x70 [28502.942177] __lookup_slow+0x81/0x140 [28502.942178] walk_component+0x141/0x1b0 [28502.942181] link_path_walk.part.0+0x22d/0x360 [28502.942182] ? path_init+0x2ab/0x390 [28502.942184] path_lookupat.isra.0+0x3e/0x150 [28502.942185] ? raw_spin_rq_unlock+0x10/0x30 [28502.942191] filename_lookup+0xcf/0x1a0 [28502.942192] ? __check_object_size+0x13f/0x150 [28502.942196] ? strncpy_from_user+0x44/0x150 [28502.942201] ? getname_flags+0x6f/0x1f0 [28502.942203] user_path_at_empty+0x3f/0x60 [28502.942204] vfs_statx+0x73/0x110 [28502.942207] __do_sys_newstat+0x3e/0x80 [28502.942209] ? __x64_sys_futex+0x7b/0x1b0 [28502.942212] ? exit_to_user_mode_prepare+0x3d/0x1c0 [28502.942218] ? syscall_exit_to_user_mode+0x27/0x50 [28502.942222] ? do_syscall_64+0x69/0xc0 [28502.942226] ? syscall_exit_to_user_mode+0x27/0x50 [28502.942228] __x64_sys_newstat+0x16/0x20 [28502.942229] do_syscall_64+0x59/0xc0 [28502.942231] ? exit_to_user_mode_prepare+0x3d/0x1c0 [28502.942233] ? schedule+0x55/0xd0 [28502.942236] ? syscall_exit_to_user_mode+0x27/0x50 [28502.942238] ? __do_sys_sched_yield+0xe/0x20 [28502.942240] ? do_syscall_64+0x69/0xc0 [28502.942242] ? do_syscall_64+0x69/0xc0 [28502.942244] ? do_syscall_64+0x69/0xc0 [28502.942245] ? do_syscall_64+0x69/0xc0 [28502.942247] ? do_syscall_64+0x69/0xc0 [28502.942248] ? asm_sysvec_apic_timer_interrupt+0xa/0x20 [28502.942250] entry_SYSCALL_64_after_hwframe+0x44/0xae [28502.942252] RIP: 0033:0x7fdb95a9baa5 [28502.942255] Code: 00 00 00 75 05 48 83 c4 18 c3 e8 26 0d 02 00 66 0f 1f 44 00 00 48 89 f0 83 ff 01 77 30 48 89 c7 48 89 d6 b8 04 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 03 c3 66 90 48 8b 15 b9 13 0d 00 f7 d8 64 89 [28502.942257] RSP: 002b:00007fda6cff54b8 EFLAGS: 00000246 ORIG_RAX: 0000000000000004 [28502.942259] RAX: ffffffffffffffda RBX: 00007fda6cff5620 RCX: 00007fdb95a9baa5 [28502.942260] RDX: 00007fda6cff54c0 RSI: 00007fda6cff54c0 RDI: 00007fda6cff5620 [28502.942261] RBP: 00007fda6cff5610 R08: 00007fda6cff5570 R09: 0000000000000047 [28502.942262] R10: 00007fdb1c4794d8 R11: 0000000000000246 R12: 00007fdae82c30c0 [28502.942263] R13: 0000000000000001 R14: 00007fda6cff5780 R15: 00007fda6cff54c0 [28502.942265] </TASK> ```
1.0
PANIC at zfs_znode.c zfs_znode_sa_init() - Regression in closed issue #10971 - ### System information Type | Version/Name --- | --- Distribution Name | Pop!_OS | Distribution Version | 20.04 & 22.04 | Kernel Version | 5.16.15 & 5.17.5 | Architecture | x86_64 | OpenZFS Version | zfs-2.1.4-1pop1~1650059719~20.04~9934f01 | ### Describe the problem you're observing Regression of bug in #10971 On 20.04 it used to work very well for many months, days ago I upgraded to 22.04 and this started popping up whenever pods in my k8s cluster access files in the raid array. There was a firmware update for my MOBO so I did that. To downgrade the OS I did a clean install, and `zpool import -f`. Would deleting the pool and recover from backup data help? It's many TBs so I would like to avoid this. ### Describe how to reproduce the problem Use Pop!_OS 22.04 on an encrypted pool ### Include any warning/errors/backtraces from the system logs ``` [70099.854321] CPU: 3 PID: 3404665 Comm: .NET ThreadPool Tainted: P OE 5.17.5-76051705-generic #202204271406~1651504840~22.04~63e51bd [70099.854323] Hardware name: To Be Filled By O.E.M. To Be Filled By O.E.M./Z590 Steel Legend WiFi 6E, BIOS P1.80 07/22/2021 [70099.854324] Call Trace: [70099.854325] <TASK> [70099.854327] show_stack+0x52/0x59 [70099.854330] dump_stack_lvl+0x4c/0x64 [70099.854333] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed [70099.854334] dump_stack+0x10/0x13 [70099.854335] PANIC at zfs_znode.c:339:zfs_znode_sa_init() [70099.854336] Showing stack for process 3417543 [70099.854335] spl_dumpstack+0x29/0x2c [spl] [70099.854341] spl_panic+0xd1/0xe9 [spl] [70099.854345] ? sa_cache_constructor+0x27/0x50 [zfs] [70099.854410] ? sa_build_index+0x2c1/0x310 [zfs] [70099.854451] ? sa_build_index+0x25d/0x310 [zfs] [70099.854491] ? dmu_buf_set_user_ie+0x62/0x80 [zfs] [70099.854518] zfs_znode_sa_init+0xdf/0xf0 [zfs] [70099.854569] zfs_znode_alloc+0x101/0x750 [zfs] [70099.854617] ? __kmalloc_node+0x1c4/0x3e0 [70099.854620] ? arc_untransform+0x23/0x80 [zfs] [70099.854644] ? aggsum_add+0x1a2/0x1c0 [zfs] [70099.854666] ? dmu_buf_unlock_parent+0x39/0x80 [zfs] [70099.854691] ? dbuf_read_impl.constprop.0+0x5a1/0x7f0 [zfs] [70099.854717] ? spl_kmem_cache_alloc+0x9d/0x780 [spl] [70099.854720] ? spl_kmem_cache_alloc+0x9d/0x780 [spl] [70099.854723] ? aggsum_add+0x1a2/0x1c0 [zfs] [70099.854745] ? aggsum_add+0x1a2/0x1c0 [zfs] [70099.854768] ? arc_space_consume+0x54/0x120 [zfs] [70099.854791] ? dbuf_create+0x5aa/0x5e0 [zfs] [70099.854816] ? dbuf_read+0x11b/0x620 [zfs] [70099.854843] ? dnode_rele_and_unlock+0x68/0xf0 [zfs] [70099.854875] ? dmu_object_info_from_dnode+0x82/0xb0 [zfs] [70099.854902] zfs_zget+0x249/0x290 [zfs] [70099.854949] zfs_dirent_lock+0x4a7/0x670 [zfs] [70099.854996] zfs_dirlook+0x91/0x2e0 [zfs] [70099.855044] ? zfs_zaccess+0x216/0x420 [zfs] [70099.855089] zfs_lookup+0x25b/0x410 [zfs] [70099.855132] zpl_lookup+0xcf/0x1e0 [zfs] [70099.855175] __lookup_slow+0x81/0x150 [70099.855178] walk_component+0x142/0x1c0 [70099.855179] link_path_walk.part.0.constprop.0+0x24b/0x3d0 [70099.855180] ? path_init+0x2c2/0x3f0 [70099.855181] path_lookupat+0x3e/0x1b0 [70099.855182] filename_lookup+0xcf/0x1d0 [70099.855183] ? __check_object_size+0x1a/0x20 [70099.855185] ? strncpy_from_user+0x44/0x140 [70099.855187] ? getname_flags.part.0+0x4c/0x1b0 [70099.855188] user_path_at_empty+0x3f/0x60 [70099.855189] vfs_statx+0x7a/0x120 [70099.855191] __do_sys_newlstat+0x3e/0x70 [70099.855192] ? restore_fpregs_from_fpstate+0x47/0xb0 [70099.855194] ? switch_fpu_return+0x4e/0xd0 [70099.855196] ? exit_to_user_mode_prepare+0x37/0xb0 [70099.855197] __x64_sys_newlstat+0x16/0x20 [70099.855199] do_syscall_64+0x59/0xc0 [70099.855200] ? exit_to_user_mode_prepare+0x37/0xb0 [70099.855201] ? syscall_exit_to_user_mode+0x27/0x50 [70099.855203] ? __do_sys_geteuid+0x28/0x30 [70099.855204] ? do_syscall_64+0x69/0xc0 [70099.855204] ? exit_to_user_mode_prepare+0x37/0xb0 [70099.855206] ? syscall_exit_to_user_mode+0x27/0x50 [70099.855207] ? do_syscall_64+0x69/0xc0 [70099.855208] ? do_syscall_64+0x69/0xc0 [70099.855208] entry_SYSCALL_64_after_hwframe+0x44/0xae [70099.855210] RIP: 0033:0x7f7ce4111b45 [70099.855211] Code: 51 13 0d 00 64 c7 00 16 00 00 00 b8 ff ff ff ff c3 0f 1f 40 00 48 89 f0 83 ff 01 77 30 48 89 c7 48 89 d6 b8 06 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 03 c3 66 90 48 8b 15 19 13 0d 00 f7 d8 64 89 [70099.855212] RSP: 002b:00007f7bebffdc68 EFLAGS: 00000246 ORIG_RAX: 0000000000000006 [70099.855214] RAX: ffffffffffffffda RBX: 00007f7c35f786c8 RCX: 00007f7ce4111b45 [70099.855215] RDX: 00007f7bebffdc70 RSI: 00007f7bebffdc70 RDI: 00007f7bebffddc0 [70099.855215] RBP: 00007f7bebffddb0 R08: 00007f7bebffdd10 R09: 0000000000000047 [70099.855216] R10: 00007f7c6ab09538 R11: 0000000000000246 R12: 0000000000000000 [70099.855217] R13: 0000000000000001 R14: 00007f7c35f786c8 R15: 00007f7be0001770 [70099.855218] </TASK> ``` After downgrading back to 20.04 ``` [28502.940773] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed [28502.940781] PANIC at zfs_znode.c:339:zfs_znode_sa_init() [28502.940784] Showing stack for process 2986375 [28502.940787] CPU: 10 PID: 2986375 Comm: .NET ThreadPool Tainted: P OE 5.16.15-76051615-generic #202203161444~1647964027~20.04~e706226-Ubuntu [28502.940789] Hardware name: To Be Filled By O.E.M. Z590 Steel Legend WiFi 6E/Z590 Steel Legend WiFi 6E, BIOS P2.00 10/21/2021 [28502.940791] Call Trace: [28502.940794] <TASK> [28502.940800] dump_stack_lvl+0x4c/0x63 [28502.940811] dump_stack+0x10/0x12 [28502.940816] spl_dumpstack+0x29/0x2b [spl] [28502.940828] spl_panic+0xd1/0xe9 [spl] [28502.940833] ? sa_cache_constructor+0x27/0x50 [zfs] [28502.940976] ? spl_kmem_cache_alloc+0xa3/0x7e0 [spl] [28502.940986] ? dmu_buf_replace_user+0x60/0x80 [zfs] [28502.941041] ? dmu_buf_set_user_ie+0x1a/0x20 [zfs] [28502.941094] zfs_znode_sa_init.isra.0+0xdd/0xf0 [zfs] [28502.941175] zfs_znode_alloc+0x102/0x720 [zfs] [28502.941273] ? __kmalloc_node+0x1db/0x540 [28502.941281] ? aggsum_add+0x1aa/0x1c0 [zfs] [28502.941351] ? dmu_buf_unlock_parent+0x38/0x80 [zfs] [28502.941412] ? dbuf_read_impl.constprop.0+0x654/0x760 [zfs] [28502.941464] ? spl_kmem_cache_alloc+0xa3/0x7e0 [spl] [28502.941471] ? arc_buf_access+0x266/0x280 [zfs] [28502.941520] ? aggsum_add+0x1aa/0x1c0 [zfs] [28502.941567] ? arc_space_consume+0x54/0x120 [zfs] [28502.941615] ? dbuf_read+0x1ed/0x5f0 [zfs] [28502.941664] ? dnode_rele_and_unlock+0x68/0xf0 [zfs] [28502.941718] ? dmu_object_info_from_dnode+0x82/0xb0 [zfs] [28502.941769] zfs_zget+0x24a/0x290 [zfs] [28502.941838] zfs_dirent_lock+0x347/0x680 [zfs] [28502.941906] zfs_dirlook+0x90/0x2e0 [zfs] [28502.941972] ? zfs_zaccess+0x224/0x410 [zfs] [28502.942039] zfs_lookup+0x25c/0x410 [zfs] [28502.942104] zpl_lookup+0xc9/0x1e0 [zfs] [28502.942173] ? __legitimize_path.isra.0+0x31/0x70 [28502.942177] __lookup_slow+0x81/0x140 [28502.942178] walk_component+0x141/0x1b0 [28502.942181] link_path_walk.part.0+0x22d/0x360 [28502.942182] ? path_init+0x2ab/0x390 [28502.942184] path_lookupat.isra.0+0x3e/0x150 [28502.942185] ? raw_spin_rq_unlock+0x10/0x30 [28502.942191] filename_lookup+0xcf/0x1a0 [28502.942192] ? __check_object_size+0x13f/0x150 [28502.942196] ? strncpy_from_user+0x44/0x150 [28502.942201] ? getname_flags+0x6f/0x1f0 [28502.942203] user_path_at_empty+0x3f/0x60 [28502.942204] vfs_statx+0x73/0x110 [28502.942207] __do_sys_newstat+0x3e/0x80 [28502.942209] ? __x64_sys_futex+0x7b/0x1b0 [28502.942212] ? exit_to_user_mode_prepare+0x3d/0x1c0 [28502.942218] ? syscall_exit_to_user_mode+0x27/0x50 [28502.942222] ? do_syscall_64+0x69/0xc0 [28502.942226] ? syscall_exit_to_user_mode+0x27/0x50 [28502.942228] __x64_sys_newstat+0x16/0x20 [28502.942229] do_syscall_64+0x59/0xc0 [28502.942231] ? exit_to_user_mode_prepare+0x3d/0x1c0 [28502.942233] ? schedule+0x55/0xd0 [28502.942236] ? syscall_exit_to_user_mode+0x27/0x50 [28502.942238] ? __do_sys_sched_yield+0xe/0x20 [28502.942240] ? do_syscall_64+0x69/0xc0 [28502.942242] ? do_syscall_64+0x69/0xc0 [28502.942244] ? do_syscall_64+0x69/0xc0 [28502.942245] ? do_syscall_64+0x69/0xc0 [28502.942247] ? do_syscall_64+0x69/0xc0 [28502.942248] ? asm_sysvec_apic_timer_interrupt+0xa/0x20 [28502.942250] entry_SYSCALL_64_after_hwframe+0x44/0xae [28502.942252] RIP: 0033:0x7fdb95a9baa5 [28502.942255] Code: 00 00 00 75 05 48 83 c4 18 c3 e8 26 0d 02 00 66 0f 1f 44 00 00 48 89 f0 83 ff 01 77 30 48 89 c7 48 89 d6 b8 04 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 03 c3 66 90 48 8b 15 b9 13 0d 00 f7 d8 64 89 [28502.942257] RSP: 002b:00007fda6cff54b8 EFLAGS: 00000246 ORIG_RAX: 0000000000000004 [28502.942259] RAX: ffffffffffffffda RBX: 00007fda6cff5620 RCX: 00007fdb95a9baa5 [28502.942260] RDX: 00007fda6cff54c0 RSI: 00007fda6cff54c0 RDI: 00007fda6cff5620 [28502.942261] RBP: 00007fda6cff5610 R08: 00007fda6cff5570 R09: 0000000000000047 [28502.942262] R10: 00007fdb1c4794d8 R11: 0000000000000246 R12: 00007fdae82c30c0 [28502.942263] R13: 0000000000000001 R14: 00007fda6cff5780 R15: 00007fda6cff54c0 [28502.942265] </TASK> ```
defect
panic at zfs znode c zfs znode sa init regression in closed issue system information type version name distribution name pop os distribution version kernel version architecture openzfs version zfs describe the problem you re observing regression of bug in on it used to work very well for many months days ago i upgraded to and this started popping up whenever pods in my cluster access files in the raid array there was a firmware update for my mobo so i did that to downgrade the os i did a clean install and zpool import f would deleting the pool and recover from backup data help it s many tbs so i would like to avoid this describe how to reproduce the problem use pop os on an encrypted pool include any warning errors backtraces from the system logs cpu pid comm net threadpool tainted p oe generic hardware name to be filled by o e m to be filled by o e m steel legend wifi bios call trace show stack dump stack lvl verify sa handle get from db zfsvfs z os db zp sa hdl shared zp z sa hdl failed dump stack panic at zfs znode c zfs znode sa init showing stack for process spl dumpstack spl panic sa cache constructor sa build index sa build index dmu buf set user ie zfs znode sa init zfs znode alloc kmalloc node arc untransform aggsum add dmu buf unlock parent dbuf read impl constprop spl kmem cache alloc spl kmem cache alloc aggsum add aggsum add arc space consume dbuf create dbuf read dnode rele and unlock dmu object info from dnode zfs zget zfs dirent lock zfs dirlook zfs zaccess zfs lookup zpl lookup lookup slow walk component link path walk part constprop path init path lookupat filename lookup check object size strncpy from user getname flags part user path at empty vfs statx do sys newlstat restore fpregs from fpstate switch fpu return exit to user mode prepare sys newlstat do syscall exit to user mode prepare syscall exit to user mode do sys geteuid do syscall exit to user mode prepare syscall exit to user mode do syscall do syscall entry syscall after hwframe rip code ff ff ff ff ff ff ff rsp eflags orig rax rax ffffffffffffffda rbx rcx rdx rsi rdi rbp after downgrading back to verify sa handle get from db zfsvfs z os db zp sa hdl shared zp z sa hdl failed panic at zfs znode c zfs znode sa init showing stack for process cpu pid comm net threadpool tainted p oe generic ubuntu hardware name to be filled by o e m steel legend wifi steel legend wifi bios call trace dump stack lvl dump stack spl dumpstack spl panic sa cache constructor spl kmem cache alloc dmu buf replace user dmu buf set user ie zfs znode sa init isra zfs znode alloc kmalloc node aggsum add dmu buf unlock parent dbuf read impl constprop spl kmem cache alloc arc buf access aggsum add arc space consume dbuf read dnode rele and unlock dmu object info from dnode zfs zget zfs dirent lock zfs dirlook zfs zaccess zfs lookup zpl lookup legitimize path isra lookup slow walk component link path walk part path init path lookupat isra raw spin rq unlock filename lookup check object size strncpy from user getname flags user path at empty vfs statx do sys newstat sys futex exit to user mode prepare syscall exit to user mode do syscall syscall exit to user mode sys newstat do syscall exit to user mode prepare schedule syscall exit to user mode do sys sched yield do syscall do syscall do syscall do syscall do syscall asm sysvec apic timer interrupt entry syscall after hwframe rip code ff ff ff rsp eflags orig rax rax ffffffffffffffda rbx rcx rdx rsi rdi rbp
1
292,666
25,228,368,838
IssuesEvent
2022-11-14 17:39:27
mxsdev/ts-type-explorer
https://api.github.com/repos/mxsdev/ts-type-explorer
closed
Test Errors
bug tests
As of now running tests on the lib will always result in 2 errors. This errors come from the `module.ts` file and are a result of how the tests are writen, i dont think this is a huge issue but at least one that should be mentioned on the contributing.md file. Error: ![Screenshot 2022-11-14 at 10 31 32](https://user-images.githubusercontent.com/80978939/201638407-d2cf087d-4138-4a4d-95e9-4a4f78b5803f.png)
1.0
Test Errors - As of now running tests on the lib will always result in 2 errors. This errors come from the `module.ts` file and are a result of how the tests are writen, i dont think this is a huge issue but at least one that should be mentioned on the contributing.md file. Error: ![Screenshot 2022-11-14 at 10 31 32](https://user-images.githubusercontent.com/80978939/201638407-d2cf087d-4138-4a4d-95e9-4a4f78b5803f.png)
non_defect
test errors as of now running tests on the lib will always result in errors this errors come from the module ts file and are a result of how the tests are writen i dont think this is a huge issue but at least one that should be mentioned on the contributing md file error
0
51,728
7,725,511,610
IssuesEvent
2018-05-24 18:12:25
smartdevicelink/sdl_server
https://api.github.com/repos/smartdevicelink/sdl_server
closed
Allow More Flexible States for App Approvals
documentation enhancement
### Enhancement Allow more options for whether the app is approved or denied permissions in both staging and production modes. Update documentation to reflect new changes.
1.0
Allow More Flexible States for App Approvals - ### Enhancement Allow more options for whether the app is approved or denied permissions in both staging and production modes. Update documentation to reflect new changes.
non_defect
allow more flexible states for app approvals enhancement allow more options for whether the app is approved or denied permissions in both staging and production modes update documentation to reflect new changes
0
31,006
6,396,123,000
IssuesEvent
2017-08-04 14:49:36
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
opened
__dx() throws an IntlException for strings containing {{...}}
Defect i18n
This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.4 * Platform and Target: Windows 10, Nginx ### What you did ```php __dx('wa', 'time_span', '{{startDate}} ... {{endDate}}'); ``` ### What happened > IntlException: Constructor failed It works just fine when I'm using `__d(`)` or `__dx()` without the curly brackets. ### What you expected to happen Not getting an exception.
1.0
__dx() throws an IntlException for strings containing {{...}} - This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.4 * Platform and Target: Windows 10, Nginx ### What you did ```php __dx('wa', 'time_span', '{{startDate}} ... {{endDate}}'); ``` ### What happened > IntlException: Constructor failed It works just fine when I'm using `__d(`)` or `__dx()` without the curly brackets. ### What you expected to happen Not getting an exception.
defect
dx throws an intlexception for strings containing this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target windows nginx what you did php dx wa time span startdate enddate what happened intlexception constructor failed it works just fine when i m using d or dx without the curly brackets what you expected to happen not getting an exception
1
24,149
4,059,433,763
IssuesEvent
2016-05-25 09:33:18
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
EntryListenerOnReconnectTest.testListenersSmartRoutingMultipleServer
Team: Core Type: Test-Failure
``` org.junit.runners.model.TestTimedOutException: test timed out after 120000 milliseconds at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:315) at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:156) at com.hazelcast.client.spi.ClientProxy.invokeOnPartition(ClientProxy.java:153) at com.hazelcast.client.spi.ClientProxy.invoke(ClientProxy.java:147) at com.hazelcast.client.proxy.ClientMapProxy.putInternal(ClientMapProxy.java:452) at com.hazelcast.client.proxy.ClientMapProxy.put(ClientMapProxy.java:446) at com.hazelcast.client.proxy.ClientMapProxy.put(ClientMapProxy.java:247) at com.hazelcast.client.listeners.EntryListenerOnReconnectTest.produceEvent(EntryListenerOnReconnectTest.java:49) ``` https://hazelcast-l337.ci.cloudbees.com/view/Official%20Builds/job/Hazelcast-3.x-IbmJDK1.8/com.hazelcast$hazelcast-client/228/testReport/junit/com.hazelcast.client.listeners/EntryListenerOnReconnectTest/testListenersSmartRoutingMultipleServer/
1.0
EntryListenerOnReconnectTest.testListenersSmartRoutingMultipleServer - ``` org.junit.runners.model.TestTimedOutException: test timed out after 120000 milliseconds at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:315) at com.hazelcast.spi.impl.AbstractInvocationFuture.get(AbstractInvocationFuture.java:156) at com.hazelcast.client.spi.ClientProxy.invokeOnPartition(ClientProxy.java:153) at com.hazelcast.client.spi.ClientProxy.invoke(ClientProxy.java:147) at com.hazelcast.client.proxy.ClientMapProxy.putInternal(ClientMapProxy.java:452) at com.hazelcast.client.proxy.ClientMapProxy.put(ClientMapProxy.java:446) at com.hazelcast.client.proxy.ClientMapProxy.put(ClientMapProxy.java:247) at com.hazelcast.client.listeners.EntryListenerOnReconnectTest.produceEvent(EntryListenerOnReconnectTest.java:49) ``` https://hazelcast-l337.ci.cloudbees.com/view/Official%20Builds/job/Hazelcast-3.x-IbmJDK1.8/com.hazelcast$hazelcast-client/228/testReport/junit/com.hazelcast.client.listeners/EntryListenerOnReconnectTest/testListenersSmartRoutingMultipleServer/
non_defect
entrylisteneronreconnecttest testlistenerssmartroutingmultipleserver org junit runners model testtimedoutexception test timed out after milliseconds at sun misc unsafe park native method at java util concurrent locks locksupport park locksupport java at com hazelcast spi impl abstractinvocationfuture get abstractinvocationfuture java at com hazelcast client spi clientproxy invokeonpartition clientproxy java at com hazelcast client spi clientproxy invoke clientproxy java at com hazelcast client proxy clientmapproxy putinternal clientmapproxy java at com hazelcast client proxy clientmapproxy put clientmapproxy java at com hazelcast client proxy clientmapproxy put clientmapproxy java at com hazelcast client listeners entrylisteneronreconnecttest produceevent entrylisteneronreconnecttest java
0
28,613
5,310,725,772
IssuesEvent
2017-02-12 22:23:37
networkx/networkx
https://api.github.com/repos/networkx/networkx
closed
eulerian_circuit() is not compatible with MultiGraph
Defect
When calculating the Eulerian circuit on a MultiGraph (which is necessary when you convert a non-Eulerian graph to an Eulerian graph), the edges are returned as a 2-tuple, even if there is a key. This means all the data related to the edges is lost
1.0
eulerian_circuit() is not compatible with MultiGraph - When calculating the Eulerian circuit on a MultiGraph (which is necessary when you convert a non-Eulerian graph to an Eulerian graph), the edges are returned as a 2-tuple, even if there is a key. This means all the data related to the edges is lost
defect
eulerian circuit is not compatible with multigraph when calculating the eulerian circuit on a multigraph which is necessary when you convert a non eulerian graph to an eulerian graph the edges are returned as a tuple even if there is a key this means all the data related to the edges is lost
1
66,682
20,514,660,307
IssuesEvent
2022-03-01 10:27:53
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Various rooms seem to repeat chains of state events and read receipts
T-Defect A-Timeline A-Read-Receipts
### Steps to reproduce 1. Look at some rooms, perhaps scroll up a bit (I am not quite sure what it takes) ### Outcome #### What did you expect? State events and read receipts should only appear in one place in the timeline. Event level "..." menus should open when clicked. #### What happened instead? As seen below, some state events seem to repeat in a loop, as well as read receipts down the side. Also, the "..." menu on the state events does not want to open. <img width="577" alt="image" src="https://user-images.githubusercontent.com/279572/156151985-50ad1f52-13cb-4d94-aa6e-c3619c6247ba.png"> ### Operating system macOS ### Browser information Firefox 99 ### URL for webapp develop.element.io ### Application version Element version: de9e65de2b18-react-36adba101caf-js-54e815085f4d, Olm version: 3.2.8 ### Homeserver matrix.org ### Will you send logs? Yes
1.0
Various rooms seem to repeat chains of state events and read receipts - ### Steps to reproduce 1. Look at some rooms, perhaps scroll up a bit (I am not quite sure what it takes) ### Outcome #### What did you expect? State events and read receipts should only appear in one place in the timeline. Event level "..." menus should open when clicked. #### What happened instead? As seen below, some state events seem to repeat in a loop, as well as read receipts down the side. Also, the "..." menu on the state events does not want to open. <img width="577" alt="image" src="https://user-images.githubusercontent.com/279572/156151985-50ad1f52-13cb-4d94-aa6e-c3619c6247ba.png"> ### Operating system macOS ### Browser information Firefox 99 ### URL for webapp develop.element.io ### Application version Element version: de9e65de2b18-react-36adba101caf-js-54e815085f4d, Olm version: 3.2.8 ### Homeserver matrix.org ### Will you send logs? Yes
defect
various rooms seem to repeat chains of state events and read receipts steps to reproduce look at some rooms perhaps scroll up a bit i am not quite sure what it takes outcome what did you expect state events and read receipts should only appear in one place in the timeline event level menus should open when clicked what happened instead as seen below some state events seem to repeat in a loop as well as read receipts down the side also the menu on the state events does not want to open img width alt image src operating system macos browser information firefox url for webapp develop element io application version element version react js olm version homeserver matrix org will you send logs yes
1
5,794
2,610,215,569
IssuesEvent
2015-02-26 19:08:46
chrsmith/somefinders
https://api.github.com/repos/chrsmith/somefinders
opened
решебник по русскому языку 6 класс
auto-migrated Priority-Medium Type-Defect
``` '''Антуан Блинов''' Привет всем не подскажите где можно найти .решебник по русскому языку 6 класс. как то выкладывали уже '''Авдей Петров''' Качай тут http://bit.ly/16CYUOP '''Гайдар Горшков''' Спасибо вроде то но просит телефон вводить '''Арий Павлов''' Неа все ок у меня ничего не списало '''Адам Кошелев''' Не это не влияет на баланс Информация о файле: решебник по русскому языку 6 класс Загружен: В этом месяце Скачан раз: 1002 Рейтинг: 508 Средняя скорость скачивания: 947 Похожих файлов: 12 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 4:53
1.0
решебник по русскому языку 6 класс - ``` '''Антуан Блинов''' Привет всем не подскажите где можно найти .решебник по русскому языку 6 класс. как то выкладывали уже '''Авдей Петров''' Качай тут http://bit.ly/16CYUOP '''Гайдар Горшков''' Спасибо вроде то но просит телефон вводить '''Арий Павлов''' Неа все ок у меня ничего не списало '''Адам Кошелев''' Не это не влияет на баланс Информация о файле: решебник по русскому языку 6 класс Загружен: В этом месяце Скачан раз: 1002 Рейтинг: 508 Средняя скорость скачивания: 947 Похожих файлов: 12 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 4:53
defect
решебник по русскому языку класс антуан блинов привет всем не подскажите где можно найти решебник по русскому языку класс как то выкладывали уже авдей петров качай тут гайдар горшков спасибо вроде то но просит телефон вводить арий павлов неа все ок у меня ничего не списало адам кошелев не это не влияет на баланс информация о файле решебник по русскому языку класс загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
1
65,416
19,488,000,663
IssuesEvent
2021-12-26 19:29:08
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
Test failures on macOS Big Sur (Version 11.4) with Apple M1 chip
defect scipy.special
<!-- Thank you for taking the time to file a bug report. Please fill in the fields below, deleting the sections that don't apply to your issue. You can view the final output by clicking the preview button above. Note: This is a comment, and won't appear in the output. --> My issue is about ... #### Reproducing code example: <!-- If you place your code between the triple backticks below, it will be rendered as a code block. --> ```python import scipy scipy.test() ``` #### Error message: <!-- If any, paste the *full* error message inside a code block as above (starting from line Traceback) --> ``` =================================== FAILURES =================================== _________________________ TestSmirnovi.test_round_trip _________________________ scipy/special/tests/test_kolmogorov.py:169: in test_round_trip FuncData(_sm_smi, dataset, (0, 1), 2, rtol=_rtol).check(dtypes=[int, float, float]) E AssertionError: E Max |adiff|: 0.484375 E Max |rdiff|: 1337.21 E Bad results (2 out of 7) for the following points (in output 0): E 3 0.125 => 0.609375 != 0.125 (rdiff 3.875) E 10 0.0001 => 0.13382053040755335 != 0.0001 (rdiff 1337.2053040755336) _sm_smi = <function TestSmirnovi.test_round_trip.<locals>._sm_smi at 0x17ab99d30> _smc_smci = <function TestSmirnovi.test_round_trip.<locals>._smc_smci at 0x17ab99820> dataset = array([[1.00e+00, 4.00e-01, 4.00e-01], [1.00e+00, 6.00e-01, 6.00e-01], [2.00e+00, 8.75e-01, 8.75e-01], ..., [3.00e+00, 1.25e-01, 1.25e-01], [1.00e+01, 9.99e-01, 9.99e-01], [1.00e+01, 1.00e-04, 1.00e-04]]) self = <scipy.special.tests.test_kolmogorov.TestSmirnovi object at 0x17ab86b20> ______________________ TestSmirnovi.test_x_equals_0point5 ______________________ scipy/special/tests/test_kolmogorov.py:186: in test_x_equals_0point5 FuncData(smirnovi, dataset, (0, 1), 2, rtol=_rtol).check(dtypes=[int, float, float]) E AssertionError: E Max |adiff|: 0.13456 E Max |rdiff|: 0.708588 E Bad results (7 out of 11) for the following points (in output 0): E 4 0.5 => 0.125 != 0.255520481121 (rdiff 0.5108024239324789) E 5 0.5 => 0.1 != 0.234559536069 (rdiff 0.5736690067012106) E 6 0.5 => 0.08333333333333333 != 0.21715965898 (rdiff 0.6162577629530716) E 7 0.5 => 0.07142857142857142 != 0.202722580034 (rdiff 0.647653599236002) E 8 0.5 => 0.0625 != 0.190621765256 (rdiff 0.67212558379121) E 9 0.5 => 0.05555555555555555 != 0.180363501362 (rdiff 0.6919800561863548) E 10 0.5 => 0.05 != 0.17157867006 (rdiff 0.7085884860716352) dataset = array([[ 1. , 0.5 , 0.5 ], [ 2. , 0.5 , 0.3660254 ], [ 2. , ... , 0.19062177], [ 9. , 0.5 , 0.1803635 ], [10. , 0.5 , 0.17157867]]) self = <scipy.special.tests.test_kolmogorov.TestSmirnovi object at 0x17ac6fe50> =========================== short test summary info ============================ FAILED scipy/special/tests/test_kolmogorov.py::TestSmirnovi::test_round_trip FAILED scipy/special/tests/test_kolmogorov.py::TestSmirnovi::test_x_equals_0point5 ``` #### Scipy/Numpy/Python version information: ``` 1.8.0.dev0+1551.d92e048 1.21.1 sys.version_info(major=3, minor=9, micro=6, releaselevel='final', serial=0) ``` I am using macOS Big Sur (Version 11.4) with Apple M1 chip. Though the conda environment is having Intel software and hence using Rosetta to run on the hardware.
1.0
Test failures on macOS Big Sur (Version 11.4) with Apple M1 chip - <!-- Thank you for taking the time to file a bug report. Please fill in the fields below, deleting the sections that don't apply to your issue. You can view the final output by clicking the preview button above. Note: This is a comment, and won't appear in the output. --> My issue is about ... #### Reproducing code example: <!-- If you place your code between the triple backticks below, it will be rendered as a code block. --> ```python import scipy scipy.test() ``` #### Error message: <!-- If any, paste the *full* error message inside a code block as above (starting from line Traceback) --> ``` =================================== FAILURES =================================== _________________________ TestSmirnovi.test_round_trip _________________________ scipy/special/tests/test_kolmogorov.py:169: in test_round_trip FuncData(_sm_smi, dataset, (0, 1), 2, rtol=_rtol).check(dtypes=[int, float, float]) E AssertionError: E Max |adiff|: 0.484375 E Max |rdiff|: 1337.21 E Bad results (2 out of 7) for the following points (in output 0): E 3 0.125 => 0.609375 != 0.125 (rdiff 3.875) E 10 0.0001 => 0.13382053040755335 != 0.0001 (rdiff 1337.2053040755336) _sm_smi = <function TestSmirnovi.test_round_trip.<locals>._sm_smi at 0x17ab99d30> _smc_smci = <function TestSmirnovi.test_round_trip.<locals>._smc_smci at 0x17ab99820> dataset = array([[1.00e+00, 4.00e-01, 4.00e-01], [1.00e+00, 6.00e-01, 6.00e-01], [2.00e+00, 8.75e-01, 8.75e-01], ..., [3.00e+00, 1.25e-01, 1.25e-01], [1.00e+01, 9.99e-01, 9.99e-01], [1.00e+01, 1.00e-04, 1.00e-04]]) self = <scipy.special.tests.test_kolmogorov.TestSmirnovi object at 0x17ab86b20> ______________________ TestSmirnovi.test_x_equals_0point5 ______________________ scipy/special/tests/test_kolmogorov.py:186: in test_x_equals_0point5 FuncData(smirnovi, dataset, (0, 1), 2, rtol=_rtol).check(dtypes=[int, float, float]) E AssertionError: E Max |adiff|: 0.13456 E Max |rdiff|: 0.708588 E Bad results (7 out of 11) for the following points (in output 0): E 4 0.5 => 0.125 != 0.255520481121 (rdiff 0.5108024239324789) E 5 0.5 => 0.1 != 0.234559536069 (rdiff 0.5736690067012106) E 6 0.5 => 0.08333333333333333 != 0.21715965898 (rdiff 0.6162577629530716) E 7 0.5 => 0.07142857142857142 != 0.202722580034 (rdiff 0.647653599236002) E 8 0.5 => 0.0625 != 0.190621765256 (rdiff 0.67212558379121) E 9 0.5 => 0.05555555555555555 != 0.180363501362 (rdiff 0.6919800561863548) E 10 0.5 => 0.05 != 0.17157867006 (rdiff 0.7085884860716352) dataset = array([[ 1. , 0.5 , 0.5 ], [ 2. , 0.5 , 0.3660254 ], [ 2. , ... , 0.19062177], [ 9. , 0.5 , 0.1803635 ], [10. , 0.5 , 0.17157867]]) self = <scipy.special.tests.test_kolmogorov.TestSmirnovi object at 0x17ac6fe50> =========================== short test summary info ============================ FAILED scipy/special/tests/test_kolmogorov.py::TestSmirnovi::test_round_trip FAILED scipy/special/tests/test_kolmogorov.py::TestSmirnovi::test_x_equals_0point5 ``` #### Scipy/Numpy/Python version information: ``` 1.8.0.dev0+1551.d92e048 1.21.1 sys.version_info(major=3, minor=9, micro=6, releaselevel='final', serial=0) ``` I am using macOS Big Sur (Version 11.4) with Apple M1 chip. Though the conda environment is having Intel software and hence using Rosetta to run on the hardware.
defect
test failures on macos big sur version with apple chip thank you for taking the time to file a bug report please fill in the fields below deleting the sections that don t apply to your issue you can view the final output by clicking the preview button above note this is a comment and won t appear in the output my issue is about reproducing code example if you place your code between the triple backticks below it will be rendered as a code block python import scipy scipy test error message if any paste the full error message inside a code block as above starting from line traceback failures testsmirnovi test round trip scipy special tests test kolmogorov py in test round trip funcdata sm smi dataset rtol rtol check dtypes e assertionerror e max adiff e max rdiff e bad results out of for the following points in output e rdiff e rdiff sm smi sm smi at smc smci smc smci at dataset array self testsmirnovi test x equals scipy special tests test kolmogorov py in test x equals funcdata smirnovi dataset rtol rtol check dtypes e assertionerror e max adiff e max rdiff e bad results out of for the following points in output e rdiff e rdiff e rdiff e rdiff e rdiff e rdiff e rdiff dataset array self short test summary info failed scipy special tests test kolmogorov py testsmirnovi test round trip failed scipy special tests test kolmogorov py testsmirnovi test x equals scipy numpy python version information sys version info major minor micro releaselevel final serial i am using macos big sur version with apple chip though the conda environment is having intel software and hence using rosetta to run on the hardware
1
181,546
21,664,385,954
IssuesEvent
2022-05-07 01:11:28
rgordon95/ecommerce-react-redux-saga-demo
https://api.github.com/repos/rgordon95/ecommerce-react-redux-saga-demo
opened
CVE-2022-29167 (High) detected in hawk-3.1.3.tgz
security vulnerability
## CVE-2022-29167 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hawk-3.1.3.tgz</b></p></summary> <p>HTTP Hawk Authentication Scheme</p> <p>Library home page: <a href="https://registry.npmjs.org/hawk/-/hawk-3.1.3.tgz">https://registry.npmjs.org/hawk/-/hawk-3.1.3.tgz</a></p> <p>Path to dependency file: /ecommerce-react-redux-saga-demo/redux-saga-cart/package.json</p> <p>Path to vulnerable library: /redux-saga-cart/node_modules/hawk/package.json</p> <p> Dependency Hierarchy: - webpack-1.15.0.tgz (Root Library) - watchpack-0.2.9.tgz - chokidar-1.7.0.tgz - fsevents-1.1.1.tgz - node-pre-gyp-0.6.33.tgz - request-2.79.0.tgz - :x: **hawk-3.1.3.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Hawk is an HTTP authentication scheme providing mechanisms for making authenticated HTTP requests with partial cryptographic verification of the request and response, covering the HTTP method, request URI, host, and optionally the request payload. Hawk used a regular expression to parse `Host` HTTP header (`Hawk.utils.parseHost()`), which was subject to regular expression DoS attack - meaning each added character in the attacker's input increases the computation time exponentially. `parseHost()` was patched in `9.0.1` to use built-in `URL` class to parse hostname instead. `Hawk.authenticate()` accepts `options` argument. If that contains `host` and `port`, those would be used instead of a call to `utils.parseHost()`. <p>Publish Date: 2022-05-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29167>CVE-2022-29167</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mozilla/hawk/security/advisories/GHSA-44pw-h2cw-w3vq">https://github.com/mozilla/hawk/security/advisories/GHSA-44pw-h2cw-w3vq</a></p> <p>Release Date: 2022-05-05</p> <p>Fix Resolution: hawk - 9.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-29167 (High) detected in hawk-3.1.3.tgz - ## CVE-2022-29167 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hawk-3.1.3.tgz</b></p></summary> <p>HTTP Hawk Authentication Scheme</p> <p>Library home page: <a href="https://registry.npmjs.org/hawk/-/hawk-3.1.3.tgz">https://registry.npmjs.org/hawk/-/hawk-3.1.3.tgz</a></p> <p>Path to dependency file: /ecommerce-react-redux-saga-demo/redux-saga-cart/package.json</p> <p>Path to vulnerable library: /redux-saga-cart/node_modules/hawk/package.json</p> <p> Dependency Hierarchy: - webpack-1.15.0.tgz (Root Library) - watchpack-0.2.9.tgz - chokidar-1.7.0.tgz - fsevents-1.1.1.tgz - node-pre-gyp-0.6.33.tgz - request-2.79.0.tgz - :x: **hawk-3.1.3.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Hawk is an HTTP authentication scheme providing mechanisms for making authenticated HTTP requests with partial cryptographic verification of the request and response, covering the HTTP method, request URI, host, and optionally the request payload. Hawk used a regular expression to parse `Host` HTTP header (`Hawk.utils.parseHost()`), which was subject to regular expression DoS attack - meaning each added character in the attacker's input increases the computation time exponentially. `parseHost()` was patched in `9.0.1` to use built-in `URL` class to parse hostname instead. `Hawk.authenticate()` accepts `options` argument. If that contains `host` and `port`, those would be used instead of a call to `utils.parseHost()`. <p>Publish Date: 2022-05-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29167>CVE-2022-29167</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mozilla/hawk/security/advisories/GHSA-44pw-h2cw-w3vq">https://github.com/mozilla/hawk/security/advisories/GHSA-44pw-h2cw-w3vq</a></p> <p>Release Date: 2022-05-05</p> <p>Fix Resolution: hawk - 9.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in hawk tgz cve high severity vulnerability vulnerable library hawk tgz http hawk authentication scheme library home page a href path to dependency file ecommerce react redux saga demo redux saga cart package json path to vulnerable library redux saga cart node modules hawk package json dependency hierarchy webpack tgz root library watchpack tgz chokidar tgz fsevents tgz node pre gyp tgz request tgz x hawk tgz vulnerable library vulnerability details hawk is an http authentication scheme providing mechanisms for making authenticated http requests with partial cryptographic verification of the request and response covering the http method request uri host and optionally the request payload hawk used a regular expression to parse host http header hawk utils parsehost which was subject to regular expression dos attack meaning each added character in the attacker s input increases the computation time exponentially parsehost was patched in to use built in url class to parse hostname instead hawk authenticate accepts options argument if that contains host and port those would be used instead of a call to utils parsehost publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hawk step up your open source security game with whitesource
0
61,229
14,619,108,739
IssuesEvent
2020-12-22 17:17:17
kenferrara/calcite
https://api.github.com/repos/kenferrara/calcite
opened
CVE-2012-5783 (Medium) detected in commons-httpclient-3.1.jar
security vulnerability
## CVE-2012-5783 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-httpclient-3.1.jar</b></p></summary> <p>The HttpClient component supports the client-side of RFC 1945 (HTTP/1.0) and RFC 2616 (HTTP/1.1) , several related specifications (RFC 2109 (Cookies) , RFC 2617 (HTTP Authentication) , etc.), and provides a framework by which new request types (methods) or HTTP extensions can be created easily.</p> <p>Path to dependency file: calcite/pig/build.gradle.kts</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-httpclient/commons-httpclient/3.1/964cd74171f427720480efdec40a7c7f6e58426a/commons-httpclient-3.1.jar</p> <p> Dependency Hierarchy: - pig-0.16.0.jar (Root Library) - :x: **commons-httpclient-3.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kenferrara/calcite/commit/5e2bb91de24b843d35df04b29fe7158fad414ee0">5e2bb91de24b843d35df04b29fe7158fad414ee0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Commons HttpClient 3.x, as used in Amazon Flexible Payments Service (FPS) merchant Java SDK and other products, does not verify that the server hostname matches a domain name in the subject's Common Name (CN) or subjectAltName field of the X.509 certificate, which allows man-in-the-middle attackers to spoof SSL servers via an arbitrary valid certificate. <p>Publish Date: 2012-11-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-5783>CVE-2012-5783</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.8</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://xforce.iss.net/xforce/xfdb/79984">http://xforce.iss.net/xforce/xfdb/79984</a></p> <p>Release Date: 2017-12-31</p> <p>Fix Resolution: Apply the appropriate patch for your system. See References.</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-httpclient","packageName":"commons-httpclient","packageVersion":"3.1","isTransitiveDependency":true,"dependencyTree":"org.apache.pig:pig:0.16.0;commons-httpclient:commons-httpclient:3.1","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2012-5783","vulnerabilityDetails":"Apache Commons HttpClient 3.x, as used in Amazon Flexible Payments Service (FPS) merchant Java SDK and other products, does not verify that the server hostname matches a domain name in the subject\u0027s Common Name (CN) or subjectAltName field of the X.509 certificate, which allows man-in-the-middle attackers to spoof SSL servers via an arbitrary valid certificate.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-5783","cvss2Severity":"medium","cvss2Score":"5.8","extraData":{}}</REMEDIATE> -->
True
CVE-2012-5783 (Medium) detected in commons-httpclient-3.1.jar - ## CVE-2012-5783 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-httpclient-3.1.jar</b></p></summary> <p>The HttpClient component supports the client-side of RFC 1945 (HTTP/1.0) and RFC 2616 (HTTP/1.1) , several related specifications (RFC 2109 (Cookies) , RFC 2617 (HTTP Authentication) , etc.), and provides a framework by which new request types (methods) or HTTP extensions can be created easily.</p> <p>Path to dependency file: calcite/pig/build.gradle.kts</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-httpclient/commons-httpclient/3.1/964cd74171f427720480efdec40a7c7f6e58426a/commons-httpclient-3.1.jar</p> <p> Dependency Hierarchy: - pig-0.16.0.jar (Root Library) - :x: **commons-httpclient-3.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kenferrara/calcite/commit/5e2bb91de24b843d35df04b29fe7158fad414ee0">5e2bb91de24b843d35df04b29fe7158fad414ee0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Commons HttpClient 3.x, as used in Amazon Flexible Payments Service (FPS) merchant Java SDK and other products, does not verify that the server hostname matches a domain name in the subject's Common Name (CN) or subjectAltName field of the X.509 certificate, which allows man-in-the-middle attackers to spoof SSL servers via an arbitrary valid certificate. <p>Publish Date: 2012-11-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-5783>CVE-2012-5783</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.8</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://xforce.iss.net/xforce/xfdb/79984">http://xforce.iss.net/xforce/xfdb/79984</a></p> <p>Release Date: 2017-12-31</p> <p>Fix Resolution: Apply the appropriate patch for your system. See References.</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-httpclient","packageName":"commons-httpclient","packageVersion":"3.1","isTransitiveDependency":true,"dependencyTree":"org.apache.pig:pig:0.16.0;commons-httpclient:commons-httpclient:3.1","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2012-5783","vulnerabilityDetails":"Apache Commons HttpClient 3.x, as used in Amazon Flexible Payments Service (FPS) merchant Java SDK and other products, does not verify that the server hostname matches a domain name in the subject\u0027s Common Name (CN) or subjectAltName field of the X.509 certificate, which allows man-in-the-middle attackers to spoof SSL servers via an arbitrary valid certificate.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-5783","cvss2Severity":"medium","cvss2Score":"5.8","extraData":{}}</REMEDIATE> -->
non_defect
cve medium detected in commons httpclient jar cve medium severity vulnerability vulnerable library commons httpclient jar the httpclient component supports the client side of rfc http and rfc http several related specifications rfc cookies rfc http authentication etc and provides a framework by which new request types methods or http extensions can be created easily path to dependency file calcite pig build gradle kts path to vulnerable library home wss scanner gradle caches modules files commons httpclient commons httpclient commons httpclient jar dependency hierarchy pig jar root library x commons httpclient jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons httpclient x as used in amazon flexible payments service fps merchant java sdk and other products does not verify that the server hostname matches a domain name in the subject s common name cn or subjectaltname field of the x certificate which allows man in the middle attackers to spoof ssl servers via an arbitrary valid certificate publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution apply the appropriate patch for your system see references isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails apache commons httpclient x as used in amazon flexible payments service fps merchant java sdk and other products does not verify that the server hostname matches a domain name in the subject common name cn or subjectaltname field of the x certificate which allows man in the middle attackers to spoof ssl servers via an arbitrary valid certificate vulnerabilityurl
0
79,707
28,498,450,255
IssuesEvent
2023-04-18 15:36:31
vector-im/element-call
https://api.github.com/repos/vector-im/element-call
opened
Join with muted video and muted audio
T-Defect
### Steps to reproduce 1. Join a call with User A no muted audio and no muted video 2. Join same call with User B muted audio and muted video x. I bug not happen, than switch and join with User A muted and User B not muted. Than Bug will happen. ![Bildschirmfoto 2023-04-18 um 17 04 39](https://user-images.githubusercontent.com/1362436/232828772-cf67fef0-7822-430f-8bd9-f1e979bb51f6.png) ### Outcome #### What did you expect? User B can hear and See User A and User A can not see and hear User B. #### What happened instead? User A and User B can not see each other. ### Operating system macOS ### Browser information Chrome lates version ### URL for webapp call.element.io ### Will you send logs? No
1.0
Join with muted video and muted audio - ### Steps to reproduce 1. Join a call with User A no muted audio and no muted video 2. Join same call with User B muted audio and muted video x. I bug not happen, than switch and join with User A muted and User B not muted. Than Bug will happen. ![Bildschirmfoto 2023-04-18 um 17 04 39](https://user-images.githubusercontent.com/1362436/232828772-cf67fef0-7822-430f-8bd9-f1e979bb51f6.png) ### Outcome #### What did you expect? User B can hear and See User A and User A can not see and hear User B. #### What happened instead? User A and User B can not see each other. ### Operating system macOS ### Browser information Chrome lates version ### URL for webapp call.element.io ### Will you send logs? No
defect
join with muted video and muted audio steps to reproduce join a call with user a no muted audio and no muted video join same call with user b muted audio and muted video x i bug not happen than switch and join with user a muted and user b not muted than bug will happen outcome what did you expect user b can hear and see user a and user a can not see and hear user b what happened instead user a and user b can not see each other operating system macos browser information chrome lates version url for webapp call element io will you send logs no
1
21,861
3,573,522,106
IssuesEvent
2016-01-27 07:02:03
hunnee/jsontoken
https://api.github.com/repos/hunnee/jsontoken
closed
Cannot check out the project in eclipse
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Import project to Eclipse 2. Specify the url of repository: http://jsontoken.googlecode.com/svn/ 3. Start check out What is the expected output? What do you see instead? the project is expected to be imported in Eclipse.Instead the check out process get stuck in the file "FakeClock.java" What version of the product are you using? On what operating system? Windows 7, eclipse indigo service release 2 Please provide any additional information below. ``` Original issue reported on code.google.com by `panipsi...@gmail.com` on 13 Jun 2013 at 12:00
1.0
Cannot check out the project in eclipse - ``` What steps will reproduce the problem? 1. Import project to Eclipse 2. Specify the url of repository: http://jsontoken.googlecode.com/svn/ 3. Start check out What is the expected output? What do you see instead? the project is expected to be imported in Eclipse.Instead the check out process get stuck in the file "FakeClock.java" What version of the product are you using? On what operating system? Windows 7, eclipse indigo service release 2 Please provide any additional information below. ``` Original issue reported on code.google.com by `panipsi...@gmail.com` on 13 Jun 2013 at 12:00
defect
cannot check out the project in eclipse what steps will reproduce the problem import project to eclipse specify the url of repository start check out what is the expected output what do you see instead the project is expected to be imported in eclipse instead the check out process get stuck in the file fakeclock java what version of the product are you using on what operating system windows eclipse indigo service release please provide any additional information below original issue reported on code google com by panipsi gmail com on jun at
1
277,843
8,633,367,834
IssuesEvent
2018-11-22 13:40:19
geosolutions-it/pyfulcrum
https://api.github.com/repos/geosolutions-it/pyfulcrum
closed
PyBackup entrypoint
Priority: Medium Task review
Develop a specific API to send Fulcrum's Records to the PyBackup module. An asynchronous mechanims needs to be included at this stage in order to grant high availability of the endpoint service, and reduce (throttle) downstream requests toward Fulcrum APIs which are needed to access and validate webhooks payloads. Needed functionalities: - [x] Module API ~~- [ ] Asynchronous mechanism to manage the incoming records and grant high availability~~ (commented out by https://github.com/geosolutions-it/pyfulcrum/issues/3#issuecomment-431051535)
1.0
PyBackup entrypoint - Develop a specific API to send Fulcrum's Records to the PyBackup module. An asynchronous mechanims needs to be included at this stage in order to grant high availability of the endpoint service, and reduce (throttle) downstream requests toward Fulcrum APIs which are needed to access and validate webhooks payloads. Needed functionalities: - [x] Module API ~~- [ ] Asynchronous mechanism to manage the incoming records and grant high availability~~ (commented out by https://github.com/geosolutions-it/pyfulcrum/issues/3#issuecomment-431051535)
non_defect
pybackup entrypoint develop a specific api to send fulcrum s records to the pybackup module an asynchronous mechanims needs to be included at this stage in order to grant high availability of the endpoint service and reduce throttle downstream requests toward fulcrum apis which are needed to access and validate webhooks payloads needed functionalities module api asynchronous mechanism to manage the incoming records and grant high availability commented out by
0
64,495
18,717,058,353
IssuesEvent
2021-11-03 07:05:36
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
closed
[🐛 Bug]: AttributeError: 'dict' object has no attribute 'send_keys'
I-defect needs-triaging
### What happened? I just start the browser and find en element using Python: ```python driver = webdriver.Chrome() driver.implicitly_wait(10) driver.get('https://account.xiaomi.com/fe/service/login/password') el = driver.find_element(By.NAME, 'account') el.send_keys('demo') ``` el should be an webelement object or NoSuchElementException, but I got this: `AttributeError: 'dict' object has no attribute 'send_keys'` ![image](https://user-images.githubusercontent.com/17780176/139525034-313b80c7-3a84-4ea3-b34d-ae0eddd45a60.png) ### How can we reproduce the issue? ```shell above ``` ### Relevant log output ```shell FAILED [100%] test_selenum.py:12 (test_selenium) def test_selenium(): """测试selenium是否可用""" driver = webdriver.Chrome() driver.implicitly_wait(10) # 访问被测试网址 driver.get('https://account.xiaomi.com/fe/service/login/password') el = driver.find_element(By.NAME, 'account') > el.send_keys('aaa') E AttributeError: 'dict' object has no attribute 'send_keys' test_selenum.py:20: AttributeError ``` ### Operating System Windows10 ### Selenium version 4.0.0 ### What are the browser(s) and version(s) where you see this issue? Chrome90 ### What are the browser driver(s) and version(s) where you see this issue? 90 ### Are you using Selenium Grid? _No response_
1.0
[🐛 Bug]: AttributeError: 'dict' object has no attribute 'send_keys' - ### What happened? I just start the browser and find en element using Python: ```python driver = webdriver.Chrome() driver.implicitly_wait(10) driver.get('https://account.xiaomi.com/fe/service/login/password') el = driver.find_element(By.NAME, 'account') el.send_keys('demo') ``` el should be an webelement object or NoSuchElementException, but I got this: `AttributeError: 'dict' object has no attribute 'send_keys'` ![image](https://user-images.githubusercontent.com/17780176/139525034-313b80c7-3a84-4ea3-b34d-ae0eddd45a60.png) ### How can we reproduce the issue? ```shell above ``` ### Relevant log output ```shell FAILED [100%] test_selenum.py:12 (test_selenium) def test_selenium(): """测试selenium是否可用""" driver = webdriver.Chrome() driver.implicitly_wait(10) # 访问被测试网址 driver.get('https://account.xiaomi.com/fe/service/login/password') el = driver.find_element(By.NAME, 'account') > el.send_keys('aaa') E AttributeError: 'dict' object has no attribute 'send_keys' test_selenum.py:20: AttributeError ``` ### Operating System Windows10 ### Selenium version 4.0.0 ### What are the browser(s) and version(s) where you see this issue? Chrome90 ### What are the browser driver(s) and version(s) where you see this issue? 90 ### Are you using Selenium Grid? _No response_
defect
attributeerror dict object has no attribute send keys what happened i just start the browser and find en element using python python driver webdriver chrome driver implicitly wait driver get el driver find element by name account el send keys demo el should be an webelement object or nosuchelementexception, but i got this attributeerror dict object has no attribute send keys how can we reproduce the issue shell above relevant log output shell failed test selenum py test selenium def test selenium 测试selenium是否可用 driver webdriver chrome driver implicitly wait 访问被测试网址 driver get el driver find element by name account el send keys aaa e attributeerror dict object has no attribute send keys test selenum py attributeerror operating system selenium version what are the browser s and version s where you see this issue what are the browser driver s and version s where you see this issue are you using selenium grid no response
1
97,403
11,012,717,866
IssuesEvent
2019-12-04 18:53:14
SUSE/kubecf
https://api.github.com/repos/SUSE/kubecf
opened
Reusing an external database after reinstalling kubecf does not work
Priority: High Type: Bug Type: Documentation
**Describe the bug** On an attempt to reuse the external database, a kubecf reinstall won't work. **To Reproduce** - Deploy kubecf using an external database. - Run smoke-tests. - Delete kubecf installation. - Deploy kubecf again without changing any values. - Run smoke-tests - it will fail. **Expected behavior** Given the proper values, the smoke-tests on the reinstallation should work.
1.0
Reusing an external database after reinstalling kubecf does not work - **Describe the bug** On an attempt to reuse the external database, a kubecf reinstall won't work. **To Reproduce** - Deploy kubecf using an external database. - Run smoke-tests. - Delete kubecf installation. - Deploy kubecf again without changing any values. - Run smoke-tests - it will fail. **Expected behavior** Given the proper values, the smoke-tests on the reinstallation should work.
non_defect
reusing an external database after reinstalling kubecf does not work describe the bug on an attempt to reuse the external database a kubecf reinstall won t work to reproduce deploy kubecf using an external database run smoke tests delete kubecf installation deploy kubecf again without changing any values run smoke tests it will fail expected behavior given the proper values the smoke tests on the reinstallation should work
0
4,882
2,610,159,596
IssuesEvent
2015-02-26 18:50:39
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Text
auto-migrated Priority-Medium Type-Defect
``` Tech 1 text screen display is missing. Check tech levels for the GC and check the code and text file ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 31 Jan 2011 at 1:34
1.0
Text - ``` Tech 1 text screen display is missing. Check tech levels for the GC and check the code and text file ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 31 Jan 2011 at 1:34
defect
text tech text screen display is missing check tech levels for the gc and check the code and text file original issue reported on code google com by gmail com on jan at
1
22,236
15,055,896,462
IssuesEvent
2021-02-03 19:25:45
commitdev/zero
https://api.github.com/repos/commitdev/zero
opened
bug: user-auth namespace already exist
bug infrastructure
when terraform applies, maybe because of the secret creation the user-auth namespace is already created terraform errors out because it also tries to create it, we need to either change the https://github.com/commitdev/zero-aws-eks-stack/blob/main/templates/kubernetes/terraform/modules/kubernetes/user_auth.tf#L21 to be a `data` instead of `resource`
1.0
bug: user-auth namespace already exist - when terraform applies, maybe because of the secret creation the user-auth namespace is already created terraform errors out because it also tries to create it, we need to either change the https://github.com/commitdev/zero-aws-eks-stack/blob/main/templates/kubernetes/terraform/modules/kubernetes/user_auth.tf#L21 to be a `data` instead of `resource`
non_defect
bug user auth namespace already exist when terraform applies maybe because of the secret creation the user auth namespace is already created terraform errors out because it also tries to create it we need to either change the to be a data instead of resource
0
228,841
25,262,809,724
IssuesEvent
2022-11-16 00:45:38
AlexRogalskiy/roadmap
https://api.github.com/repos/AlexRogalskiy/roadmap
closed
CVE-2019-16943 (High) detected in jackson-databind-2.1.3.jar
security vulnerability no-issue-activity
## CVE-2019-16943 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.1.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Path to dependency file: /modules/roadmap-all/pom.xml</p> <p>Path to vulnerable library: /tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar,/tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar,/tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar</p> <p> Dependency Hierarchy: - logback-jackson-0.1.5.jar (Root Library) - :x: **jackson-databind-2.1.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/roadmap/commit/fff78511c51c60993515e1063ec6b51834154cd1">fff78511c51c60993515e1063ec6b51834154cd1</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the p6spy (3.8.6) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of com.p6spy.engine.spy.P6DataSource mishandling. <p>Publish Date: 2019-10-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-16943>CVE-2019-16943</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943</a></p> <p>Release Date: 2019-10-01</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-16943 (High) detected in jackson-databind-2.1.3.jar - ## CVE-2019-16943 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.1.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Path to dependency file: /modules/roadmap-all/pom.xml</p> <p>Path to vulnerable library: /tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar,/tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar,/tmp/ws-ua_20221108020925_RXRGLZ/downloadResource_PJGKYN/20221108021017/jackson-databind-2.1.3.jar</p> <p> Dependency Hierarchy: - logback-jackson-0.1.5.jar (Root Library) - :x: **jackson-databind-2.1.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/roadmap/commit/fff78511c51c60993515e1063ec6b51834154cd1">fff78511c51c60993515e1063ec6b51834154cd1</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the p6spy (3.8.6) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of com.p6spy.engine.spy.P6DataSource mishandling. <p>Publish Date: 2019-10-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-16943>CVE-2019-16943</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943</a></p> <p>Release Date: 2019-10-01</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to dependency file modules roadmap all pom xml path to vulnerable library tmp ws ua rxrglz downloadresource pjgkyn jackson databind jar tmp ws ua rxrglz downloadresource pjgkyn jackson databind jar tmp ws ua rxrglz downloadresource pjgkyn jackson databind jar dependency hierarchy logback jackson jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the jar in the classpath and an attacker can find an rmi service endpoint to access it is possible to make the service execute a malicious payload this issue exists because of com engine spy mishandling publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with mend
0
337
2,532,788,858
IssuesEvent
2015-01-23 18:30:11
google/error-prone
https://api.github.com/repos/google/error-prone
closed
IJ plugin compilation fails with GC overhead limit exceeded
migrated Priority-Medium Type-Defect
_[Original issue](https://code.google.com/p/error-prone/issues/detail?id=250) created by **alexeagle@google.com** on 2014-04-28 at 06:23 PM_ --- <b>What steps will reproduce the problem?</b> 1. Set up an IntelliJ IDEA project with a very large number of sources 2. Use error-prone compiler plugin Information:Using error-prone compiler to compile java sources Information:Module "java" was fully rebuilt due to project configuration/dependencies changes Information:Compilation completed with 1 error and 70 warnings in 1 min 34 sec Information:1 error Information:70 warnings Error:java: java.lang.OutOfMemoryError: GC overhead limit exceeded
1.0
IJ plugin compilation fails with GC overhead limit exceeded - _[Original issue](https://code.google.com/p/error-prone/issues/detail?id=250) created by **alexeagle@google.com** on 2014-04-28 at 06:23 PM_ --- <b>What steps will reproduce the problem?</b> 1. Set up an IntelliJ IDEA project with a very large number of sources 2. Use error-prone compiler plugin Information:Using error-prone compiler to compile java sources Information:Module "java" was fully rebuilt due to project configuration/dependencies changes Information:Compilation completed with 1 error and 70 warnings in 1 min 34 sec Information:1 error Information:70 warnings Error:java: java.lang.OutOfMemoryError: GC overhead limit exceeded
defect
ij plugin compilation fails with gc overhead limit exceeded created by alexeagle google com on at pm what steps will reproduce the problem set up an intellij idea project with a very large number of sources use error prone compiler plugin information using error prone compiler to compile java sources information module java was fully rebuilt due to project configuration dependencies changes information compilation completed with error and warnings in min sec information error information warnings error java java lang outofmemoryerror gc overhead limit exceeded
1
40,865
10,198,144,353
IssuesEvent
2019-08-13 03:47:54
Psionics-ToEE/PsionicFiles
https://api.github.com/repos/Psionics-ToEE/PsionicFiles
opened
Mind blade weapon feat double dipping and dual wield bonus apply incorrectly.
defect low weight
Due to the workaround involving using the sickle type as the proto weapon type for mind blades, you can "double dip" on weapon feats like weapon focus to take the sickle versions for extra bonuses to the mind blade, when only the mind blade versions should apply. Detecting those sickle versions and not applying the mind blade versions could be done, but most people wouldn't waste a feat on a measly +1 attack bonus or on sickle since there is no indication of this in game. Additionally, due to sickle classification, two weapon fighting bonus for light weapon in offhand works correctly still if the offhand is mind blade, however if it happens to be the LS or BS version of the mind blade (shape mind blade), the bonus would still apply when it should not. Depending on implementation of shape mind blade, it may be impossible to get a LS or BS mind blade in the secondary hand anyway.
1.0
Mind blade weapon feat double dipping and dual wield bonus apply incorrectly. - Due to the workaround involving using the sickle type as the proto weapon type for mind blades, you can "double dip" on weapon feats like weapon focus to take the sickle versions for extra bonuses to the mind blade, when only the mind blade versions should apply. Detecting those sickle versions and not applying the mind blade versions could be done, but most people wouldn't waste a feat on a measly +1 attack bonus or on sickle since there is no indication of this in game. Additionally, due to sickle classification, two weapon fighting bonus for light weapon in offhand works correctly still if the offhand is mind blade, however if it happens to be the LS or BS version of the mind blade (shape mind blade), the bonus would still apply when it should not. Depending on implementation of shape mind blade, it may be impossible to get a LS or BS mind blade in the secondary hand anyway.
defect
mind blade weapon feat double dipping and dual wield bonus apply incorrectly due to the workaround involving using the sickle type as the proto weapon type for mind blades you can double dip on weapon feats like weapon focus to take the sickle versions for extra bonuses to the mind blade when only the mind blade versions should apply detecting those sickle versions and not applying the mind blade versions could be done but most people wouldn t waste a feat on a measly attack bonus or on sickle since there is no indication of this in game additionally due to sickle classification two weapon fighting bonus for light weapon in offhand works correctly still if the offhand is mind blade however if it happens to be the ls or bs version of the mind blade shape mind blade the bonus would still apply when it should not depending on implementation of shape mind blade it may be impossible to get a ls or bs mind blade in the secondary hand anyway
1
10,680
2,622,180,257
IssuesEvent
2015-03-04 00:18:30
byzhang/leveldb
https://api.github.com/repos/byzhang/leveldb
closed
The C API lacks a binary safe leveldb_get() function.
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? My C program that is using LevelDB for storing binary blobs and I ran into the following problem with the LevelDB C API: 1)Leveldb_put() allows the programmer to pass a char* and a length for both the "key" and "val", which allows the programmer to store binary data, which may contain null bytes. 2)leveldb_get() returns a char* without a size parameter... Digging into ./leveldb/c.cc you can see that it is using a "std:string" to return this char*... Not only is this not binary safe, but std::string? Having an efficient slice, like the one provided in the C++ api would be much preferred. So in short; using the C API its impossible to put() items without ever being able to get() them back... The solution is to provide a leveldb_get_binary() which allows the programmer to pass in a reference to char* and size param. The call convention should be consistent with the leveldb_put() function. What is the expected output? What do you see instead? I expected to obtain an unmolested blob of binary, I got a null terminated string. What version of the product are you using? On what operating system? libleveldb-dev:amd64 0+20120530.gitdd0d562-2 installed with apt, under Ubuntu. ``` Original issue reported on code.google.com by `firealwa...@gmail.com` on 26 Feb 2013 at 9:24
1.0
The C API lacks a binary safe leveldb_get() function. - ``` What steps will reproduce the problem? My C program that is using LevelDB for storing binary blobs and I ran into the following problem with the LevelDB C API: 1)Leveldb_put() allows the programmer to pass a char* and a length for both the "key" and "val", which allows the programmer to store binary data, which may contain null bytes. 2)leveldb_get() returns a char* without a size parameter... Digging into ./leveldb/c.cc you can see that it is using a "std:string" to return this char*... Not only is this not binary safe, but std::string? Having an efficient slice, like the one provided in the C++ api would be much preferred. So in short; using the C API its impossible to put() items without ever being able to get() them back... The solution is to provide a leveldb_get_binary() which allows the programmer to pass in a reference to char* and size param. The call convention should be consistent with the leveldb_put() function. What is the expected output? What do you see instead? I expected to obtain an unmolested blob of binary, I got a null terminated string. What version of the product are you using? On what operating system? libleveldb-dev:amd64 0+20120530.gitdd0d562-2 installed with apt, under Ubuntu. ``` Original issue reported on code.google.com by `firealwa...@gmail.com` on 26 Feb 2013 at 9:24
defect
the c api lacks a binary safe leveldb get function what steps will reproduce the problem my c program that is using leveldb for storing binary blobs and i ran into the following problem with the leveldb c api leveldb put allows the programmer to pass a char and a length for both the key and val which allows the programmer to store binary data which may contain null bytes leveldb get returns a char without a size parameter digging into leveldb c cc you can see that it is using a std string to return this char not only is this not binary safe but std string having an efficient slice like the one provided in the c api would be much preferred so in short using the c api its impossible to put items without ever being able to get them back the solution is to provide a leveldb get binary which allows the programmer to pass in a reference to char and size param the call convention should be consistent with the leveldb put function what is the expected output what do you see instead i expected to obtain an unmolested blob of binary i got a null terminated string what version of the product are you using on what operating system libleveldb dev installed with apt under ubuntu original issue reported on code google com by firealwa gmail com on feb at
1
6,665
9,682,769,815
IssuesEvent
2019-05-23 09:53:38
goharbor/harbor
https://api.github.com/repos/goharbor/harbor
closed
Allow Robot users to upload Helm Charts
candidate/1.9.0 kind/requirement
**Is your feature request related to a problem? Please describe.** I've integrated Harbor into my CI workflow, and automatically push Images via the Robot user. I also create Helm charts in CI, but to upload those, I have to have a second user with another set of credentials, etc. **Describe the solution you'd like** An option to allow a Robot user to upload Helm Charts. **Describe the main design/architecture of your solution** A simple mockup: ![image](https://user-images.githubusercontent.com/1932513/56902701-1e292980-6a9b-11e9-959b-673687938365.png)
1.0
Allow Robot users to upload Helm Charts - **Is your feature request related to a problem? Please describe.** I've integrated Harbor into my CI workflow, and automatically push Images via the Robot user. I also create Helm charts in CI, but to upload those, I have to have a second user with another set of credentials, etc. **Describe the solution you'd like** An option to allow a Robot user to upload Helm Charts. **Describe the main design/architecture of your solution** A simple mockup: ![image](https://user-images.githubusercontent.com/1932513/56902701-1e292980-6a9b-11e9-959b-673687938365.png)
non_defect
allow robot users to upload helm charts is your feature request related to a problem please describe i ve integrated harbor into my ci workflow and automatically push images via the robot user i also create helm charts in ci but to upload those i have to have a second user with another set of credentials etc describe the solution you d like an option to allow a robot user to upload helm charts describe the main design architecture of your solution a simple mockup
0
685,252
23,449,562,884
IssuesEvent
2022-08-16 00:13:32
NuGet/Home
https://api.github.com/repos/NuGet/Home
closed
[MacOS, Linux] NuGet writes the NuGet.Config in ~/.config folder, but dotnet restore reads from ~/.nuget
Priority:2 Type:DCR Product:NuGet.exe Platform:Xplat Area:Settings Pipeline:In Progress Category:Customer Sprint
When using NuGet to add a private feed, `NuGet Sources Add` writes in `~/.config/NuGet/NuGet.Config` file. But when using `dotnet restore` the `~/.nuget/NuGet/NuGet.Config` file is used to read the `packageSources` and `packageSourceCredentials`. This issue forces users to manually copy and paste all necessary information instead of having that generated automatically by the NuGet CLI. DotNetCore repo issue: https://github.com/dotnet/core/issues/453
1.0
[MacOS, Linux] NuGet writes the NuGet.Config in ~/.config folder, but dotnet restore reads from ~/.nuget - When using NuGet to add a private feed, `NuGet Sources Add` writes in `~/.config/NuGet/NuGet.Config` file. But when using `dotnet restore` the `~/.nuget/NuGet/NuGet.Config` file is used to read the `packageSources` and `packageSourceCredentials`. This issue forces users to manually copy and paste all necessary information instead of having that generated automatically by the NuGet CLI. DotNetCore repo issue: https://github.com/dotnet/core/issues/453
non_defect
nuget writes the nuget config in config folder but dotnet restore reads from nuget when using nuget to add a private feed nuget sources add writes in config nuget nuget config file but when using dotnet restore the nuget nuget nuget config file is used to read the packagesources and packagesourcecredentials this issue forces users to manually copy and paste all necessary information instead of having that generated automatically by the nuget cli dotnetcore repo issue
0
28,352
13,646,320,128
IssuesEvent
2020-09-25 22:50:21
microsoft/STL
https://api.github.com/repos/microsoft/STL
closed
Update OS minimum version and remove XP runtime detection
performance work in progress
Currently STL has to target Vista as minimum OS version for x86 and x64 Update `filesystem.cpp` https://github.com/microsoft/STL/blob/87dc1d3383a052216822fab1a37c8fe2794b0cdb/stl/src/filesystem.cpp#L56-L99 Update awint.hpp / winapisupp.cpp *Note*: the ABI compatibility for XP toolset should stay. The XP toolset should build; it just doesn't need to support Windows XP anymore.
True
Update OS minimum version and remove XP runtime detection - Currently STL has to target Vista as minimum OS version for x86 and x64 Update `filesystem.cpp` https://github.com/microsoft/STL/blob/87dc1d3383a052216822fab1a37c8fe2794b0cdb/stl/src/filesystem.cpp#L56-L99 Update awint.hpp / winapisupp.cpp *Note*: the ABI compatibility for XP toolset should stay. The XP toolset should build; it just doesn't need to support Windows XP anymore.
non_defect
update os minimum version and remove xp runtime detection currently stl has to target vista as minimum os version for and update filesystem cpp update awint hpp winapisupp cpp note the abi compatibility for xp toolset should stay the xp toolset should build it just doesn t need to support windows xp anymore
0
6,004
2,610,219,278
IssuesEvent
2015-02-26 19:09:42
chrsmith/somefinders
https://api.github.com/repos/chrsmith/somefinders
opened
озвучка avast
auto-migrated Priority-Medium Type-Defect
``` '''Богдан Овчинников''' Привет всем не подскажите где можно найти .озвучка avast. как то выкладывали уже '''Арам Котов''' Качай тут http://bit.ly/18TFMsK '''Алим Лобанов''' Спасибо вроде то но просит телефон вводить '''Баян Зыков''' Не это не влияет на баланс '''Арсений Родионов''' Не это не влияет на баланс Информация о файле: озвучка avast Загружен: В этом месяце Скачан раз: 932 Рейтинг: 1045 Средняя скорость скачивания: 1282 Похожих файлов: 39 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 12:37
1.0
озвучка avast - ``` '''Богдан Овчинников''' Привет всем не подскажите где можно найти .озвучка avast. как то выкладывали уже '''Арам Котов''' Качай тут http://bit.ly/18TFMsK '''Алим Лобанов''' Спасибо вроде то но просит телефон вводить '''Баян Зыков''' Не это не влияет на баланс '''Арсений Родионов''' Не это не влияет на баланс Информация о файле: озвучка avast Загружен: В этом месяце Скачан раз: 932 Рейтинг: 1045 Средняя скорость скачивания: 1282 Похожих файлов: 39 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 12:37
defect
озвучка avast богдан овчинников привет всем не подскажите где можно найти озвучка avast как то выкладывали уже арам котов качай тут алим лобанов спасибо вроде то но просит телефон вводить баян зыков не это не влияет на баланс арсений родионов не это не влияет на баланс информация о файле озвучка avast загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
1
149,934
23,550,923,210
IssuesEvent
2022-08-21 20:22:51
ParadoxGameConverters/Vic3ToHoI4
https://api.github.com/repos/ParadoxGameConverters/Vic3ToHoI4
opened
Convert state categories
enhancement coding design
Vic2 to HoI4 mostly just made them big enough for the existing industry. This should be re-examined, and should handle cases like islands. Maybe population based?
1.0
Convert state categories - Vic2 to HoI4 mostly just made them big enough for the existing industry. This should be re-examined, and should handle cases like islands. Maybe population based?
non_defect
convert state categories to mostly just made them big enough for the existing industry this should be re examined and should handle cases like islands maybe population based
0
25,585
7,727,653,109
IssuesEvent
2018-05-25 04:03:45
Microsoft/MixedRealityToolkit-Unity
https://api.github.com/repos/Microsoft/MixedRealityToolkit-Unity
reopened
Unable to Deploy from Build Window
Build Tools HoloLens
## Overview Unable to deploy to Hololens. ## Expected Behavior After successfully building a barebones project and building the Appx, the program deploys correctly to the Hololens ## Actual Behavior. Builds project correctly. Appears to connect to Hololens. Does not deploy to the Hololens when pressing "Install" ## Steps to reproduce The correct WiFi IP address is provided and when I press "Connect" under deploy options, I get the correct Hololens device name. When I press "Open Device Portal" - IE opens the Hololens Device Portal correctly without issue. When I press "Install" - I get a Network Error: Request timeout. Then "Failed to install [APP].appx on [DEVICE]. See screenshot. ![image](https://user-images.githubusercontent.com/13334553/37381755-2571fbba-26fc-11e8-9dde-a8afd6516688.png) ## Unity Editor Version 2017.3.1f1 ## Mixed Reality Toolkit Release Version 2017.2.1.2 ## Hololens OS Version 10.0.14393.2068 ## Windows OS Version (Development PC) 10.0.16299 Build 16299
1.0
Unable to Deploy from Build Window - ## Overview Unable to deploy to Hololens. ## Expected Behavior After successfully building a barebones project and building the Appx, the program deploys correctly to the Hololens ## Actual Behavior. Builds project correctly. Appears to connect to Hololens. Does not deploy to the Hololens when pressing "Install" ## Steps to reproduce The correct WiFi IP address is provided and when I press "Connect" under deploy options, I get the correct Hololens device name. When I press "Open Device Portal" - IE opens the Hololens Device Portal correctly without issue. When I press "Install" - I get a Network Error: Request timeout. Then "Failed to install [APP].appx on [DEVICE]. See screenshot. ![image](https://user-images.githubusercontent.com/13334553/37381755-2571fbba-26fc-11e8-9dde-a8afd6516688.png) ## Unity Editor Version 2017.3.1f1 ## Mixed Reality Toolkit Release Version 2017.2.1.2 ## Hololens OS Version 10.0.14393.2068 ## Windows OS Version (Development PC) 10.0.16299 Build 16299
non_defect
unable to deploy from build window overview unable to deploy to hololens expected behavior after successfully building a barebones project and building the appx the program deploys correctly to the hololens actual behavior builds project correctly appears to connect to hololens does not deploy to the hololens when pressing install steps to reproduce the correct wifi ip address is provided and when i press connect under deploy options i get the correct hololens device name when i press open device portal ie opens the hololens device portal correctly without issue when i press install i get a network error request timeout then failed to install appx on see screenshot unity editor version mixed reality toolkit release version hololens os version windows os version development pc build
0
57,962
16,194,482,020
IssuesEvent
2021-05-04 13:03:24
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
BigQuery cannot reference exprsessions from SELECT DISTINCT in ORDER BY
C: DB: BigQuery C: Functionality E: Enterprise Edition P: Medium T: Defect
This doesn't work in BigQuery: ```sql select distinct lower(TITLE) from TEST.T_BOOK order by lower(TITLE); ``` The error is: > SQL Error [100032] [HY000]: [Simba][BigQueryJDBCDriver](100032) Error executing query job. Message: ORDER BY clause expression references column TITLE which is not visible after SELECT DISTINCT at [3:16] A workaround could be to replace all `ORDER BY` expressions by the column index: ```sql select distinct lower(TITLE) from TEST.T_BOOK order by 1; ``` But that only works if the expressions match directly. It doesn't work for inputs like that: ```sql select distinct lower(TITLE) from TEST.T_BOOK order by reverse(lower(TITLE)); ```
1.0
BigQuery cannot reference exprsessions from SELECT DISTINCT in ORDER BY - This doesn't work in BigQuery: ```sql select distinct lower(TITLE) from TEST.T_BOOK order by lower(TITLE); ``` The error is: > SQL Error [100032] [HY000]: [Simba][BigQueryJDBCDriver](100032) Error executing query job. Message: ORDER BY clause expression references column TITLE which is not visible after SELECT DISTINCT at [3:16] A workaround could be to replace all `ORDER BY` expressions by the column index: ```sql select distinct lower(TITLE) from TEST.T_BOOK order by 1; ``` But that only works if the expressions match directly. It doesn't work for inputs like that: ```sql select distinct lower(TITLE) from TEST.T_BOOK order by reverse(lower(TITLE)); ```
defect
bigquery cannot reference exprsessions from select distinct in order by this doesn t work in bigquery sql select distinct lower title from test t book order by lower title the error is sql error error executing query job message order by clause expression references column title which is not visible after select distinct at a workaround could be to replace all order by expressions by the column index sql select distinct lower title from test t book order by but that only works if the expressions match directly it doesn t work for inputs like that sql select distinct lower title from test t book order by reverse lower title
1
7,323
2,610,363,453
IssuesEvent
2015-02-26 19:57:28
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
tumblr authentication failing?
auto-migrated Priority-Medium tumblr Type-Defect
``` What's the problem? In the past week or so, when I try to post to my tumblr blog via scribefire, I get an 'authentication failure' I haven't changed my username/password and verified that the correct credentials are cached. What browser are you using? Chrome ; but repros on firefox as well What version of ScribeFire are you running? ``` ----- Original issue reported on code.google.com by `vinod.va...@gmail.com` on 7 Sep 2012 at 4:47 * Merged into: #766
1.0
tumblr authentication failing? - ``` What's the problem? In the past week or so, when I try to post to my tumblr blog via scribefire, I get an 'authentication failure' I haven't changed my username/password and verified that the correct credentials are cached. What browser are you using? Chrome ; but repros on firefox as well What version of ScribeFire are you running? ``` ----- Original issue reported on code.google.com by `vinod.va...@gmail.com` on 7 Sep 2012 at 4:47 * Merged into: #766
defect
tumblr authentication failing what s the problem in the past week or so when i try to post to my tumblr blog via scribefire i get an authentication failure i haven t changed my username password and verified that the correct credentials are cached what browser are you using chrome but repros on firefox as well what version of scribefire are you running original issue reported on code google com by vinod va gmail com on sep at merged into
1
57,838
16,100,064,387
IssuesEvent
2021-04-27 08:10:55
line/armeria
https://api.github.com/repos/line/armeria
closed
'Failed to force-initialize 'io.netty.channel.unix.Socket' message while using the armeria webflux sample project.
defect
Hello, I am inquiring with an error while using the armeria webflux sample project. [https://github.com/line/armeria-examples/tree/master/spring-boot-webflux](https://github.com/line/armeria-examples/tree/master/spring-boot-webflux) When this sample project is executed without modification, 'Failed to force-initialize 'io.netty.channel.unix.Socket' message is displayed. ![image](https://user-images.githubusercontent.com/18299676/114808584-8c2f9e00-9de3-11eb-8602-4edaf0b77a4f.png) [armeria_log.txt](https://github.com/line/armeria/files/6314962/armeria_log.txt) There seems to be no problem with the use, but is this okay? I am trying to use armeria for the first time, but I am anxious because this message is printed from the beginning. If there is a problem, I hope it will be fixed.
1.0
'Failed to force-initialize 'io.netty.channel.unix.Socket' message while using the armeria webflux sample project. - Hello, I am inquiring with an error while using the armeria webflux sample project. [https://github.com/line/armeria-examples/tree/master/spring-boot-webflux](https://github.com/line/armeria-examples/tree/master/spring-boot-webflux) When this sample project is executed without modification, 'Failed to force-initialize 'io.netty.channel.unix.Socket' message is displayed. ![image](https://user-images.githubusercontent.com/18299676/114808584-8c2f9e00-9de3-11eb-8602-4edaf0b77a4f.png) [armeria_log.txt](https://github.com/line/armeria/files/6314962/armeria_log.txt) There seems to be no problem with the use, but is this okay? I am trying to use armeria for the first time, but I am anxious because this message is printed from the beginning. If there is a problem, I hope it will be fixed.
defect
failed to force initialize io netty channel unix socket message while using the armeria webflux sample project hello i am inquiring with an error while using the armeria webflux sample project when this sample project is executed without modification failed to force initialize io netty channel unix socket message is displayed there seems to be no problem with the use but is this okay i am trying to use armeria for the first time but i am anxious because this message is printed from the beginning if there is a problem i hope it will be fixed
1
26,560
4,759,670,767
IssuesEvent
2016-10-24 23:30:40
sukona/Grapevine
https://api.github.com/repos/sukona/Grapevine
reopened
ContentType is not correctly detected
defect Hacktoberfest Server
Hello, While testing this excellent framework, I ran into some issues with content type detection. Upon some introspection, I found the culprit: contentType might come richer than expected (either the mime type followed by the encoding or mime type coming as doubled). My fix was a one-liner in the ContentType.cs file. Maybe it helps others ... `public static ContentType FromString(this ContentType ct, string contentType)` `{` ` return string.IsNullOrWhiteSpace(contentType)` ` ? ContentType.DEFAULT` `//: Enum.GetValues(typeof(ContentType)).Cast<ContentType>().FirstOrDefault(t => t.ToValue().Equals(contentType));` ` : Enum.GetValues(typeof(ContentType)).Cast<ContentType>().FirstOrDefault(t => contentType.Contains(t.ToValue()));` `}` Best regards, Marian
1.0
ContentType is not correctly detected - Hello, While testing this excellent framework, I ran into some issues with content type detection. Upon some introspection, I found the culprit: contentType might come richer than expected (either the mime type followed by the encoding or mime type coming as doubled). My fix was a one-liner in the ContentType.cs file. Maybe it helps others ... `public static ContentType FromString(this ContentType ct, string contentType)` `{` ` return string.IsNullOrWhiteSpace(contentType)` ` ? ContentType.DEFAULT` `//: Enum.GetValues(typeof(ContentType)).Cast<ContentType>().FirstOrDefault(t => t.ToValue().Equals(contentType));` ` : Enum.GetValues(typeof(ContentType)).Cast<ContentType>().FirstOrDefault(t => contentType.Contains(t.ToValue()));` `}` Best regards, Marian
defect
contenttype is not correctly detected hello while testing this excellent framework i ran into some issues with content type detection upon some introspection i found the culprit contenttype might come richer than expected either the mime type followed by the encoding or mime type coming as doubled my fix was a one liner in the contenttype cs file maybe it helps others public static contenttype fromstring this contenttype ct string contenttype return string isnullorwhitespace contenttype contenttype default enum getvalues typeof contenttype cast firstordefault t t tovalue equals contenttype enum getvalues typeof contenttype cast firstordefault t contenttype contains t tovalue best regards marian
1
12,325
3,601,040,255
IssuesEvent
2016-02-03 09:35:10
mayuso/Datuk
https://api.github.com/repos/mayuso/Datuk
reopened
Examples
documentation
Datuk needs examples on how to use the API. Examples needed: * Excel + Data. (Missing) * Txt + Data. (Missing) * Classification. (Missing) * Clustering. (Missing) * Charts. (BarChart and XY done, PieChart missing).
1.0
Examples - Datuk needs examples on how to use the API. Examples needed: * Excel + Data. (Missing) * Txt + Data. (Missing) * Classification. (Missing) * Clustering. (Missing) * Charts. (BarChart and XY done, PieChart missing).
non_defect
examples datuk needs examples on how to use the api examples needed excel data missing txt data missing classification missing clustering missing charts barchart and xy done piechart missing
0
9,109
2,615,131,038
IssuesEvent
2015-03-01 06:01:01
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
closed
Error parsing a json response
auto-migrated Priority-Medium Type-Defect
``` Version of google-api-java-client (e.g. 1.5.0-beta)? google-oauth-client-java7 -> 1.13.1-beta google-http-client-gson -> 1.13.1-beta Java environment (e.g. Java 6, Android 2.3, App Engine)? Java 7u10 Describe the problem. i'm playing around with Dropbox Api, and i'm getting a json response like this one { "referral_link": "https://www.dropbox.com/referrals/r1a2n3d4m5s6t7", "display_name": "John P. User", "uid": 12345678, "country": "US", "quota_info": { "shared": 253738410565, "quota": 107374182400000, "normal": 680031877871 } } no problem sending/singing the req or getting the input stream from the response, is when i try to parse the response using something like this DropboxAccountInfo accountInfo = response.parseAs(DropboxAccountInfo.class); it actually does return a new instance of the DropboxAccountInfo class but all the field inside of it are null. this is my DropboxAccountInfo class: public class DropboxAccountInfo { private String uid; private String display_name; private String country; private String referral_link; private DropboxQuota quota_info; //Getters and setters not who do not necessary match with the variables names private class DropboxQuota { private long quota; private long normal; private long shared; //Getters and setters not who do not necessary match with the variables names } now, i'm using a GsonObjectParser and it's working fine, the problem seems to be on the ClassInfo.class when the constructor is running, the nameToFieldInfoMap map is never filled up and i think to that's the root of the entire problem. this map is not filled up because on the line 174 FieldInfo fieldInfo = FieldInfo.of(field); is always getting null, and this is because inside of the "of" method the variable field is not an enum and then check for the @Key annotation line (93), and it doesn't have that one cuz... just don't really need to, anyways, i don't see why if it doesn't contain the @key annotation this method should return null. i hope have explained me well enough How would you expect it to be fixed? The parseAs() method should return a new instance of the class but with all the values populated according to the json IN ADVANCE THANKS SO MUCH AND SORRY FOR MY BAD ENGLISH, if you need more information (or want to pair to track this down or whatever) please do not no hesitate on contact me (lcc.julio@gmail.com) ``` Original issue reported on code.google.com by `lcc.ju...@gmail.com` on 7 Jan 2013 at 5:21
1.0
Error parsing a json response - ``` Version of google-api-java-client (e.g. 1.5.0-beta)? google-oauth-client-java7 -> 1.13.1-beta google-http-client-gson -> 1.13.1-beta Java environment (e.g. Java 6, Android 2.3, App Engine)? Java 7u10 Describe the problem. i'm playing around with Dropbox Api, and i'm getting a json response like this one { "referral_link": "https://www.dropbox.com/referrals/r1a2n3d4m5s6t7", "display_name": "John P. User", "uid": 12345678, "country": "US", "quota_info": { "shared": 253738410565, "quota": 107374182400000, "normal": 680031877871 } } no problem sending/singing the req or getting the input stream from the response, is when i try to parse the response using something like this DropboxAccountInfo accountInfo = response.parseAs(DropboxAccountInfo.class); it actually does return a new instance of the DropboxAccountInfo class but all the field inside of it are null. this is my DropboxAccountInfo class: public class DropboxAccountInfo { private String uid; private String display_name; private String country; private String referral_link; private DropboxQuota quota_info; //Getters and setters not who do not necessary match with the variables names private class DropboxQuota { private long quota; private long normal; private long shared; //Getters and setters not who do not necessary match with the variables names } now, i'm using a GsonObjectParser and it's working fine, the problem seems to be on the ClassInfo.class when the constructor is running, the nameToFieldInfoMap map is never filled up and i think to that's the root of the entire problem. this map is not filled up because on the line 174 FieldInfo fieldInfo = FieldInfo.of(field); is always getting null, and this is because inside of the "of" method the variable field is not an enum and then check for the @Key annotation line (93), and it doesn't have that one cuz... just don't really need to, anyways, i don't see why if it doesn't contain the @key annotation this method should return null. i hope have explained me well enough How would you expect it to be fixed? The parseAs() method should return a new instance of the class but with all the values populated according to the json IN ADVANCE THANKS SO MUCH AND SORRY FOR MY BAD ENGLISH, if you need more information (or want to pair to track this down or whatever) please do not no hesitate on contact me (lcc.julio@gmail.com) ``` Original issue reported on code.google.com by `lcc.ju...@gmail.com` on 7 Jan 2013 at 5:21
defect
error parsing a json response version of google api java client e g beta google oauth client beta google http client gson beta java environment e g java android app engine java describe the problem i m playing around with dropbox api and i m getting a json response like this one referral link display name john p user uid country us quota info shared quota normal no problem sending singing the req or getting the input stream from the response is when i try to parse the response using something like this dropboxaccountinfo accountinfo response parseas dropboxaccountinfo class it actually does return a new instance of the dropboxaccountinfo class but all the field inside of it are null this is my dropboxaccountinfo class public class dropboxaccountinfo private string uid private string display name private string country private string referral link private dropboxquota quota info getters and setters not who do not necessary match with the variables names private class dropboxquota private long quota private long normal private long shared getters and setters not who do not necessary match with the variables names now i m using a gsonobjectparser and it s working fine the problem seems to be on the classinfo class when the constructor is running the nametofieldinfomap map is never filled up and i think to that s the root of the entire problem this map is not filled up because on the line fieldinfo fieldinfo fieldinfo of field is always getting null and this is because inside of the of method the variable field is not an enum and then check for the key annotation line and it doesn t have that one cuz just don t really need to anyways i don t see why if it doesn t contain the key annotation this method should return null i hope have explained me well enough how would you expect it to be fixed the parseas method should return a new instance of the class but with all the values populated according to the json in advance thanks so much and sorry for my bad english if you need more information or want to pair to track this down or whatever please do not no hesitate on contact me lcc julio gmail com original issue reported on code google com by lcc ju gmail com on jan at
1
79,713
28,500,180,645
IssuesEvent
2023-04-18 16:44:13
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
SelectCheckboxMenu: emptyLabel not considered as overlay-target and opens behind dialog
:lady_beetle: defect
### Describe the bug Dialog `bindEvents` checks for whether `mousedown` event target was `primefaces-overlay-target` and if not moves to top. This causes issues with clicks on `selectCheckboxMenu` `emptyLabel` which opens the menu/dropdown but dialog moves on top of it. When clicking on dropdown arrow or actual data label it works fine. `SelectCheckboxMenu` `emptyLabel` fails this check on `line 379`: https://github.com/primefaces/primefaces/blob/4e4430b40a2d1573137a975071edad1b3f953b66/primefaces/src/main/resources/META-INF/resources/primefaces/dialog/dialog.js#L377-L382 ### Reproducer Place selectCheckboxMenu inside popup and click on emptyLabel instead of dropdown arrow. Menu opens but container dialog moves on top right after ### Expected behavior Menu should open on top of the dialog ### PrimeFaces edition Elite ### PrimeFaces version 12.0.3 ### Theme _No response_ ### JSF implementation Mojarra ### JSF version 2.3 ### Java version 8 ### Browser(s) _No response_
1.0
SelectCheckboxMenu: emptyLabel not considered as overlay-target and opens behind dialog - ### Describe the bug Dialog `bindEvents` checks for whether `mousedown` event target was `primefaces-overlay-target` and if not moves to top. This causes issues with clicks on `selectCheckboxMenu` `emptyLabel` which opens the menu/dropdown but dialog moves on top of it. When clicking on dropdown arrow or actual data label it works fine. `SelectCheckboxMenu` `emptyLabel` fails this check on `line 379`: https://github.com/primefaces/primefaces/blob/4e4430b40a2d1573137a975071edad1b3f953b66/primefaces/src/main/resources/META-INF/resources/primefaces/dialog/dialog.js#L377-L382 ### Reproducer Place selectCheckboxMenu inside popup and click on emptyLabel instead of dropdown arrow. Menu opens but container dialog moves on top right after ### Expected behavior Menu should open on top of the dialog ### PrimeFaces edition Elite ### PrimeFaces version 12.0.3 ### Theme _No response_ ### JSF implementation Mojarra ### JSF version 2.3 ### Java version 8 ### Browser(s) _No response_
defect
selectcheckboxmenu emptylabel not considered as overlay target and opens behind dialog describe the bug dialog bindevents checks for whether mousedown event target was primefaces overlay target and if not moves to top this causes issues with clicks on selectcheckboxmenu emptylabel which opens the menu dropdown but dialog moves on top of it when clicking on dropdown arrow or actual data label it works fine selectcheckboxmenu emptylabel fails this check on line reproducer place selectcheckboxmenu inside popup and click on emptylabel instead of dropdown arrow menu opens but container dialog moves on top right after expected behavior menu should open on top of the dialog primefaces edition elite primefaces version theme no response jsf implementation mojarra jsf version java version browser s no response
1
75,173
25,568,397,449
IssuesEvent
2022-11-30 15:51:01
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
CMS sitewide alerts not visible to editors
Critical defect ⭐️ Sitewide CMS
## Describe the defect Admins are able to see sitewide alerts. However, our shields and deploy warnings are not visible to everyone else. Slack conversation https://dsva.slack.com/archives/CDHBKAL9W/p1669675816587709 This is incorrectly permed It should be set for anonymous AND authenticated ![image](https://user-images.githubusercontent.com/5752113/204650526-dc35f5d1-6d2c-4224-af40-dd7d1965e2a7.png) ## To Reproduce Steps to reproduce the behavior: 1. As an admin set a sitewide alert to display 2. Logout and login as a non-admin 3. See that you see no sitewide alert ## AC / Expected behavior Sitewide alerts must be visible to ALL users! ## Additional context Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days). ## Desktop (please complete the following information if relevant, or delete) - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [x] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
1.0
CMS sitewide alerts not visible to editors - ## Describe the defect Admins are able to see sitewide alerts. However, our shields and deploy warnings are not visible to everyone else. Slack conversation https://dsva.slack.com/archives/CDHBKAL9W/p1669675816587709 This is incorrectly permed It should be set for anonymous AND authenticated ![image](https://user-images.githubusercontent.com/5752113/204650526-dc35f5d1-6d2c-4224-af40-dd7d1965e2a7.png) ## To Reproduce Steps to reproduce the behavior: 1. As an admin set a sitewide alert to display 2. Logout and login as a non-admin 3. See that you see no sitewide alert ## AC / Expected behavior Sitewide alerts must be visible to ALL users! ## Additional context Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days). ## Desktop (please complete the following information if relevant, or delete) - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [x] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
defect
cms sitewide alerts not visible to editors describe the defect admins are able to see sitewide alerts however our shields and deploy warnings are not visible to everyone else slack conversation this is incorrectly permed it should be set for anonymous and authenticated to reproduce steps to reproduce the behavior as an admin set a sitewide alert to display logout and login as a non admin see that you see no sitewide alert ac expected behavior sitewide alerts must be visible to all users additional context add any other context about the problem here reach out to the product managers to determine if it should be escalated as critical prevents users from accomplishing their work with no known workaround and needs to be addressed within business days desktop please complete the following information if relevant or delete os browser version labels you can delete this section once it s complete issue type red defaults to defect cms subsystem green cms practice area blue cms workstream orange not needed for bug tickets cms supported product black cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
1
34,047
7,330,874,230
IssuesEvent
2018-03-05 11:27:07
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
Atribute visible from MenuItem is not work in SplitButton
defect
``` [ x ] bug report ``` **What is the motivation / use case for changing the behavior?** Atribute visible from MenuItem is not work in SplitButton **Please tell us about your environment:** Windows 10 IDE: Webstorm * **Angular version:** 5.2.1 * **PrimeNG version:** 5.2.0 * **Browser:** Versão 63.0.3239.132 (Versão oficial) 64 bits * **Language:** "typescript": "~2.6.1" "target": "es5", * **Node (for AoT issues):** v9.2.1
1.0
Atribute visible from MenuItem is not work in SplitButton - ``` [ x ] bug report ``` **What is the motivation / use case for changing the behavior?** Atribute visible from MenuItem is not work in SplitButton **Please tell us about your environment:** Windows 10 IDE: Webstorm * **Angular version:** 5.2.1 * **PrimeNG version:** 5.2.0 * **Browser:** Versão 63.0.3239.132 (Versão oficial) 64 bits * **Language:** "typescript": "~2.6.1" "target": "es5", * **Node (for AoT issues):** v9.2.1
defect
atribute visible from menuitem is not work in splitbutton bug report what is the motivation use case for changing the behavior atribute visible from menuitem is not work in splitbutton please tell us about your environment windows ide webstorm angular version primeng version browser versão versão oficial bits language typescript target node for aot issues
1
78,978
27,849,345,489
IssuesEvent
2023-03-20 17:32:42
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
Tagged residual being accumulated twice
C: Framework T: defect P: normal
## Bug Description MOOSE has the capability to save residual contributions from specific kernels using the [Tagging System](https://mooseframework.inl.gov/framework_development/tagging.html). You can accumulate the contribution for a specific tag by calling `NonlinearSystemBase::computeResidualTags(const std::set<TagID> & tags)`. This allows only the kernels associated with the given tags to be computed. However, it seems that tags that aren't requested are being accumulated by the `computeResidualTags` call. This is causing some unexpected behavior. ## Steps to Reproduce Let's say we have two calls to computeResidual with different requested tags: ```C++ _fe_problem.computeResidualTags({residual_tag_1}); _fe_problem.computeResidualTags({residual_tag_2}); ``` You will see that the vectors associated with `residual_tag_1` and `residual_tag_2` are accumulated in both of these calls. This is especially troublesome since we only zero and close the requested tags before and after the residual evaluation. ## Impact Preventing Griffin from converting `save_in` functionality to vector tagging.
1.0
Tagged residual being accumulated twice - ## Bug Description MOOSE has the capability to save residual contributions from specific kernels using the [Tagging System](https://mooseframework.inl.gov/framework_development/tagging.html). You can accumulate the contribution for a specific tag by calling `NonlinearSystemBase::computeResidualTags(const std::set<TagID> & tags)`. This allows only the kernels associated with the given tags to be computed. However, it seems that tags that aren't requested are being accumulated by the `computeResidualTags` call. This is causing some unexpected behavior. ## Steps to Reproduce Let's say we have two calls to computeResidual with different requested tags: ```C++ _fe_problem.computeResidualTags({residual_tag_1}); _fe_problem.computeResidualTags({residual_tag_2}); ``` You will see that the vectors associated with `residual_tag_1` and `residual_tag_2` are accumulated in both of these calls. This is especially troublesome since we only zero and close the requested tags before and after the residual evaluation. ## Impact Preventing Griffin from converting `save_in` functionality to vector tagging.
defect
tagged residual being accumulated twice bug description moose has the capability to save residual contributions from specific kernels using the you can accumulate the contribution for a specific tag by calling nonlinearsystembase computeresidualtags const std set tags this allows only the kernels associated with the given tags to be computed however it seems that tags that aren t requested are being accumulated by the computeresidualtags call this is causing some unexpected behavior steps to reproduce let s say we have two calls to computeresidual with different requested tags c fe problem computeresidualtags residual tag fe problem computeresidualtags residual tag you will see that the vectors associated with residual tag and residual tag are accumulated in both of these calls this is especially troublesome since we only zero and close the requested tags before and after the residual evaluation impact preventing griffin from converting save in functionality to vector tagging
1
60,055
17,023,322,800
IssuesEvent
2021-07-03 01:25:45
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Potlatch incorrectly handle GPX timestamps with fractions of second
Component: potlatch (flash editor) Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 10.26pm, Saturday, 15th November 2008]** I have uploaded a GPX file with track point timestamps going up to fractions of second. The file was imported with no errors, however when I press edit on the web site, Potlatch shows it to me as if I have the first point in the GPX being somewhere far away (sometimes north-west, sometimes north-east, depending on track but constant for a track). I checked manually, the first point is where it should be (no such big error). The GPX view (clicking on the file name in the list of GPX tracks) shows the track correctly, just the Potlatch does it wrong (though I haven't tested with JOSM or similar). For a sample look e.g. at this track: http://www.openstreetmap.org/user/dtbow/traces/250801 I have fixed in the meantime my application to not provide fractions of second, but in general, the GPX standard doesn't forbid doing it.
1.0
Potlatch incorrectly handle GPX timestamps with fractions of second - **[Submitted to the original trac issue database at 10.26pm, Saturday, 15th November 2008]** I have uploaded a GPX file with track point timestamps going up to fractions of second. The file was imported with no errors, however when I press edit on the web site, Potlatch shows it to me as if I have the first point in the GPX being somewhere far away (sometimes north-west, sometimes north-east, depending on track but constant for a track). I checked manually, the first point is where it should be (no such big error). The GPX view (clicking on the file name in the list of GPX tracks) shows the track correctly, just the Potlatch does it wrong (though I haven't tested with JOSM or similar). For a sample look e.g. at this track: http://www.openstreetmap.org/user/dtbow/traces/250801 I have fixed in the meantime my application to not provide fractions of second, but in general, the GPX standard doesn't forbid doing it.
defect
potlatch incorrectly handle gpx timestamps with fractions of second i have uploaded a gpx file with track point timestamps going up to fractions of second the file was imported with no errors however when i press edit on the web site potlatch shows it to me as if i have the first point in the gpx being somewhere far away sometimes north west sometimes north east depending on track but constant for a track i checked manually the first point is where it should be no such big error the gpx view clicking on the file name in the list of gpx tracks shows the track correctly just the potlatch does it wrong though i haven t tested with josm or similar for a sample look e g at this track i have fixed in the meantime my application to not provide fractions of second but in general the gpx standard doesn t forbid doing it
1
17,427
3,004,869,415
IssuesEvent
2015-07-26 11:47:53
emukidid/swiss-gc
https://api.github.com/repos/emukidid/swiss-gc
closed
Star fox Adventures no audio and wrong subtitles
auto-migrated Priority-Medium Type-Defect
``` What step(s) will reproduce the problem? 1.Play the game on Swiss. 2.Select Story. 3.Watch the cutscenes. What console are you running Swiss on (Wii/GC)? Wii What version of Swiss are you using? Swiss r298 What disc was in the drive if any? Wii Sports What device was the data being loaded from (SD/HDD/DVD/DVD-R)? SD Please provide any additional information below. When I play the story mode the voices of the character is mute and the subtitles are wrong. ``` Original issue reported on code.google.com by `diego.so...@gmail.com` on 28 Feb 2015 at 12:26
1.0
Star fox Adventures no audio and wrong subtitles - ``` What step(s) will reproduce the problem? 1.Play the game on Swiss. 2.Select Story. 3.Watch the cutscenes. What console are you running Swiss on (Wii/GC)? Wii What version of Swiss are you using? Swiss r298 What disc was in the drive if any? Wii Sports What device was the data being loaded from (SD/HDD/DVD/DVD-R)? SD Please provide any additional information below. When I play the story mode the voices of the character is mute and the subtitles are wrong. ``` Original issue reported on code.google.com by `diego.so...@gmail.com` on 28 Feb 2015 at 12:26
defect
star fox adventures no audio and wrong subtitles what step s will reproduce the problem play the game on swiss select story watch the cutscenes what console are you running swiss on wii gc wii what version of swiss are you using swiss what disc was in the drive if any wii sports what device was the data being loaded from sd hdd dvd dvd r sd please provide any additional information below when i play the story mode the voices of the character is mute and the subtitles are wrong original issue reported on code google com by diego so gmail com on feb at
1
31,676
13,613,139,636
IssuesEvent
2020-09-23 11:24:44
dessant/buster
https://api.github.com/repos/dessant/buster
closed
hCaptcha image is not being solved
wontfix-service
The extension doesn't solve hCaptcha's. I've tried it multiple times on the page that required it and it just doesn't work. All times I had to manually do it myself. EDIT: please add hCaptcha to the extension.
1.0
hCaptcha image is not being solved - The extension doesn't solve hCaptcha's. I've tried it multiple times on the page that required it and it just doesn't work. All times I had to manually do it myself. EDIT: please add hCaptcha to the extension.
non_defect
hcaptcha image is not being solved the extension doesn t solve hcaptcha s i ve tried it multiple times on the page that required it and it just doesn t work all times i had to manually do it myself edit please add hcaptcha to the extension
0