Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
89,638
3,798,181,480
IssuesEvent
2016-03-23 11:21:45
FAC7/amazon2.0
https://api.github.com/repos/FAC7/amazon2.0
closed
Display a list of different items in a category
high priority User
**Description** As a customer I want to be able to browse a list of items in a category so that I can see a range of items to choose from. **Acceptance Criteria** When I go to a category I can see a list of items from that category with information on item title and price. For @eleanorreem, @Virginie-T and @andrewMacmurray @RobStallion
1.0
Display a list of different items in a category - **Description** As a customer I want to be able to browse a list of items in a category so that I can see a range of items to choose from. **Acceptance Criteria** When I go to a category I can see a list of items from that category with information on item title and price. For @eleanorreem, @Virginie-T and @andrewMacmurray @RobStallion
priority
display a list of different items in a category description as a customer i want to be able to browse a list of items in a category so that i can see a range of items to choose from acceptance criteria when i go to a category i can see a list of items from that category with information on item title and price for eleanorreem virginie t and andrewmacmurray robstallion
1
736,565
25,478,721,282
IssuesEvent
2022-11-25 17:16:41
CrowdDotDev/crowd.dev
https://api.github.com/repos/CrowdDotDev/crowd.dev
closed
[C-164] Read-only role should allow to access listing pages
Bug High priority
Currently, if a user has a **read-only** role, they are not able to access the Members list page. Instead, an error message is shown to the user that Something went wrong. The expected behaviour is for **admin** roles to be able to access all pages, and **read-only** roles to be able to access all listing pages, such as the Members list page. <sub>From [SyncLinear.com](https://synclinear.com) | [C-164](https://linear.app/crowddotdev/issue/C-164/read-only-role-should-allow-to-access-listing-pages)</sub>
1.0
[C-164] Read-only role should allow to access listing pages - Currently, if a user has a **read-only** role, they are not able to access the Members list page. Instead, an error message is shown to the user that Something went wrong. The expected behaviour is for **admin** roles to be able to access all pages, and **read-only** roles to be able to access all listing pages, such as the Members list page. <sub>From [SyncLinear.com](https://synclinear.com) | [C-164](https://linear.app/crowddotdev/issue/C-164/read-only-role-should-allow-to-access-listing-pages)</sub>
priority
read only role should allow to access listing pages currently if a user has a read only role they are not able to access the members list page instead an error message is shown to the user that something went wrong the expected behaviour is for admin roles to be able to access all pages and read only roles to be able to access all listing pages such as the members list page from
1
698,859
23,994,385,855
IssuesEvent
2022-09-14 06:05:27
stackabletech/t2
https://api.github.com/repos/stackabletech/t2
closed
Test K8s connectivity in testdriver
priority/high
The operator integration tests are somewhat flaky and we do not really see a reason. It is assumed that we have issues with the connectivity to a running K8s cluster. In this task, we should find a way to test/monitor the K8s connectivity during a long running test. - [x] testdriver script refactored and released from feature branch - [x] update the docs - [x] if nightly tests successful, set up PR - [x] merge into develop (no T2 release required)
1.0
Test K8s connectivity in testdriver - The operator integration tests are somewhat flaky and we do not really see a reason. It is assumed that we have issues with the connectivity to a running K8s cluster. In this task, we should find a way to test/monitor the K8s connectivity during a long running test. - [x] testdriver script refactored and released from feature branch - [x] update the docs - [x] if nightly tests successful, set up PR - [x] merge into develop (no T2 release required)
priority
test connectivity in testdriver the operator integration tests are somewhat flaky and we do not really see a reason it is assumed that we have issues with the connectivity to a running cluster in this task we should find a way to test monitor the connectivity during a long running test testdriver script refactored and released from feature branch update the docs if nightly tests successful set up pr merge into develop no release required
1
788,920
27,772,980,739
IssuesEvent
2023-03-16 15:31:15
AY2223S2-CS2103-F10-2/tp
https://api.github.com/repos/AY2223S2-CS2103-F10-2/tp
closed
As a user, I can clear all modules
type.Story priority.High
so that I can remove obsolete modules quickly after a semester is over.
1.0
As a user, I can clear all modules - so that I can remove obsolete modules quickly after a semester is over.
priority
as a user i can clear all modules so that i can remove obsolete modules quickly after a semester is over
1
141,876
5,446,246,426
IssuesEvent
2017-03-07 10:06:26
CS2103JAN2017-W14-B3/main
https://api.github.com/repos/CS2103JAN2017-W14-B3/main
opened
As a user I want to be able tag/categorise a floating task
priority.high type.story
So that I can know what to do when I am in the current context
1.0
As a user I want to be able tag/categorise a floating task - So that I can know what to do when I am in the current context
priority
as a user i want to be able tag categorise a floating task so that i can know what to do when i am in the current context
1
166,976
6,329,606,743
IssuesEvent
2017-07-26 03:45:29
NCAR/chords
https://api.github.com/repos/NCAR/chords
closed
Create "Archives" MVCs
feature HIGH PRIORITY
Add a new MVC for "Archives" the will list supported archives, starting with CUAHSI
1.0
Create "Archives" MVCs - Add a new MVC for "Archives" the will list supported archives, starting with CUAHSI
priority
create archives mvcs add a new mvc for archives the will list supported archives starting with cuahsi
1
506
2,497,481,819
IssuesEvent
2015-01-07 07:26:32
orientechnologies/orientdb
https://api.github.com/repos/orientechnologies/orientdb
closed
Mass insertion of edges slows down over time
core high priority
Running a script that inserts millions of edges starts of fast and then slows down to crawl (1 edge/second) after hours. Disk I/O is goes down though CPU use is normal. If I restart the server, it exits fairly quickly as normal. I can immediately restart the server and resume the import script, and edges are added very quickly again. I'm not seeing leaks in the import script (it would just OOM if the heap was >256MB) itself. Heap is 2GB, and diskcache is 8GB. I'm using concurrent Mark-Sweep GC if that matters. Note that some edges have indexes. I also have -Dstorage.wal.syncOnPageFlush=false set. I didn't have this problem in 1.7.10, and I'm not sure how to work around it other than restarting things every few hours. Are there any known issues that make this problem likely? Would disabling the level1 cache help?
1.0
Mass insertion of edges slows down over time - Running a script that inserts millions of edges starts of fast and then slows down to crawl (1 edge/second) after hours. Disk I/O is goes down though CPU use is normal. If I restart the server, it exits fairly quickly as normal. I can immediately restart the server and resume the import script, and edges are added very quickly again. I'm not seeing leaks in the import script (it would just OOM if the heap was >256MB) itself. Heap is 2GB, and diskcache is 8GB. I'm using concurrent Mark-Sweep GC if that matters. Note that some edges have indexes. I also have -Dstorage.wal.syncOnPageFlush=false set. I didn't have this problem in 1.7.10, and I'm not sure how to work around it other than restarting things every few hours. Are there any known issues that make this problem likely? Would disabling the level1 cache help?
priority
mass insertion of edges slows down over time running a script that inserts millions of edges starts of fast and then slows down to crawl edge second after hours disk i o is goes down though cpu use is normal if i restart the server it exits fairly quickly as normal i can immediately restart the server and resume the import script and edges are added very quickly again i m not seeing leaks in the import script it would just oom if the heap was itself heap is and diskcache is i m using concurrent mark sweep gc if that matters note that some edges have indexes i also have dstorage wal synconpageflush false set i didn t have this problem in and i m not sure how to work around it other than restarting things every few hours are there any known issues that make this problem likely would disabling the cache help
1
629,730
20,051,588,073
IssuesEvent
2022-02-03 07:20:14
dnd-side-project/dnd-6th-1-iOS
https://api.github.com/repos/dnd-side-project/dnd-6th-1-iOS
opened
커뮤니티 홈 뷰 MVVM구조로 변환
윤경 Priority: High Chore
## 기능 설명 RxSwift + MVVM 구조로 변환 ## 완료 조건 - [ ] CommunityVC 수정 - [ ] CommunityMV 구현
1.0
커뮤니티 홈 뷰 MVVM구조로 변환 - ## 기능 설명 RxSwift + MVVM 구조로 변환 ## 완료 조건 - [ ] CommunityVC 수정 - [ ] CommunityMV 구현
priority
커뮤니티 홈 뷰 mvvm구조로 변환 기능 설명 rxswift mvvm 구조로 변환 완료 조건 communityvc 수정 communitymv 구현
1
62,727
3,192,992,471
IssuesEvent
2015-09-30 00:45:17
fusioninventory/fusioninventory-for-glpi
https://api.github.com/repos/fusioninventory/fusioninventory-for-glpi
closed
Migration ocs - fusion : Big problem with OS like software
Category: Computer inventory Component: For junior contributor Priority: High Status: Closed Tracker: Feature
--- Author Name: **Xavier Caillaud** (@tsmr) Original Redmine Issue: 1063, http://forge.fusioninventory.org/issues/1063 Original Date: 2011-08-11 --- Migration ocs - fusion : Big problem with OS like software : Fusion don't add OS like a software and so the licensing use of GLPI is broken. Link with oem license are broken for all existent computers :(
1.0
Migration ocs - fusion : Big problem with OS like software - --- Author Name: **Xavier Caillaud** (@tsmr) Original Redmine Issue: 1063, http://forge.fusioninventory.org/issues/1063 Original Date: 2011-08-11 --- Migration ocs - fusion : Big problem with OS like software : Fusion don't add OS like a software and so the licensing use of GLPI is broken. Link with oem license are broken for all existent computers :(
priority
migration ocs fusion big problem with os like software author name xavier caillaud tsmr original redmine issue original date migration ocs fusion big problem with os like software fusion don t add os like a software and so the licensing use of glpi is broken link with oem license are broken for all existent computers
1
374,349
11,088,379,626
IssuesEvent
2019-12-14 10:35:20
ahmedkaludi/accelerated-mobile-pages
https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages
closed
Give default styling for <small> tag.
NEXT UPDATE [Priority: HIGH] bug
We have no default styling for `<small>` tag, as it can be used anywhere in the content.
1.0
Give default styling for <small> tag. - We have no default styling for `<small>` tag, as it can be used anywhere in the content.
priority
give default styling for tag we have no default styling for tag as it can be used anywhere in the content
1
148,337
5,673,143,699
IssuesEvent
2017-04-12 05:00:08
brunnopleffken/addictive-community
https://api.github.com/repos/brunnopleffken/addictive-community
closed
bug with Cyrillic
bug high-priority
When I try to let my topic Cyrillic displays this error: Fatal error: Class 'Transliterator' not found in C:\xampp\htdocs\kernel\Text.php on line 172
1.0
bug with Cyrillic - When I try to let my topic Cyrillic displays this error: Fatal error: Class 'Transliterator' not found in C:\xampp\htdocs\kernel\Text.php on line 172
priority
bug with cyrillic when i try to let my topic cyrillic displays this error fatal error class transliterator not found in c xampp htdocs kernel text php on line
1
398,866
11,742,442,533
IssuesEvent
2020-03-12 00:48:36
magda-io/magda
https://api.github.com/repos/magda-io/magda
closed
Apply authorisation to dereferenced records when requested via /record/id API
Authorisation priority: high refined
### Problem description Building on #2364, this applies authorisation to aspects retrieved through links when `dereference=true` is specified through the [/record/{id}](https://dev.magda.io/api/v0/apidocs/#api-Registry_Record_Service-GetV0RegistryRecordsId) endpoint in the registry. ### Acceptance Criteria (assuming `dereference=true` is specified) - If a user is allowed to see a requested record aspect, and that aspect links to another record that the user is allowed to see, the inner aspect should be returned as part of the outer one, just as it was before - If the user is allowed to see an outer aspect but not the inner aspect it links to, there should be no inner aspect present - as if the inner record aspect row didn't exist. **(e.g. if the dataset-distributions schema links to a single distribution record but the user isn't allowed to see that distribution, the `distributions` property should be `[]` but you should still be able to see the dataset-distributions aspect.** - If the user is not allowed to see the outer aspect but is allowed to see the inner one, they still shouldn't be able to see either when dereferencing. **(with `deference=false`)** - **The id of dereferenced records that the user doesn't have access to should be suppressed... e.g. in the example above where a user has access to a dataset record but not the distribution, the `distributions` property should also be `[]`. ** ### Technical Notes - This should be achievable using similar logic to getting records without dereferencing. Dereferencing works by examining the aspect definition, pulling out linked aspects and then querying for them. When the linked aspects are queries, it should be possible to look up the possible policies from OPA, turn them into SQL and include them in the query for the inner aspect. ### Out of Scope - Aspect-level authorisation
1.0
Apply authorisation to dereferenced records when requested via /record/id API - ### Problem description Building on #2364, this applies authorisation to aspects retrieved through links when `dereference=true` is specified through the [/record/{id}](https://dev.magda.io/api/v0/apidocs/#api-Registry_Record_Service-GetV0RegistryRecordsId) endpoint in the registry. ### Acceptance Criteria (assuming `dereference=true` is specified) - If a user is allowed to see a requested record aspect, and that aspect links to another record that the user is allowed to see, the inner aspect should be returned as part of the outer one, just as it was before - If the user is allowed to see an outer aspect but not the inner aspect it links to, there should be no inner aspect present - as if the inner record aspect row didn't exist. **(e.g. if the dataset-distributions schema links to a single distribution record but the user isn't allowed to see that distribution, the `distributions` property should be `[]` but you should still be able to see the dataset-distributions aspect.** - If the user is not allowed to see the outer aspect but is allowed to see the inner one, they still shouldn't be able to see either when dereferencing. **(with `deference=false`)** - **The id of dereferenced records that the user doesn't have access to should be suppressed... e.g. in the example above where a user has access to a dataset record but not the distribution, the `distributions` property should also be `[]`. ** ### Technical Notes - This should be achievable using similar logic to getting records without dereferencing. Dereferencing works by examining the aspect definition, pulling out linked aspects and then querying for them. When the linked aspects are queries, it should be possible to look up the possible policies from OPA, turn them into SQL and include them in the query for the inner aspect. ### Out of Scope - Aspect-level authorisation
priority
apply authorisation to dereferenced records when requested via record id api problem description building on this applies authorisation to aspects retrieved through links when dereference true is specified through the endpoint in the registry acceptance criteria assuming dereference true is specified if a user is allowed to see a requested record aspect and that aspect links to another record that the user is allowed to see the inner aspect should be returned as part of the outer one just as it was before if the user is allowed to see an outer aspect but not the inner aspect it links to there should be no inner aspect present as if the inner record aspect row didn t exist e g if the dataset distributions schema links to a single distribution record but the user isn t allowed to see that distribution the distributions property should be but you should still be able to see the dataset distributions aspect if the user is not allowed to see the outer aspect but is allowed to see the inner one they still shouldn t be able to see either when dereferencing with deference false the id of dereferenced records that the user doesn t have access to should be suppressed e g in the example above where a user has access to a dataset record but not the distribution the distributions property should also be technical notes this should be achievable using similar logic to getting records without dereferencing dereferencing works by examining the aspect definition pulling out linked aspects and then querying for them when the linked aspects are queries it should be possible to look up the possible policies from opa turn them into sql and include them in the query for the inner aspect out of scope aspect level authorisation
1
750,541
26,205,468,407
IssuesEvent
2023-01-03 22:04:11
gamefreedomgit/Maelstrom
https://api.github.com/repos/gamefreedomgit/Maelstrom
opened
[Quest][Cata] Parting Packages + Of Utmost Importance + Of Utmost Importance
Quest - Cataclysm (80+) Quest - Event Priority: High Status: Confirmed
**How to reproduce:** there is a phasing issue with these quests after completing parting packages both npcs which is called patch and ticker phase out which makes it impossible for the players to complete or take the following quests ![image](https://user-images.githubusercontent.com/36106057/210447927-e9b93f5c-8164-4f8f-8e3e-b9aca7fb405e.png) **How it should work:** they should not phase out
1.0
[Quest][Cata] Parting Packages + Of Utmost Importance + Of Utmost Importance - **How to reproduce:** there is a phasing issue with these quests after completing parting packages both npcs which is called patch and ticker phase out which makes it impossible for the players to complete or take the following quests ![image](https://user-images.githubusercontent.com/36106057/210447927-e9b93f5c-8164-4f8f-8e3e-b9aca7fb405e.png) **How it should work:** they should not phase out
priority
parting packages of utmost importance of utmost importance how to reproduce there is a phasing issue with these quests after completing parting packages both npcs which is called patch and ticker phase out which makes it impossible for the players to complete or take the following quests how it should work they should not phase out
1
671,004
22,718,490,072
IssuesEvent
2022-07-06 05:55:30
Valdes-Tresanco-MS/gmx_MMPBSA
https://api.github.com/repos/Valdes-Tresanco-MS/gmx_MMPBSA
closed
[Bug-gmx_MMPBSA]: Inconsistent decomposition output
bug high priority
### Bug summary The decomposition results only show the first value if a set of residues is selected. This is because the frame index is not changed correctly since the residue list of the parts does not match the residue list selected with `print_res`. ### Terminal output ```bash N/A ``` ### gmx_MMPBSA.log N/A ### Operating system _No response_ ### gmx_MMPBSA Version 1.5.5 ### Python version _No response_ ### Installation _No response_
1.0
[Bug-gmx_MMPBSA]: Inconsistent decomposition output - ### Bug summary The decomposition results only show the first value if a set of residues is selected. This is because the frame index is not changed correctly since the residue list of the parts does not match the residue list selected with `print_res`. ### Terminal output ```bash N/A ``` ### gmx_MMPBSA.log N/A ### Operating system _No response_ ### gmx_MMPBSA Version 1.5.5 ### Python version _No response_ ### Installation _No response_
priority
inconsistent decomposition output bug summary the decomposition results only show the first value if a set of residues is selected this is because the frame index is not changed correctly since the residue list of the parts does not match the residue list selected with print res terminal output bash n a gmx mmpbsa log n a operating system no response gmx mmpbsa version python version no response installation no response
1
153,209
5,886,997,136
IssuesEvent
2017-05-17 05:42:42
ThoughtWorksInc/treadmill
https://api.github.com/repos/ThoughtWorksInc/treadmill
closed
Use the CLI to destroy a Treadmill cell
Done Feature-AWS Support Priority-High Role-Administrator Size-Medium (M)
So that : I can later script this into some kind of automated flow. Tasks: Extend the existing CLI to destroy a cell. "Assumption: Story '5c4533' has been played."
1.0
Use the CLI to destroy a Treadmill cell - So that : I can later script this into some kind of automated flow. Tasks: Extend the existing CLI to destroy a cell. "Assumption: Story '5c4533' has been played."
priority
use the cli to destroy a treadmill cell so that i can later script this into some kind of automated flow tasks extend the existing cli to destroy a cell assumption story has been played
1
222,644
7,434,732,075
IssuesEvent
2018-03-26 12:07:05
wso2/docker-is
https://api.github.com/repos/wso2/docker-is
closed
Update Docker Compose Templates to use new IS Docker Image
Priority/High Type/Improvement
**Description:** Currently, Identity Server Docker Compose templates use a separate set of Dockerfiles for building its Docker images. This process can be improved by using the Identity Server Docker image built with the default Dockerfile and using a volume mount to provide the configurations. As a result, Docker Compose template would not need to build new product Docker images. **Affected Product Version:** 5.3.0, 5.4.0 **Related Issues:** https://github.com/wso2/docker-ei/issues/13, https://github.com/wso2/docker-apim/issues/63, https://github.com/wso2/docker-das/issues/3
1.0
Update Docker Compose Templates to use new IS Docker Image - **Description:** Currently, Identity Server Docker Compose templates use a separate set of Dockerfiles for building its Docker images. This process can be improved by using the Identity Server Docker image built with the default Dockerfile and using a volume mount to provide the configurations. As a result, Docker Compose template would not need to build new product Docker images. **Affected Product Version:** 5.3.0, 5.4.0 **Related Issues:** https://github.com/wso2/docker-ei/issues/13, https://github.com/wso2/docker-apim/issues/63, https://github.com/wso2/docker-das/issues/3
priority
update docker compose templates to use new is docker image description currently identity server docker compose templates use a separate set of dockerfiles for building its docker images this process can be improved by using the identity server docker image built with the default dockerfile and using a volume mount to provide the configurations as a result docker compose template would not need to build new product docker images affected product version related issues
1
236,931
7,753,690,850
IssuesEvent
2018-05-31 02:11:37
Gloirin/m2gTest
https://api.github.com/repos/Gloirin/m2gTest
closed
0007004: "dead lock detected" when trying to uninstall ActiveSync
Setup bug high priority
**Reported by pschuele on 27 Aug 2012 09:10** **Version:** Milan (2012.03.5) &quot;dead lock detected&quot; when trying to uninstall ActiveSync - installed: 5.3, available: 5.8 **Additional information:** dead lock detected oldCount: 2 .../Setup/Controller.php(1272): Setup_Controller-&gt;_uninstallApplication() .../Setup/Frontend/Json.php(134): Setup_Controller-&gt;uninstallApplications() [internal function]: Setup_Frontend_Json-&gt;uninstallApplications() .../library/Zend/Server/Abstract.php(232): call_user_func_array() .../Zend/Json/Server.php(558): Zend_Server_Abstract-&gt;_dispatch() .../Zend/Json/Server.php(197): Zend_Json_Server-&gt;_handle() .../Setup/Server/Json.php(77): Zend_Json_Server-&gt;handle()
1.0
0007004: "dead lock detected" when trying to uninstall ActiveSync - **Reported by pschuele on 27 Aug 2012 09:10** **Version:** Milan (2012.03.5) &quot;dead lock detected&quot; when trying to uninstall ActiveSync - installed: 5.3, available: 5.8 **Additional information:** dead lock detected oldCount: 2 .../Setup/Controller.php(1272): Setup_Controller-&gt;_uninstallApplication() .../Setup/Frontend/Json.php(134): Setup_Controller-&gt;uninstallApplications() [internal function]: Setup_Frontend_Json-&gt;uninstallApplications() .../library/Zend/Server/Abstract.php(232): call_user_func_array() .../Zend/Json/Server.php(558): Zend_Server_Abstract-&gt;_dispatch() .../Zend/Json/Server.php(197): Zend_Json_Server-&gt;_handle() .../Setup/Server/Json.php(77): Zend_Json_Server-&gt;handle()
priority
dead lock detected when trying to uninstall activesync reported by pschuele on aug version milan quot dead lock detected quot when trying to uninstall activesync installed available additional information dead lock detected oldcount setup controller php setup controller gt uninstallapplication setup frontend json php setup controller gt uninstallapplications setup frontend json gt uninstallapplications library zend server abstract php call user func array zend json server php zend server abstract gt dispatch zend json server php zend json server gt handle setup server json php zend json server gt handle
1
605,093
18,724,709,537
IssuesEvent
2021-11-03 15:13:27
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
reopened
pfsnmpd issue: systemd service file not correct
Type: Bug Priority: High
Right now: ```` [Service] StartLimitBurst=3 StartLimitInterval=10 Type=notify ExecStartPre=/usr/bin/perl -I/usr/local/pf/lib -I/usr/local/pf/lib_perl/lib/perl5 '-Mpf::services::manager::snmptrapd' -e 'pf::services::manager::snmptrapd->new()->generateConfig()' ExecStart=/usr/sbin/snmptrapd -f -n -c /usr/local/pf/var/conf/snmptrapd.conf -C -A -Lf /usr/local/pf/logs/snmptrapd.log -p /usr/local/pf/var/run/snmptrapd.pid -On ExecReload=/bin/kill -HUP $MAINPID Slice=packetfence.slice Restart=on-failure ```` Should be: ```` [Service] StartLimitBurst=3 StartLimitInterval=10 Type=simple ExecStartPre=/usr/bin/perl -I/usr/local/pf/lib -I/usr/local/pf/lib_perl/lib/perl5 '-Mpf::services::manager::snmptrapd' -e 'pf::services::manager::snmptrapd->new()->generateConfig()' ExecStart=/usr/sbin/snmptrapd -f -n -c /usr/local/pf/var/conf/snmptrapd.conf -C -A -Lf /usr/local/pf/logs/snmptrapd.log -p /usr/local/pf/var/run/snmptrapd.pid -On ExecReload=/bin/kill -HUP $MAINPID Slice=packetfence.slice Restart=on-failure ````
1.0
pfsnmpd issue: systemd service file not correct - Right now: ```` [Service] StartLimitBurst=3 StartLimitInterval=10 Type=notify ExecStartPre=/usr/bin/perl -I/usr/local/pf/lib -I/usr/local/pf/lib_perl/lib/perl5 '-Mpf::services::manager::snmptrapd' -e 'pf::services::manager::snmptrapd->new()->generateConfig()' ExecStart=/usr/sbin/snmptrapd -f -n -c /usr/local/pf/var/conf/snmptrapd.conf -C -A -Lf /usr/local/pf/logs/snmptrapd.log -p /usr/local/pf/var/run/snmptrapd.pid -On ExecReload=/bin/kill -HUP $MAINPID Slice=packetfence.slice Restart=on-failure ```` Should be: ```` [Service] StartLimitBurst=3 StartLimitInterval=10 Type=simple ExecStartPre=/usr/bin/perl -I/usr/local/pf/lib -I/usr/local/pf/lib_perl/lib/perl5 '-Mpf::services::manager::snmptrapd' -e 'pf::services::manager::snmptrapd->new()->generateConfig()' ExecStart=/usr/sbin/snmptrapd -f -n -c /usr/local/pf/var/conf/snmptrapd.conf -C -A -Lf /usr/local/pf/logs/snmptrapd.log -p /usr/local/pf/var/run/snmptrapd.pid -On ExecReload=/bin/kill -HUP $MAINPID Slice=packetfence.slice Restart=on-failure ````
priority
pfsnmpd issue systemd service file not correct right now startlimitburst startlimitinterval type notify execstartpre usr bin perl i usr local pf lib i usr local pf lib perl lib mpf services manager snmptrapd e pf services manager snmptrapd new generateconfig execstart usr sbin snmptrapd f n c usr local pf var conf snmptrapd conf c a lf usr local pf logs snmptrapd log p usr local pf var run snmptrapd pid on execreload bin kill hup mainpid slice packetfence slice restart on failure should be startlimitburst startlimitinterval type simple execstartpre usr bin perl i usr local pf lib i usr local pf lib perl lib mpf services manager snmptrapd e pf services manager snmptrapd new generateconfig execstart usr sbin snmptrapd f n c usr local pf var conf snmptrapd conf c a lf usr local pf logs snmptrapd log p usr local pf var run snmptrapd pid on execreload bin kill hup mainpid slice packetfence slice restart on failure
1
136,988
5,291,191,961
IssuesEvent
2017-02-08 21:52:13
urbit/urbit
https://api.github.com/repos/urbit/urbit
closed
dojo should print all error messages, as well as EVERY command you enter in
bug dojo priority high
got unlinked from :talk (see below). This was terrible as I did not realize this until later when my errors weren't getting printed out, and it took me about 30m to find out why. I can't even tell you what command I typed because dojo stupidly hides the commands you've typed. I tried to reproduce this, but frankly have no idea what I did. Will keep trying However, this issue does highlight some things that should be done that are within our power to do: 1. dojo should print every command you make, so that it's clear what output was caused by what. I can't really see any argument against this being reasonable. 2. error messages should all come back to me in one place. Why do my errors come back through :talk, :dojo, and other places ([mangled](https://github.com/urbit/urbit/issues/584))? Please just give them all back to me through dojo. It seems like we're trying to be too cute by making this more complicated than it needs to be. Talk and dojo are in the same place on the terminal anyways. Moreover, if I'm on web talk, do I really want my programming errors coming through here? ``` ? -2 0vetufl.vhadi.vantg.43vk7.vb2e9 at ~2016.2.25..00.36.24..1c60 ~zod to : [drum]: /~zod/home/0/app/gmail:<[32 1].[273 3]> --------------| ;26 ? 26 0v7.r8bbb.nhpj2.p8acj.p4h01.255q5 at ~2016.2.25..00.36.24..1c60 ~zod to : [drum]: /~zod/home/0/app/gmail:<[86 3].[105 5]> [unlinked from [p=~zod q=%talk]] /~zod/home/0/app/talk:<[2.065 3].[2.065 27]> /~zod/home/0/app/talk:<[2.065 11].[2.065 27]> /~zod/home/0/app/talk:<[1.225 5].[1.231 40]> /~zod/home/0/app/talk:<[1.226 5].[1.231 40]> /~zod/home/0/app/talk:<[1.227 5].[1.231 40]> /~zod/home/0/app/talk:<[1.231 5].[1.231 40]> /~zod/home/0/app/talk:<[1.231 13].[1.231 40]> /~zod/home/0/app/talk:<[1.165 7].[1.170 9]> /~zod/home/0/app/talk:<[1.166 7].[1.170 9]> /~zod/home/0/app/talk:<[1.169 15].[1.169 22]> /~zod/home/0/app/talk:<[1.145 7].[1.161 9]> /~zod/home/0/app/talk:<[1.146 7].[1.161 9]> /~zod/home/0/app/talk:<[1.148 7].[1.161 9]> /~zod/home/0/app/talk:<[1.149 7].[1.161 9]> /~zod/home/0/app/talk:<[1.150 7].[1.161 9]> /~zod/home/0/app/talk:<[887 7].[1.142 9]> /~zod/home/0/app/talk:<[888 7].[1.142 9]> /~zod/home/0/app/talk:<[889 7].[1.142 9]> /~zod/home/0/app/talk:<[889 11].[889 15]> /~zod/home/0/app/talk:<[892 9].[912 11]> /~zod/home/0/app/talk:<[893 20].[893 34]> /~zod/home/0/app/talk:<[1.097 9].[1.111 78]> /~zod/home/0/app/talk:<[1.098 9].[1.111 78]> /~zod/home/0/app/talk:<[1.099 9].[1.111 78]> /~zod/home/0/app/talk:<[1.100 9].[1.111 78]> /~zod/home/0/app/talk:<[1.105 9].[1.111 78]> /~zod/home/0/app/talk:<[1.106 11].[1.110 61]> /~zod/home/0/app/talk:<[1.108 11].[1.110 61]> /~zod/home/0/app/talk:<[1.108 15].[1.108 45]> /~zod/home/0/app/talk:<[1.108 19].[1.108 45]> /~zod/home/0/app/talk:<[1.114 39].[1.118 42]> /~zod/home/0/app/talk:<[1.115 9].[1.118 42]> /~zod/home/0/app/talk:<[1.116 9].[1.118 42]> /~zod/home/0/app/talk:<[1.117 9].[1.118 42]> /~zod/home/0/app/talk:<[1.117 13].[1.117 42]> /~zod/home/0/app/talk:<[1.117 30].[1.117 41]> subtract-underflow ```
1.0
dojo should print all error messages, as well as EVERY command you enter in - got unlinked from :talk (see below). This was terrible as I did not realize this until later when my errors weren't getting printed out, and it took me about 30m to find out why. I can't even tell you what command I typed because dojo stupidly hides the commands you've typed. I tried to reproduce this, but frankly have no idea what I did. Will keep trying However, this issue does highlight some things that should be done that are within our power to do: 1. dojo should print every command you make, so that it's clear what output was caused by what. I can't really see any argument against this being reasonable. 2. error messages should all come back to me in one place. Why do my errors come back through :talk, :dojo, and other places ([mangled](https://github.com/urbit/urbit/issues/584))? Please just give them all back to me through dojo. It seems like we're trying to be too cute by making this more complicated than it needs to be. Talk and dojo are in the same place on the terminal anyways. Moreover, if I'm on web talk, do I really want my programming errors coming through here? ``` ? -2 0vetufl.vhadi.vantg.43vk7.vb2e9 at ~2016.2.25..00.36.24..1c60 ~zod to : [drum]: /~zod/home/0/app/gmail:<[32 1].[273 3]> --------------| ;26 ? 26 0v7.r8bbb.nhpj2.p8acj.p4h01.255q5 at ~2016.2.25..00.36.24..1c60 ~zod to : [drum]: /~zod/home/0/app/gmail:<[86 3].[105 5]> [unlinked from [p=~zod q=%talk]] /~zod/home/0/app/talk:<[2.065 3].[2.065 27]> /~zod/home/0/app/talk:<[2.065 11].[2.065 27]> /~zod/home/0/app/talk:<[1.225 5].[1.231 40]> /~zod/home/0/app/talk:<[1.226 5].[1.231 40]> /~zod/home/0/app/talk:<[1.227 5].[1.231 40]> /~zod/home/0/app/talk:<[1.231 5].[1.231 40]> /~zod/home/0/app/talk:<[1.231 13].[1.231 40]> /~zod/home/0/app/talk:<[1.165 7].[1.170 9]> /~zod/home/0/app/talk:<[1.166 7].[1.170 9]> /~zod/home/0/app/talk:<[1.169 15].[1.169 22]> /~zod/home/0/app/talk:<[1.145 7].[1.161 9]> /~zod/home/0/app/talk:<[1.146 7].[1.161 9]> /~zod/home/0/app/talk:<[1.148 7].[1.161 9]> /~zod/home/0/app/talk:<[1.149 7].[1.161 9]> /~zod/home/0/app/talk:<[1.150 7].[1.161 9]> /~zod/home/0/app/talk:<[887 7].[1.142 9]> /~zod/home/0/app/talk:<[888 7].[1.142 9]> /~zod/home/0/app/talk:<[889 7].[1.142 9]> /~zod/home/0/app/talk:<[889 11].[889 15]> /~zod/home/0/app/talk:<[892 9].[912 11]> /~zod/home/0/app/talk:<[893 20].[893 34]> /~zod/home/0/app/talk:<[1.097 9].[1.111 78]> /~zod/home/0/app/talk:<[1.098 9].[1.111 78]> /~zod/home/0/app/talk:<[1.099 9].[1.111 78]> /~zod/home/0/app/talk:<[1.100 9].[1.111 78]> /~zod/home/0/app/talk:<[1.105 9].[1.111 78]> /~zod/home/0/app/talk:<[1.106 11].[1.110 61]> /~zod/home/0/app/talk:<[1.108 11].[1.110 61]> /~zod/home/0/app/talk:<[1.108 15].[1.108 45]> /~zod/home/0/app/talk:<[1.108 19].[1.108 45]> /~zod/home/0/app/talk:<[1.114 39].[1.118 42]> /~zod/home/0/app/talk:<[1.115 9].[1.118 42]> /~zod/home/0/app/talk:<[1.116 9].[1.118 42]> /~zod/home/0/app/talk:<[1.117 9].[1.118 42]> /~zod/home/0/app/talk:<[1.117 13].[1.117 42]> /~zod/home/0/app/talk:<[1.117 30].[1.117 41]> subtract-underflow ```
priority
dojo should print all error messages as well as every command you enter in got unlinked from talk see below this was terrible as i did not realize this until later when my errors weren t getting printed out and it took me about to find out why i can t even tell you what command i typed because dojo stupidly hides the commands you ve typed i tried to reproduce this but frankly have no idea what i did will keep trying however this issue does highlight some things that should be done that are within our power to do dojo should print every command you make so that it s clear what output was caused by what i can t really see any argument against this being reasonable error messages should all come back to me in one place why do my errors come back through talk dojo and other places please just give them all back to me through dojo it seems like we re trying to be too cute by making this more complicated than it needs to be talk and dojo are in the same place on the terminal anyways moreover if i m on web talk do i really want my programming errors coming through here vhadi vantg at zod to zod home app gmail at zod to zod home app gmail zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk zod home app talk subtract underflow
1
245,749
7,890,500,741
IssuesEvent
2018-06-28 09:00:55
coq/coq
https://api.github.com/repos/coq/coq
closed
Mutually coinductive records are not handled
part: kernel priority: high
Mutually coinductive records are currently not handled by the kernel. The reason is quite stupid if I understand correctly, as it is due to the fact extraction cannot handle them properly. This is a limitation that prevents the transition to coinductive negative types, so it should be swiftly lifted.
1.0
Mutually coinductive records are not handled - Mutually coinductive records are currently not handled by the kernel. The reason is quite stupid if I understand correctly, as it is due to the fact extraction cannot handle them properly. This is a limitation that prevents the transition to coinductive negative types, so it should be swiftly lifted.
priority
mutually coinductive records are not handled mutually coinductive records are currently not handled by the kernel the reason is quite stupid if i understand correctly as it is due to the fact extraction cannot handle them properly this is a limitation that prevents the transition to coinductive negative types so it should be swiftly lifted
1
305,909
9,378,189,558
IssuesEvent
2019-04-04 12:18:40
IBM/watson-assistant-workbench
https://api.github.com/repos/IBM/watson-assistant-workbench
closed
Add support for contextual (open) entities for csv2json direction
Priority: high XML customer request enhancement
Implement conversion from WAW inline csv intent notation `I'd like <color>red</color> dress` to WA json structure ``` { "text": "I'd like red dress", "mentions": [ { "entity": "color", "location": [ 9, 12 ] } ] } ```
1.0
Add support for contextual (open) entities for csv2json direction - Implement conversion from WAW inline csv intent notation `I'd like <color>red</color> dress` to WA json structure ``` { "text": "I'd like red dress", "mentions": [ { "entity": "color", "location": [ 9, 12 ] } ] } ```
priority
add support for contextual open entities for direction implement conversion from waw inline csv intent notation i d like red dress to wa json structure text i d like red dress mentions entity color location
1
425,271
12,337,969,385
IssuesEvent
2020-05-14 15:44:52
cb-geo/mpm
https://api.github.com/repos/cb-geo/mpm
opened
Semi-Implicit Navier-Stokes solver for incompressible flow
Priority: High Status: Review needed Type: Discussion
## Summary This RFC is to propose the Navier-Stokes solver in branch `solver/navier-stokes`. The NS solver utilized semi-implicit Chorin's projection scheme which is mainly to be used for modeling incompressible fluids. ## Motivation To extend the capability of treating internal incompressibility constraint more accurately particularly for simulations of fluid flows. The proposed method has been commonly studied in CFD community and proven to be one of the simplest yet efficient. ## Design Detail The following classes and functionality were implemented: 1. `MPMSemiImplicitNavierStokes` solver class (this at the moment only works with point 2.) 2. A `FluidParticle` derived from `Particle` with some specialized class derivation. 3. Linear solvers and assembler classes based on Eigen. 4. Free-surface detection algorithms (both cheap and expensive detection were implemented) in `Mesh` class. 5. Some boundary conditions were implemented, e.g. pressure BC. 6. Some work on level-set signed distance computation if accurate free-surface detection method is considered. ## Drawbacks No drawbacks in performance at the moment. It will enrich the capability of the software in many ways. However, the implementation might cause the function derivatives to be a bit messy, particularly in `Cell` class where we need to store and construct some local element matrices. Further thoughts and suggestions are welcome. ## Rationale and Alternatives After the merge of the NS solver, we can also extend the solver to be used for any incompressible Solid with minimum modification. Some further thoughts on including purely Implicit solver class is also possible for future development. They utilize very much the same structure and the capability can be extended from the proposed work. ## Prior Art Referring a lot to the work of @srhgk2: 1. Kularathna, S., & Soga, K. (2017). Implicit formulation of material point method for analysis of incompressible materials. Computer Methods in Applied Mechanics and Engineering, 313, 673-686.
1.0
Semi-Implicit Navier-Stokes solver for incompressible flow - ## Summary This RFC is to propose the Navier-Stokes solver in branch `solver/navier-stokes`. The NS solver utilized semi-implicit Chorin's projection scheme which is mainly to be used for modeling incompressible fluids. ## Motivation To extend the capability of treating internal incompressibility constraint more accurately particularly for simulations of fluid flows. The proposed method has been commonly studied in CFD community and proven to be one of the simplest yet efficient. ## Design Detail The following classes and functionality were implemented: 1. `MPMSemiImplicitNavierStokes` solver class (this at the moment only works with point 2.) 2. A `FluidParticle` derived from `Particle` with some specialized class derivation. 3. Linear solvers and assembler classes based on Eigen. 4. Free-surface detection algorithms (both cheap and expensive detection were implemented) in `Mesh` class. 5. Some boundary conditions were implemented, e.g. pressure BC. 6. Some work on level-set signed distance computation if accurate free-surface detection method is considered. ## Drawbacks No drawbacks in performance at the moment. It will enrich the capability of the software in many ways. However, the implementation might cause the function derivatives to be a bit messy, particularly in `Cell` class where we need to store and construct some local element matrices. Further thoughts and suggestions are welcome. ## Rationale and Alternatives After the merge of the NS solver, we can also extend the solver to be used for any incompressible Solid with minimum modification. Some further thoughts on including purely Implicit solver class is also possible for future development. They utilize very much the same structure and the capability can be extended from the proposed work. ## Prior Art Referring a lot to the work of @srhgk2: 1. Kularathna, S., & Soga, K. (2017). Implicit formulation of material point method for analysis of incompressible materials. Computer Methods in Applied Mechanics and Engineering, 313, 673-686.
priority
semi implicit navier stokes solver for incompressible flow summary this rfc is to propose the navier stokes solver in branch solver navier stokes the ns solver utilized semi implicit chorin s projection scheme which is mainly to be used for modeling incompressible fluids motivation to extend the capability of treating internal incompressibility constraint more accurately particularly for simulations of fluid flows the proposed method has been commonly studied in cfd community and proven to be one of the simplest yet efficient design detail the following classes and functionality were implemented mpmsemiimplicitnavierstokes solver class this at the moment only works with point a fluidparticle derived from particle with some specialized class derivation linear solvers and assembler classes based on eigen free surface detection algorithms both cheap and expensive detection were implemented in mesh class some boundary conditions were implemented e g pressure bc some work on level set signed distance computation if accurate free surface detection method is considered drawbacks no drawbacks in performance at the moment it will enrich the capability of the software in many ways however the implementation might cause the function derivatives to be a bit messy particularly in cell class where we need to store and construct some local element matrices further thoughts and suggestions are welcome rationale and alternatives after the merge of the ns solver we can also extend the solver to be used for any incompressible solid with minimum modification some further thoughts on including purely implicit solver class is also possible for future development they utilize very much the same structure and the capability can be extended from the proposed work prior art referring a lot to the work of kularathna s soga k implicit formulation of material point method for analysis of incompressible materials computer methods in applied mechanics and engineering
1
110,554
4,434,188,880
IssuesEvent
2016-08-18 01:02:08
codeforamerica/intake
https://api.github.com/repos/codeforamerica/intake
opened
Tie submissions to orgs based on eligibility
priority-high size-less than a day
Given a submission to a county, have the county & orgs to determine the correct orgs to send the submission to, and add those orgs to a many to many field
1.0
Tie submissions to orgs based on eligibility - Given a submission to a county, have the county & orgs to determine the correct orgs to send the submission to, and add those orgs to a many to many field
priority
tie submissions to orgs based on eligibility given a submission to a county have the county orgs to determine the correct orgs to send the submission to and add those orgs to a many to many field
1
336,147
10,172,275,036
IssuesEvent
2019-08-08 10:16:34
pmem/issues
https://api.github.com/repos/pmem/issues
closed
Test: pmempool_transform/TEST[11, 14, 15, 18] fail with valgrind
Exposure: Medium OS: Windows Priority: 2 high Type: Bug
<!-- Before creating new issue, ensure that similar issue wasn't already created * Search: https://github.com/pmem/issues/issues Note that if you do not provide enough information to reproduce the issue, we may not be able to take action on your report. Remember this is just a minimal template. You can extend it with data you think may be useful. --> # ISSUE: <!-- fill the title of issue --> ## Environment Information - PMDK package version(s): - OS(es) version(s): Fedora30 - ndctl version(s): 65 - kernel version(s): 5.1.17-300.fc30.x86_64 <!-- fill in also other useful environment data --> ## Please provide a reproduction of the bug: ``` ./RUNTESTS pmempool_transform -s TEST11 -d force-enable -t all ./RUNTESTS pmempool_transform -s TEST15 -d force-enable -t all ./RUNTESTS pmempool_transform -s TEST11 -e force-enable -t all ./RUNTESTS pmempool_transform -s TEST14 -e force-enable -t all ./RUNTESTS pmempool_transform -s TEST15 -e force-enable -t all ./RUNTESTS pmempool_transform -s TEST18 -e force-enable -t all ``` ## How often bug is revealed: (always, often, rare): always <!-- describe special circumstances in section above --> ## Actual behavior: ``` ./RUNTESTS pmempool_transform -s TEST11 -d force-enable -t all pmempool_transform/TEST11: SETUP (all/pmem/debug/drd) --74101:0: aspacem <<< SHOW_SEGMENTS: out_of_memory (177 segments) --74101:0: aspacem 25 segment names in 25 slots --74101:0: aspacem freelist is empty --74101:0: aspacem (0,4,9) /usr/local/lib/valgrind/drd-amd64-linux --74101:0: aspacem (1,48,8) /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool --74101:0: aspacem (2,104,8) /usr/lib64/ld-2.29.so --74101:0: aspacem (3,130,1) /tmp/vgdb-pipe-shared-mem-vgdb-74101-by-jenkins-on-localhost.localdomain --74101:0: aspacem (4,207,8) /usr/local/lib/valgrind/vgpreload_core-amd64-linux.so --74101:0: aspacem (5,265,8) /usr/local/lib/valgrind/vgpreload_drd-amd64-linux.so --74101:0: aspacem (6,322,9) /home/jenkins/greg/pmdk/src/debug/libpmempool.so.1.0.0 --74101:0: aspacem (7,381,8) /home/jenkins/greg/pmdk/src/debug/libpmemblk.so.1.0.0 --74101:0: aspacem (8,439,8) /home/jenkins/greg/pmdk/src/debug/libpmemlog.so.1.0.0 --74101:0: aspacem (9,497,9) /home/jenkins/greg/pmdk/src/debug/libpmemobj.so.1.0.0 --74101:0: aspacem (10,555,8) /home/jenkins/greg/pmdk/src/debug/libpmem.so.1.0.0 --74101:0: aspacem (11,610,8) /usr/lib64/libndctl.so.6.14.0 --74101:0: aspacem (12,644,7) /usr/lib64/libdaxctl.so.1.2.1 --74101:0: aspacem (13,678,8) /usr/lib64/libdl-2.29.so --74101:0: aspacem (14,707,8) /usr/lib64/libpthread-2.29.so --74101:0: aspacem (15,741,9) /usr/lib64/libc-2.29.so --74101:0: aspacem (16,769,9) /usr/lib64/libudev.so.1.6.13 --74101:0: aspacem (17,802,7) /usr/lib64/libuuid.so.1.3.0 --74101:0: aspacem (18,834,8) /usr/lib64/libkmod.so.2.3.3 --74101:0: aspacem (19,866,8) /usr/lib64/librt-2.29.so --74101:0: aspacem (20,895,8) /usr/lib64/libgcc_s-9-20190503.so.1 --74101:0: aspacem (21,935,8) /usr/lib64/liblzma.so.5.2.4 --74101:0: aspacem (22,967,7) /usr/lib64/libz.so.1.2.11 --74101:0: aspacem (23,997,10) /dev/dax1.0 --74101:0: aspacem (24,1013,5) /dev/dax1.3 --74101:0: aspacem 0: RSVN 0000000000-00003fffff 4194304 ----- SmFixed --74101:0: aspacem 1: file 0000400000-0000404fff 20480 r---- d=0xfd00 i=28113057 o=0 (1,48) --74101:0: aspacem 2: file 0000405000-000045afff 352256 r-xT- d=0xfd00 i=28113057 o=20480 (1,48) --74101:0: aspacem 3: file 000045b000-000047dfff 143360 r---- d=0xfd00 i=28113057 o=372736 (1,48) --74101:0: aspacem 4: RSVN 000047e000-000047efff 4096 ----- SmFixed --74101:0: aspacem 5: file 000047f000-0000480fff 8192 r---- d=0xfd00 i=28113057 o=516096 (1,48) --74101:0: aspacem 6: file 0000481000-0000481fff 4096 rw--- d=0xfd00 i=28113057 o=524288 (1,48) --74101:0: aspacem 7: anon 0000482000-0000488fff 28672 rw--- --74101:0: aspacem 8: RSVN 0000489000-0003ffffff 59m ----- SmFixed --74101:0: aspacem 9: file 0004000000-0004000fff 4096 r---- d=0xfd00 i=25168806 o=0 (2,104) --74101:0: aspacem 10: file 0004001000-0004020fff 131072 r-xT- d=0xfd00 i=25168806 o=4096 (2,104) --74101:0: aspacem 11: file 0004021000-0004028fff 32768 r---- d=0xfd00 i=25168806 o=135168 (2,104) --74101:0: aspacem 12: 0004029000-0004029fff 4096 --74101:0: aspacem 13: file 000402a000-000402afff 4096 r---- d=0xfd00 i=25168806 o=167936 (2,104) --74101:0: aspacem 14: file 000402b000-000402bfff 4096 rw--- d=0xfd00 i=25168806 o=172032 (2,104) --74101:0: aspacem 15: anon 000402c000-000402cfff 4096 rw--- --74101:0: aspacem 16: anon 000402d000-000402dfff 4096 rwx-- --74101:0: aspacem 17: RSVN 000402e000-000482cfff 8384512 ----- SmLower --74101:0: aspacem 18: file 000482d000-000482dfff 4096 r---- d=0xfd00 i=17991823 o=0 (4,207) --74101:0: aspacem 19: file 000482e000-000482efff 4096 r-xT- d=0xfd00 i=17991823 o=4096 (4,207) --74101:0: aspacem 20: file 000482f000-000482ffff 4096 r---- d=0xfd00 i=17991823 o=8192 (4,207) --74101:0: aspacem 21: file 0004830000-0004830fff 4096 r---- d=0xfd00 i=17991823 o=8192 (4,207) --74101:0: aspacem 22: file 0004831000-0004831fff 4096 rw--- d=0xfd00 i=17991823 o=12288 (4,207) --74101:0: aspacem 23: anon 0004832000-0004833fff 8192 rw--- --74101:0: aspacem 24: file 0004834000-000483afff 28672 r---- d=0xfd00 i=18008249 o=0 (5,265) --74101:0: aspacem 25: file 000483b000-0004851fff 94208 r-xT- d=0xfd00 i=18008249 o=28672 (5,265) --74101:0: aspacem 26: file 0004852000-0004856fff 20480 r---- d=0xfd00 i=18008249 o=122880 (5,265) --74101:0: aspacem 27: file 0004857000-0004857fff 4096 r---- d=0xfd00 i=18008249 o=139264 (5,265) --74101:0: aspacem 28: file 0004858000-0004858fff 4096 rw--- d=0xfd00 i=18008249 o=143360 (5,265) --74101:0: aspacem 29: file 0004859000-000485efff 24576 r---- d=0xfd00 i=11224907 o=0 (6,322) --74101:0: aspacem 30: file 000485f000-00048b1fff 339968 r-xT- d=0xfd00 i=11224907 o=24576 (6,322) --74101:0: aspacem 31: file 00048b2000-00048d3fff 139264 r---- d=0xfd00 i=11224907 o=364544 (6,322) --74101:0: aspacem 32: file 00048d4000-00048d4fff 4096 ----- d=0xfd00 i=11224907 o=503808 (6,322) --74101:0: aspacem 33: file 00048d5000-00048d5fff 4096 r---- d=0xfd00 i=11224907 o=503808 (6,322) --74101:0: aspacem 34: file 00048d6000-00048d6fff 4096 rw--- d=0xfd00 i=11224907 o=507904 (6,322) --74101:0: aspacem 35: anon 00048d7000-00048dcfff 24576 rw--- --74101:0: aspacem 36: file 00048dd000-00048e1fff 20480 r---- d=0xfd00 i=11239161 o=0 (7,381) --74101:0: aspacem 37: file 00048e2000-0004906fff 151552 r-xT- d=0xfd00 i=11239161 o=20480 (7,381) --74101:0: aspacem 38: file 0004907000-0004914fff 57344 r---- d=0xfd00 i=11239161 o=172032 (7,381) --74101:0: aspacem 39: file 0004915000-0004915fff 4096 r---- d=0xfd00 i=11239161 o=225280 (7,381) --74101:0: aspacem 40: file 0004916000-0004916fff 4096 rw--- d=0xfd00 i=11239161 o=229376 (7,381) --74101:0: aspacem 41: anon 0004917000-0004918fff 8192 rw--- --74101:0: aspacem 42: file 0004919000-000491dfff 20480 r---- d=0xfd00 i=11224899 o=0 (8,439) --74101:0: aspacem 43: file 000491e000-000493efff 135168 r-xT- d=0xfd00 i=11224899 o=20480 (8,439) --74101:0: aspacem 44: file 000493f000-000494bfff 53248 r---- d=0xfd00 i=11224899 o=155648 (8,439) --74101:0: aspacem 45: file 000494c000-000494cfff 4096 r---- d=0xfd00 i=11224899 o=204800 (8,439) --74101:0: aspacem 46: file 000494d000-000494dfff 4096 rw--- d=0xfd00 i=11224899 o=208896 (8,439) --74101:0: aspacem 47: anon 000494e000-000494ffff 8192 rw--- --74101:0: aspacem 48: file 0004950000-0004956fff 28672 r---- d=0xfd00 i=11224903 o=0 (9,497) --74101:0: aspacem 49: file 0004957000-00049a1fff 307200 r-xT- d=0xfd00 i=11224903 o=28672 (9,497) --74101:0: aspacem 50: file 00049a2000-00049b9fff 98304 r---- d=0xfd00 i=11224903 o=335872 (9,497) --74101:0: aspacem 51: file 00049ba000-00049bafff 4096 ----- d=0xfd00 i=11224903 o=434176 (9,497) --74101:0: aspacem 52: file 00049bb000-00049bcfff 8192 r---- d=0xfd00 i=11224903 o=434176 (9,497) --74101:0: aspacem 53: file 00049bd000-00049bdfff 4096 rw--- d=0xfd00 i=11224903 o=442368 (9,497) --74101:0: aspacem 54: anon 00049be000-00049c2fff 20480 rw--- --74101:0: aspacem 55: file 00049c3000-00049c5fff 12288 r---- d=0xfd00 i=11239755 o=0 (10,555) --74101:0: aspacem 56: file 00049c6000-0004a68fff 667648 r-xT- d=0xfd00 i=11239755 o=12288 (10,555) --74101:0: aspacem 57: file 0004a69000-0004a71fff 36864 r---- d=0xfd00 i=11239755 o=679936 (10,555) --74101:0: aspacem 58: file 0004a72000-0004a72fff 4096 r---- d=0xfd00 i=11239755 o=712704 (10,555) --74101:0: aspacem 59: file 0004a73000-0004a73fff 4096 rw--- d=0xfd00 i=11239755 o=716800 (10,555) --74101:0: aspacem 60: anon 0004a74000-0004a74fff 4096 rw--- --74101:0: aspacem 61: file 0004a75000-0004a7dfff 36864 r---- d=0xfd00 i=25170280 o=0 (11,610) --74101:0: aspacem 62: file 0004a7e000-0004a8ffff 73728 r-xT- d=0xfd00 i=25170280 o=36864 (11,610) --74101:0: aspacem 63: file 0004a90000-0004a97fff 32768 r---- d=0xfd00 i=25170280 o=110592 (11,610) --74101:0: aspacem 64: file 0004a98000-0004a98fff 4096 r---- d=0xfd00 i=25170280 o=139264 (11,610) --74101:0: aspacem 65: file 0004a99000-0004a99fff 4096 rw--- d=0xfd00 i=25170280 o=143360 (11,610) --74101:0: aspacem 66: file 0004a9a000-0004a9bfff 8192 r---- d=0xfd00 i=25170278 o=0 (12,644) --74101:0: aspacem 67: file 0004a9c000-0004a9dfff 8192 r-xT- d=0xfd00 i=25170278 o=8192 (12,644) --74101:0: aspacem 68: file 0004a9e000-0004a9efff 4096 r---- d=0xfd00 i=25170278 o=16384 (12,644) --74101:0: aspacem 69: file 0004a9f000-0004a9ffff 4096 r---- d=0xfd00 i=25170278 o=16384 (12,644) --74101:0: aspacem 70: anon 0004aa0000-0004aa0fff 4096 rw--- --74101:0: aspacem 71: file 0004aa1000-0004aa1fff 4096 r---- d=0xfd00 i=25488751 o=0 (13,678) --74101:0: aspacem 72: file 0004aa2000-0004aa3fff 8192 r-xT- d=0xfd00 i=25488751 o=4096 (13,678) --74101:0: aspacem 73: file 0004aa4000-0004aa4fff 4096 r---- d=0xfd00 i=25488751 o=12288 (13,678) --74101:0: aspacem 74: file 0004aa5000-0004aa5fff 4096 r---- d=0xfd00 i=25488751 o=12288 (13,678) --74101:0: aspacem 75: file 0004aa6000-0004aa6fff 4096 rw--- d=0xfd00 i=25488751 o=16384 (13,678) --74101:0: aspacem 76: file 0004aa7000-0004aacfff 24576 r---- d=0xfd00 i=25488945 o=0 (14,707) --74101:0: aspacem 77: file 0004aad000-0004abbfff 61440 r-xT- d=0xfd00 i=25488945 o=24576 (14,707) --74101:0: aspacem 78: file 0004abc000-0004ac1fff 24576 r---- d=0xfd00 i=25488945 o=86016 (14,707) --74101:0: aspacem 79: file 0004ac2000-0004ac2fff 4096 r---- d=0xfd00 i=25488945 o=106496 (14,707) --74101:0: aspacem 80: file 0004ac3000-0004ac3fff 4096 rw--- d=0xfd00 i=25488945 o=110592 (14,707) --74101:0: aspacem 81: anon 0004ac4000-0004ac7fff 16384 rw--- --74101:0: aspacem 82: file 0004ac8000-0004ae9fff 139264 r---- d=0xfd00 i=25168814 o=0 (15,741) --74101:0: aspacem 83: file 0004aea000-0004c36fff 1363968 r-xT- d=0xfd00 i=25168814 o=139264 (15,741) --74101:0: aspacem 84: file 0004c37000-0004c82fff 311296 r---- d=0xfd00 i=25168814 o=1503232 (15,741) --74101:0: aspacem 85: file 0004c83000-0004c83fff 4096 ----- d=0xfd00 i=25168814 o=1814528 (15,741) --74101:0: aspacem 86: file 0004c84000-0004c87fff 16384 r---- d=0xfd00 i=25168814 o=1814528 (15,741) --74101:0: aspacem 87: file 0004c88000-0004c89fff 8192 rw--- d=0xfd00 i=25168814 o=1830912 (15,741) --74101:0: aspacem 88: anon 0004c8a000-0004c8ffff 24576 rw--- --74101:0: aspacem 89: file 0004c90000-0004c94fff 20480 r---- d=0xfd00 i=25183443 o=0 (16,769) --74101:0: aspacem 90: file 0004c95000-0004caefff 106496 r-xT- d=0xfd00 i=25183443 o=20480 (16,769) --74101:0: aspacem 91: file 0004caf000-0004cb8fff 40960 r---- d=0xfd00 i=25183443 o=126976 (16,769) --74101:0: aspacem 92: file 0004cb9000-0004cb9fff 4096 ----- d=0xfd00 i=25183443 o=167936 (16,769) --74101:0: aspacem 93: file 0004cba000-0004cbafff 4096 r---- d=0xfd00 i=25183443 o=167936 (16,769) --74101:0: aspacem 94: file 0004cbb000-0004cbbfff 4096 rw--- d=0xfd00 i=25183443 o=172032 (16,769) --74101:0: aspacem 95: file 0004cbc000-0004cbdfff 8192 r---- d=0xfd00 i=25169119 o=0 (17,802) --74101:0: aspacem 96: file 0004cbe000-0004cc2fff 20480 r-xT- d=0xfd00 i=25169119 o=8192 (17,802) --74101:0: aspacem 97: file 0004cc3000-0004cc3fff 4096 r---- d=0xfd00 i=25169119 o=28672 (17,802) --74101:0: aspacem 98: file 0004cc4000-0004cc4fff 4096 r---- d=0xfd00 i=25169119 o=28672 (17,802) --74101:0: aspacem 99: anon 0004cc5000-0004cc5fff 4096 rw--- --74101:0: aspacem 100: file 0004cc6000-0004cc8fff 12288 r---- d=0xfd00 i=25169145 o=0 (18,834) --74101:0: aspacem 101: file 0004cc9000-0004cd7fff 61440 r-xT- d=0xfd00 i=25169145 o=12288 (18,834) --74101:0: aspacem 102: file 0004cd8000-0004cdcfff 20480 r---- d=0xfd00 i=25169145 o=73728 (18,834) --74101:0: aspacem 103: file 0004cdd000-0004cddfff 4096 r---- d=0xfd00 i=25169145 o=90112 (18,834) --74101:0: aspacem 104: file 0004cde000-0004cdefff 4096 rw--- d=0xfd00 i=25169145 o=94208 (18,834) --74101:0: aspacem 105: file 0004cdf000-0004ce0fff 8192 r---- d=0xfd00 i=25488951 o=0 (19,866) --74101:0: aspacem 106: file 0004ce1000-0004ce4fff 16384 r-xT- d=0xfd00 i=25488951 o=8192 (19,866) --74101:0: aspacem 107: file 0004ce5000-0004ce6fff 8192 r---- d=0xfd00 i=25488951 o=24576 (19,866) --74101:0: aspacem 108: file 0004ce7000-0004ce7fff 4096 r---- d=0xfd00 i=25488951 o=28672 (19,866) --74101:0: aspacem 109: file 0004ce8000-0004ce8fff 4096 rw--- d=0xfd00 i=25488951 o=32768 (19,866) --74101:0: aspacem 110: file 0004ce9000-0004cebfff 12288 r---- d=0xfd00 i=25168754 o=0 (20,895) --74101:0: aspacem 111: file 0004cec000-0004cfcfff 69632 r-xT- d=0xfd00 i=25168754 o=12288 (20,895) --74101:0: aspacem 112: file 0004cfd000-0004d00fff 16384 r---- d=0xfd00 i=25168754 o=81920 (20,895) --74101:0: aspacem 113: file 0004d01000-0004d01fff 4096 r---- d=0xfd00 i=25168754 o=94208 (20,895) --74101:0: aspacem 114: file 0004d02000-0004d02fff 4096 rw--- d=0xfd00 i=25168754 o=98304 (20,895) --74101:0: aspacem 115: file 0004d03000-0004d05fff 12288 r---- d=0xfd00 i=25169099 o=0 (21,935) --74101:0: aspacem 116: file 0004d06000-0004d1dfff 98304 r-xT- d=0xfd00 i=25169099 o=12288 (21,935) --74101:0: aspacem 117: file 0004d1e000-0004d28fff 45056 r---- d=0xfd00 i=25169099 o=110592 (21,935) --74101:0: aspacem 118: file 0004d29000-0004d29fff 4096 ----- d=0xfd00 i=25169099 o=155648 (21,935) --74101:0: aspacem 119: file 0004d2a000-0004d2afff 4096 r---- d=0xfd00 i=25169099 o=155648 (21,935) --74101:0: aspacem 120: anon 0004d2b000-0004d2dfff 12288 rw--- --74101:0: aspacem 121: file 0004d2e000-0004d30fff 12288 r---- d=0xfd00 i=25169023 o=0 (22,967) --74101:0: aspacem 122: file 0004d31000-0004d3efff 57344 r-xT- d=0xfd00 i=25169023 o=12288 (22,967) --74101:0: aspacem 123: file 0004d3f000-0004d45fff 28672 r---- d=0xfd00 i=25169023 o=69632 (22,967) --74101:0: aspacem 124: file 0004d46000-0004d46fff 4096 r---- d=0xfd00 i=25169023 o=94208 (22,967) --74101:0: aspacem 125: anon 0004d47000-0004d4cfff 24576 rw--- --74101:0: aspacem 126: anon 0004d4d000-000514cfff 4194304 rwx-H --74101:0: aspacem 127: file 000514d000-000514dfff 4096 rw--- d=0x006 i=16516 o=0 (23,997) --74101:0: aspacem 128: file 000514e000-000514efff 4096 rw--- d=0x006 i=16519 o=0 (24,1013) --74101:0: aspacem 129: file 000514f000-000514ffff 4096 rw--- d=0x006 i=16516 o=0 (23,997) --74101:0: aspacem 130: ANON 0005150000-0057d4ffff 1324m rwx-- --74101:0: aspacem 131: 0057d50000-0057ffffff 2818048 --74101:0: aspacem 132: FILE 0058000000-0058000fff 4096 r---- d=0xfd00 i=18008248 o=0 (0,4) --74101:0: aspacem 133: FILE 0058001000-005809bfff 634880 r-x-- d=0xfd00 i=18008248 o=4096 (0,4) --74101:0: aspacem 134: file 005809c000-005809cfff 4096 r-x-- d=0xfd00 i=18008248 o=638976 (0,4) --74101:0: aspacem 135: FILE 005809d000-00581befff 1187840 r-x-- d=0xfd00 i=18008248 o=643072 (0,4) --74101:0: aspacem 136: FILE 00581bf000-0058267fff 692224 r---- d=0xfd00 i=18008248 o=1830912 (0,4) --74101:0: aspacem 137: 0058268000-0058268fff 4096 --74101:0: aspacem 138: FILE 0058269000-005826bfff 12288 rw--- d=0xfd00 i=18008248 o=2523136 (0,4) --74101:0: aspacem 139: ANON 005826c000-0058c5efff 9m rw--- --74101:0: aspacem 140: ANON 0058c5f000-007fc5efff 624m rwx-- --74101:0: aspacem 141: 007fc5f000-007fffffff 3805184 --74101:0: aspacem 142: file 0080000000-02f5dfffff 10078m rw--- d=0x006 i=16516 o=0 (23,997) --74101:0: aspacem 143: file 02f5e00000-0f43bfefff 50397m rw--- d=0x006 i=16519 o=4096 (24,1013) --74101:0: aspacem 144: ANON 0f43bff000-1001ffefff 3044m rwx-- --74101:0: aspacem 145: 1001fff000-1001ffffff 4096 --74101:0: aspacem 146: RSVN 1002000000-1002000fff 4096 ----- SmFixed --74101:0: aspacem 147: ANON 1002001000-100268bfff 6860800 rwx-- --74101:0: aspacem 148: ANON 100268c000-100268dfff 8192 ----- --74101:0: aspacem 149: ANON 100268e000-100278dfff 1048576 rwx-- --74101:0: aspacem 150: ANON 100278e000-100278ffff 8192 ----- --74101:0: aspacem 151: FILE 1002790000-1002790fff 4096 rw--- d=0x02b i=14406966 o=0 (3,130) --74101:0: aspacem 152: ANON 1002791000-10027e0fff 327680 rwx-- --74101:0: aspacem 153: 10027e1000-10027f6fff 90112 --74101:0: aspacem 154: ANON 10027f7000-1002816fff 131072 rwx-- --74101:0: aspacem 155: 1002817000-100288bfff 479232 --74101:0: aspacem 156: ANON 100288c000-1002bd8fff 3461120 rwx-- --74101:0: aspacem 157: 1002bd9000-1002c80fff 688128 --74101:0: aspacem 158: ANON 1002c81000-10033d5fff 7688192 rwx-- --74101:0: aspacem 159: 10033d6000-1003469fff 606208 --74101:0: aspacem 160: ANON 100346a000-103fd13fff 968m rwx-- --74101:0: aspacem 161: 103fd14000-103fffffff 3063808 --74101:0: aspacem 162: file 1040000000-12b5dfffff 10078m rw--- d=0x006 i=16516 o=0 (23,997) --74101:0: aspacem 163: file 12b5e00000-1f03bfffff 50398m rw--- d=0x006 i=16519 o=0 (24,1013) --74101:0: aspacem 164: ANON 1f03c00000-1ffe7fffff 4012m rwx-- --74101:0: aspacem 165: 1ffe800000-1ffe800fff 4096 --74101:0: aspacem 166: RSVN 1ffe801000-1ffeff8fff 8355840 ----- SmUpper --74101:0: aspacem 167: anon 1ffeff9000-1fff000fff 32768 rw--- --74101:0: aspacem 168: ANON 1fff001000-1fffc00fff 12m rwx-- --74101:0: aspacem 169: 1fffc01000-1fffffffff 4190208 --74101:0: aspacem 170: RSVN 2000000000-7ffe930eefff 130938g ----- SmFixed --74101:0: aspacem 171: ANON 7ffe930ef000-7ffe93110fff 139264 rw--- --74101:0: aspacem 172: RSVN 7ffe93111000-7ffe9312dfff 118784 ----- SmFixed --74101:0: aspacem 173: ANON 7ffe9312e000-7ffe93130fff 12288 r---- --74101:0: aspacem 174: RSVN 7ffe93131000-ffffffffff5fffff 16383e ----- SmFixed --74101:0: aspacem 175: ANON ffffffffff600000-ffffffffff600fff 4096 r-x-- --74101:0: aspacem 176: RSVN ffffffffff601000-ffffffffffffffff 9m ----- SmFixed --74101:0: aspacem >>> pmempool_transform/TEST11 failed with exit code 1. Last 30 lines of drd11.log below (whole file has 95 lines). pmempool_transform/TEST11 drd11.log ==74101== by 0x48932E5: pmempool_transformU (replica.c:2462) pmempool_transform/TEST11 drd11.log ==74101== by 0x48933A5: pmempool_transform (replica.c:2492) pmempool_transform/TEST11 drd11.log ==74101== by 0x40F164: pmempool_transform_func (in /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool) pmempool_transform/TEST11 drd11.log ==74101== by 0x406AD3: main (in /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool) pmempool_transform/TEST11 drd11.log client stack range: [0x1FFEFF9000 0x1FFF000FFF] client SP: 0x1FFEFFD4B0 pmempool_transform/TEST11 drd11.log valgrind stack range: [0x100268E000 0x100278DFFF] top usage: 12072 of 1048576 pmempool_transform/TEST11 drd11.log pmempool_transform/TEST11 drd11.log ==74101== pmempool_transform/TEST11 drd11.log ==74101== Valgrind's memory management: out of memory: pmempool_transform/TEST11 drd11.log ==74101== newSuperblock's request for 4194304 bytes failed. pmempool_transform/TEST11 drd11.log ==74101== 10,504,224,768 bytes have already been mmap-ed ANONYMOUS. pmempool_transform/TEST11 drd11.log ==74101== Valgrind cannot continue. Sorry. pmempool_transform/TEST11 drd11.log ==74101== pmempool_transform/TEST11 drd11.log ==74101== There are several possible reasons for this. pmempool_transform/TEST11 drd11.log ==74101== - You have some kind of memory limit in place. Look at the pmempool_transform/TEST11 drd11.log ==74101== output of 'ulimit -a'. Is there a limit on the size of pmempool_transform/TEST11 drd11.log ==74101== virtual memory or address space? pmempool_transform/TEST11 drd11.log ==74101== - You have run out of swap space. pmempool_transform/TEST11 drd11.log ==74101== - Valgrind has a bug. If you think this is the case or you are pmempool_transform/TEST11 drd11.log ==74101== not sure, please let us know and we'll try to fix it. pmempool_transform/TEST11 drd11.log ==74101== Please note that programs can take substantially more memory than pmempool_transform/TEST11 drd11.log ==74101== normal when running under Valgrind tools, eg. up to twice or pmempool_transform/TEST11 drd11.log ==74101== more, depending on the tool. On a 64-bit machine, Valgrind pmempool_transform/TEST11 drd11.log ==74101== should be able to make use of up 32GB memory. On a 32-bit pmempool_transform/TEST11 drd11.log ==74101== machine, Valgrind should be able to use all the memory available pmempool_transform/TEST11 drd11.log ==74101== to a single process, up to 4GB if that's how you have your pmempool_transform/TEST11 drd11.log ==74101== kernel configured. Most 32-bit Linux setups allow a maximum of pmempool_transform/TEST11 drd11.log ==74101== 3GB per process. pmempool_transform/TEST11 drd11.log ==74101== pmempool_transform/TEST11 drd11.log ==74101== Whatever the reason, Valgrind cannot continue. Sorry. out11.log below. pmem11.log below. pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:235 out_init] pid 74101: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:238 out_init] libpmem version 1.1 pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:242 out_init] src version: 1.6+git99.gb19d5cc6b pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+ pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [mmap.c:67 util_mmap_init] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [libpmem.c:56 libpmem_init] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [pmem.c:784 pmem_init] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:419 pmem_init_funcs] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:368 pmem_cpuinfo_to_funcs] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:372 pmem_cpuinfo_to_funcs] clflush supported pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:281 use_avx_memcpy_memset] avx supported pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:285 use_avx_memcpy_memset] PMEM_AVX not set or not == 1 pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [pmem.c:216 pmem_has_auto_flush] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [os_auto_flush_linux.c:114 check_domain_in_region] region_path: /sys/bus/nd/devices/region0 pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [os_auto_flush_linux.c:59 check_cpu_cache] domain_path: /sys/bus/nd/devices/region0/persistence_domain pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:472 pmem_init_funcs] Flushing CPU cache pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:487 pmem_init_funcs] using clflush pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:501 pmem_init_funcs] using movnt SSE2 pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [pmem_posix.c:107 pmem_os_init] pmemblk11.log below. pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:235 out_init] pid 74101: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:238 out_init] libpmemblk version 1.1 pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:242 out_init] src version: 1.6+git99.gb19d5cc6b pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:265 out_init] compiled with support for Valgrind drd pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:270 out_init] compiled with support for shutdown state pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:275 out_init] compiled with libndctl 63+ pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <3> [mmap.c:67 util_mmap_init] pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <3> [libpmemblk.c:118 libpmemblk_init] pmemlog11.log below. pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:235 out_init] pid 74101: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:238 out_init] libpmemlog version 1.1 pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:242 out_init] src version: 1.6+git99.gb19d5cc6b pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:265 out_init] compiled with support for Valgrind drd pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:270 out_init] compiled with support for shutdown state pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:275 out_init] compiled with libndctl 63+ pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <3> [mmap.c:67 util_mmap_init] pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <3> [libpmemlog.c:118 libpmemlog_init] pmemobj11.log below. pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:235 out_init] pid 74101: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:238 out_init] libpmemobj version 2.4 pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:242 out_init] src version: 1.6+git99.gb19d5cc6b pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:265 out_init] compiled with support for Valgrind drd pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:270 out_init] compiled with support for shutdown state pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:275 out_init] compiled with libndctl 63+ pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [mmap.c:67 util_mmap_init] pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [libpmemobj.c:52 libpmemobj_init] pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [obj.c:283 obj_init] pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [obj.c:183 obj_ctl_init_and_load] pop (nil) pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [ctl.c:424 ctl_load_config_from_string] ctl (nil) ctx (nil) cfg_string "fallocate.at_create=0;" pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [ctl.c:300 ctl_query] ctl (nil) ctx (nil) source 2 name fallocate.at_create type 1 arg 0x4d4f524 pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [ctl.c:79 ctl_find_node] nodes 0x49bd900 name fallocate.at_create indexes 0x1ffefff588 pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [set.c:124 util_remote_init] Last 30 lines of pmempool11.log below (whole file has 357 lines). pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:67 device_dax_size] path "/dev/dax1.0" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:1781 util_part_open] part 0x4d55708 minsize 0 create 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:131 util_file_exists] path "/dev/dax1.3" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:558 util_file_open] path "/dev/dax1.3" size 0x1ffefff408 minsize 0 flags 2 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:258 util_file_get_size] path "/dev/dax1.3" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.3" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:131 util_file_exists] path "/dev/dax1.3" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:67 device_dax_size] path "/dev/dax1.3" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3547 util_replica_open] set 0x4d554a0 repidx 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3367 util_replica_open_local] set 0x4d554a0 repidx 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [mmap_posix.c:153 util_map_hint] len 63413678080 req_align 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:439 util_map_part] part 0x4d55688 addr 0x80000000 size 63413678080 offset 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:364 util_map_hdr] part 0x4d55688 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [mmap_posix.c:153 util_map_hint] len 4096 req_align 4096 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:364 util_map_hdr] part 0x4d55708 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [mmap_posix.c:153 util_map_hint] len 4096 req_align 4096 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:439 util_map_part] part 0x4d55708 addr 0x2f5e00000 size 0 offset 4096 flags 17 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:1130 util_replica_check_map_sync] set 0x4d554a0 repidx 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3488 util_replica_open_local] replica addr 0x80000000 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3547 util_replica_open] set 0x4d557b0 repidx 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3367 util_replica_open_local] set 0x4d557b0 repidx 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [mmap_posix.c:153 util_map_hint] len 63413682176 req_align 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:439 util_map_part] part 0x4d55968 addr 0x1040000000 size 63413682176 offset 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:364 util_map_hdr] part 0x4d55968 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [mmap_posix.c:153 util_map_hint] len 4096 req_align 4096 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:439 util_map_part] part 0x4d559e8 addr 0x12b5e00000 size 0 offset 0 flags 17 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:1130 util_replica_check_map_sync] set 0x4d557b0 repidx 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3488 util_replica_open_local] replica addr 0x1040000000 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [transform.c:584 copy_replica_data_fw] set_in 0x4d554a0, set_out 0x4d557b0, repn 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [replica.c:2117 replica_get_pool_size] set 0x4d554a0, repn 0 rpmem11.log below. pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:235 out_init] pid 74069: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:238 out_init] librpmem version 1.2 pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:242 out_init] src version: 1.6+git99.gb19d5cc6b pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+ pmempool_transform/TEST11 rpmem11.log <librpmem>: <3> [librpmem.c:61 librpmem_init] pmempool_transform/TEST11 rpmem11.log <librpmem>: <3> [librpmem.c:76 librpmem_fini] RUNTESTS: stopping: pmempool_transform/TEST11 failed, TEST=all FS=any BUILD=debug ``` ## Expected behavior: Tests should pass. ## Details [Logs.zip](https://github.com/pmem/issues/files/3467373/Logs.zip) ## Additional information about Priority and Help Requested: Are you willing to submit a pull request with a proposed change? (Yes, No) <!-- check one if possible --> Requested priority: (Showstopper, High, Medium, Low) <!-- check one if possible -->
1.0
Test: pmempool_transform/TEST[11, 14, 15, 18] fail with valgrind - <!-- Before creating new issue, ensure that similar issue wasn't already created * Search: https://github.com/pmem/issues/issues Note that if you do not provide enough information to reproduce the issue, we may not be able to take action on your report. Remember this is just a minimal template. You can extend it with data you think may be useful. --> # ISSUE: <!-- fill the title of issue --> ## Environment Information - PMDK package version(s): - OS(es) version(s): Fedora30 - ndctl version(s): 65 - kernel version(s): 5.1.17-300.fc30.x86_64 <!-- fill in also other useful environment data --> ## Please provide a reproduction of the bug: ``` ./RUNTESTS pmempool_transform -s TEST11 -d force-enable -t all ./RUNTESTS pmempool_transform -s TEST15 -d force-enable -t all ./RUNTESTS pmempool_transform -s TEST11 -e force-enable -t all ./RUNTESTS pmempool_transform -s TEST14 -e force-enable -t all ./RUNTESTS pmempool_transform -s TEST15 -e force-enable -t all ./RUNTESTS pmempool_transform -s TEST18 -e force-enable -t all ``` ## How often bug is revealed: (always, often, rare): always <!-- describe special circumstances in section above --> ## Actual behavior: ``` ./RUNTESTS pmempool_transform -s TEST11 -d force-enable -t all pmempool_transform/TEST11: SETUP (all/pmem/debug/drd) --74101:0: aspacem <<< SHOW_SEGMENTS: out_of_memory (177 segments) --74101:0: aspacem 25 segment names in 25 slots --74101:0: aspacem freelist is empty --74101:0: aspacem (0,4,9) /usr/local/lib/valgrind/drd-amd64-linux --74101:0: aspacem (1,48,8) /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool --74101:0: aspacem (2,104,8) /usr/lib64/ld-2.29.so --74101:0: aspacem (3,130,1) /tmp/vgdb-pipe-shared-mem-vgdb-74101-by-jenkins-on-localhost.localdomain --74101:0: aspacem (4,207,8) /usr/local/lib/valgrind/vgpreload_core-amd64-linux.so --74101:0: aspacem (5,265,8) /usr/local/lib/valgrind/vgpreload_drd-amd64-linux.so --74101:0: aspacem (6,322,9) /home/jenkins/greg/pmdk/src/debug/libpmempool.so.1.0.0 --74101:0: aspacem (7,381,8) /home/jenkins/greg/pmdk/src/debug/libpmemblk.so.1.0.0 --74101:0: aspacem (8,439,8) /home/jenkins/greg/pmdk/src/debug/libpmemlog.so.1.0.0 --74101:0: aspacem (9,497,9) /home/jenkins/greg/pmdk/src/debug/libpmemobj.so.1.0.0 --74101:0: aspacem (10,555,8) /home/jenkins/greg/pmdk/src/debug/libpmem.so.1.0.0 --74101:0: aspacem (11,610,8) /usr/lib64/libndctl.so.6.14.0 --74101:0: aspacem (12,644,7) /usr/lib64/libdaxctl.so.1.2.1 --74101:0: aspacem (13,678,8) /usr/lib64/libdl-2.29.so --74101:0: aspacem (14,707,8) /usr/lib64/libpthread-2.29.so --74101:0: aspacem (15,741,9) /usr/lib64/libc-2.29.so --74101:0: aspacem (16,769,9) /usr/lib64/libudev.so.1.6.13 --74101:0: aspacem (17,802,7) /usr/lib64/libuuid.so.1.3.0 --74101:0: aspacem (18,834,8) /usr/lib64/libkmod.so.2.3.3 --74101:0: aspacem (19,866,8) /usr/lib64/librt-2.29.so --74101:0: aspacem (20,895,8) /usr/lib64/libgcc_s-9-20190503.so.1 --74101:0: aspacem (21,935,8) /usr/lib64/liblzma.so.5.2.4 --74101:0: aspacem (22,967,7) /usr/lib64/libz.so.1.2.11 --74101:0: aspacem (23,997,10) /dev/dax1.0 --74101:0: aspacem (24,1013,5) /dev/dax1.3 --74101:0: aspacem 0: RSVN 0000000000-00003fffff 4194304 ----- SmFixed --74101:0: aspacem 1: file 0000400000-0000404fff 20480 r---- d=0xfd00 i=28113057 o=0 (1,48) --74101:0: aspacem 2: file 0000405000-000045afff 352256 r-xT- d=0xfd00 i=28113057 o=20480 (1,48) --74101:0: aspacem 3: file 000045b000-000047dfff 143360 r---- d=0xfd00 i=28113057 o=372736 (1,48) --74101:0: aspacem 4: RSVN 000047e000-000047efff 4096 ----- SmFixed --74101:0: aspacem 5: file 000047f000-0000480fff 8192 r---- d=0xfd00 i=28113057 o=516096 (1,48) --74101:0: aspacem 6: file 0000481000-0000481fff 4096 rw--- d=0xfd00 i=28113057 o=524288 (1,48) --74101:0: aspacem 7: anon 0000482000-0000488fff 28672 rw--- --74101:0: aspacem 8: RSVN 0000489000-0003ffffff 59m ----- SmFixed --74101:0: aspacem 9: file 0004000000-0004000fff 4096 r---- d=0xfd00 i=25168806 o=0 (2,104) --74101:0: aspacem 10: file 0004001000-0004020fff 131072 r-xT- d=0xfd00 i=25168806 o=4096 (2,104) --74101:0: aspacem 11: file 0004021000-0004028fff 32768 r---- d=0xfd00 i=25168806 o=135168 (2,104) --74101:0: aspacem 12: 0004029000-0004029fff 4096 --74101:0: aspacem 13: file 000402a000-000402afff 4096 r---- d=0xfd00 i=25168806 o=167936 (2,104) --74101:0: aspacem 14: file 000402b000-000402bfff 4096 rw--- d=0xfd00 i=25168806 o=172032 (2,104) --74101:0: aspacem 15: anon 000402c000-000402cfff 4096 rw--- --74101:0: aspacem 16: anon 000402d000-000402dfff 4096 rwx-- --74101:0: aspacem 17: RSVN 000402e000-000482cfff 8384512 ----- SmLower --74101:0: aspacem 18: file 000482d000-000482dfff 4096 r---- d=0xfd00 i=17991823 o=0 (4,207) --74101:0: aspacem 19: file 000482e000-000482efff 4096 r-xT- d=0xfd00 i=17991823 o=4096 (4,207) --74101:0: aspacem 20: file 000482f000-000482ffff 4096 r---- d=0xfd00 i=17991823 o=8192 (4,207) --74101:0: aspacem 21: file 0004830000-0004830fff 4096 r---- d=0xfd00 i=17991823 o=8192 (4,207) --74101:0: aspacem 22: file 0004831000-0004831fff 4096 rw--- d=0xfd00 i=17991823 o=12288 (4,207) --74101:0: aspacem 23: anon 0004832000-0004833fff 8192 rw--- --74101:0: aspacem 24: file 0004834000-000483afff 28672 r---- d=0xfd00 i=18008249 o=0 (5,265) --74101:0: aspacem 25: file 000483b000-0004851fff 94208 r-xT- d=0xfd00 i=18008249 o=28672 (5,265) --74101:0: aspacem 26: file 0004852000-0004856fff 20480 r---- d=0xfd00 i=18008249 o=122880 (5,265) --74101:0: aspacem 27: file 0004857000-0004857fff 4096 r---- d=0xfd00 i=18008249 o=139264 (5,265) --74101:0: aspacem 28: file 0004858000-0004858fff 4096 rw--- d=0xfd00 i=18008249 o=143360 (5,265) --74101:0: aspacem 29: file 0004859000-000485efff 24576 r---- d=0xfd00 i=11224907 o=0 (6,322) --74101:0: aspacem 30: file 000485f000-00048b1fff 339968 r-xT- d=0xfd00 i=11224907 o=24576 (6,322) --74101:0: aspacem 31: file 00048b2000-00048d3fff 139264 r---- d=0xfd00 i=11224907 o=364544 (6,322) --74101:0: aspacem 32: file 00048d4000-00048d4fff 4096 ----- d=0xfd00 i=11224907 o=503808 (6,322) --74101:0: aspacem 33: file 00048d5000-00048d5fff 4096 r---- d=0xfd00 i=11224907 o=503808 (6,322) --74101:0: aspacem 34: file 00048d6000-00048d6fff 4096 rw--- d=0xfd00 i=11224907 o=507904 (6,322) --74101:0: aspacem 35: anon 00048d7000-00048dcfff 24576 rw--- --74101:0: aspacem 36: file 00048dd000-00048e1fff 20480 r---- d=0xfd00 i=11239161 o=0 (7,381) --74101:0: aspacem 37: file 00048e2000-0004906fff 151552 r-xT- d=0xfd00 i=11239161 o=20480 (7,381) --74101:0: aspacem 38: file 0004907000-0004914fff 57344 r---- d=0xfd00 i=11239161 o=172032 (7,381) --74101:0: aspacem 39: file 0004915000-0004915fff 4096 r---- d=0xfd00 i=11239161 o=225280 (7,381) --74101:0: aspacem 40: file 0004916000-0004916fff 4096 rw--- d=0xfd00 i=11239161 o=229376 (7,381) --74101:0: aspacem 41: anon 0004917000-0004918fff 8192 rw--- --74101:0: aspacem 42: file 0004919000-000491dfff 20480 r---- d=0xfd00 i=11224899 o=0 (8,439) --74101:0: aspacem 43: file 000491e000-000493efff 135168 r-xT- d=0xfd00 i=11224899 o=20480 (8,439) --74101:0: aspacem 44: file 000493f000-000494bfff 53248 r---- d=0xfd00 i=11224899 o=155648 (8,439) --74101:0: aspacem 45: file 000494c000-000494cfff 4096 r---- d=0xfd00 i=11224899 o=204800 (8,439) --74101:0: aspacem 46: file 000494d000-000494dfff 4096 rw--- d=0xfd00 i=11224899 o=208896 (8,439) --74101:0: aspacem 47: anon 000494e000-000494ffff 8192 rw--- --74101:0: aspacem 48: file 0004950000-0004956fff 28672 r---- d=0xfd00 i=11224903 o=0 (9,497) --74101:0: aspacem 49: file 0004957000-00049a1fff 307200 r-xT- d=0xfd00 i=11224903 o=28672 (9,497) --74101:0: aspacem 50: file 00049a2000-00049b9fff 98304 r---- d=0xfd00 i=11224903 o=335872 (9,497) --74101:0: aspacem 51: file 00049ba000-00049bafff 4096 ----- d=0xfd00 i=11224903 o=434176 (9,497) --74101:0: aspacem 52: file 00049bb000-00049bcfff 8192 r---- d=0xfd00 i=11224903 o=434176 (9,497) --74101:0: aspacem 53: file 00049bd000-00049bdfff 4096 rw--- d=0xfd00 i=11224903 o=442368 (9,497) --74101:0: aspacem 54: anon 00049be000-00049c2fff 20480 rw--- --74101:0: aspacem 55: file 00049c3000-00049c5fff 12288 r---- d=0xfd00 i=11239755 o=0 (10,555) --74101:0: aspacem 56: file 00049c6000-0004a68fff 667648 r-xT- d=0xfd00 i=11239755 o=12288 (10,555) --74101:0: aspacem 57: file 0004a69000-0004a71fff 36864 r---- d=0xfd00 i=11239755 o=679936 (10,555) --74101:0: aspacem 58: file 0004a72000-0004a72fff 4096 r---- d=0xfd00 i=11239755 o=712704 (10,555) --74101:0: aspacem 59: file 0004a73000-0004a73fff 4096 rw--- d=0xfd00 i=11239755 o=716800 (10,555) --74101:0: aspacem 60: anon 0004a74000-0004a74fff 4096 rw--- --74101:0: aspacem 61: file 0004a75000-0004a7dfff 36864 r---- d=0xfd00 i=25170280 o=0 (11,610) --74101:0: aspacem 62: file 0004a7e000-0004a8ffff 73728 r-xT- d=0xfd00 i=25170280 o=36864 (11,610) --74101:0: aspacem 63: file 0004a90000-0004a97fff 32768 r---- d=0xfd00 i=25170280 o=110592 (11,610) --74101:0: aspacem 64: file 0004a98000-0004a98fff 4096 r---- d=0xfd00 i=25170280 o=139264 (11,610) --74101:0: aspacem 65: file 0004a99000-0004a99fff 4096 rw--- d=0xfd00 i=25170280 o=143360 (11,610) --74101:0: aspacem 66: file 0004a9a000-0004a9bfff 8192 r---- d=0xfd00 i=25170278 o=0 (12,644) --74101:0: aspacem 67: file 0004a9c000-0004a9dfff 8192 r-xT- d=0xfd00 i=25170278 o=8192 (12,644) --74101:0: aspacem 68: file 0004a9e000-0004a9efff 4096 r---- d=0xfd00 i=25170278 o=16384 (12,644) --74101:0: aspacem 69: file 0004a9f000-0004a9ffff 4096 r---- d=0xfd00 i=25170278 o=16384 (12,644) --74101:0: aspacem 70: anon 0004aa0000-0004aa0fff 4096 rw--- --74101:0: aspacem 71: file 0004aa1000-0004aa1fff 4096 r---- d=0xfd00 i=25488751 o=0 (13,678) --74101:0: aspacem 72: file 0004aa2000-0004aa3fff 8192 r-xT- d=0xfd00 i=25488751 o=4096 (13,678) --74101:0: aspacem 73: file 0004aa4000-0004aa4fff 4096 r---- d=0xfd00 i=25488751 o=12288 (13,678) --74101:0: aspacem 74: file 0004aa5000-0004aa5fff 4096 r---- d=0xfd00 i=25488751 o=12288 (13,678) --74101:0: aspacem 75: file 0004aa6000-0004aa6fff 4096 rw--- d=0xfd00 i=25488751 o=16384 (13,678) --74101:0: aspacem 76: file 0004aa7000-0004aacfff 24576 r---- d=0xfd00 i=25488945 o=0 (14,707) --74101:0: aspacem 77: file 0004aad000-0004abbfff 61440 r-xT- d=0xfd00 i=25488945 o=24576 (14,707) --74101:0: aspacem 78: file 0004abc000-0004ac1fff 24576 r---- d=0xfd00 i=25488945 o=86016 (14,707) --74101:0: aspacem 79: file 0004ac2000-0004ac2fff 4096 r---- d=0xfd00 i=25488945 o=106496 (14,707) --74101:0: aspacem 80: file 0004ac3000-0004ac3fff 4096 rw--- d=0xfd00 i=25488945 o=110592 (14,707) --74101:0: aspacem 81: anon 0004ac4000-0004ac7fff 16384 rw--- --74101:0: aspacem 82: file 0004ac8000-0004ae9fff 139264 r---- d=0xfd00 i=25168814 o=0 (15,741) --74101:0: aspacem 83: file 0004aea000-0004c36fff 1363968 r-xT- d=0xfd00 i=25168814 o=139264 (15,741) --74101:0: aspacem 84: file 0004c37000-0004c82fff 311296 r---- d=0xfd00 i=25168814 o=1503232 (15,741) --74101:0: aspacem 85: file 0004c83000-0004c83fff 4096 ----- d=0xfd00 i=25168814 o=1814528 (15,741) --74101:0: aspacem 86: file 0004c84000-0004c87fff 16384 r---- d=0xfd00 i=25168814 o=1814528 (15,741) --74101:0: aspacem 87: file 0004c88000-0004c89fff 8192 rw--- d=0xfd00 i=25168814 o=1830912 (15,741) --74101:0: aspacem 88: anon 0004c8a000-0004c8ffff 24576 rw--- --74101:0: aspacem 89: file 0004c90000-0004c94fff 20480 r---- d=0xfd00 i=25183443 o=0 (16,769) --74101:0: aspacem 90: file 0004c95000-0004caefff 106496 r-xT- d=0xfd00 i=25183443 o=20480 (16,769) --74101:0: aspacem 91: file 0004caf000-0004cb8fff 40960 r---- d=0xfd00 i=25183443 o=126976 (16,769) --74101:0: aspacem 92: file 0004cb9000-0004cb9fff 4096 ----- d=0xfd00 i=25183443 o=167936 (16,769) --74101:0: aspacem 93: file 0004cba000-0004cbafff 4096 r---- d=0xfd00 i=25183443 o=167936 (16,769) --74101:0: aspacem 94: file 0004cbb000-0004cbbfff 4096 rw--- d=0xfd00 i=25183443 o=172032 (16,769) --74101:0: aspacem 95: file 0004cbc000-0004cbdfff 8192 r---- d=0xfd00 i=25169119 o=0 (17,802) --74101:0: aspacem 96: file 0004cbe000-0004cc2fff 20480 r-xT- d=0xfd00 i=25169119 o=8192 (17,802) --74101:0: aspacem 97: file 0004cc3000-0004cc3fff 4096 r---- d=0xfd00 i=25169119 o=28672 (17,802) --74101:0: aspacem 98: file 0004cc4000-0004cc4fff 4096 r---- d=0xfd00 i=25169119 o=28672 (17,802) --74101:0: aspacem 99: anon 0004cc5000-0004cc5fff 4096 rw--- --74101:0: aspacem 100: file 0004cc6000-0004cc8fff 12288 r---- d=0xfd00 i=25169145 o=0 (18,834) --74101:0: aspacem 101: file 0004cc9000-0004cd7fff 61440 r-xT- d=0xfd00 i=25169145 o=12288 (18,834) --74101:0: aspacem 102: file 0004cd8000-0004cdcfff 20480 r---- d=0xfd00 i=25169145 o=73728 (18,834) --74101:0: aspacem 103: file 0004cdd000-0004cddfff 4096 r---- d=0xfd00 i=25169145 o=90112 (18,834) --74101:0: aspacem 104: file 0004cde000-0004cdefff 4096 rw--- d=0xfd00 i=25169145 o=94208 (18,834) --74101:0: aspacem 105: file 0004cdf000-0004ce0fff 8192 r---- d=0xfd00 i=25488951 o=0 (19,866) --74101:0: aspacem 106: file 0004ce1000-0004ce4fff 16384 r-xT- d=0xfd00 i=25488951 o=8192 (19,866) --74101:0: aspacem 107: file 0004ce5000-0004ce6fff 8192 r---- d=0xfd00 i=25488951 o=24576 (19,866) --74101:0: aspacem 108: file 0004ce7000-0004ce7fff 4096 r---- d=0xfd00 i=25488951 o=28672 (19,866) --74101:0: aspacem 109: file 0004ce8000-0004ce8fff 4096 rw--- d=0xfd00 i=25488951 o=32768 (19,866) --74101:0: aspacem 110: file 0004ce9000-0004cebfff 12288 r---- d=0xfd00 i=25168754 o=0 (20,895) --74101:0: aspacem 111: file 0004cec000-0004cfcfff 69632 r-xT- d=0xfd00 i=25168754 o=12288 (20,895) --74101:0: aspacem 112: file 0004cfd000-0004d00fff 16384 r---- d=0xfd00 i=25168754 o=81920 (20,895) --74101:0: aspacem 113: file 0004d01000-0004d01fff 4096 r---- d=0xfd00 i=25168754 o=94208 (20,895) --74101:0: aspacem 114: file 0004d02000-0004d02fff 4096 rw--- d=0xfd00 i=25168754 o=98304 (20,895) --74101:0: aspacem 115: file 0004d03000-0004d05fff 12288 r---- d=0xfd00 i=25169099 o=0 (21,935) --74101:0: aspacem 116: file 0004d06000-0004d1dfff 98304 r-xT- d=0xfd00 i=25169099 o=12288 (21,935) --74101:0: aspacem 117: file 0004d1e000-0004d28fff 45056 r---- d=0xfd00 i=25169099 o=110592 (21,935) --74101:0: aspacem 118: file 0004d29000-0004d29fff 4096 ----- d=0xfd00 i=25169099 o=155648 (21,935) --74101:0: aspacem 119: file 0004d2a000-0004d2afff 4096 r---- d=0xfd00 i=25169099 o=155648 (21,935) --74101:0: aspacem 120: anon 0004d2b000-0004d2dfff 12288 rw--- --74101:0: aspacem 121: file 0004d2e000-0004d30fff 12288 r---- d=0xfd00 i=25169023 o=0 (22,967) --74101:0: aspacem 122: file 0004d31000-0004d3efff 57344 r-xT- d=0xfd00 i=25169023 o=12288 (22,967) --74101:0: aspacem 123: file 0004d3f000-0004d45fff 28672 r---- d=0xfd00 i=25169023 o=69632 (22,967) --74101:0: aspacem 124: file 0004d46000-0004d46fff 4096 r---- d=0xfd00 i=25169023 o=94208 (22,967) --74101:0: aspacem 125: anon 0004d47000-0004d4cfff 24576 rw--- --74101:0: aspacem 126: anon 0004d4d000-000514cfff 4194304 rwx-H --74101:0: aspacem 127: file 000514d000-000514dfff 4096 rw--- d=0x006 i=16516 o=0 (23,997) --74101:0: aspacem 128: file 000514e000-000514efff 4096 rw--- d=0x006 i=16519 o=0 (24,1013) --74101:0: aspacem 129: file 000514f000-000514ffff 4096 rw--- d=0x006 i=16516 o=0 (23,997) --74101:0: aspacem 130: ANON 0005150000-0057d4ffff 1324m rwx-- --74101:0: aspacem 131: 0057d50000-0057ffffff 2818048 --74101:0: aspacem 132: FILE 0058000000-0058000fff 4096 r---- d=0xfd00 i=18008248 o=0 (0,4) --74101:0: aspacem 133: FILE 0058001000-005809bfff 634880 r-x-- d=0xfd00 i=18008248 o=4096 (0,4) --74101:0: aspacem 134: file 005809c000-005809cfff 4096 r-x-- d=0xfd00 i=18008248 o=638976 (0,4) --74101:0: aspacem 135: FILE 005809d000-00581befff 1187840 r-x-- d=0xfd00 i=18008248 o=643072 (0,4) --74101:0: aspacem 136: FILE 00581bf000-0058267fff 692224 r---- d=0xfd00 i=18008248 o=1830912 (0,4) --74101:0: aspacem 137: 0058268000-0058268fff 4096 --74101:0: aspacem 138: FILE 0058269000-005826bfff 12288 rw--- d=0xfd00 i=18008248 o=2523136 (0,4) --74101:0: aspacem 139: ANON 005826c000-0058c5efff 9m rw--- --74101:0: aspacem 140: ANON 0058c5f000-007fc5efff 624m rwx-- --74101:0: aspacem 141: 007fc5f000-007fffffff 3805184 --74101:0: aspacem 142: file 0080000000-02f5dfffff 10078m rw--- d=0x006 i=16516 o=0 (23,997) --74101:0: aspacem 143: file 02f5e00000-0f43bfefff 50397m rw--- d=0x006 i=16519 o=4096 (24,1013) --74101:0: aspacem 144: ANON 0f43bff000-1001ffefff 3044m rwx-- --74101:0: aspacem 145: 1001fff000-1001ffffff 4096 --74101:0: aspacem 146: RSVN 1002000000-1002000fff 4096 ----- SmFixed --74101:0: aspacem 147: ANON 1002001000-100268bfff 6860800 rwx-- --74101:0: aspacem 148: ANON 100268c000-100268dfff 8192 ----- --74101:0: aspacem 149: ANON 100268e000-100278dfff 1048576 rwx-- --74101:0: aspacem 150: ANON 100278e000-100278ffff 8192 ----- --74101:0: aspacem 151: FILE 1002790000-1002790fff 4096 rw--- d=0x02b i=14406966 o=0 (3,130) --74101:0: aspacem 152: ANON 1002791000-10027e0fff 327680 rwx-- --74101:0: aspacem 153: 10027e1000-10027f6fff 90112 --74101:0: aspacem 154: ANON 10027f7000-1002816fff 131072 rwx-- --74101:0: aspacem 155: 1002817000-100288bfff 479232 --74101:0: aspacem 156: ANON 100288c000-1002bd8fff 3461120 rwx-- --74101:0: aspacem 157: 1002bd9000-1002c80fff 688128 --74101:0: aspacem 158: ANON 1002c81000-10033d5fff 7688192 rwx-- --74101:0: aspacem 159: 10033d6000-1003469fff 606208 --74101:0: aspacem 160: ANON 100346a000-103fd13fff 968m rwx-- --74101:0: aspacem 161: 103fd14000-103fffffff 3063808 --74101:0: aspacem 162: file 1040000000-12b5dfffff 10078m rw--- d=0x006 i=16516 o=0 (23,997) --74101:0: aspacem 163: file 12b5e00000-1f03bfffff 50398m rw--- d=0x006 i=16519 o=0 (24,1013) --74101:0: aspacem 164: ANON 1f03c00000-1ffe7fffff 4012m rwx-- --74101:0: aspacem 165: 1ffe800000-1ffe800fff 4096 --74101:0: aspacem 166: RSVN 1ffe801000-1ffeff8fff 8355840 ----- SmUpper --74101:0: aspacem 167: anon 1ffeff9000-1fff000fff 32768 rw--- --74101:0: aspacem 168: ANON 1fff001000-1fffc00fff 12m rwx-- --74101:0: aspacem 169: 1fffc01000-1fffffffff 4190208 --74101:0: aspacem 170: RSVN 2000000000-7ffe930eefff 130938g ----- SmFixed --74101:0: aspacem 171: ANON 7ffe930ef000-7ffe93110fff 139264 rw--- --74101:0: aspacem 172: RSVN 7ffe93111000-7ffe9312dfff 118784 ----- SmFixed --74101:0: aspacem 173: ANON 7ffe9312e000-7ffe93130fff 12288 r---- --74101:0: aspacem 174: RSVN 7ffe93131000-ffffffffff5fffff 16383e ----- SmFixed --74101:0: aspacem 175: ANON ffffffffff600000-ffffffffff600fff 4096 r-x-- --74101:0: aspacem 176: RSVN ffffffffff601000-ffffffffffffffff 9m ----- SmFixed --74101:0: aspacem >>> pmempool_transform/TEST11 failed with exit code 1. Last 30 lines of drd11.log below (whole file has 95 lines). pmempool_transform/TEST11 drd11.log ==74101== by 0x48932E5: pmempool_transformU (replica.c:2462) pmempool_transform/TEST11 drd11.log ==74101== by 0x48933A5: pmempool_transform (replica.c:2492) pmempool_transform/TEST11 drd11.log ==74101== by 0x40F164: pmempool_transform_func (in /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool) pmempool_transform/TEST11 drd11.log ==74101== by 0x406AD3: main (in /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool) pmempool_transform/TEST11 drd11.log client stack range: [0x1FFEFF9000 0x1FFF000FFF] client SP: 0x1FFEFFD4B0 pmempool_transform/TEST11 drd11.log valgrind stack range: [0x100268E000 0x100278DFFF] top usage: 12072 of 1048576 pmempool_transform/TEST11 drd11.log pmempool_transform/TEST11 drd11.log ==74101== pmempool_transform/TEST11 drd11.log ==74101== Valgrind's memory management: out of memory: pmempool_transform/TEST11 drd11.log ==74101== newSuperblock's request for 4194304 bytes failed. pmempool_transform/TEST11 drd11.log ==74101== 10,504,224,768 bytes have already been mmap-ed ANONYMOUS. pmempool_transform/TEST11 drd11.log ==74101== Valgrind cannot continue. Sorry. pmempool_transform/TEST11 drd11.log ==74101== pmempool_transform/TEST11 drd11.log ==74101== There are several possible reasons for this. pmempool_transform/TEST11 drd11.log ==74101== - You have some kind of memory limit in place. Look at the pmempool_transform/TEST11 drd11.log ==74101== output of 'ulimit -a'. Is there a limit on the size of pmempool_transform/TEST11 drd11.log ==74101== virtual memory or address space? pmempool_transform/TEST11 drd11.log ==74101== - You have run out of swap space. pmempool_transform/TEST11 drd11.log ==74101== - Valgrind has a bug. If you think this is the case or you are pmempool_transform/TEST11 drd11.log ==74101== not sure, please let us know and we'll try to fix it. pmempool_transform/TEST11 drd11.log ==74101== Please note that programs can take substantially more memory than pmempool_transform/TEST11 drd11.log ==74101== normal when running under Valgrind tools, eg. up to twice or pmempool_transform/TEST11 drd11.log ==74101== more, depending on the tool. On a 64-bit machine, Valgrind pmempool_transform/TEST11 drd11.log ==74101== should be able to make use of up 32GB memory. On a 32-bit pmempool_transform/TEST11 drd11.log ==74101== machine, Valgrind should be able to use all the memory available pmempool_transform/TEST11 drd11.log ==74101== to a single process, up to 4GB if that's how you have your pmempool_transform/TEST11 drd11.log ==74101== kernel configured. Most 32-bit Linux setups allow a maximum of pmempool_transform/TEST11 drd11.log ==74101== 3GB per process. pmempool_transform/TEST11 drd11.log ==74101== pmempool_transform/TEST11 drd11.log ==74101== Whatever the reason, Valgrind cannot continue. Sorry. out11.log below. pmem11.log below. pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:235 out_init] pid 74101: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:238 out_init] libpmem version 1.1 pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:242 out_init] src version: 1.6+git99.gb19d5cc6b pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state pmempool_transform/TEST11 pmem11.log <libpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+ pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [mmap.c:67 util_mmap_init] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [libpmem.c:56 libpmem_init] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [pmem.c:784 pmem_init] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:419 pmem_init_funcs] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:368 pmem_cpuinfo_to_funcs] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:372 pmem_cpuinfo_to_funcs] clflush supported pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:281 use_avx_memcpy_memset] avx supported pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:285 use_avx_memcpy_memset] PMEM_AVX not set or not == 1 pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [pmem.c:216 pmem_has_auto_flush] pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [os_auto_flush_linux.c:114 check_domain_in_region] region_path: /sys/bus/nd/devices/region0 pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [os_auto_flush_linux.c:59 check_cpu_cache] domain_path: /sys/bus/nd/devices/region0/persistence_domain pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:472 pmem_init_funcs] Flushing CPU cache pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:487 pmem_init_funcs] using clflush pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [init.c:501 pmem_init_funcs] using movnt SSE2 pmempool_transform/TEST11 pmem11.log <libpmem>: <3> [pmem_posix.c:107 pmem_os_init] pmemblk11.log below. pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:235 out_init] pid 74101: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:238 out_init] libpmemblk version 1.1 pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:242 out_init] src version: 1.6+git99.gb19d5cc6b pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:265 out_init] compiled with support for Valgrind drd pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:270 out_init] compiled with support for shutdown state pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <1> [out.c:275 out_init] compiled with libndctl 63+ pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <3> [mmap.c:67 util_mmap_init] pmempool_transform/TEST11 pmemblk11.log <libpmemblk>: <3> [libpmemblk.c:118 libpmemblk_init] pmemlog11.log below. pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:235 out_init] pid 74101: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:238 out_init] libpmemlog version 1.1 pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:242 out_init] src version: 1.6+git99.gb19d5cc6b pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:265 out_init] compiled with support for Valgrind drd pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:270 out_init] compiled with support for shutdown state pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <1> [out.c:275 out_init] compiled with libndctl 63+ pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <3> [mmap.c:67 util_mmap_init] pmempool_transform/TEST11 pmemlog11.log <libpmemlog>: <3> [libpmemlog.c:118 libpmemlog_init] pmemobj11.log below. pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:235 out_init] pid 74101: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:238 out_init] libpmemobj version 2.4 pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:242 out_init] src version: 1.6+git99.gb19d5cc6b pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:265 out_init] compiled with support for Valgrind drd pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:270 out_init] compiled with support for shutdown state pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <1> [out.c:275 out_init] compiled with libndctl 63+ pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [mmap.c:67 util_mmap_init] pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [libpmemobj.c:52 libpmemobj_init] pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [obj.c:283 obj_init] pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [obj.c:183 obj_ctl_init_and_load] pop (nil) pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [ctl.c:424 ctl_load_config_from_string] ctl (nil) ctx (nil) cfg_string "fallocate.at_create=0;" pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [ctl.c:300 ctl_query] ctl (nil) ctx (nil) source 2 name fallocate.at_create type 1 arg 0x4d4f524 pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [ctl.c:79 ctl_find_node] nodes 0x49bd900 name fallocate.at_create indexes 0x1ffefff588 pmempool_transform/TEST11 pmemobj11.log <libpmemobj>: <3> [set.c:124 util_remote_init] Last 30 lines of pmempool11.log below (whole file has 357 lines). pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:67 device_dax_size] path "/dev/dax1.0" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:1781 util_part_open] part 0x4d55708 minsize 0 create 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:131 util_file_exists] path "/dev/dax1.3" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:558 util_file_open] path "/dev/dax1.3" size 0x1ffefff408 minsize 0 flags 2 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:258 util_file_get_size] path "/dev/dax1.3" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:223 util_file_get_type] path "/dev/dax1.3" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:131 util_file_exists] path "/dev/dax1.3" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [file.c:67 device_dax_size] path "/dev/dax1.3" pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3547 util_replica_open] set 0x4d554a0 repidx 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3367 util_replica_open_local] set 0x4d554a0 repidx 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [mmap_posix.c:153 util_map_hint] len 63413678080 req_align 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:439 util_map_part] part 0x4d55688 addr 0x80000000 size 63413678080 offset 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:364 util_map_hdr] part 0x4d55688 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [mmap_posix.c:153 util_map_hint] len 4096 req_align 4096 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:364 util_map_hdr] part 0x4d55708 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [mmap_posix.c:153 util_map_hint] len 4096 req_align 4096 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:439 util_map_part] part 0x4d55708 addr 0x2f5e00000 size 0 offset 4096 flags 17 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:1130 util_replica_check_map_sync] set 0x4d554a0 repidx 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3488 util_replica_open_local] replica addr 0x80000000 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3547 util_replica_open] set 0x4d557b0 repidx 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3367 util_replica_open_local] set 0x4d557b0 repidx 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [mmap_posix.c:153 util_map_hint] len 63413682176 req_align 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:439 util_map_part] part 0x4d55968 addr 0x1040000000 size 63413682176 offset 0 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:364 util_map_hdr] part 0x4d55968 flags 1 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [mmap_posix.c:153 util_map_hint] len 4096 req_align 4096 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:439 util_map_part] part 0x4d559e8 addr 0x12b5e00000 size 0 offset 0 flags 17 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:1130 util_replica_check_map_sync] set 0x4d557b0 repidx 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [set.c:3488 util_replica_open_local] replica addr 0x1040000000 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [transform.c:584 copy_replica_data_fw] set_in 0x4d554a0, set_out 0x4d557b0, repn 0 pmempool_transform/TEST11 pmempool11.log <libpmempool>: <3> [replica.c:2117 replica_get_pool_size] set 0x4d554a0, repn 0 rpmem11.log below. pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:235 out_init] pid 74069: program: /home/jenkins/greg/pmdk/src/tools/pmempool/pmempool.static-debug pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:238 out_init] librpmem version 1.2 pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:242 out_init] src version: 1.6+git99.gb19d5cc6b pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:250 out_init] compiled with support for Valgrind pmemcheck pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:255 out_init] compiled with support for Valgrind helgrind pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:260 out_init] compiled with support for Valgrind memcheck pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:265 out_init] compiled with support for Valgrind drd pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:270 out_init] compiled with support for shutdown state pmempool_transform/TEST11 rpmem11.log <librpmem>: <1> [out.c:275 out_init] compiled with libndctl 63+ pmempool_transform/TEST11 rpmem11.log <librpmem>: <3> [librpmem.c:61 librpmem_init] pmempool_transform/TEST11 rpmem11.log <librpmem>: <3> [librpmem.c:76 librpmem_fini] RUNTESTS: stopping: pmempool_transform/TEST11 failed, TEST=all FS=any BUILD=debug ``` ## Expected behavior: Tests should pass. ## Details [Logs.zip](https://github.com/pmem/issues/files/3467373/Logs.zip) ## Additional information about Priority and Help Requested: Are you willing to submit a pull request with a proposed change? (Yes, No) <!-- check one if possible --> Requested priority: (Showstopper, High, Medium, Low) <!-- check one if possible -->
priority
test pmempool transform test fail with valgrind before creating new issue ensure that similar issue wasn t already created search note that if you do not provide enough information to reproduce the issue we may not be able to take action on your report remember this is just a minimal template you can extend it with data you think may be useful issue environment information pmdk package version s os es version s ndctl version s kernel version s please provide a reproduction of the bug runtests pmempool transform s d force enable t all runtests pmempool transform s d force enable t all runtests pmempool transform s e force enable t all runtests pmempool transform s e force enable t all runtests pmempool transform s e force enable t all runtests pmempool transform s e force enable t all how often bug is revealed always often rare always actual behavior runtests pmempool transform s d force enable t all pmempool transform setup all pmem debug drd aspacem show segments out of memory segments aspacem segment names in slots aspacem freelist is empty aspacem usr local lib valgrind drd linux aspacem home jenkins greg pmdk src tools pmempool pmempool aspacem usr ld so aspacem tmp vgdb pipe shared mem vgdb by jenkins on localhost localdomain aspacem usr local lib valgrind vgpreload core linux so aspacem usr local lib valgrind vgpreload drd linux so aspacem home jenkins greg pmdk src debug libpmempool so aspacem home jenkins greg pmdk src debug libpmemblk so aspacem home jenkins greg pmdk src debug libpmemlog so aspacem home jenkins greg pmdk src debug libpmemobj so aspacem home jenkins greg pmdk src debug libpmem so aspacem usr libndctl so aspacem usr libdaxctl so aspacem usr libdl so aspacem usr libpthread so aspacem usr libc so aspacem usr libudev so aspacem usr libuuid so aspacem usr libkmod so aspacem usr librt so aspacem usr libgcc s so aspacem usr liblzma so aspacem usr libz so aspacem dev aspacem dev aspacem rsvn smfixed aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem rsvn smfixed aspacem file r d i o aspacem file rw d i o aspacem anon rw aspacem rsvn smfixed aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem aspacem file r d i o aspacem file rw d i o aspacem anon rw aspacem anon rwx aspacem rsvn smlower aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem file rw d i o aspacem anon rw aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem file rw d i o aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file d i o aspacem file r d i o aspacem file rw d i o aspacem anon rw aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem file rw d i o aspacem anon rw aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem file rw d i o aspacem anon rw aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file d i o aspacem file r d i o aspacem file rw d i o aspacem anon rw aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem file rw d i o aspacem anon rw aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem file rw d i o aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem anon rw aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem file rw d i o aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem file rw d i o aspacem anon rw aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file d i o aspacem file r d i o aspacem file rw d i o aspacem anon rw aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file d i o aspacem file r d i o aspacem file rw d i o aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem anon rw aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem file rw d i o aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem file rw d i o aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem file rw d i o aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file d i o aspacem file r d i o aspacem anon rw aspacem file r d i o aspacem file r xt d i o aspacem file r d i o aspacem file r d i o aspacem anon rw aspacem anon rwx h aspacem file rw d i o aspacem file rw d i o aspacem file rw d i o aspacem anon rwx aspacem aspacem file r d i o aspacem file r x d i o aspacem file r x d i o aspacem file r x d i o aspacem file r d i o aspacem aspacem file rw d i o aspacem anon rw aspacem anon rwx aspacem aspacem file rw d i o aspacem file rw d i o aspacem anon rwx aspacem aspacem rsvn smfixed aspacem anon rwx aspacem anon aspacem anon rwx aspacem anon aspacem file rw d i o aspacem anon rwx aspacem aspacem anon rwx aspacem aspacem anon rwx aspacem aspacem anon rwx aspacem aspacem anon rwx aspacem aspacem file rw d i o aspacem file rw d i o aspacem anon rwx aspacem aspacem rsvn smupper aspacem anon rw aspacem anon rwx aspacem aspacem rsvn smfixed aspacem anon rw aspacem rsvn smfixed aspacem anon r aspacem rsvn smfixed aspacem anon r x aspacem rsvn ffffffffffffffff smfixed aspacem pmempool transform failed with exit code last lines of log below whole file has lines pmempool transform log by pmempool transformu replica c pmempool transform log by pmempool transform replica c pmempool transform log by pmempool transform func in home jenkins greg pmdk src tools pmempool pmempool pmempool transform log by main in home jenkins greg pmdk src tools pmempool pmempool pmempool transform log client stack range client sp pmempool transform log valgrind stack range top usage of pmempool transform log pmempool transform log pmempool transform log valgrind s memory management out of memory pmempool transform log newsuperblock s request for bytes failed pmempool transform log bytes have already been mmap ed anonymous pmempool transform log valgrind cannot continue sorry pmempool transform log pmempool transform log there are several possible reasons for this pmempool transform log you have some kind of memory limit in place look at the pmempool transform log output of ulimit a is there a limit on the size of pmempool transform log virtual memory or address space pmempool transform log you have run out of swap space pmempool transform log valgrind has a bug if you think this is the case or you are pmempool transform log not sure please let us know and we ll try to fix it pmempool transform log please note that programs can take substantially more memory than pmempool transform log normal when running under valgrind tools eg up to twice or pmempool transform log more depending on the tool on a bit machine valgrind pmempool transform log should be able to make use of up memory on a bit pmempool transform log machine valgrind should be able to use all the memory available pmempool transform log to a single process up to if that s how you have your pmempool transform log kernel configured most bit linux setups allow a maximum of pmempool transform log per process pmempool transform log pmempool transform log whatever the reason valgrind cannot continue sorry log below log below pmempool transform log pid program home jenkins greg pmdk src tools pmempool pmempool pmempool transform log libpmem version pmempool transform log src version pmempool transform log compiled with support for valgrind pmemcheck pmempool transform log compiled with support for valgrind helgrind pmempool transform log compiled with support for valgrind memcheck pmempool transform log compiled with support for valgrind drd pmempool transform log compiled with support for shutdown state pmempool transform log compiled with libndctl pmempool transform log pmempool transform log pmempool transform log pmempool transform log pmempool transform log pmempool transform log clflush supported pmempool transform log avx supported pmempool transform log pmem avx not set or not pmempool transform log pmempool transform log region path sys bus nd devices pmempool transform log domain path sys bus nd devices persistence domain pmempool transform log flushing cpu cache pmempool transform log using clflush pmempool transform log using movnt pmempool transform log log below pmempool transform log pid program home jenkins greg pmdk src tools pmempool pmempool pmempool transform log libpmemblk version pmempool transform log src version pmempool transform log compiled with support for valgrind pmemcheck pmempool transform log compiled with support for valgrind helgrind pmempool transform log compiled with support for valgrind memcheck pmempool transform log compiled with support for valgrind drd pmempool transform log compiled with support for shutdown state pmempool transform log compiled with libndctl pmempool transform log pmempool transform log log below pmempool transform log pid program home jenkins greg pmdk src tools pmempool pmempool pmempool transform log libpmemlog version pmempool transform log src version pmempool transform log compiled with support for valgrind pmemcheck pmempool transform log compiled with support for valgrind helgrind pmempool transform log compiled with support for valgrind memcheck pmempool transform log compiled with support for valgrind drd pmempool transform log compiled with support for shutdown state pmempool transform log compiled with libndctl pmempool transform log pmempool transform log log below pmempool transform log pid program home jenkins greg pmdk src tools pmempool pmempool pmempool transform log libpmemobj version pmempool transform log src version pmempool transform log compiled with support for valgrind pmemcheck pmempool transform log compiled with support for valgrind helgrind pmempool transform log compiled with support for valgrind memcheck pmempool transform log compiled with support for valgrind drd pmempool transform log compiled with support for shutdown state pmempool transform log compiled with libndctl pmempool transform log pmempool transform log pmempool transform log pmempool transform log pop nil pmempool transform log ctl nil ctx nil cfg string fallocate at create pmempool transform log ctl nil ctx nil source name fallocate at create type arg pmempool transform log nodes name fallocate at create indexes pmempool transform log last lines of log below whole file has lines pmempool transform log path dev pmempool transform log part minsize create pmempool transform log path dev pmempool transform log path dev size minsize flags pmempool transform log path dev pmempool transform log path dev pmempool transform log path dev pmempool transform log path dev pmempool transform log set repidx flags pmempool transform log set repidx flags pmempool transform log len req align pmempool transform log part addr size offset flags pmempool transform log part flags pmempool transform log len req align pmempool transform log part flags pmempool transform log len req align pmempool transform log part addr size offset flags pmempool transform log set repidx pmempool transform log replica addr pmempool transform log set repidx flags pmempool transform log set repidx flags pmempool transform log len req align pmempool transform log part addr size offset flags pmempool transform log part flags pmempool transform log len req align pmempool transform log part addr size offset flags pmempool transform log set repidx pmempool transform log replica addr pmempool transform log set in set out repn pmempool transform log set repn log below pmempool transform log pid program home jenkins greg pmdk src tools pmempool pmempool static debug pmempool transform log librpmem version pmempool transform log src version pmempool transform log compiled with support for valgrind pmemcheck pmempool transform log compiled with support for valgrind helgrind pmempool transform log compiled with support for valgrind memcheck pmempool transform log compiled with support for valgrind drd pmempool transform log compiled with support for shutdown state pmempool transform log compiled with libndctl pmempool transform log pmempool transform log runtests stopping pmempool transform failed test all fs any build debug expected behavior tests should pass details additional information about priority and help requested are you willing to submit a pull request with a proposed change yes no requested priority showstopper high medium low
1
384,682
11,396,641,053
IssuesEvent
2020-01-30 13:56:31
input-output-hk/jormungandr
https://api.github.com/repos/input-output-hk/jormungandr
closed
Tokio Thread Panicked: Validated Block Must Be Unique
Priority - High bug subsys-storage
**Describe the bug** Shortly after bootstrapping I received: thread 'tokio-runtime-worker-1' panicked at 'validated block must be unique', src/libcore/option.rs:1185:5 note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace. thread 'main' panicked at 'internal error: entered unreachable code', /root/.cargo/registry/src/github.com-1ecc6299db9ec823/tokio-0.1.22/src/runtime/threadpool/mod.rs:296:26 **Mandatory Information** 1. `jcli --full-version` output; jcli 0.8.8 (master-4687999, release, linux [x86_64]) - [rustc 1.40.0 (73528e339 2019-12-16)] 2. `jormungandr --full-version` output; jormungandr 0.8.8 (master-4687999, release, linux [x86_64]) - [rustc 1.40.0 (73528e339 2019-12-16)] **To Reproduce** Steps to reproduce the behavior: 1. Upgrade to the above version of Jormungandr 2. Bootstrap 3. See error **Expected behavior** Don't Panic! **Additional context** Running CentOS 7
1.0
Tokio Thread Panicked: Validated Block Must Be Unique - **Describe the bug** Shortly after bootstrapping I received: thread 'tokio-runtime-worker-1' panicked at 'validated block must be unique', src/libcore/option.rs:1185:5 note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace. thread 'main' panicked at 'internal error: entered unreachable code', /root/.cargo/registry/src/github.com-1ecc6299db9ec823/tokio-0.1.22/src/runtime/threadpool/mod.rs:296:26 **Mandatory Information** 1. `jcli --full-version` output; jcli 0.8.8 (master-4687999, release, linux [x86_64]) - [rustc 1.40.0 (73528e339 2019-12-16)] 2. `jormungandr --full-version` output; jormungandr 0.8.8 (master-4687999, release, linux [x86_64]) - [rustc 1.40.0 (73528e339 2019-12-16)] **To Reproduce** Steps to reproduce the behavior: 1. Upgrade to the above version of Jormungandr 2. Bootstrap 3. See error **Expected behavior** Don't Panic! **Additional context** Running CentOS 7
priority
tokio thread panicked validated block must be unique describe the bug shortly after bootstrapping i received thread tokio runtime worker panicked at validated block must be unique src libcore option rs note run with rust backtrace environment variable to display a backtrace thread main panicked at internal error entered unreachable code root cargo registry src github com tokio src runtime threadpool mod rs mandatory information jcli full version output jcli master release linux jormungandr full version output jormungandr master release linux to reproduce steps to reproduce the behavior upgrade to the above version of jormungandr bootstrap see error expected behavior don t panic additional context running centos
1
53,407
3,039,971,182
IssuesEvent
2015-08-07 12:51:24
Itseez/opencv
https://api.github.com/repos/Itseez/opencv
closed
cv2.imshow() doesn't work correctly without a named window
auto-transferred bug category: highgui-gui priority: normal
Transferred from http://code.opencv.org/issues/2146 ``` || Gareth Webber on 2012-07-06 08:00 || Priority: Normal || Affected: None || Category: highgui-gui || Tracker: Bug || Difficulty: None || PR: None || Platform: None / None ``` cv2.imshow() doesn't work correctly without a named window ----------- ``` Hi, I'm using opencv 2.4.2, python 2.7 and installed the bindings from "here":http://www.lfd.uci.edu/~gohlke/pythonlibs/#opencv The following code created a window of the correct size, but its contents is just blank (grey): <pre> img = cv2.imread(r'D:\opencv\samples\c\box.png') #cv2.namedWindow("gray", 1) cv2.imshow("gray", img) cv2.waitKey() </pre> It only works as expected with a named window being created first, which was not the case in previous versions. This stops certain samples from working for me as well, such as stereo_match.py. ``` History ------- ##### Andrey Kamaev on 2012-08-16 15:36 ``` - Category changed from highgui-images to highgui-gui ``` ##### Vadim Pisarevsky on 2012-10-05 10:19 ``` - Target version deleted () - Assignee deleted (Vadim Pisarevsky) ``` ##### Kirill Kornyakov on 2012-10-24 08:38 ``` - Target version set to Next Hackathon ``` ##### Emily Dunham on 2013-06-05 09:16 ``` Tested this and found that it does not happen with the latest OpenCV installed from source on Linux, but is still a problem in Windows 7 with the closest version (opencv 2.4.5) from the site linked in the description. ```
1.0
cv2.imshow() doesn't work correctly without a named window - Transferred from http://code.opencv.org/issues/2146 ``` || Gareth Webber on 2012-07-06 08:00 || Priority: Normal || Affected: None || Category: highgui-gui || Tracker: Bug || Difficulty: None || PR: None || Platform: None / None ``` cv2.imshow() doesn't work correctly without a named window ----------- ``` Hi, I'm using opencv 2.4.2, python 2.7 and installed the bindings from "here":http://www.lfd.uci.edu/~gohlke/pythonlibs/#opencv The following code created a window of the correct size, but its contents is just blank (grey): <pre> img = cv2.imread(r'D:\opencv\samples\c\box.png') #cv2.namedWindow("gray", 1) cv2.imshow("gray", img) cv2.waitKey() </pre> It only works as expected with a named window being created first, which was not the case in previous versions. This stops certain samples from working for me as well, such as stereo_match.py. ``` History ------- ##### Andrey Kamaev on 2012-08-16 15:36 ``` - Category changed from highgui-images to highgui-gui ``` ##### Vadim Pisarevsky on 2012-10-05 10:19 ``` - Target version deleted () - Assignee deleted (Vadim Pisarevsky) ``` ##### Kirill Kornyakov on 2012-10-24 08:38 ``` - Target version set to Next Hackathon ``` ##### Emily Dunham on 2013-06-05 09:16 ``` Tested this and found that it does not happen with the latest OpenCV installed from source on Linux, but is still a problem in Windows 7 with the closest version (opencv 2.4.5) from the site linked in the description. ```
priority
imshow doesn t work correctly without a named window transferred from gareth webber on priority normal affected none category highgui gui tracker bug difficulty none pr none platform none none imshow doesn t work correctly without a named window hi i m using opencv python and installed the bindings from here the following code created a window of the correct size but its contents is just blank grey img imread r d opencv samples c box png namedwindow gray imshow gray img waitkey it only works as expected with a named window being created first which was not the case in previous versions this stops certain samples from working for me as well such as stereo match py history andrey kamaev on category changed from highgui images to highgui gui vadim pisarevsky on target version deleted assignee deleted vadim pisarevsky kirill kornyakov on target version set to next hackathon emily dunham on tested this and found that it does not happen with the latest opencv installed from source on linux but is still a problem in windows with the closest version opencv from the site linked in the description
1
657,602
21,797,787,188
IssuesEvent
2022-05-15 21:49:26
okTurtles/group-income
https://api.github.com/repos/okTurtles/group-income
opened
Find out if we should be saving data in localstorage using JSON.stringify
Note:Up-for-grabs App:Frontend Priority:High Note:Research Kind:Core
### Problem Currently we are storing the Vuex state (which includes all app settings + all contract state) in localstorage directly, and not calling `JSON.stringify` on it. Somehow it is working OK. However, I've read online people saying that data in localstorage should be a string. ### Solution Find out: - Whether what we're doing is OK or if we should convert `'state/vuex/save'` and `'gi.actions/identity/login'` to save & restore the settings using `JSON.stringify` & `JSON.parse` - If the answer to above is "yes", find out and explain below why it's currently working OK without this
1.0
Find out if we should be saving data in localstorage using JSON.stringify - ### Problem Currently we are storing the Vuex state (which includes all app settings + all contract state) in localstorage directly, and not calling `JSON.stringify` on it. Somehow it is working OK. However, I've read online people saying that data in localstorage should be a string. ### Solution Find out: - Whether what we're doing is OK or if we should convert `'state/vuex/save'` and `'gi.actions/identity/login'` to save & restore the settings using `JSON.stringify` & `JSON.parse` - If the answer to above is "yes", find out and explain below why it's currently working OK without this
priority
find out if we should be saving data in localstorage using json stringify problem currently we are storing the vuex state which includes all app settings all contract state in localstorage directly and not calling json stringify on it somehow it is working ok however i ve read online people saying that data in localstorage should be a string solution find out whether what we re doing is ok or if we should convert state vuex save and gi actions identity login to save restore the settings using json stringify json parse if the answer to above is yes find out and explain below why it s currently working ok without this
1
236,388
7,748,998,581
IssuesEvent
2018-05-30 10:00:04
Gloirin/m2gTest
https://api.github.com/repos/Gloirin/m2gTest
closed
0003134: No body element found
Felamimail bug high priority
**Reported by pschuele on 19 Oct 2010 14:25** **Version:** Mialena (2010-03-7) No body element found -&gt; email with empty body has been sent **Additional information:** 311bc Admin Account, Tine 2.0 - 2010-10-19T14:13:52+00:00 NOTICE (5): Tinebase_Server_Json::_handleException::165 exception &#039;Felamimail_Exception&#039; with message &#039;No body element found!&#039; in /opt/workspace/tinetrunk/tine20/Felamimail/Model/Message.php:481 Stack trace: #0 /opt/workspace/tinetrunk/tine20/Felamimail/Controller/Message.php(748): Felamimail_Model_Message-&gt;getPlainTextBody() #1 /opt/workspace/tinetrunk/tine20/Felamimail/Controller/Message.php(557): Felamimail_Controller_Message-&gt;_createMailForSending(Object(Felamimail_Model_Message), Object(Felamimail_Model_Account), NULL, NULL) #2 /opt/workspace/tinetrunk/tine20/Felamimail/Frontend/Json.php(223): Felamimail_Controller_Message-&gt;sendMessage(Object(Felamimail_Model_Message)) #3 [internal function]: Felamimail_Frontend_Json-&gt;saveMessage(Array) #4 /opt/workspace/tinetrunk/tine20/library/Zend/Server/Abstract.php(232): call_user_func_array(Array, Array) #5 /opt/workspace/tinetrunk/tine20/Zend/Json/Server.php(558): Zend_Server_Abstract-&gt;_dispatch(Object(Zend_Server_Method_Definition), Array) #6 /opt/workspace/tinetrunk/tine20/Zend/Json/Server.php(197): Zend_Json_Server-&gt;_handle() #7 /opt/workspace/tinetrunk/tine20/Tinebase/Server/Json.php(131): Zend_Json_Server-&gt;handle(Object(Zend_Json_Server_Request)) #8 /opt/workspace/tinetrunk/tine20/Tinebase/Server/Json.php(71): Tinebase_Server_Json-&gt;_handle(Object(Zend_Json_Server), Object(Zend_Json_Server_Request)) #9 /opt/workspace/tinetrunk/tine20/Tinebase/Core.php(203): Tinebase_Server_Json-&gt;handle() #10 /opt/workspace/tinetrunk/tine20/index.php(56): Tinebase_Core::dispatchRequest() #11 {main}
1.0
0003134: No body element found - **Reported by pschuele on 19 Oct 2010 14:25** **Version:** Mialena (2010-03-7) No body element found -&gt; email with empty body has been sent **Additional information:** 311bc Admin Account, Tine 2.0 - 2010-10-19T14:13:52+00:00 NOTICE (5): Tinebase_Server_Json::_handleException::165 exception &#039;Felamimail_Exception&#039; with message &#039;No body element found!&#039; in /opt/workspace/tinetrunk/tine20/Felamimail/Model/Message.php:481 Stack trace: #0 /opt/workspace/tinetrunk/tine20/Felamimail/Controller/Message.php(748): Felamimail_Model_Message-&gt;getPlainTextBody() #1 /opt/workspace/tinetrunk/tine20/Felamimail/Controller/Message.php(557): Felamimail_Controller_Message-&gt;_createMailForSending(Object(Felamimail_Model_Message), Object(Felamimail_Model_Account), NULL, NULL) #2 /opt/workspace/tinetrunk/tine20/Felamimail/Frontend/Json.php(223): Felamimail_Controller_Message-&gt;sendMessage(Object(Felamimail_Model_Message)) #3 [internal function]: Felamimail_Frontend_Json-&gt;saveMessage(Array) #4 /opt/workspace/tinetrunk/tine20/library/Zend/Server/Abstract.php(232): call_user_func_array(Array, Array) #5 /opt/workspace/tinetrunk/tine20/Zend/Json/Server.php(558): Zend_Server_Abstract-&gt;_dispatch(Object(Zend_Server_Method_Definition), Array) #6 /opt/workspace/tinetrunk/tine20/Zend/Json/Server.php(197): Zend_Json_Server-&gt;_handle() #7 /opt/workspace/tinetrunk/tine20/Tinebase/Server/Json.php(131): Zend_Json_Server-&gt;handle(Object(Zend_Json_Server_Request)) #8 /opt/workspace/tinetrunk/tine20/Tinebase/Server/Json.php(71): Tinebase_Server_Json-&gt;_handle(Object(Zend_Json_Server), Object(Zend_Json_Server_Request)) #9 /opt/workspace/tinetrunk/tine20/Tinebase/Core.php(203): Tinebase_Server_Json-&gt;handle() #10 /opt/workspace/tinetrunk/tine20/index.php(56): Tinebase_Core::dispatchRequest() #11 {main}
priority
no body element found reported by pschuele on oct version mialena no body element found gt email with empty body has been sent additional information admin account tine notice tinebase server json handleexception exception felamimail exception with message no body element found in opt workspace tinetrunk felamimail model message php stack trace opt workspace tinetrunk felamimail controller message php felamimail model message gt getplaintextbody opt workspace tinetrunk felamimail controller message php felamimail controller message gt createmailforsending object felamimail model message object felamimail model account null null opt workspace tinetrunk felamimail frontend json php felamimail controller message gt sendmessage object felamimail model message felamimail frontend json gt savemessage array opt workspace tinetrunk library zend server abstract php call user func array array array opt workspace tinetrunk zend json server php zend server abstract gt dispatch object zend server method definition array opt workspace tinetrunk zend json server php zend json server gt handle opt workspace tinetrunk tinebase server json php zend json server gt handle object zend json server request opt workspace tinetrunk tinebase server json php tinebase server json gt handle object zend json server object zend json server request opt workspace tinetrunk tinebase core php tinebase server json gt handle opt workspace tinetrunk index php tinebase core dispatchrequest main
1
664,258
22,263,144,679
IssuesEvent
2022-06-10 03:44:58
wso2/docs-apim
https://api.github.com/repos/wso2/docs-apim
closed
Review demo script
Priority/Highest CEXP API-M-4.1.0
**Description:** Review the Gartner demo script. - [x] Review the first half of the doc. - Mari - [x] Review the second half of the doc. - Ruthryi - [x] Come up with a 5s closing - Mari - [x] Shorten the intro - Mari
1.0
Review demo script - **Description:** Review the Gartner demo script. - [x] Review the first half of the doc. - Mari - [x] Review the second half of the doc. - Ruthryi - [x] Come up with a 5s closing - Mari - [x] Shorten the intro - Mari
priority
review demo script description review the gartner demo script review the first half of the doc mari review the second half of the doc ruthryi come up with a closing mari shorten the intro mari
1
93,385
3,899,473,710
IssuesEvent
2016-04-17 19:20:33
TranslationWMcs435/TranslationWMcs435
https://api.github.com/repos/TranslationWMcs435/TranslationWMcs435
closed
Create checker interface for checking the JSON file prior to translating
High Priority New Feature
This will check to see that every action has its necessary components to work. If an action does not work then it'll make a note in the console and should write a comment into the output file.
1.0
Create checker interface for checking the JSON file prior to translating - This will check to see that every action has its necessary components to work. If an action does not work then it'll make a note in the console and should write a comment into the output file.
priority
create checker interface for checking the json file prior to translating this will check to see that every action has its necessary components to work if an action does not work then it ll make a note in the console and should write a comment into the output file
1
366,578
10,824,479,513
IssuesEvent
2019-11-09 09:37:42
AY1920S1-CS2113T-F09-3/main
https://api.github.com/repos/AY1920S1-CS2113T-F09-3/main
closed
As a lab tech, I want to be able to easily edit the number of parts that I have in inventory
priority.High type.Story
so that I can update the inventory quickly when new stock comes in.
1.0
As a lab tech, I want to be able to easily edit the number of parts that I have in inventory - so that I can update the inventory quickly when new stock comes in.
priority
as a lab tech i want to be able to easily edit the number of parts that i have in inventory so that i can update the inventory quickly when new stock comes in
1
474,770
13,676,066,547
IssuesEvent
2020-09-29 13:28:31
genetics-statistics/GEMMA
https://api.github.com/repos/genetics-statistics/GEMMA
closed
Travis build for MacOS is failing
bug high priority
A recent update on Travis is causing gemma builds to fail: https://travis-ci.org/genenetwork/GEMMA/jobs/403562335 Notable gcc got upgraded to v8 and openblas to 0.3.1. I suspect the latter since I have seen other failures. I will try the new openblas on Linux soon.
1.0
Travis build for MacOS is failing - A recent update on Travis is causing gemma builds to fail: https://travis-ci.org/genenetwork/GEMMA/jobs/403562335 Notable gcc got upgraded to v8 and openblas to 0.3.1. I suspect the latter since I have seen other failures. I will try the new openblas on Linux soon.
priority
travis build for macos is failing a recent update on travis is causing gemma builds to fail notable gcc got upgraded to and openblas to i suspect the latter since i have seen other failures i will try the new openblas on linux soon
1
101,199
4,109,010,305
IssuesEvent
2016-06-06 18:02:47
VirtoCommerce/vc-community
https://api.github.com/repos/VirtoCommerce/vc-community
opened
No localization after sample data is selected
backend bug high priority
<img width="1113" alt="screen shot 2016-06-06 at 11 02 01 am" src="https://cloud.githubusercontent.com/assets/1566470/15832246/281856ca-2bd6-11e6-882c-d50a66a2390a.png">
1.0
No localization after sample data is selected - <img width="1113" alt="screen shot 2016-06-06 at 11 02 01 am" src="https://cloud.githubusercontent.com/assets/1566470/15832246/281856ca-2bd6-11e6-882c-d50a66a2390a.png">
priority
no localization after sample data is selected img width alt screen shot at am src
1
351,231
10,514,515,418
IssuesEvent
2019-09-28 01:16:09
astropy/astropy
https://api.github.com/repos/astropy/astropy
closed
Implement Issue and PR templates
Docs Feature Request Priority-High sprint
As suggested by @brechmos-stsci , it might be beneficial to give users a check list when they open a new issue or a new pull request (the latter should be consistent with existing dev guide). A first draft could be obtained from existing sources, for example https://www.talater.com/open-source-templates . For instance, choosing new issue, bug + feature request, and backend brought me to https://www.talater.com/open-source-templates/#/page/98 ``` <!--- Provide a general summary of the issue in the Title above --> ## Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> ## Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> ## Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1. 2. 3. 4. ## Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used: * Environment name and version (e.g. PHP 5.4 on nginx 1.9.1): * Server type and version: * Operating System and version: * Link to your project: ```
1.0
Implement Issue and PR templates - As suggested by @brechmos-stsci , it might be beneficial to give users a check list when they open a new issue or a new pull request (the latter should be consistent with existing dev guide). A first draft could be obtained from existing sources, for example https://www.talater.com/open-source-templates . For instance, choosing new issue, bug + feature request, and backend brought me to https://www.talater.com/open-source-templates/#/page/98 ``` <!--- Provide a general summary of the issue in the Title above --> ## Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> ## Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> ## Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1. 2. 3. 4. ## Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used: * Environment name and version (e.g. PHP 5.4 on nginx 1.9.1): * Server type and version: * Operating System and version: * Link to your project: ```
priority
implement issue and pr templates as suggested by brechmos stsci it might be beneficial to give users a check list when they open a new issue or a new pull request the latter should be consistent with existing dev guide a first draft could be obtained from existing sources for example for instance choosing new issue bug feature request and backend brought me to expected behavior current behavior possible solution steps to reproduce for bugs context your environment version used environment name and version e g php on nginx server type and version operating system and version link to your project
1
400,638
11,778,139,930
IssuesEvent
2020-03-16 15:51:36
alphagov/govuk-design-system
https://api.github.com/repos/alphagov/govuk-design-system
closed
Publish 'Recover from errors (validation)' guidance in the Design System
Effort: hours Priority: high
## What 'Recover from errors (validation)' pattern has been approved by the working group. Feedback has been reviewed by the contributor and pattern iterated accordingly. The contribution is now ready to be published. ## Why So Design System users can access the pattern. ## Who needs to know about this Stephen, contributor, developer (for review) ## Done when - [ ] Final review completed - [ ] Pattern published - [ ] New pattern promoted on GDS and cross-government Slack channels
1.0
Publish 'Recover from errors (validation)' guidance in the Design System - ## What 'Recover from errors (validation)' pattern has been approved by the working group. Feedback has been reviewed by the contributor and pattern iterated accordingly. The contribution is now ready to be published. ## Why So Design System users can access the pattern. ## Who needs to know about this Stephen, contributor, developer (for review) ## Done when - [ ] Final review completed - [ ] Pattern published - [ ] New pattern promoted on GDS and cross-government Slack channels
priority
publish recover from errors validation guidance in the design system what recover from errors validation pattern has been approved by the working group feedback has been reviewed by the contributor and pattern iterated accordingly the contribution is now ready to be published why so design system users can access the pattern who needs to know about this stephen contributor developer for review done when final review completed pattern published new pattern promoted on gds and cross government slack channels
1
112,358
4,525,217,922
IssuesEvent
2016-09-07 03:26:43
StargateMC/StargateMCPublic
https://api.github.com/repos/StargateMC/StargateMCPublic
opened
Server Crash: Clearing Unregistered NPCs
Bug High Priority StargateMC: CoreMod
Unregistered NPC clearing needs to do things in two phases, just like registered npcs. First they need to be flagged for removal, then they need to be removed on the next tick. Crash report below; ``` Time: 06/09/16 22:45 Description: Exception in server tick loop java.util.ConcurrentModificationException at java.util.ArrayList$Itr.checkForComodification(ArrayList.java:901) at java.util.ArrayList$Itr.next(ArrayList.java:851) at thermos.chaud.HashedArrayList$HashedArrayIterator.next(HashedArrayList.java:215) at com.stargatemc.mod.providers.npcs.npcRegistry.clearUnregistered(npcRegistry.java:141) at com.stargatemc.mod.providers.npcs.npcRegistry.tickNPCs(npcRegistry.java:130) at com.stargatemc.mod.StargateMCMod.onWorldTick(StargateMCMod.java:110) at cpw.mods.fml.common.eventhandler.ASMEventHandler_131_StargateMCMod_onWorldTick_WorldTickEvent.invoke(.dynamic) at cpw.mods.fml.common.eventhandler.ASMEventHandler.invoke(ASMEventHandler.java:54) at cpw.mods.fml.common.eventhandler.EventBus.post(EventBus.java:140) at cpw.mods.fml.common.FMLCommonHandler.onPreWorldTick(FMLCommonHandler.java:273) at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:957) at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:432) at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:841) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:693) at java.lang.Thread.run(Thread.java:745) ```
1.0
Server Crash: Clearing Unregistered NPCs - Unregistered NPC clearing needs to do things in two phases, just like registered npcs. First they need to be flagged for removal, then they need to be removed on the next tick. Crash report below; ``` Time: 06/09/16 22:45 Description: Exception in server tick loop java.util.ConcurrentModificationException at java.util.ArrayList$Itr.checkForComodification(ArrayList.java:901) at java.util.ArrayList$Itr.next(ArrayList.java:851) at thermos.chaud.HashedArrayList$HashedArrayIterator.next(HashedArrayList.java:215) at com.stargatemc.mod.providers.npcs.npcRegistry.clearUnregistered(npcRegistry.java:141) at com.stargatemc.mod.providers.npcs.npcRegistry.tickNPCs(npcRegistry.java:130) at com.stargatemc.mod.StargateMCMod.onWorldTick(StargateMCMod.java:110) at cpw.mods.fml.common.eventhandler.ASMEventHandler_131_StargateMCMod_onWorldTick_WorldTickEvent.invoke(.dynamic) at cpw.mods.fml.common.eventhandler.ASMEventHandler.invoke(ASMEventHandler.java:54) at cpw.mods.fml.common.eventhandler.EventBus.post(EventBus.java:140) at cpw.mods.fml.common.FMLCommonHandler.onPreWorldTick(FMLCommonHandler.java:273) at net.minecraft.server.MinecraftServer.func_71190_q(MinecraftServer.java:957) at net.minecraft.server.dedicated.DedicatedServer.func_71190_q(DedicatedServer.java:432) at net.minecraft.server.MinecraftServer.func_71217_p(MinecraftServer.java:841) at net.minecraft.server.MinecraftServer.run(MinecraftServer.java:693) at java.lang.Thread.run(Thread.java:745) ```
priority
server crash clearing unregistered npcs unregistered npc clearing needs to do things in two phases just like registered npcs first they need to be flagged for removal then they need to be removed on the next tick crash report below time description exception in server tick loop java util concurrentmodificationexception at java util arraylist itr checkforcomodification arraylist java at java util arraylist itr next arraylist java at thermos chaud hashedarraylist hashedarrayiterator next hashedarraylist java at com stargatemc mod providers npcs npcregistry clearunregistered npcregistry java at com stargatemc mod providers npcs npcregistry ticknpcs npcregistry java at com stargatemc mod stargatemcmod onworldtick stargatemcmod java at cpw mods fml common eventhandler asmeventhandler stargatemcmod onworldtick worldtickevent invoke dynamic at cpw mods fml common eventhandler asmeventhandler invoke asmeventhandler java at cpw mods fml common eventhandler eventbus post eventbus java at cpw mods fml common fmlcommonhandler onpreworldtick fmlcommonhandler java at net minecraft server minecraftserver func q minecraftserver java at net minecraft server dedicated dedicatedserver func q dedicatedserver java at net minecraft server minecraftserver func p minecraftserver java at net minecraft server minecraftserver run minecraftserver java at java lang thread run thread java
1
494,334
14,248,829,496
IssuesEvent
2020-11-19 13:29:03
abpframework/abp
https://api.github.com/repos/abpframework/abp
closed
Module template angular authentication problem
bug effort-0.5 priority:high ui-angular
To reproduce; * Switch to rel-4.0 branch. * Run `dotnet build /graphBuild` in D:\Github\abp\templates\module\aspnet-core * Delete existing database(s): MyProjectName_Main * Run migrations of the IdentityServer project: ![image](https://user-images.githubusercontent.com/1210527/99495776-306ffe80-2984-11eb-87c8-a2f55fab5476.png) * Run IdentityServer project * Run HttpApi.Host project * Run `yarn`, then `yarn start` on the Angular application. * Try login with admin / 1q2w3E* It seems logging in, but can not get the username and permissions; User menu: ![image](https://user-images.githubusercontent.com/1210527/99495909-6ca35f00-2984-11eb-9429-390321822024.png) Application configuration endpoint call result: ![image](https://user-images.githubusercontent.com/1210527/99496090-b55b1800-2984-11eb-99ac-7c0516e76292.png) I see a potential problem: It is requesting to `https://localhost:44300/api/abp/application-configuration`, while the environment.ts contains; ````js apis: { default: { url: 'https://localhost:44301', rootNamespace: 'MyCompanyName.MyProjectName', }, MyProjectName: { url: 'https://localhost:44300', rootNamespace: 'MyCompanyName.MyProjectName', }, }, ```` It should use the port `44301` since it is the default port. `44300` should only be used for `MyProjectName` API.
1.0
Module template angular authentication problem - To reproduce; * Switch to rel-4.0 branch. * Run `dotnet build /graphBuild` in D:\Github\abp\templates\module\aspnet-core * Delete existing database(s): MyProjectName_Main * Run migrations of the IdentityServer project: ![image](https://user-images.githubusercontent.com/1210527/99495776-306ffe80-2984-11eb-87c8-a2f55fab5476.png) * Run IdentityServer project * Run HttpApi.Host project * Run `yarn`, then `yarn start` on the Angular application. * Try login with admin / 1q2w3E* It seems logging in, but can not get the username and permissions; User menu: ![image](https://user-images.githubusercontent.com/1210527/99495909-6ca35f00-2984-11eb-9429-390321822024.png) Application configuration endpoint call result: ![image](https://user-images.githubusercontent.com/1210527/99496090-b55b1800-2984-11eb-99ac-7c0516e76292.png) I see a potential problem: It is requesting to `https://localhost:44300/api/abp/application-configuration`, while the environment.ts contains; ````js apis: { default: { url: 'https://localhost:44301', rootNamespace: 'MyCompanyName.MyProjectName', }, MyProjectName: { url: 'https://localhost:44300', rootNamespace: 'MyCompanyName.MyProjectName', }, }, ```` It should use the port `44301` since it is the default port. `44300` should only be used for `MyProjectName` API.
priority
module template angular authentication problem to reproduce switch to rel branch run dotnet build graphbuild in d github abp templates module aspnet core delete existing database s myprojectname main run migrations of the identityserver project run identityserver project run httpapi host project run yarn then yarn start on the angular application try login with admin it seems logging in but can not get the username and permissions user menu application configuration endpoint call result i see a potential problem it is requesting to while the environment ts contains js apis default url rootnamespace mycompanyname myprojectname myprojectname url rootnamespace mycompanyname myprojectname it should use the port since it is the default port should only be used for myprojectname api
1
222,944
7,444,756,136
IssuesEvent
2018-03-28 00:20:47
cilium/cilium
https://api.github.com/repos/cilium/cilium
closed
Delete map on open if metadata mismatches
area/datapath priority/1.0-blocker priority/high
Delete map prior to opening with a warning if metadata such as size or type does not match existing map. This allows to change the format of maps at the cost of losing some state.
2.0
Delete map on open if metadata mismatches - Delete map prior to opening with a warning if metadata such as size or type does not match existing map. This allows to change the format of maps at the cost of losing some state.
priority
delete map on open if metadata mismatches delete map prior to opening with a warning if metadata such as size or type does not match existing map this allows to change the format of maps at the cost of losing some state
1
308,189
9,435,962,540
IssuesEvent
2019-04-13 01:45:19
MitMaro/git-interactive-rebase-tool
https://api.github.com/repos/MitMaro/git-interactive-rebase-tool
closed
Add support for break instruction
enhancement high priority
Git 2.20 has been released with a new `break` command that can be inserted between list items: > * "git rebase -i" learned a new [instruction], 'break', that the user can > insert in the to-do list. Upon hitting it, the command returns > control back to the user. — [git/Documentation/RelNotes/2.20.0.txt](https://github.com/git/git/blob/v2.20.0/Documentation/RelNotes/2.20.0.txt#L143-L145)
1.0
Add support for break instruction - Git 2.20 has been released with a new `break` command that can be inserted between list items: > * "git rebase -i" learned a new [instruction], 'break', that the user can > insert in the to-do list. Upon hitting it, the command returns > control back to the user. — [git/Documentation/RelNotes/2.20.0.txt](https://github.com/git/git/blob/v2.20.0/Documentation/RelNotes/2.20.0.txt#L143-L145)
priority
add support for break instruction git has been released with a new break command that can be inserted between list items git rebase i learned a new break that the user can insert in the to do list upon hitting it the command returns control back to the user —
1
327,076
9,966,047,414
IssuesEvent
2019-07-08 10:10:47
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
opened
Sliceviewer crashes on windows with Matrix workspaces
Component: Workbench Priority: High
### Actual behavior The sliceviewer crashes the workbench when trying to open the sliceviewer on a "normal" matrix workspace on windows. It seems to start ok for MDWorkspaces. and it seems to be ok on linux. ### Steps to reproduce the behavior 1. open workbench 1. Load a worspace 1. right click and select sliceviewer 1. The window will appear white for a couple of seconds and then crash out to the error reporter. The following file definately crashes, but it seems to be consistent for any matrix workspace. SXD23767.raw (can be found in the training data) ### Platforms affected Windows 8 and 10, although linux seems to be ok.
1.0
Sliceviewer crashes on windows with Matrix workspaces - ### Actual behavior The sliceviewer crashes the workbench when trying to open the sliceviewer on a "normal" matrix workspace on windows. It seems to start ok for MDWorkspaces. and it seems to be ok on linux. ### Steps to reproduce the behavior 1. open workbench 1. Load a worspace 1. right click and select sliceviewer 1. The window will appear white for a couple of seconds and then crash out to the error reporter. The following file definately crashes, but it seems to be consistent for any matrix workspace. SXD23767.raw (can be found in the training data) ### Platforms affected Windows 8 and 10, although linux seems to be ok.
priority
sliceviewer crashes on windows with matrix workspaces actual behavior the sliceviewer crashes the workbench when trying to open the sliceviewer on a normal matrix workspace on windows it seems to start ok for mdworkspaces and it seems to be ok on linux steps to reproduce the behavior open workbench load a worspace right click and select sliceviewer the window will appear white for a couple of seconds and then crash out to the error reporter the following file definately crashes but it seems to be consistent for any matrix workspace raw can be found in the training data platforms affected windows and although linux seems to be ok
1
280,979
8,689,016,073
IssuesEvent
2018-12-03 17:31:13
AnalyticalGraphicsInc/cesium
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
opened
Failing: Scene/GlobeSurfaceTileProvider can render more imagery layers than the available texture units
priority - high type - bug
`Scene/GlobeSurfaceTileProvider can render more imagery layers than the available texture units` is failing in Firefox: http://localhost:8080/Specs/SpecRunner.html?spec=Scene%2FGlobeSurfaceTileProvider%20can%20render%20more%20imagery%20layers%20than%20the%20available%20texture%20units > Failed: promise rejected: RuntimeError: Program failed to link. Link log: Programs with more than 16 samplers are disallowed on Mesa drivers to avoid crashing.
1.0
Failing: Scene/GlobeSurfaceTileProvider can render more imagery layers than the available texture units - `Scene/GlobeSurfaceTileProvider can render more imagery layers than the available texture units` is failing in Firefox: http://localhost:8080/Specs/SpecRunner.html?spec=Scene%2FGlobeSurfaceTileProvider%20can%20render%20more%20imagery%20layers%20than%20the%20available%20texture%20units > Failed: promise rejected: RuntimeError: Program failed to link. Link log: Programs with more than 16 samplers are disallowed on Mesa drivers to avoid crashing.
priority
failing scene globesurfacetileprovider can render more imagery layers than the available texture units scene globesurfacetileprovider can render more imagery layers than the available texture units is failing in firefox failed promise rejected runtimeerror program failed to link link log programs with more than samplers are disallowed on mesa drivers to avoid crashing
1
185,514
6,724,707,429
IssuesEvent
2017-10-17 00:14:42
Allsteel/awesome
https://api.github.com/repos/Allsteel/awesome
opened
Can't place an overhead on the edge of the panel
bug High Priority
## Steps 1. Create a panel run with three 30"W panels 2. Bring an overhead from Involve into the world 3. Try placing it on the edge of the leftmost panel It will automatically snap on the other side of the panel and it seems that it only happens on the left side of the panel run ![overhead](https://user-images.githubusercontent.com/111954/31640808-3d790930-b2a6-11e7-96ef-8f1bfbe68540.gif)
1.0
Can't place an overhead on the edge of the panel - ## Steps 1. Create a panel run with three 30"W panels 2. Bring an overhead from Involve into the world 3. Try placing it on the edge of the leftmost panel It will automatically snap on the other side of the panel and it seems that it only happens on the left side of the panel run ![overhead](https://user-images.githubusercontent.com/111954/31640808-3d790930-b2a6-11e7-96ef-8f1bfbe68540.gif)
priority
can t place an overhead on the edge of the panel steps create a panel run with three w panels bring an overhead from involve into the world try placing it on the edge of the leftmost panel it will automatically snap on the other side of the panel and it seems that it only happens on the left side of the panel run
1
806,384
29,819,789,200
IssuesEvent
2023-06-17 00:10:35
Faithful-Resource-Pack/Discord-Bot
https://api.github.com/repos/Faithful-Resource-Pack/Discord-Bot
closed
[REQUEST] Adding optional version tag to the /texture command
feature high priority request
It would be good to have a version tag for the `/texture` command, to view older textures of the game. For example, if you would want to see the old 1.16 ore textures, then you could use: ``` /texture 16 diamond_ore --ver 1.16 ``` If you would want to see the newest one, then you could just not specify the tag.
1.0
[REQUEST] Adding optional version tag to the /texture command - It would be good to have a version tag for the `/texture` command, to view older textures of the game. For example, if you would want to see the old 1.16 ore textures, then you could use: ``` /texture 16 diamond_ore --ver 1.16 ``` If you would want to see the newest one, then you could just not specify the tag.
priority
adding optional version tag to the texture command it would be good to have a version tag for the texture command to view older textures of the game for example if you would want to see the old ore textures then you could use texture diamond ore ver if you would want to see the newest one then you could just not specify the tag
1
623,589
19,673,164,907
IssuesEvent
2022-01-11 09:36:21
airbytehq/airbyte
https://api.github.com/repos/airbytehq/airbyte
closed
Source Salesforce: Support oauth'ing into sandbox environment
type/enhancement area/connectors priority/high
## Tell us about the problem you're trying to solve When oauth'ing a salesforce user into a sandbox environment, we need to send them to `test.salesforce.com` rather than `login.salesforce.com`. Our current server implementation of salesforce oauth always assumes they are not trying to use a sandbox environment and redirects them to `login.salesforce.com`. Unfortunately, many enterprise users are not allowed to use production environments for testing new tools. Therefore, they would not be able to adopt Airbyte Cloud without being able to access the sandbox environment. Ideally, we would wait for #6971 to be implemented and use the functionality offered there to implement this flow. However, since that could take some time, we need to come up with a way to allow new Airbyte Cloud customers to start Once #6971 is implemented, we need to look at the user's input config and determine which oauth configuration to use. ## Describe the solution you’d like This fix has the following components: 1. Disable the `sandbox` option of the Salesforce connector and always assume we are using the production environment (remove it from the spec but not from the code, and allow additional properties = true. if the undocumented option is passed then the connector should use sandbox) 2. Create a fork of the salesforce connector, `source-salesforce-sandbox` (similar to the `source-file-secure` fork of the `source-file` connector) which always assumes we are using the salesforce sandbox environment. By the end we should have two salesforce connectors, `source-salesforce` which always assumes NOT sandbox, and `source-salesforce-sandbox` which always assumes sandbox environment. Publish both connectors normally like we do with any connector. 3. Create a new oauth flow in the server which can log the user in for the sandbox environment. For testing this flow, use the `Salesforce (Sandbox)` Lastpass credentials. The only difference between this flow and the current salesforce flow is that the current salesforce flow uses the `login.salesforce.com` endpoint whereas the new flow should use `test.salesforce.com`. That's it. Otherwise, it can use the same credentials. 4. Once you've tested the flow, create a PR and merge as normal on salesforce. 5. Create a ticket to revert 1 & 2 once #6971 is implemented. That is, once that ticket is implemented, we should go back to having a single salesforce connector with a sandbox toggle whose oauth flow is controlled based on the toggle.
1.0
Source Salesforce: Support oauth'ing into sandbox environment - ## Tell us about the problem you're trying to solve When oauth'ing a salesforce user into a sandbox environment, we need to send them to `test.salesforce.com` rather than `login.salesforce.com`. Our current server implementation of salesforce oauth always assumes they are not trying to use a sandbox environment and redirects them to `login.salesforce.com`. Unfortunately, many enterprise users are not allowed to use production environments for testing new tools. Therefore, they would not be able to adopt Airbyte Cloud without being able to access the sandbox environment. Ideally, we would wait for #6971 to be implemented and use the functionality offered there to implement this flow. However, since that could take some time, we need to come up with a way to allow new Airbyte Cloud customers to start Once #6971 is implemented, we need to look at the user's input config and determine which oauth configuration to use. ## Describe the solution you’d like This fix has the following components: 1. Disable the `sandbox` option of the Salesforce connector and always assume we are using the production environment (remove it from the spec but not from the code, and allow additional properties = true. if the undocumented option is passed then the connector should use sandbox) 2. Create a fork of the salesforce connector, `source-salesforce-sandbox` (similar to the `source-file-secure` fork of the `source-file` connector) which always assumes we are using the salesforce sandbox environment. By the end we should have two salesforce connectors, `source-salesforce` which always assumes NOT sandbox, and `source-salesforce-sandbox` which always assumes sandbox environment. Publish both connectors normally like we do with any connector. 3. Create a new oauth flow in the server which can log the user in for the sandbox environment. For testing this flow, use the `Salesforce (Sandbox)` Lastpass credentials. The only difference between this flow and the current salesforce flow is that the current salesforce flow uses the `login.salesforce.com` endpoint whereas the new flow should use `test.salesforce.com`. That's it. Otherwise, it can use the same credentials. 4. Once you've tested the flow, create a PR and merge as normal on salesforce. 5. Create a ticket to revert 1 & 2 once #6971 is implemented. That is, once that ticket is implemented, we should go back to having a single salesforce connector with a sandbox toggle whose oauth flow is controlled based on the toggle.
priority
source salesforce support oauth ing into sandbox environment tell us about the problem you re trying to solve when oauth ing a salesforce user into a sandbox environment we need to send them to test salesforce com rather than login salesforce com our current server implementation of salesforce oauth always assumes they are not trying to use a sandbox environment and redirects them to login salesforce com unfortunately many enterprise users are not allowed to use production environments for testing new tools therefore they would not be able to adopt airbyte cloud without being able to access the sandbox environment ideally we would wait for to be implemented and use the functionality offered there to implement this flow however since that could take some time we need to come up with a way to allow new airbyte cloud customers to start once is implemented we need to look at the user s input config and determine which oauth configuration to use describe the solution you’d like this fix has the following components disable the sandbox option of the salesforce connector and always assume we are using the production environment remove it from the spec but not from the code and allow additional properties true if the undocumented option is passed then the connector should use sandbox create a fork of the salesforce connector source salesforce sandbox similar to the source file secure fork of the source file connector which always assumes we are using the salesforce sandbox environment by the end we should have two salesforce connectors source salesforce which always assumes not sandbox and source salesforce sandbox which always assumes sandbox environment publish both connectors normally like we do with any connector create a new oauth flow in the server which can log the user in for the sandbox environment for testing this flow use the salesforce sandbox lastpass credentials the only difference between this flow and the current salesforce flow is that the current salesforce flow uses the login salesforce com endpoint whereas the new flow should use test salesforce com that s it otherwise it can use the same credentials once you ve tested the flow create a pr and merge as normal on salesforce create a ticket to revert once is implemented that is once that ticket is implemented we should go back to having a single salesforce connector with a sandbox toggle whose oauth flow is controlled based on the toggle
1
815,573
30,562,591,423
IssuesEvent
2023-07-20 15:29:50
CredentialEngine/CredentialRegistry
https://api.github.com/repos/CredentialEngine/CredentialRegistry
closed
Include the last modified date in the bulk download files
enhancement High Priority
@excelsior Would it be possible to include the last modified date from the envelope in the files included in the bulk download? re: GET {community}/envelopes/download. An additional "feature" could be to append the last modified date (yyyy-mm-dd) to the file name (after the CTID).
1.0
Include the last modified date in the bulk download files - @excelsior Would it be possible to include the last modified date from the envelope in the files included in the bulk download? re: GET {community}/envelopes/download. An additional "feature" could be to append the last modified date (yyyy-mm-dd) to the file name (after the CTID).
priority
include the last modified date in the bulk download files excelsior would it be possible to include the last modified date from the envelope in the files included in the bulk download re get community envelopes download an additional feature could be to append the last modified date yyyy mm dd to the file name after the ctid
1
56,667
3,080,841,313
IssuesEvent
2015-08-22 03:51:49
cdnjs/cdnjs
https://api.github.com/repos/cdnjs/cdnjs
opened
Support fix package.json format on certain lib
Feature - High Priority Functionality - Nice to Have
Feature of [fixFormat.js](https://github.com/cdnjs/cdnjs/blob/master/tools/fixFormat.js). So that the whole progress will be faster if we already know a certain target.
1.0
Support fix package.json format on certain lib - Feature of [fixFormat.js](https://github.com/cdnjs/cdnjs/blob/master/tools/fixFormat.js). So that the whole progress will be faster if we already know a certain target.
priority
support fix package json format on certain lib feature of so that the whole progress will be faster if we already know a certain target
1
80,555
3,563,777,597
IssuesEvent
2016-01-25 06:58:00
sa31fh/ims
https://api.github.com/repos/sa31fh/ims
closed
Improve login checks
enhancement High Priority
Currently the website only handles login info via the main page. Other pages can be accessed directly even if the user is not logged in. Simple 'if' checks should be added at the top of each page and an error displayed if the user is not logged in. Don't allow the code to display any more information.
1.0
Improve login checks - Currently the website only handles login info via the main page. Other pages can be accessed directly even if the user is not logged in. Simple 'if' checks should be added at the top of each page and an error displayed if the user is not logged in. Don't allow the code to display any more information.
priority
improve login checks currently the website only handles login info via the main page other pages can be accessed directly even if the user is not logged in simple if checks should be added at the top of each page and an error displayed if the user is not logged in don t allow the code to display any more information
1
344,031
10,339,411,403
IssuesEvent
2019-09-03 19:16:17
emory-libraries/ezpaarse-platforms
https://api.github.com/repos/emory-libraries/ezpaarse-platforms
closed
SAGE Knowledge (sk)
High Priority Stakeholder Priority Update Parser
### Example:star::star: : http://sk.sagepub.com.proxy.library.emory.edu ### Priority: High ### Subscriber (Library): Woodruff ### ezPAARSE Analysis: http://ang.couperin.org/platforms/59f165e8884aa7daec163f6e
2.0
SAGE Knowledge (sk) - ### Example:star::star: : http://sk.sagepub.com.proxy.library.emory.edu ### Priority: High ### Subscriber (Library): Woodruff ### ezPAARSE Analysis: http://ang.couperin.org/platforms/59f165e8884aa7daec163f6e
priority
sage knowledge sk example star star priority high subscriber library woodruff ezpaarse analysis
1
59,194
3,103,916,051
IssuesEvent
2015-08-31 13:15:43
cs2103aug2015-w15-3j/main
https://api.github.com/repos/cs2103aug2015-w15-3j/main
opened
Construct user stories table for Appendix A
priority.high status.ongoing type.task
List finalized user stories in the following format: Task ID, Functionality, Value
1.0
Construct user stories table for Appendix A - List finalized user stories in the following format: Task ID, Functionality, Value
priority
construct user stories table for appendix a list finalized user stories in the following format task id functionality value
1
803,851
29,191,765,465
IssuesEvent
2023-05-19 20:47:42
unlock-protocol/unlock
https://api.github.com/repos/unlock-protocol/unlock
closed
Bug when showing membership details
🚨 High Priority
Not sure what is wrong but clearly something is wrong. Also please do NOT show the `Key Manager` if it is unset or is the same address as the `Key Owner` <img width="871" alt="Screenshot 2023-05-19 at 2 59 13 PM" src="https://github.com/unlock-protocol/unlock/assets/17735/f0a78db5-02f9-443b-b228-32134c618850">
1.0
Bug when showing membership details - Not sure what is wrong but clearly something is wrong. Also please do NOT show the `Key Manager` if it is unset or is the same address as the `Key Owner` <img width="871" alt="Screenshot 2023-05-19 at 2 59 13 PM" src="https://github.com/unlock-protocol/unlock/assets/17735/f0a78db5-02f9-443b-b228-32134c618850">
priority
bug when showing membership details not sure what is wrong but clearly something is wrong also please do not show the key manager if it is unset or is the same address as the key owner img width alt screenshot at pm src
1
335,317
10,151,844,079
IssuesEvent
2019-08-05 21:28:06
ilakeful/LakeBot
https://api.github.com/repos/ilakeful/LakeBot
opened
Fix of the music module
bug changes: patch priority: high type: experimental type: feature or enhancement type: issue
The feature implies fixing the music module being stuck at the end of the currently playing track.
1.0
Fix of the music module - The feature implies fixing the music module being stuck at the end of the currently playing track.
priority
fix of the music module the feature implies fixing the music module being stuck at the end of the currently playing track
1
464,044
13,305,267,007
IssuesEvent
2020-08-25 18:15:48
dkfans/keeperfx
https://api.github.com/repos/dkfans/keeperfx
closed
newdig: Computer Player should use pre-dug out rooms.
Component-CompPlayer Component-NewDig Priority-High Status-Started Type-Enhancement
Originally reported on Google Code with ID 627 ``` In r1850 with the newdig option when a mapmaker gives a CP a few empty rooms to start out with, the CP should use them and fill them with rooms. Instead he just decides to dig new spaces. See attached map for behavior. ``` Reported by `Loobinex` on 2015-05-19 00:44:33 <hr> - _Attachment: [TC627_1.7z](https://storage.googleapis.com/google-code-attachments/keeperfx/issue-627/comment-0/TC627_1.7z)_
1.0
newdig: Computer Player should use pre-dug out rooms. - Originally reported on Google Code with ID 627 ``` In r1850 with the newdig option when a mapmaker gives a CP a few empty rooms to start out with, the CP should use them and fill them with rooms. Instead he just decides to dig new spaces. See attached map for behavior. ``` Reported by `Loobinex` on 2015-05-19 00:44:33 <hr> - _Attachment: [TC627_1.7z](https://storage.googleapis.com/google-code-attachments/keeperfx/issue-627/comment-0/TC627_1.7z)_
priority
newdig computer player should use pre dug out rooms originally reported on google code with id in with the newdig option when a mapmaker gives a cp a few empty rooms to start out with the cp should use them and fill them with rooms instead he just decides to dig new spaces see attached map for behavior reported by loobinex on attachment
1
140,628
5,413,541,805
IssuesEvent
2017-03-01 16:56:49
qutebrowser/qutebrowser
https://api.github.com/repos/qutebrowser/qutebrowser
closed
Do not select the content of a search field
component: hints priority: 0 - high
With the upgrade from 0.9 to 0.10 the contend of a search field gets selected on focusing. Example: ![scr](http://i.imgur.com/vU8lbn7.png) I can't find the way to disable this new thing. Is it possible to keep the content not selected as it used to? I'm on QtWebKit backend.
1.0
Do not select the content of a search field - With the upgrade from 0.9 to 0.10 the contend of a search field gets selected on focusing. Example: ![scr](http://i.imgur.com/vU8lbn7.png) I can't find the way to disable this new thing. Is it possible to keep the content not selected as it used to? I'm on QtWebKit backend.
priority
do not select the content of a search field with the upgrade from to the contend of a search field gets selected on focusing example i can t find the way to disable this new thing is it possible to keep the content not selected as it used to i m on qtwebkit backend
1
638,256
20,720,024,998
IssuesEvent
2022-03-13 08:31:10
AY2122S2-CS2103T-T13-3/tp
https://api.github.com/repos/AY2122S2-CS2103T-T13-3/tp
closed
Add a new meeting entry
type.Story priority.High
As a user, I want to add a new meeting entry containing the following: 1. Meeting link 2. Meeting name 3. Date & time 4. Tags (if any) Depends on #35 #38
1.0
Add a new meeting entry - As a user, I want to add a new meeting entry containing the following: 1. Meeting link 2. Meeting name 3. Date & time 4. Tags (if any) Depends on #35 #38
priority
add a new meeting entry as a user i want to add a new meeting entry containing the following meeting link meeting name date time tags if any depends on
1
678,015
23,183,692,972
IssuesEvent
2022-08-01 06:16:35
CAFECA-IO/TideBitEx
https://api.github.com/repos/CAFECA-IO/TideBitEx
closed
[BUG][Frontend][Tempory] 定期用 api reset library book 裡面的資料
bug 2 high priority
---- WORKAROUND ---- in src/modal/middleman.js add sync function,在 - sync function - 等待 1min - 呼叫 depthBook api 、orders api 、trades api 來更新 library book - 呼叫自己
1.0
[BUG][Frontend][Tempory] 定期用 api reset library book 裡面的資料 - ---- WORKAROUND ---- in src/modal/middleman.js add sync function,在 - sync function - 等待 1min - 呼叫 depthBook api 、orders api 、trades api 來更新 library book - 呼叫自己
priority
定期用 api reset library book 裡面的資料 workaround in src modal middleman js add sync function,在 sync function 等待 呼叫 depthbook api 、orders api 、trades api 來更新 library book 呼叫自己
1
416,418
12,146,126,768
IssuesEvent
2020-04-24 10:33:01
dirkwhoffmann/vAmiga
https://api.github.com/repos/dirkwhoffmann/vAmiga
opened
Ghost'n Goblins broken
Priority-High bug
vAmiga v0.67: <img width="938" alt="Bildschirmfoto 2020-04-24 um 12 29 05" src="https://user-images.githubusercontent.com/12561945/80203381-59ffbc00-8627-11ea-988a-9f64d161960f.png"> vAmiga v0.68 and later: 🙈 <img width="938" alt="Bildschirmfoto 2020-04-24 um 12 23 19" src="https://user-images.githubusercontent.com/12561945/80203399-6126ca00-8627-11ea-980d-a07e8d7de561.png">
1.0
Ghost'n Goblins broken - vAmiga v0.67: <img width="938" alt="Bildschirmfoto 2020-04-24 um 12 29 05" src="https://user-images.githubusercontent.com/12561945/80203381-59ffbc00-8627-11ea-988a-9f64d161960f.png"> vAmiga v0.68 and later: 🙈 <img width="938" alt="Bildschirmfoto 2020-04-24 um 12 23 19" src="https://user-images.githubusercontent.com/12561945/80203399-6126ca00-8627-11ea-980d-a07e8d7de561.png">
priority
ghost n goblins broken vamiga img width alt bildschirmfoto um src vamiga and later 🙈 img width alt bildschirmfoto um src
1
700,922
24,078,727,457
IssuesEvent
2022-09-19 02:54:12
oasis-engine/engine
https://api.github.com/repos/oasis-engine/engine
closed
【物理系统】创建共享shape后,再更新shape的localPosition会报错
bug physical high priority
**Describe the bug** PhysXColliderShape.ts中创建shape使用的共享模式PhysXPhysics._pxPhysics.createShape( this._pxGeometry, material._pxMaterial, **false**, new PhysXPhysics._physX.PxShapeFlags(this._shapeFlags) ); 对于共享模式的shape是不可以再更改其属性的,然后该shape会被修改localPose, this._pxShape.setLocalPose(transform); 这种行为是不正确的,在physX debug模式下会看到很多报错 ![image](https://user-images.githubusercontent.com/2437648/185032353-06f23a77-718e-4f6d-b1af-a578ba8894e3.png) 一旦将创建shape的时候使用独享模式,该warning就会被解除。 **那么新的问题来了,如何设计共享模式以提高物理系统性能?** **bug重现方式** 重现该demo,使用本地代理使用debug版本physicX即可: https://oasisengine.cn/0.8/examples#lite-collision-detection
1.0
【物理系统】创建共享shape后,再更新shape的localPosition会报错 - **Describe the bug** PhysXColliderShape.ts中创建shape使用的共享模式PhysXPhysics._pxPhysics.createShape( this._pxGeometry, material._pxMaterial, **false**, new PhysXPhysics._physX.PxShapeFlags(this._shapeFlags) ); 对于共享模式的shape是不可以再更改其属性的,然后该shape会被修改localPose, this._pxShape.setLocalPose(transform); 这种行为是不正确的,在physX debug模式下会看到很多报错 ![image](https://user-images.githubusercontent.com/2437648/185032353-06f23a77-718e-4f6d-b1af-a578ba8894e3.png) 一旦将创建shape的时候使用独享模式,该warning就会被解除。 **那么新的问题来了,如何设计共享模式以提高物理系统性能?** **bug重现方式** 重现该demo,使用本地代理使用debug版本physicX即可: https://oasisengine.cn/0.8/examples#lite-collision-detection
priority
【物理系统】创建共享shape后,再更新shape的localposition会报错 describe the bug physxcollidershape ts中创建shape使用的共享模式physxphysics pxphysics createshape this pxgeometry material pxmaterial false new physxphysics physx pxshapeflags this shapeflags 对于共享模式的shape是不可以再更改其属性的,然后该shape会被修改localpose this pxshape setlocalpose transform 这种行为是不正确的,在physx debug模式下会看到很多报错 一旦将创建shape的时候使用独享模式,该warning就会被解除。 那么新的问题来了,如何设计共享模式以提高物理系统性能? bug重现方式 重现该demo,使用本地代理使用debug版本physicx即可
1
672,579
22,832,006,739
IssuesEvent
2022-07-12 13:41:22
openmm/openmmforcefields
https://api.github.com/repos/openmm/openmmforcefields
closed
Atom typing always uses GAFF even when GAFF2 specified
bug high-priority
When parameterising with GAFF2, openmmforcefields uses the following antechamber command: `antechamber -i in.mdl -fi mdl -o out.mol2 -fo mol2 -s 0 -at 2 -dr no` However, the options available to the 'at' flag are gaff, gaff2, amber, bcc, and sybyl. With the current argument of '2', antechamber is defaulting to using GAFF atom types. Sometimes this makes no difference as the correct GAFF2 parameters are used by parmchk2, but GAFF2 contains some atom types that GAFF does not, so sometimes it does matter. For example, two of the nitrogens in the attached molecule ([A_MOL_1_0.sdf.txt](https://github.com/openmm/openmmforcefields/files/9082915/A_MOL_1_0.sdf.txt)) are given atom type n by GAFF, but are given atom type ns by GAFF2.
1.0
Atom typing always uses GAFF even when GAFF2 specified - When parameterising with GAFF2, openmmforcefields uses the following antechamber command: `antechamber -i in.mdl -fi mdl -o out.mol2 -fo mol2 -s 0 -at 2 -dr no` However, the options available to the 'at' flag are gaff, gaff2, amber, bcc, and sybyl. With the current argument of '2', antechamber is defaulting to using GAFF atom types. Sometimes this makes no difference as the correct GAFF2 parameters are used by parmchk2, but GAFF2 contains some atom types that GAFF does not, so sometimes it does matter. For example, two of the nitrogens in the attached molecule ([A_MOL_1_0.sdf.txt](https://github.com/openmm/openmmforcefields/files/9082915/A_MOL_1_0.sdf.txt)) are given atom type n by GAFF, but are given atom type ns by GAFF2.
priority
atom typing always uses gaff even when specified when parameterising with openmmforcefields uses the following antechamber command antechamber i in mdl fi mdl o out fo s at dr no however the options available to the at flag are gaff amber bcc and sybyl with the current argument of antechamber is defaulting to using gaff atom types sometimes this makes no difference as the correct parameters are used by but contains some atom types that gaff does not so sometimes it does matter for example two of the nitrogens in the attached molecule are given atom type n by gaff but are given atom type ns by
1
394,148
11,632,690,868
IssuesEvent
2020-02-28 06:04:28
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.0 satging-1337] Web Election: no candidates for title election
Category: Web Priority: High Status: Fixed
1. Create title 2. Start Election for title ![bandicam 2020-01-15 15-09-01-891](https://user-images.githubusercontent.com/27898520/72426637-edfdf400-37ab-11ea-9a22-211186b0fd73.jpg) 3. Enter the election like a candidate ![bandicam 2020-01-15 15-09-11-504](https://user-images.githubusercontent.com/27898520/72426658-fa824c80-37ab-11ea-8eea-12400ae5b77f.jpg) ![bandicam 2020-01-15 15-09-13-198](https://user-images.githubusercontent.com/27898520/72426659-fc4c1000-37ab-11ea-907c-ec2441d0a77e.jpg) 4. Open site for voting No candidates ![bandicam 2020-01-15 15-09-38-593](https://user-images.githubusercontent.com/27898520/72426683-0a019580-37ac-11ea-9f6d-39d7f0dd73f4.jpg)
1.0
[0.9.0 satging-1337] Web Election: no candidates for title election - 1. Create title 2. Start Election for title ![bandicam 2020-01-15 15-09-01-891](https://user-images.githubusercontent.com/27898520/72426637-edfdf400-37ab-11ea-9a22-211186b0fd73.jpg) 3. Enter the election like a candidate ![bandicam 2020-01-15 15-09-11-504](https://user-images.githubusercontent.com/27898520/72426658-fa824c80-37ab-11ea-8eea-12400ae5b77f.jpg) ![bandicam 2020-01-15 15-09-13-198](https://user-images.githubusercontent.com/27898520/72426659-fc4c1000-37ab-11ea-907c-ec2441d0a77e.jpg) 4. Open site for voting No candidates ![bandicam 2020-01-15 15-09-38-593](https://user-images.githubusercontent.com/27898520/72426683-0a019580-37ac-11ea-9f6d-39d7f0dd73f4.jpg)
priority
web election no candidates for title election create title start election for title enter the election like a candidate open site for voting no candidates
1
293,455
8,990,608,701
IssuesEvent
2019-02-01 05:57:38
ooni/probe-android
https://api.github.com/repos/ooni/probe-android
closed
Report ID field is empty in the raw data section
High Priority bug
## Expected Behavior When I go into the view data section of the app I should see in the `report_id` populated with a value. This does not happen on probe-ios. ## Actual Behavior The `report_id` field contains the empty strings `""`. See: ![signal-attachment-2019-01-31-104928](https://user-images.githubusercontent.com/424620/52016089-88880780-2548-11e9-8171-22e4f22b398f.jpeg) ## Steps to Reproduce the Problem 1. Run any experiment on android 2. Go the view data section of the measurement via the option menu 3. Notice how the report_id field is empty ## Specifications - Version: 2.0.0 - Platform: android
1.0
Report ID field is empty in the raw data section - ## Expected Behavior When I go into the view data section of the app I should see in the `report_id` populated with a value. This does not happen on probe-ios. ## Actual Behavior The `report_id` field contains the empty strings `""`. See: ![signal-attachment-2019-01-31-104928](https://user-images.githubusercontent.com/424620/52016089-88880780-2548-11e9-8171-22e4f22b398f.jpeg) ## Steps to Reproduce the Problem 1. Run any experiment on android 2. Go the view data section of the measurement via the option menu 3. Notice how the report_id field is empty ## Specifications - Version: 2.0.0 - Platform: android
priority
report id field is empty in the raw data section expected behavior when i go into the view data section of the app i should see in the report id populated with a value this does not happen on probe ios actual behavior the report id field contains the empty strings see steps to reproduce the problem run any experiment on android go the view data section of the measurement via the option menu notice how the report id field is empty specifications version platform android
1
333,372
10,121,128,126
IssuesEvent
2019-07-31 15:01:32
bbc/simorgh
https://api.github.com/repos/bbc/simorgh
closed
Allow importing of specific moment timezones
high priority simorgh-core-stream
**Is your feature request related to a problem? Please describe.** Our current timestamp support wont scale, it currently filters out all timezones except London, this wont work for multiple services as they're in the vendor bundles, meaning any added are downloaded by all services. Doing this for multiple services means we need to import timezones into the service bundles. Easy? lol nope. Moments timezone data is a GIANT json file, making it so you cant just import a single one (ikr its dumb). We most likely need a solution that allows us to build timezone data into our service bundles at build time and then attach it to moment. **Describe the solution you'd like** Find a way of splitting timezone data and only importing a single one into each service bundle **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Testing notes** Timezones should work as expected, bundle size shouldnt increase atall (or by a tiny amount) Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc **Additional context** Add any other context or screenshots about the feature request here.
1.0
Allow importing of specific moment timezones - **Is your feature request related to a problem? Please describe.** Our current timestamp support wont scale, it currently filters out all timezones except London, this wont work for multiple services as they're in the vendor bundles, meaning any added are downloaded by all services. Doing this for multiple services means we need to import timezones into the service bundles. Easy? lol nope. Moments timezone data is a GIANT json file, making it so you cant just import a single one (ikr its dumb). We most likely need a solution that allows us to build timezone data into our service bundles at build time and then attach it to moment. **Describe the solution you'd like** Find a way of splitting timezone data and only importing a single one into each service bundle **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Testing notes** Timezones should work as expected, bundle size shouldnt increase atall (or by a tiny amount) Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc **Additional context** Add any other context or screenshots about the feature request here.
priority
allow importing of specific moment timezones is your feature request related to a problem please describe our current timestamp support wont scale it currently filters out all timezones except london this wont work for multiple services as they re in the vendor bundles meaning any added are downloaded by all services doing this for multiple services means we need to import timezones into the service bundles easy lol nope moments timezone data is a giant json file making it so you cant just import a single one ikr its dumb we most likely need a solution that allows us to build timezone data into our service bundles at build time and then attach it to moment describe the solution you d like find a way of splitting timezone data and only importing a single one into each service bundle describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered testing notes timezones should work as expected bundle size shouldnt increase atall or by a tiny amount dev insight will cypress tests be required or are unit tests sufficient will there be any potential regression etc additional context add any other context or screenshots about the feature request here
1
815,275
30,544,105,844
IssuesEvent
2023-07-20 01:23:06
KoseiN38/pj_study
https://api.github.com/repos/KoseiN38/pj_study
closed
[書籍] 『統計的因果推論の理論と実装』から既存のnotebook(document)改修をする
3 priority: high
// Template name issue template // About Here is a template for creating an issue // Template content ## Overview Please provide an overview of the issue you are creating. ex) I want to adapt the style of the header. ## Purpose Describe the purpose of this issue ex) To style it. ## Task Break down and manage your tasks. - [x] 傾向スコアに関する分析手法を調査 - [x] 操作変数法に関する分析手法を調査 - [x] ドキュメント、Notebookの作成
1.0
[書籍] 『統計的因果推論の理論と実装』から既存のnotebook(document)改修をする - // Template name issue template // About Here is a template for creating an issue // Template content ## Overview Please provide an overview of the issue you are creating. ex) I want to adapt the style of the header. ## Purpose Describe the purpose of this issue ex) To style it. ## Task Break down and manage your tasks. - [x] 傾向スコアに関する分析手法を調査 - [x] 操作変数法に関する分析手法を調査 - [x] ドキュメント、Notebookの作成
priority
『統計的因果推論の理論と実装』から既存のnotebook document 改修をする template name issue template about here is a template for creating an issue template content overview please provide an overview of the issue you are creating ex i want to adapt the style of the header purpose describe the purpose of this issue ex to style it task break down and manage your tasks 傾向スコアに関する分析手法を調査 操作変数法に関する分析手法を調査 ドキュメント、notebookの作成
1
242,060
7,837,462,369
IssuesEvent
2018-06-18 06:17:51
telerik/kendo-ui-core
https://api.github.com/repos/telerik/kendo-ui-core
closed
Input text is not visible in Combobox in Flat theme
Appearance Bug C: ComboBox Kendo2 Priority 5 SEV: High
### Bug report When Flat theme is used for ComboBox, the text in the input field is not visible. ### Reproduction of the problem 1. Open the [Demo](https://demos.telerik.com/kendo-ui/combobox/index) 2. Try to select an item ### Current behavior The item is selected correctly, but the text is not visible as the background color and the text color are the same. ### Expected/desired behavior The text in combobox should be visible ### Workaround Add color to the text in the ComboBox input `<style> .k-combobox .k-input{ color: #10c4b2; } </style>` ### Environment * **Kendo UI version:** 2018.2.516 * **Browser:** [all ]
1.0
Input text is not visible in Combobox in Flat theme - ### Bug report When Flat theme is used for ComboBox, the text in the input field is not visible. ### Reproduction of the problem 1. Open the [Demo](https://demos.telerik.com/kendo-ui/combobox/index) 2. Try to select an item ### Current behavior The item is selected correctly, but the text is not visible as the background color and the text color are the same. ### Expected/desired behavior The text in combobox should be visible ### Workaround Add color to the text in the ComboBox input `<style> .k-combobox .k-input{ color: #10c4b2; } </style>` ### Environment * **Kendo UI version:** 2018.2.516 * **Browser:** [all ]
priority
input text is not visible in combobox in flat theme bug report when flat theme is used for combobox the text in the input field is not visible reproduction of the problem open the try to select an item current behavior the item is selected correctly but the text is not visible as the background color and the text color are the same expected desired behavior the text in combobox should be visible workaround add color to the text in the combobox input k combobox k input color environment kendo ui version browser
1
788,280
27,749,479,088
IssuesEvent
2023-03-15 19:31:36
flatironinstitute/mcmc-monitor
https://api.github.com/repos/flatironinstitute/mcmc-monitor
closed
Data refresh should not change the variable selection state
bug high-priority
In the current version, clicking "Refresh data" or having it automatically triggered from auto refresh resets the selected list of variables (to `lp__`). The selection state should be preserved across data refreshes. Confirmed this occurs on at least the Diagnostics tab and the Scatterplots tab. To confirm fix: - Navigate to Diagnostics tab - Select several variables, not limited to `lp__` - Manually refresh the data, confirm selection state does not change - Repeat test for automated update - Repeat test for 3d scatterplot
1.0
Data refresh should not change the variable selection state - In the current version, clicking "Refresh data" or having it automatically triggered from auto refresh resets the selected list of variables (to `lp__`). The selection state should be preserved across data refreshes. Confirmed this occurs on at least the Diagnostics tab and the Scatterplots tab. To confirm fix: - Navigate to Diagnostics tab - Select several variables, not limited to `lp__` - Manually refresh the data, confirm selection state does not change - Repeat test for automated update - Repeat test for 3d scatterplot
priority
data refresh should not change the variable selection state in the current version clicking refresh data or having it automatically triggered from auto refresh resets the selected list of variables to lp the selection state should be preserved across data refreshes confirmed this occurs on at least the diagnostics tab and the scatterplots tab to confirm fix navigate to diagnostics tab select several variables not limited to lp manually refresh the data confirm selection state does not change repeat test for automated update repeat test for scatterplot
1
531,440
15,497,823,517
IssuesEvent
2021-03-11 05:22:22
arduino/arduino-cli
https://api.github.com/repos/arduino/arduino-cli
closed
Dependabot can't resolve your Go dependency files
component/dependencies priority: high type: bug
Dependabot can't resolve your Go dependency files. As a result, Dependabot couldn't update your dependencies. The error Dependabot encountered was: ``` Cannot detect VCS for go.bug.st/relaxed-semver. Attempted to detect VCS because the version looks like a git revision: v0.0.0-20190922224835-391e10178d18 ``` If you think the above is an error on Dependabot's side please don't hesitate to get in touch - we'll do whatever we can to fix it. [View the update logs](https://app.dependabot.com/accounts/arduino/update-logs/71377433).
1.0
Dependabot can't resolve your Go dependency files - Dependabot can't resolve your Go dependency files. As a result, Dependabot couldn't update your dependencies. The error Dependabot encountered was: ``` Cannot detect VCS for go.bug.st/relaxed-semver. Attempted to detect VCS because the version looks like a git revision: v0.0.0-20190922224835-391e10178d18 ``` If you think the above is an error on Dependabot's side please don't hesitate to get in touch - we'll do whatever we can to fix it. [View the update logs](https://app.dependabot.com/accounts/arduino/update-logs/71377433).
priority
dependabot can t resolve your go dependency files dependabot can t resolve your go dependency files as a result dependabot couldn t update your dependencies the error dependabot encountered was cannot detect vcs for go bug st relaxed semver attempted to detect vcs because the version looks like a git revision if you think the above is an error on dependabot s side please don t hesitate to get in touch we ll do whatever we can to fix it
1
145,179
5,559,994,504
IssuesEvent
2017-03-24 18:17:27
mreishman/Log-Hog
https://api.github.com/repos/mreishman/Log-Hog
opened
Warning popup if blank watch folder value
Priority - 2 - High
show a warning popup if one of the values if blank (only for the required value, other can be blank)
1.0
Warning popup if blank watch folder value - show a warning popup if one of the values if blank (only for the required value, other can be blank)
priority
warning popup if blank watch folder value show a warning popup if one of the values if blank only for the required value other can be blank
1
816,625
30,604,840,132
IssuesEvent
2023-07-22 21:49:29
evan-palmer/blue
https://api.github.com/repos/evan-palmer/blue
closed
[FEATURE]: Rework Docker VCS imports
enhancement high priority
### Feature Type Changing existing functionality in the BlueROV2 driver ### Problem Description The existing implementation installs all dependencies by cloning the VCS imports, but then deletes these repositories. After setting everything up, the dev container then, re-imports all of these. This is redundant and requires users to manage these dependencies in their local environments. ### Feature Description Move the VCS installation to a colcon workspace in the `$HOME` directory. From here, install all dependencies and add the package sourcing to the `ros_entrypoint.sh` script or to the `.bashrc` file. ### Alternative Solutions The current solution technically works; however, this is ultimately very redundant and can be confusing for new users. ### Additional Context _No response_
1.0
[FEATURE]: Rework Docker VCS imports - ### Feature Type Changing existing functionality in the BlueROV2 driver ### Problem Description The existing implementation installs all dependencies by cloning the VCS imports, but then deletes these repositories. After setting everything up, the dev container then, re-imports all of these. This is redundant and requires users to manage these dependencies in their local environments. ### Feature Description Move the VCS installation to a colcon workspace in the `$HOME` directory. From here, install all dependencies and add the package sourcing to the `ros_entrypoint.sh` script or to the `.bashrc` file. ### Alternative Solutions The current solution technically works; however, this is ultimately very redundant and can be confusing for new users. ### Additional Context _No response_
priority
rework docker vcs imports feature type changing existing functionality in the driver problem description the existing implementation installs all dependencies by cloning the vcs imports but then deletes these repositories after setting everything up the dev container then re imports all of these this is redundant and requires users to manage these dependencies in their local environments feature description move the vcs installation to a colcon workspace in the home directory from here install all dependencies and add the package sourcing to the ros entrypoint sh script or to the bashrc file alternative solutions the current solution technically works however this is ultimately very redundant and can be confusing for new users additional context no response
1
225,222
7,479,702,065
IssuesEvent
2018-04-04 15:16:51
HeathWallace/coinface
https://api.github.com/repos/HeathWallace/coinface
closed
Yaab breaks npm run build
:bug: bug :runner: Type: Deployment :small_red_triangle: Priority: High
``` > coinface@0.1.0 build /tmp/build_82caf179bcada5d8263dc894ba080334/gh > react-scripts build Creating an optimized production build... Failed to compile. Failed to minify the code from this file: ./node_modules/yaab/yaab.js:2 Read more here: http://bit.ly/2tRViJ9 npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! coinface@0.1.0 build: `react-scripts build` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the coinface@0.1.0 build script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /app/.npm/_logs/2018-04-04T12_56_47_326Z-debug.log ```
1.0
Yaab breaks npm run build - ``` > coinface@0.1.0 build /tmp/build_82caf179bcada5d8263dc894ba080334/gh > react-scripts build Creating an optimized production build... Failed to compile. Failed to minify the code from this file: ./node_modules/yaab/yaab.js:2 Read more here: http://bit.ly/2tRViJ9 npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! coinface@0.1.0 build: `react-scripts build` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the coinface@0.1.0 build script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /app/.npm/_logs/2018-04-04T12_56_47_326Z-debug.log ```
priority
yaab breaks npm run build coinface build tmp build gh react scripts build creating an optimized production build failed to compile failed to minify the code from this file node modules yaab yaab js read more here npm err code elifecycle npm err errno npm err coinface build react scripts build npm err exit status npm err npm err failed at the coinface build script npm err this is probably not a problem with npm there is likely additional logging output above npm err a complete log of this run can be found in npm err app npm logs debug log
1
757,245
26,502,145,896
IssuesEvent
2023-01-18 11:05:39
valantic/vue-template
https://api.github.com/repos/valantic/vue-template
closed
Replace body-scroll helper with plugin
enhancement high priority
On the IV project, we replaced the body-scroll helper with a self maintained plugin, which is more reliable. Should be replaced here as well.
1.0
Replace body-scroll helper with plugin - On the IV project, we replaced the body-scroll helper with a self maintained plugin, which is more reliable. Should be replaced here as well.
priority
replace body scroll helper with plugin on the iv project we replaced the body scroll helper with a self maintained plugin which is more reliable should be replaced here as well
1
417,923
12,190,058,065
IssuesEvent
2020-04-29 08:39:05
WoWManiaUK/Blackwing-Lair
https://api.github.com/repos/WoWManiaUK/Blackwing-Lair
closed
[Worgen] RDF (phasing issue)
Fixed in Dev Priority-High
**What is happening:** - Worgens are getting phasing bug after they have joined RDF _worgen must finish the questline before moving from their startzone (we know this)_ This new issue was just reported to a GM ingame **What should happen:** Restrict worgens to be locked from quing for RDF until out of startzone _we restricted the Mall stone for the same reason_ #4156 this gives the worgen a chance to complete all their important quest level 19 and still only at Battle for Gilneas City questline ![lvl 19](https://user-images.githubusercontent.com/39439201/80379299-2da1a500-889e-11ea-9218-9a0accf869d5.jpg) (still 20 quests left to do before that zone is completed and the Worgen can leave) https://wow.gamepedia.com/Worgen_starting_experience **Other Information:** 1. This will save the worgens getting bugged/phased out 2. Less tickets for GMs
1.0
[Worgen] RDF (phasing issue) - **What is happening:** - Worgens are getting phasing bug after they have joined RDF _worgen must finish the questline before moving from their startzone (we know this)_ This new issue was just reported to a GM ingame **What should happen:** Restrict worgens to be locked from quing for RDF until out of startzone _we restricted the Mall stone for the same reason_ #4156 this gives the worgen a chance to complete all their important quest level 19 and still only at Battle for Gilneas City questline ![lvl 19](https://user-images.githubusercontent.com/39439201/80379299-2da1a500-889e-11ea-9218-9a0accf869d5.jpg) (still 20 quests left to do before that zone is completed and the Worgen can leave) https://wow.gamepedia.com/Worgen_starting_experience **Other Information:** 1. This will save the worgens getting bugged/phased out 2. Less tickets for GMs
priority
rdf phasing issue what is happening worgens are getting phasing bug after they have joined rdf worgen must finish the questline before moving from their startzone we know this this new issue was just reported to a gm ingame what should happen restrict worgens to be locked from quing for rdf until out of startzone we restricted the mall stone for the same reason this gives the worgen a chance to complete all their important quest level and still only at battle for gilneas city questline still quests left to do before that zone is completed and the worgen can leave other information this will save the worgens getting bugged phased out less tickets for gms
1
274,115
8,557,158,167
IssuesEvent
2018-11-08 15:04:31
metwork-framework/mfserv
https://api.github.com/repos/metwork-framework/mfserv
opened
when you have plenty of plugins/applications on a slow machine...
Priority: High Status: Accepted Type: Enhancement
... you can have a bad circus status during startup because of timeout problems
1.0
when you have plenty of plugins/applications on a slow machine... - ... you can have a bad circus status during startup because of timeout problems
priority
when you have plenty of plugins applications on a slow machine you can have a bad circus status during startup because of timeout problems
1
232,318
7,657,616,943
IssuesEvent
2018-05-10 20:16:49
Connexions/nebuchadnezzar
https://api.github.com/repos/Connexions/nebuchadnezzar
opened
neb publish un-bakes a previously baked collection
bug priority:high
After running `$ neb -v publish qa col11448 'aneuploid to ANEUPLOID: m44480 m44483'`, Biology (which was previously baked on qa) is no longer baked. https://qa.cnx.org/contents/GFy_h8cu@12.1:rZudN6XP@3/Introduction: ![image](https://user-images.githubusercontent.com/11298165/39891992-c37791cc-5464-11e8-8237-f5ad93c0828f.png)
1.0
neb publish un-bakes a previously baked collection - After running `$ neb -v publish qa col11448 'aneuploid to ANEUPLOID: m44480 m44483'`, Biology (which was previously baked on qa) is no longer baked. https://qa.cnx.org/contents/GFy_h8cu@12.1:rZudN6XP@3/Introduction: ![image](https://user-images.githubusercontent.com/11298165/39891992-c37791cc-5464-11e8-8237-f5ad93c0828f.png)
priority
neb publish un bakes a previously baked collection after running neb v publish qa aneuploid to aneuploid biology which was previously baked on qa is no longer baked
1
65,031
3,223,023,430
IssuesEvent
2015-10-09 07:15:33
GermanCentralLibraryForTheBlind/readium-js-viewer
https://api.github.com/repos/GermanCentralLibraryForTheBlind/readium-js-viewer
closed
reader toc view handling
bug high priority
toc open -> goto library -> reopen the book-> toc will be displayed instead of book content
1.0
reader toc view handling - toc open -> goto library -> reopen the book-> toc will be displayed instead of book content
priority
reader toc view handling toc open goto library reopen the book toc will be displayed instead of book content
1
535,723
15,697,226,092
IssuesEvent
2021-03-26 04:00:42
daleal/zum
https://api.github.com/repos/daleal/zum
closed
Support different config file names
feature high priority
It is essential to be able to use different config file names (other than `zum.toml`) to start the engine. The problem is that the CLI instance runs "_all at once_". My idea would be to use a "hidden" parser to catch some `--file` flag before running the second argument parser. Then, I could start the engine with the specified file.
1.0
Support different config file names - It is essential to be able to use different config file names (other than `zum.toml`) to start the engine. The problem is that the CLI instance runs "_all at once_". My idea would be to use a "hidden" parser to catch some `--file` flag before running the second argument parser. Then, I could start the engine with the specified file.
priority
support different config file names it is essential to be able to use different config file names other than zum toml to start the engine the problem is that the cli instance runs all at once my idea would be to use a hidden parser to catch some file flag before running the second argument parser then i could start the engine with the specified file
1
402,641
11,812,311,253
IssuesEvent
2020-03-19 19:54:53
TannerDisney/DisneyCafe-Portfolio
https://api.github.com/repos/TannerDisney/DisneyCafe-Portfolio
closed
Create HTML page for users to take orders
Feature Front-End High Priority
need to create a HTML page for users with accounts to be able to take orders and receive an invoice for said order.
1.0
Create HTML page for users to take orders - need to create a HTML page for users with accounts to be able to take orders and receive an invoice for said order.
priority
create html page for users to take orders need to create a html page for users with accounts to be able to take orders and receive an invoice for said order
1
165,915
6,288,175,203
IssuesEvent
2017-07-19 16:23:30
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Ship AVX2 mkl .so files with wheels
bug high priority
Here is the code causing the segmentation fault: ``` import torch import torch.nn as nn import torch.nn.functional as F from torch.autograd import Variable as Var #import Constants import numpy as np import random class dc_model(nn.Module): def __init__(self): super(dc_model, self).__init__() self.projection = nn.Linear(193, 2) if __name__ =='__main__': model = dc_model() variance = Var(torch.randn(3000, 193)) print(model.projection(variance)) ``` Here is the gdb trace ``` (gdb) r pytest.py Starting program: /users/pxie1/miniconda2/envs/py35/bin/python pytest.py [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1". [New Thread 0x7ffff350f700 (LWP 12359)] [New Thread 0x7ffff2d0e700 (LWP 12360)] [New Thread 0x7fffee50d700 (LWP 12361)] [New Thread 0x7fffebd0c700 (LWP 12362)] [New Thread 0x7fffe950b700 (LWP 12363)] [New Thread 0x7fffe6d0a700 (LWP 12364)] [New Thread 0x7fffe4509700 (LWP 12365)] [New Thread 0x7fffe1d08700 (LWP 12366)] [New Thread 0x7fffdf507700 (LWP 12367)] [New Thread 0x7fffdcd06700 (LWP 12368)] [New Thread 0x7fffda505700 (LWP 12369)] [New Thread 0x7fffd7d04700 (LWP 12370)] [New Thread 0x7fffd5503700 (LWP 12371)] [New Thread 0x7fffd2d02700 (LWP 12372)] [New Thread 0x7fffd0501700 (LWP 12373)] [New Thread 0x7fffcdd00700 (LWP 12374)] [New Thread 0x7fffcb4ff700 (LWP 12375)] [New Thread 0x7fffcacfe700 (LWP 12376)] [New Thread 0x7fffc84fd700 (LWP 12377)] [New Thread 0x7fffc5cfc700 (LWP 12378)] [New Thread 0x7fffc34fb700 (LWP 12379)] [New Thread 0x7fffc0cfa700 (LWP 12380)] [New Thread 0x7fffbe4f9700 (LWP 12381)] [New Thread 0x7fffbbcf8700 (LWP 12382)] [New Thread 0x7fffb94f7700 (LWP 12383)] [New Thread 0x7fffb6cf6700 (LWP 12384)] [New Thread 0x7fffb44f5700 (LWP 12385)] [New Thread 0x7fffb1cf4700 (LWP 12386)] [New Thread 0x7fffaf4f3700 (LWP 12387)] [New Thread 0x7fffaccf2700 (LWP 12388)] [New Thread 0x7fffaa4f1700 (LWP 12389)] warning: File "/users/pxie1/miniconda2/envs/py35/lib/libstdc++.so.6.0.19-gdb.py" auto-loading has been declined by your `auto-load safe-path' set to "$debugdir:$datadir/auto-load". To enable execution of this file add add-auto-load-safe-path /users/pxie1/miniconda2/envs/py35/lib/libstdc++.so.6.0.19-gdb.py line to your configuration file "/users/pxie1/.gdbinit". To completely disable this security protection add set auto-load safe-path / line to your configuration file "/users/pxie1/.gdbinit". For more information about this security protection see the "Auto-loading safe path" section in the GDB manual. E.g., run from the shell: info "(gdb)Auto-loading safe path" Thread 1 "python" received signal SIGSEGV, Segmentation fault. 0x0000000000000010 in ?? () (gdb) where #0 0x0000000000000010 in ?? () #1 0x00007fff71106782 in mkl_blas_avx2_sgemm_initialize_kernel_info () from /users/pxie1/miniconda2/envs/py35/bin/../lib/libmkl_avx2.so #2 0x00007fff8059e096 in ?? () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libmkl_intel_thread.so #3 0x00007fff8059ded9 in ?? () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libmkl_intel_thread.so #4 0x00007fff8059cff0 in ?? () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libmkl_intel_thread.so #5 0x00007fff805c9ca1 in ?? () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libmkl_intel_thread.so #6 0x00007fff81e3f145 in ?? () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libmkl_intel_lp64.so #7 0x00007fff98087f61 in THFloatBlas_gemm () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libTH.so.1 #8 0x00007fff97e24e60 in THFloatTensor_addmm () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libTH.so.1 #9 0x00007fffa1f7ed8d in THPFloatTensor_addmm_ (self=0x7fff742f5fc8, args=<optimized out>, kwargs=<optimized out>) at /b/wheel/pytorch-src/torch/csrc/generic/TensorMethods.cpp:19898 #10 0x00007ffff79a0dd1 in PyCFunction_Call (func=0x7fff742f7fc0, args=0x7fff742cfae8, kwds=<optimized out>) at Objects/methodobject.c:98 #11 0x00007ffff7a294a6 in call_function (oparg=<optimized out>, pp_stack=0x7fffffffd038) at Python/ceval.c:4720 #12 PyEval_EvalFrameEx (f=<optimized out>, throwflag=<optimized out>) at Python/ceval.c:3251 #13 0x00007ffff7a29fc9 in _PyEval_EvalCodeWithName (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=4, kws=0x0, kwcount=0, defs=0x7fff74402108, defcount=1, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at Python/ceval.c:4033 #14 0x00007ffff7a2a158 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x7fff74402108, defcount=1, kwdefs=0x0, closure=0x0) at Python/ceval.c:4054 #15 0x00007ffff797eb72 in function_call (func=0x7fff743f86a8, arg=0x7fff742cfbd8, kw=0x0) at Objects/funcobject.c:627 #16 0x00007ffff794b4c6 in PyObject_Call (func=0x7fff743f86a8, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #17 0x00007ffff7967bdc in method_call (func=0x7fff743f86a8, arg=0x7fff742cfbd8, kw=0x0) at Objects/classobject.c:330 #18 0x00007ffff794b4c6 in PyObject_Call (func=0x7ffff7f9d188, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #19 0x00007ffff7a1e5a3 in PyEval_CallObjectWithKeywords (func=0x7ffff7f9d188, arg=0x7ffff64a13f0, kw=<optimized out>) at Python/ceval.c:4595 #20 0x00007fffa2326568 in THPFunction_do_forward (self=0x7ffff7e82048, inputs=0x7ffff7e955a0) at torch/csrc/autograd/python_function.cpp:456 #21 0x00007ffff79a0d99 in PyCFunction_Call (func=0x7fff742f7f78, args=0x7ffff7e955a0, kwds=<optimized out>) at Objects/methodobject.c:109 #22 0x00007ffff794b4c6 in PyObject_Call (func=0x7fff742f7f78, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #23 0x00007ffff79bf25f in slot_tp_call (self=0x7ffff7e82048, args=0x7ffff7e955a0, kwds=0x0) at Objects/typeobject.c:6072 #24 0x00007ffff794b4c6 in PyObject_Call (func=0x7ffff7e82048, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #25 0x00007ffff7a2779c in do_call (nk=<optimized out>, na=3, pp_stack=0x7fffffffd718, func=0x7ffff7e82048) at Python/ceval.c:4951 #26 call_function (oparg=<optimized out>, pp_stack=0x7fffffffd718) at Python/ceval.c:4747 #27 PyEval_EvalFrameEx (f=<optimized out>, throwflag=<optimized out>) at Python/ceval.c:3251 #28 0x00007ffff7a29fc9 in _PyEval_EvalCodeWithName (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=2, kws=0x7ffff7fa5060, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at Python/ceval.c:4033 #29 0x00007ffff7a2a158 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0) at Python/ceval.c:4054 #30 0x00007ffff797ec91 in function_call (func=0x7fff7437cae8, arg=0x7ffff7e8f508, kw=0x7ffff64ac588) at Objects/funcobject.c:627 #31 0x00007ffff794b4c6 in PyObject_Call (func=0x7fff7437cae8, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #32 0x00007ffff7a26286 in ext_do_call (nk=-135727864, na=1, flags=<optimized out>, pp_stack=0x7fffffffda68, func=0x7fff7437cae8) at Python/ceval.c:5049 #33 PyEval_EvalFrameEx (f=<optimized out>, throwflag=<optimized out>) at Python/ceval.c:3290 #34 0x00007ffff7a29fc9 in _PyEval_EvalCodeWithName (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=2, kws=0x0, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at Python/ceval.c:4033 #35 0x00007ffff7a2a158 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0) at Python/ceval.c:4054 #36 0x00007ffff797eb72 in function_call (func=0x7fff7437bd08, arg=0x7ffff7e86608, kw=0x0) at Objects/funcobject.c:627 #37 0x00007ffff794b4c6 in PyObject_Call (func=0x7fff7437bd08, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #38 0x00007ffff7967bdc in method_call (func=0x7fff7437bd08, arg=0x7ffff7e86608, kw=0x0) at Objects/classobject.c:330 #39 0x00007ffff794b4c6 in PyObject_Call (func=0x7ffff7f94648, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #40 0x00007ffff79bf25f in slot_tp_call (self=0x7ffff64921d0, args=0x7ffff7e89198, kwds=0x0) at Objects/typeobject.c:6072 #41 0x00007ffff794b4c6 in PyObject_Call (func=0x7ffff64921d0, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #42 0x00007ffff7a2779c in do_call (nk=<optimized out>, na=1, pp_stack=0x7fffffffdec8, func=0x7ffff64921d0) at Python/ceval.c:4951 #43 call_function (oparg=<optimized out>, pp_stack=0x7fffffffdec8) at Python/ceval.c:4747 #44 PyEval_EvalFrameEx (f=<optimized out>, throwflag=<optimized out>) at Python/ceval.c:3251 #45 0x00007ffff7a29fc9 in _PyEval_EvalCodeWithName (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=0, kws=0x0, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at Python/ceval.c:4033 #46 0x00007ffff7a2a158 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0) at Python/ceval.c:4054 #47 0x00007ffff7a2a19b in PyEval_EvalCode (co=<optimized out>, globals=<optimized out>, locals=<optimized out>) at Python/ceval.c:777 #48 0x00007ffff7a4f410 in run_mod (arena=0x6a4dd0, flags=0x7fffffffe210, locals=0x7ffff7f53148, globals=0x7ffff7f53148, filename=0x7ffff670d830, mod=0x70f1d8) at Python/pythonrun.c:982 #49 PyRun_FileExFlags (fp=0x669eb0, filename_str=<optimized out>, start=<optimized out>, globals=0x7ffff7f53148, locals=0x7ffff7f53148, closeit=<optimized out>, flags=0x7fffffffe210) at Python/pythonrun.c:935 #50 0x00007ffff7a50a03 in PyRun_SimpleFileExFlags (fp=0x669eb0, filename=<optimized out>, closeit=1, flags=0x7fffffffe210) at Python/pythonrun.c:402 #51 0x00007ffff7a6bce7 in run_file (p_cf=0x7fffffffe210, filename=0x6032f0 L"pytest.py", fp=0x669eb0) at Modules/main.c:318 #52 Py_Main (argc=<optimized out>, argv=<optimized out>) at Modules/main.c:769 #53 0x0000000000400bbd in main (argc=2, argv=0x7fffffffe388) at ./Programs/python.c:65 ``` Actually, this program run smoothly on my own PC, but meet segmentation fault on the gpu server. So I'm afraid there might be wrong configuration or dependency error. Anyone can help? Thanks very much!
1.0
Ship AVX2 mkl .so files with wheels - Here is the code causing the segmentation fault: ``` import torch import torch.nn as nn import torch.nn.functional as F from torch.autograd import Variable as Var #import Constants import numpy as np import random class dc_model(nn.Module): def __init__(self): super(dc_model, self).__init__() self.projection = nn.Linear(193, 2) if __name__ =='__main__': model = dc_model() variance = Var(torch.randn(3000, 193)) print(model.projection(variance)) ``` Here is the gdb trace ``` (gdb) r pytest.py Starting program: /users/pxie1/miniconda2/envs/py35/bin/python pytest.py [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1". [New Thread 0x7ffff350f700 (LWP 12359)] [New Thread 0x7ffff2d0e700 (LWP 12360)] [New Thread 0x7fffee50d700 (LWP 12361)] [New Thread 0x7fffebd0c700 (LWP 12362)] [New Thread 0x7fffe950b700 (LWP 12363)] [New Thread 0x7fffe6d0a700 (LWP 12364)] [New Thread 0x7fffe4509700 (LWP 12365)] [New Thread 0x7fffe1d08700 (LWP 12366)] [New Thread 0x7fffdf507700 (LWP 12367)] [New Thread 0x7fffdcd06700 (LWP 12368)] [New Thread 0x7fffda505700 (LWP 12369)] [New Thread 0x7fffd7d04700 (LWP 12370)] [New Thread 0x7fffd5503700 (LWP 12371)] [New Thread 0x7fffd2d02700 (LWP 12372)] [New Thread 0x7fffd0501700 (LWP 12373)] [New Thread 0x7fffcdd00700 (LWP 12374)] [New Thread 0x7fffcb4ff700 (LWP 12375)] [New Thread 0x7fffcacfe700 (LWP 12376)] [New Thread 0x7fffc84fd700 (LWP 12377)] [New Thread 0x7fffc5cfc700 (LWP 12378)] [New Thread 0x7fffc34fb700 (LWP 12379)] [New Thread 0x7fffc0cfa700 (LWP 12380)] [New Thread 0x7fffbe4f9700 (LWP 12381)] [New Thread 0x7fffbbcf8700 (LWP 12382)] [New Thread 0x7fffb94f7700 (LWP 12383)] [New Thread 0x7fffb6cf6700 (LWP 12384)] [New Thread 0x7fffb44f5700 (LWP 12385)] [New Thread 0x7fffb1cf4700 (LWP 12386)] [New Thread 0x7fffaf4f3700 (LWP 12387)] [New Thread 0x7fffaccf2700 (LWP 12388)] [New Thread 0x7fffaa4f1700 (LWP 12389)] warning: File "/users/pxie1/miniconda2/envs/py35/lib/libstdc++.so.6.0.19-gdb.py" auto-loading has been declined by your `auto-load safe-path' set to "$debugdir:$datadir/auto-load". To enable execution of this file add add-auto-load-safe-path /users/pxie1/miniconda2/envs/py35/lib/libstdc++.so.6.0.19-gdb.py line to your configuration file "/users/pxie1/.gdbinit". To completely disable this security protection add set auto-load safe-path / line to your configuration file "/users/pxie1/.gdbinit". For more information about this security protection see the "Auto-loading safe path" section in the GDB manual. E.g., run from the shell: info "(gdb)Auto-loading safe path" Thread 1 "python" received signal SIGSEGV, Segmentation fault. 0x0000000000000010 in ?? () (gdb) where #0 0x0000000000000010 in ?? () #1 0x00007fff71106782 in mkl_blas_avx2_sgemm_initialize_kernel_info () from /users/pxie1/miniconda2/envs/py35/bin/../lib/libmkl_avx2.so #2 0x00007fff8059e096 in ?? () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libmkl_intel_thread.so #3 0x00007fff8059ded9 in ?? () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libmkl_intel_thread.so #4 0x00007fff8059cff0 in ?? () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libmkl_intel_thread.so #5 0x00007fff805c9ca1 in ?? () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libmkl_intel_thread.so #6 0x00007fff81e3f145 in ?? () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libmkl_intel_lp64.so #7 0x00007fff98087f61 in THFloatBlas_gemm () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libTH.so.1 #8 0x00007fff97e24e60 in THFloatTensor_addmm () from /users/pxie1/miniconda2/envs/py35/lib/python3.5/site-packages/torch/lib/libTH.so.1 #9 0x00007fffa1f7ed8d in THPFloatTensor_addmm_ (self=0x7fff742f5fc8, args=<optimized out>, kwargs=<optimized out>) at /b/wheel/pytorch-src/torch/csrc/generic/TensorMethods.cpp:19898 #10 0x00007ffff79a0dd1 in PyCFunction_Call (func=0x7fff742f7fc0, args=0x7fff742cfae8, kwds=<optimized out>) at Objects/methodobject.c:98 #11 0x00007ffff7a294a6 in call_function (oparg=<optimized out>, pp_stack=0x7fffffffd038) at Python/ceval.c:4720 #12 PyEval_EvalFrameEx (f=<optimized out>, throwflag=<optimized out>) at Python/ceval.c:3251 #13 0x00007ffff7a29fc9 in _PyEval_EvalCodeWithName (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=4, kws=0x0, kwcount=0, defs=0x7fff74402108, defcount=1, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at Python/ceval.c:4033 #14 0x00007ffff7a2a158 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x7fff74402108, defcount=1, kwdefs=0x0, closure=0x0) at Python/ceval.c:4054 #15 0x00007ffff797eb72 in function_call (func=0x7fff743f86a8, arg=0x7fff742cfbd8, kw=0x0) at Objects/funcobject.c:627 #16 0x00007ffff794b4c6 in PyObject_Call (func=0x7fff743f86a8, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #17 0x00007ffff7967bdc in method_call (func=0x7fff743f86a8, arg=0x7fff742cfbd8, kw=0x0) at Objects/classobject.c:330 #18 0x00007ffff794b4c6 in PyObject_Call (func=0x7ffff7f9d188, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #19 0x00007ffff7a1e5a3 in PyEval_CallObjectWithKeywords (func=0x7ffff7f9d188, arg=0x7ffff64a13f0, kw=<optimized out>) at Python/ceval.c:4595 #20 0x00007fffa2326568 in THPFunction_do_forward (self=0x7ffff7e82048, inputs=0x7ffff7e955a0) at torch/csrc/autograd/python_function.cpp:456 #21 0x00007ffff79a0d99 in PyCFunction_Call (func=0x7fff742f7f78, args=0x7ffff7e955a0, kwds=<optimized out>) at Objects/methodobject.c:109 #22 0x00007ffff794b4c6 in PyObject_Call (func=0x7fff742f7f78, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #23 0x00007ffff79bf25f in slot_tp_call (self=0x7ffff7e82048, args=0x7ffff7e955a0, kwds=0x0) at Objects/typeobject.c:6072 #24 0x00007ffff794b4c6 in PyObject_Call (func=0x7ffff7e82048, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #25 0x00007ffff7a2779c in do_call (nk=<optimized out>, na=3, pp_stack=0x7fffffffd718, func=0x7ffff7e82048) at Python/ceval.c:4951 #26 call_function (oparg=<optimized out>, pp_stack=0x7fffffffd718) at Python/ceval.c:4747 #27 PyEval_EvalFrameEx (f=<optimized out>, throwflag=<optimized out>) at Python/ceval.c:3251 #28 0x00007ffff7a29fc9 in _PyEval_EvalCodeWithName (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=2, kws=0x7ffff7fa5060, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at Python/ceval.c:4033 #29 0x00007ffff7a2a158 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0) at Python/ceval.c:4054 #30 0x00007ffff797ec91 in function_call (func=0x7fff7437cae8, arg=0x7ffff7e8f508, kw=0x7ffff64ac588) at Objects/funcobject.c:627 #31 0x00007ffff794b4c6 in PyObject_Call (func=0x7fff7437cae8, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #32 0x00007ffff7a26286 in ext_do_call (nk=-135727864, na=1, flags=<optimized out>, pp_stack=0x7fffffffda68, func=0x7fff7437cae8) at Python/ceval.c:5049 #33 PyEval_EvalFrameEx (f=<optimized out>, throwflag=<optimized out>) at Python/ceval.c:3290 #34 0x00007ffff7a29fc9 in _PyEval_EvalCodeWithName (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=2, kws=0x0, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at Python/ceval.c:4033 #35 0x00007ffff7a2a158 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0) at Python/ceval.c:4054 #36 0x00007ffff797eb72 in function_call (func=0x7fff7437bd08, arg=0x7ffff7e86608, kw=0x0) at Objects/funcobject.c:627 #37 0x00007ffff794b4c6 in PyObject_Call (func=0x7fff7437bd08, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #38 0x00007ffff7967bdc in method_call (func=0x7fff7437bd08, arg=0x7ffff7e86608, kw=0x0) at Objects/classobject.c:330 #39 0x00007ffff794b4c6 in PyObject_Call (func=0x7ffff7f94648, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #40 0x00007ffff79bf25f in slot_tp_call (self=0x7ffff64921d0, args=0x7ffff7e89198, kwds=0x0) at Objects/typeobject.c:6072 #41 0x00007ffff794b4c6 in PyObject_Call (func=0x7ffff64921d0, arg=<optimized out>, kw=<optimized out>) at Objects/abstract.c:2166 #42 0x00007ffff7a2779c in do_call (nk=<optimized out>, na=1, pp_stack=0x7fffffffdec8, func=0x7ffff64921d0) at Python/ceval.c:4951 #43 call_function (oparg=<optimized out>, pp_stack=0x7fffffffdec8) at Python/ceval.c:4747 #44 PyEval_EvalFrameEx (f=<optimized out>, throwflag=<optimized out>) at Python/ceval.c:3251 #45 0x00007ffff7a29fc9 in _PyEval_EvalCodeWithName (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=0, kws=0x0, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0, name=0x0, qualname=0x0) at Python/ceval.c:4033 #46 0x00007ffff7a2a158 in PyEval_EvalCodeEx (_co=<optimized out>, globals=<optimized out>, locals=<optimized out>, args=<optimized out>, argcount=<optimized out>, kws=<optimized out>, kwcount=0, defs=0x0, defcount=0, kwdefs=0x0, closure=0x0) at Python/ceval.c:4054 #47 0x00007ffff7a2a19b in PyEval_EvalCode (co=<optimized out>, globals=<optimized out>, locals=<optimized out>) at Python/ceval.c:777 #48 0x00007ffff7a4f410 in run_mod (arena=0x6a4dd0, flags=0x7fffffffe210, locals=0x7ffff7f53148, globals=0x7ffff7f53148, filename=0x7ffff670d830, mod=0x70f1d8) at Python/pythonrun.c:982 #49 PyRun_FileExFlags (fp=0x669eb0, filename_str=<optimized out>, start=<optimized out>, globals=0x7ffff7f53148, locals=0x7ffff7f53148, closeit=<optimized out>, flags=0x7fffffffe210) at Python/pythonrun.c:935 #50 0x00007ffff7a50a03 in PyRun_SimpleFileExFlags (fp=0x669eb0, filename=<optimized out>, closeit=1, flags=0x7fffffffe210) at Python/pythonrun.c:402 #51 0x00007ffff7a6bce7 in run_file (p_cf=0x7fffffffe210, filename=0x6032f0 L"pytest.py", fp=0x669eb0) at Modules/main.c:318 #52 Py_Main (argc=<optimized out>, argv=<optimized out>) at Modules/main.c:769 #53 0x0000000000400bbd in main (argc=2, argv=0x7fffffffe388) at ./Programs/python.c:65 ``` Actually, this program run smoothly on my own PC, but meet segmentation fault on the gpu server. So I'm afraid there might be wrong configuration or dependency error. Anyone can help? Thanks very much!
priority
ship mkl so files with wheels here is the code causing the segmentation fault import torch import torch nn as nn import torch nn functional as f from torch autograd import variable as var import constants import numpy as np import random class dc model nn module def init self super dc model self init self projection nn linear if name main model dc model variance var torch randn print model projection variance here is the gdb trace gdb r pytest py starting program users envs bin python pytest py using host libthread db library lib linux gnu libthread db so warning file users envs lib libstdc so gdb py auto loading has been declined by your auto load safe path set to debugdir datadir auto load to enable execution of this file add add auto load safe path users envs lib libstdc so gdb py line to your configuration file users gdbinit to completely disable this security protection add set auto load safe path line to your configuration file users gdbinit for more information about this security protection see the auto loading safe path section in the gdb manual e g run from the shell info gdb auto loading safe path thread python received signal sigsegv segmentation fault in gdb where in in mkl blas sgemm initialize kernel info from users envs bin lib libmkl so in from users envs lib site packages torch lib libmkl intel thread so in from users envs lib site packages torch lib libmkl intel thread so in from users envs lib site packages torch lib libmkl intel thread so in from users envs lib site packages torch lib libmkl intel thread so in from users envs lib site packages torch lib libmkl intel so in thfloatblas gemm from users envs lib site packages torch lib libth so in thfloattensor addmm from users envs lib site packages torch lib libth so in thpfloattensor addmm self args kwargs at b wheel pytorch src torch csrc generic tensormethods cpp in pycfunction call func args kwds at objects methodobject c in call function oparg pp stack at python ceval c pyeval evalframeex f throwflag at python ceval c in pyeval evalcodewithname co globals locals args argcount kws kwcount defs defcount kwdefs closure name qualname at python ceval c in pyeval evalcodeex co globals locals args argcount kws kwcount defs defcount kwdefs closure at python ceval c in function call func arg kw at objects funcobject c in pyobject call func arg kw at objects abstract c in method call func arg kw at objects classobject c in pyobject call func arg kw at objects abstract c in pyeval callobjectwithkeywords func arg kw at python ceval c in thpfunction do forward self inputs at torch csrc autograd python function cpp in pycfunction call func args kwds at objects methodobject c in pyobject call func arg kw at objects abstract c in slot tp call self args kwds at objects typeobject c in pyobject call func arg kw at objects abstract c in do call nk na pp stack func at python ceval c call function oparg pp stack at python ceval c pyeval evalframeex f throwflag at python ceval c in pyeval evalcodewithname co globals locals args argcount kws kwcount defs defcount kwdefs closure name qualname at python ceval c in pyeval evalcodeex co globals locals args argcount kws kwcount defs defcount kwdefs closure at python ceval c in function call func arg kw at objects funcobject c in pyobject call func arg kw at objects abstract c in ext do call nk na flags pp stack func at python ceval c pyeval evalframeex f throwflag at python ceval c in pyeval evalcodewithname co globals locals args argcount kws kwcount defs defcount kwdefs closure name qualname at python ceval c in pyeval evalcodeex co globals locals args argcount kws kwcount defs defcount kwdefs closure at python ceval c in function call func arg kw at objects funcobject c in pyobject call func arg kw at objects abstract c in method call func arg kw at objects classobject c in pyobject call func arg kw at objects abstract c in slot tp call self args kwds at objects typeobject c in pyobject call func arg kw at objects abstract c in do call nk na pp stack func at python ceval c call function oparg pp stack at python ceval c pyeval evalframeex f throwflag at python ceval c in pyeval evalcodewithname co globals locals args argcount kws kwcount defs defcount kwdefs closure name qualname at python ceval c in pyeval evalcodeex co globals locals args argcount kws kwcount defs defcount kwdefs closure at python ceval c in pyeval evalcode co globals locals at python ceval c in run mod arena flags locals globals filename mod at python pythonrun c pyrun fileexflags fp filename str start globals locals closeit flags at python pythonrun c in pyrun simplefileexflags fp filename closeit flags at python pythonrun c in run file p cf filename l pytest py fp at modules main c py main argc argv at modules main c in main argc argv at programs python c actually this program run smoothly on my own pc but meet segmentation fault on the gpu server so i m afraid there might be wrong configuration or dependency error anyone can help thanks very much
1
153,005
5,873,004,570
IssuesEvent
2017-05-15 13:04:22
EEA-Norway-Grants/dataviz
https://api.github.com/repos/EEA-Norway-Grants/dataviz
opened
programme areas colours are not persistent
Complexity: High Priority: Low Type: Bug
we need programme areas colours pre-generated regardless of filters.
1.0
programme areas colours are not persistent - we need programme areas colours pre-generated regardless of filters.
priority
programme areas colours are not persistent we need programme areas colours pre generated regardless of filters
1
249,014
7,948,483,077
IssuesEvent
2018-07-11 08:17:44
ballerina-platform/ballerina-message-broker
https://api.github.com/repos/ballerina-platform/ballerina-message-broker
closed
Message Broker goes OOM
Complexity/Advanced Priority/Highest Severity/Critical Type/Bug
**Description:** Message Broker goes OOM when the jms consumer is started and jms producer is publishing large (40kb) messages with average TPS of 5000 to 10000. **Affected Product Version:** message-broker-0.1.66 **OS, DB, other environment details and versions:** Ubuntu 15.10 **Steps to reproduce:** Used the following artefacts - https://github.com/sashikamw/ballerina-stabilitytest/tree/master/ServiceScenarios/Scenario03_JMS/beta17/MultipleWorkers **Observations** OOM HeapDump - https://drive.google.com/file/d/1wEhz6X6sn5pSSsFlBHRjyNvVcthhDBf0/view?usp=sharing
1.0
Message Broker goes OOM - **Description:** Message Broker goes OOM when the jms consumer is started and jms producer is publishing large (40kb) messages with average TPS of 5000 to 10000. **Affected Product Version:** message-broker-0.1.66 **OS, DB, other environment details and versions:** Ubuntu 15.10 **Steps to reproduce:** Used the following artefacts - https://github.com/sashikamw/ballerina-stabilitytest/tree/master/ServiceScenarios/Scenario03_JMS/beta17/MultipleWorkers **Observations** OOM HeapDump - https://drive.google.com/file/d/1wEhz6X6sn5pSSsFlBHRjyNvVcthhDBf0/view?usp=sharing
priority
message broker goes oom description message broker goes oom when the jms consumer is started and jms producer is publishing large messages with average tps of to affected product version message broker os db other environment details and versions ubuntu steps to reproduce used the following artefacts observations oom heapdump
1
2,251
2,524,980,633
IssuesEvent
2015-01-20 21:23:36
graybeal/ont
https://api.github.com/repos/graybeal/ont
closed
list of versions include unversioned
1 star bug imported portal Priority-High
_From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on September 29, 2009 17:51:03_ What steps will reproduce the problem? 1. go to the portal and click any registered, non-external ontology 2. click "versions" to open the list of available versions What is the expected output? All entries in the list in versioned form What do you see instead? the most recent shows in unversioned form (although with the "version" parameter in the link. Note: we will also add an entry with the "unversioned" form in this list, but this is not the current intent, that's why this is a bug. _Original issue: http://code.google.com/p/mmisw/issues/detail?id=193_
1.0
list of versions include unversioned - _From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on September 29, 2009 17:51:03_ What steps will reproduce the problem? 1. go to the portal and click any registered, non-external ontology 2. click "versions" to open the list of available versions What is the expected output? All entries in the list in versioned form What do you see instead? the most recent shows in unversioned form (although with the "version" parameter in the link. Note: we will also add an entry with the "unversioned" form in this list, but this is not the current intent, that's why this is a bug. _Original issue: http://code.google.com/p/mmisw/issues/detail?id=193_
priority
list of versions include unversioned from on september what steps will reproduce the problem go to the portal and click any registered non external ontology click versions to open the list of available versions what is the expected output all entries in the list in versioned form what do you see instead the most recent shows in unversioned form although with the version parameter in the link note we will also add an entry with the unversioned form in this list but this is not the current intent that s why this is a bug original issue
1
59,238
3,104,378,657
IssuesEvent
2015-08-31 15:33:19
wp-invoice/wp-invoice
https://api.github.com/repos/wp-invoice/wp-invoice
closed
WooComerce Subscriptions compatibility issue.
core bug High Priority
1. If the visitor navigates to a Product that is a Variable Subscription <http://www.theportlandcompany.com/product/pay-per-click-management/> (this is a type of Product created by WooCommerce Subscriptions Extension) 2. And they select an option from the Select menu. 3. They *should* be prompted with an Add to Cart button, instead they're presented with an error message that says "Sorry, this product is unavailable. Please choose a different combination.". 1. This is a WooCommerce message but it is displayed only when WP Invoice is activated. https://usabilitydynamics.zendesk.com/agent/#/tickets/816 topic on wordpress - https://wordpress.org/support/topic/bug-wp-invoice-and-woocommerce-subscription-plugin?replies=1
1.0
WooComerce Subscriptions compatibility issue. - 1. If the visitor navigates to a Product that is a Variable Subscription <http://www.theportlandcompany.com/product/pay-per-click-management/> (this is a type of Product created by WooCommerce Subscriptions Extension) 2. And they select an option from the Select menu. 3. They *should* be prompted with an Add to Cart button, instead they're presented with an error message that says "Sorry, this product is unavailable. Please choose a different combination.". 1. This is a WooCommerce message but it is displayed only when WP Invoice is activated. https://usabilitydynamics.zendesk.com/agent/#/tickets/816 topic on wordpress - https://wordpress.org/support/topic/bug-wp-invoice-and-woocommerce-subscription-plugin?replies=1
priority
woocomerce subscriptions compatibility issue if the visitor navigates to a product that is a variable subscription this is a type of product created by woocommerce subscriptions extension and they select an option from the select menu they should be prompted with an add to cart button instead they re presented with an error message that says sorry this product is unavailable please choose a different combination this is a woocommerce message but it is displayed only when wp invoice is activated topic on wordpress
1
487,608
14,049,170,010
IssuesEvent
2020-11-02 09:54:58
trezor/trezor-suite
https://api.github.com/repos/trezor/trezor-suite
closed
Support Zcash Canopy hardfork
Bug High priority
The Zcash network will undergo its next network upgrade, named Canopy, at blockheight 1046400, which is expected to be mined on approximately November 18th. We need to start sending `0xE9FF75A6` as the new consensus branch id. That should be all that's required. This PR could help: https://github.com/trezor/trezor-suite/pull/2072/files. We should include this in the upcoming release :/. ---- More info from the email: ``` This network upgrade contains changes to Zcash consensus and therefore requires all network participants to upgrade their Zcashd nodes and accompanying software to version 4.0.0 or later. We've released the Canopy Network Upgrade Guide with more information. The main changes in this upgrade include: ZIP 251 defines the Consensus Branch ID, which should be updated to 0xE9FF75A6 ZIP 207 and ZIP 214 are needed to establish a new development fund. ZIP 211 disables the ability to add new value to the Sprout value pool balance. ZIP 212 improves the security assumptions needed to ensure that diversified addresses are not linkable. ZIP 215 improves the validity criteria of Ed25519 signatures used in Sprout transactions. ```
1.0
Support Zcash Canopy hardfork - The Zcash network will undergo its next network upgrade, named Canopy, at blockheight 1046400, which is expected to be mined on approximately November 18th. We need to start sending `0xE9FF75A6` as the new consensus branch id. That should be all that's required. This PR could help: https://github.com/trezor/trezor-suite/pull/2072/files. We should include this in the upcoming release :/. ---- More info from the email: ``` This network upgrade contains changes to Zcash consensus and therefore requires all network participants to upgrade their Zcashd nodes and accompanying software to version 4.0.0 or later. We've released the Canopy Network Upgrade Guide with more information. The main changes in this upgrade include: ZIP 251 defines the Consensus Branch ID, which should be updated to 0xE9FF75A6 ZIP 207 and ZIP 214 are needed to establish a new development fund. ZIP 211 disables the ability to add new value to the Sprout value pool balance. ZIP 212 improves the security assumptions needed to ensure that diversified addresses are not linkable. ZIP 215 improves the validity criteria of Ed25519 signatures used in Sprout transactions. ```
priority
support zcash canopy hardfork the zcash network will undergo its next network upgrade named canopy at blockheight which is expected to be mined on approximately november we need to start sending as the new consensus branch id that should be all that s required this pr could help we should include this in the upcoming release more info from the email this network upgrade contains changes to zcash consensus and therefore requires all network participants to upgrade their zcashd nodes and accompanying software to version or later we ve released the canopy network upgrade guide with more information the main changes in this upgrade include zip defines the consensus branch id which should be updated to zip and zip are needed to establish a new development fund zip disables the ability to add new value to the sprout value pool balance zip improves the security assumptions needed to ensure that diversified addresses are not linkable zip improves the validity criteria of signatures used in sprout transactions
1
461,234
13,226,957,425
IssuesEvent
2020-08-18 01:36:40
fossasia/open-event-frontend
https://api.github.com/repos/fossasia/open-event-frontend
closed
Scheduler time limiting is wrong
Priority: High bug
If an event starts at 21 June 7 PM and ends at 23 June 9 PM, all days show the scheduler from 7 PM onwards only. Only the first day should be limited by 7 PM (minTime) and the last day should be limited by 9 PM (maxTime) and days in between should have all hours available
1.0
Scheduler time limiting is wrong - If an event starts at 21 June 7 PM and ends at 23 June 9 PM, all days show the scheduler from 7 PM onwards only. Only the first day should be limited by 7 PM (minTime) and the last day should be limited by 9 PM (maxTime) and days in between should have all hours available
priority
scheduler time limiting is wrong if an event starts at june pm and ends at june pm all days show the scheduler from pm onwards only only the first day should be limited by pm mintime and the last day should be limited by pm maxtime and days in between should have all hours available
1
403,449
11,841,172,015
IssuesEvent
2020-03-23 20:16:33
Pulsar4xDevs/Pulsar4x
https://api.github.com/repos/Pulsar4xDevs/Pulsar4x
opened
Flesh out Json mod files.
Discussion: Game Play |Task| High Priority
These basically describe the ingame components, and installations that can be designed, tech that can be researched, minerals, materials, etc. This is a good issue for somone who has an eye for balance and can come up with interesting gameplay mechanics that work within the hardcode framework. The game is looking to be more "hard" scifi than Aurora (i.e. we have newtonion movement and real instead of TN minerals) numbers for e.g. engine thrust, component weights, fuel use etc. should have some basis on reality. #71 could help with this but is not a requirement. (once you've messed with the json files a little bit, they're not too hard to work with raw)
1.0
Flesh out Json mod files. - These basically describe the ingame components, and installations that can be designed, tech that can be researched, minerals, materials, etc. This is a good issue for somone who has an eye for balance and can come up with interesting gameplay mechanics that work within the hardcode framework. The game is looking to be more "hard" scifi than Aurora (i.e. we have newtonion movement and real instead of TN minerals) numbers for e.g. engine thrust, component weights, fuel use etc. should have some basis on reality. #71 could help with this but is not a requirement. (once you've messed with the json files a little bit, they're not too hard to work with raw)
priority
flesh out json mod files these basically describe the ingame components and installations that can be designed tech that can be researched minerals materials etc this is a good issue for somone who has an eye for balance and can come up with interesting gameplay mechanics that work within the hardcode framework the game is looking to be more hard scifi than aurora i e we have newtonion movement and real instead of tn minerals numbers for e g engine thrust component weights fuel use etc should have some basis on reality could help with this but is not a requirement once you ve messed with the json files a little bit they re not too hard to work with raw
1
531,566
15,500,386,072
IssuesEvent
2021-03-11 09:15:15
ISISScientificComputing/autoreduce
https://api.github.com/repos/ISISScientificComputing/autoreduce
opened
Message.description cannot be default=None anymore
:clock1: High Priority
Issue raised by: [developer/user/project requirement] ### What? The default for `Message.description` is `description = attr.ib(default=None)`. This doesn't work well anymore because we *always* pass in `message.description` on `db_records.create_reduction_run_record` - see [here](https://github1s.com/ISISScientificComputing/autoreduce/blob/master/model/database/records.py)
1.0
Message.description cannot be default=None anymore - Issue raised by: [developer/user/project requirement] ### What? The default for `Message.description` is `description = attr.ib(default=None)`. This doesn't work well anymore because we *always* pass in `message.description` on `db_records.create_reduction_run_record` - see [here](https://github1s.com/ISISScientificComputing/autoreduce/blob/master/model/database/records.py)
priority
message description cannot be default none anymore issue raised by what the default for message description is description attr ib default none this doesn t work well anymore because we always pass in message description on db records create reduction run record see
1
317,014
9,659,838,451
IssuesEvent
2019-05-20 14:17:20
FundacionParaguaya/MentorApp
https://api.github.com/repos/FundacionParaguaya/MentorApp
closed
Local Detox e2e setup
enhancement high priority
We should setup e2e testing locally with Detox. This includes: 1. Make Detox run basic test 2. Write actual tests for lifemap 3. Setup some form of automation
1.0
Local Detox e2e setup - We should setup e2e testing locally with Detox. This includes: 1. Make Detox run basic test 2. Write actual tests for lifemap 3. Setup some form of automation
priority
local detox setup we should setup testing locally with detox this includes make detox run basic test write actual tests for lifemap setup some form of automation
1
681,644
23,319,473,918
IssuesEvent
2022-08-08 15:08:23
vaticle/typedb-studio
https://api.github.com/repos/vaticle/typedb-studio
opened
Preferences dialog
type: feature priority: high domain: user-experience
## Problem to Solve We have quite a few feature requests that would work better as preferences rather than having their own dedicated icons or dialogs. ## Proposed Solution Implement a preferences dialog. ## Additional Information A list of issues that would benefit from a preferences dialog: - https://github.com/vaticle/typedb-studio/issues/428 (how many past connections get remembered and auto-reconnect) - https://github.com/vaticle/typedb-studio/issues/435 (a toggle and an integer representing the limit, defaulting to 1000) - https://github.com/vaticle/typedb-studio/issues/550 - https://github.com/vaticle/typedb-studio/issues/573 - https://github.com/vaticle/typedb-studio/issues/600
1.0
Preferences dialog - ## Problem to Solve We have quite a few feature requests that would work better as preferences rather than having their own dedicated icons or dialogs. ## Proposed Solution Implement a preferences dialog. ## Additional Information A list of issues that would benefit from a preferences dialog: - https://github.com/vaticle/typedb-studio/issues/428 (how many past connections get remembered and auto-reconnect) - https://github.com/vaticle/typedb-studio/issues/435 (a toggle and an integer representing the limit, defaulting to 1000) - https://github.com/vaticle/typedb-studio/issues/550 - https://github.com/vaticle/typedb-studio/issues/573 - https://github.com/vaticle/typedb-studio/issues/600
priority
preferences dialog problem to solve we have quite a few feature requests that would work better as preferences rather than having their own dedicated icons or dialogs proposed solution implement a preferences dialog additional information a list of issues that would benefit from a preferences dialog how many past connections get remembered and auto reconnect a toggle and an integer representing the limit defaulting to
1
158,511
6,029,196,569
IssuesEvent
2017-06-08 17:27:27
littlebits/react-popover
https://api.github.com/repos/littlebits/react-popover
closed
Build folder is missing in 0.4.11
bug priority / high
Latest 0.4.11 has build folder missing. Reverting to 0.4.10 works
1.0
Build folder is missing in 0.4.11 - Latest 0.4.11 has build folder missing. Reverting to 0.4.10 works
priority
build folder is missing in latest has build folder missing reverting to works
1
826,842
31,714,475,197
IssuesEvent
2023-09-09 17:36:17
BenWestgate/Bails
https://api.github.com/repos/BenWestgate/Bails
closed
Unable to create a new receiving address
bug help wanted priority: high priority: medium
After initial sync & importing a BAILS-created wallet: Under the Receive tab, I cannot create a new receiving address to send a small test transaction (button is grayed out). I tried filling out all the forms, changing from Bech32, closing the wallet, restarting Core, and restarting Tails, but still no luck. The Window > Receiving addresses window shows no addresses.
2.0
Unable to create a new receiving address - After initial sync & importing a BAILS-created wallet: Under the Receive tab, I cannot create a new receiving address to send a small test transaction (button is grayed out). I tried filling out all the forms, changing from Bech32, closing the wallet, restarting Core, and restarting Tails, but still no luck. The Window > Receiving addresses window shows no addresses.
priority
unable to create a new receiving address after initial sync importing a bails created wallet under the receive tab i cannot create a new receiving address to send a small test transaction button is grayed out i tried filling out all the forms changing from closing the wallet restarting core and restarting tails but still no luck the window receiving addresses window shows no addresses
1
4,615
2,559,699,127
IssuesEvent
2015-02-05 03:43:04
chessmasterhong/WaterEmblem
https://api.github.com/repos/chessmasterhong/WaterEmblem
closed
Enemy turn does not end if they don't get killed
bug high priority
Currently, if enemies attack a player and aren't defeated during that battle, they're turn does not end.
1.0
Enemy turn does not end if they don't get killed - Currently, if enemies attack a player and aren't defeated during that battle, they're turn does not end.
priority
enemy turn does not end if they don t get killed currently if enemies attack a player and aren t defeated during that battle they re turn does not end
1
353,642
10,555,310,143
IssuesEvent
2019-10-03 21:32:12
OpenSRP/opensrp-client-chw
https://api.github.com/repos/OpenSRP/opensrp-client-chw
closed
Update constraint and constraint message on phone numbers for changing caregiver
bug high priority
![Screenshot_20190920-101357_CFC-RTM - Togo](https://user-images.githubusercontent.com/3488986/65307092-b087b200-db8f-11e9-96f7-e12ebd41b4b5.jpg) We need to update the phone number constraints for the changing primary caregiver action in the app (it should match the changing family head action, but it's not currently) To pass QA: - [x] Both phone number fields in the screenshot above should accept any number between 8 and 16 digits long, inclusive. - [x] The constraint messages on both phone number fields should match the constraint message in the change family head action - [x] All of the above works in both the French and English versions
1.0
Update constraint and constraint message on phone numbers for changing caregiver - ![Screenshot_20190920-101357_CFC-RTM - Togo](https://user-images.githubusercontent.com/3488986/65307092-b087b200-db8f-11e9-96f7-e12ebd41b4b5.jpg) We need to update the phone number constraints for the changing primary caregiver action in the app (it should match the changing family head action, but it's not currently) To pass QA: - [x] Both phone number fields in the screenshot above should accept any number between 8 and 16 digits long, inclusive. - [x] The constraint messages on both phone number fields should match the constraint message in the change family head action - [x] All of the above works in both the French and English versions
priority
update constraint and constraint message on phone numbers for changing caregiver we need to update the phone number constraints for the changing primary caregiver action in the app it should match the changing family head action but it s not currently to pass qa both phone number fields in the screenshot above should accept any number between and digits long inclusive the constraint messages on both phone number fields should match the constraint message in the change family head action all of the above works in both the french and english versions
1
739,306
25,590,432,090
IssuesEvent
2022-12-01 12:35:50
4paradigm/OpenMLDB
https://api.github.com/repos/4paradigm/OpenMLDB
opened
Tablet coredump
bug high-priority
version: 0.6.0 the tablet log as below: ``` W1122 15:14:52.672183 33843 tablet_impl.cc:1385] table is not exist. tid 82, pid 7 W1122 15:14:52.674259 33843 tablet_impl.cc:2140] table is not exist. tid 82, pid 2 W1122 15:14:52.712164 33844 tablet_catalog.cc:279] fail to get tablet: pk 2022112215143753985145 not exist W1122 15:14:52.712221 33844 runner.cc:3769] fail to run proxy runner with rows: subquery tablet is null W1122 15:14:52.712239 33844 runner.cc:3592] run proxy runner with rows fail, result table is null ```
1.0
Tablet coredump - version: 0.6.0 the tablet log as below: ``` W1122 15:14:52.672183 33843 tablet_impl.cc:1385] table is not exist. tid 82, pid 7 W1122 15:14:52.674259 33843 tablet_impl.cc:2140] table is not exist. tid 82, pid 2 W1122 15:14:52.712164 33844 tablet_catalog.cc:279] fail to get tablet: pk 2022112215143753985145 not exist W1122 15:14:52.712221 33844 runner.cc:3769] fail to run proxy runner with rows: subquery tablet is null W1122 15:14:52.712239 33844 runner.cc:3592] run proxy runner with rows fail, result table is null ```
priority
tablet coredump version the tablet log as below tablet impl cc table is not exist tid pid tablet impl cc table is not exist tid pid tablet catalog cc fail to get tablet pk not exist runner cc fail to run proxy runner with rows subquery tablet is null runner cc run proxy runner with rows fail result table is null
1
616,631
19,308,143,235
IssuesEvent
2021-12-13 13:44:48
bounswe/2021SpringGroup2
https://api.github.com/repos/bounswe/2021SpringGroup2
opened
[Frontend] Unit tests for map functionalities
type: enhancement priority: high state: in progress Frontend
We have to create unit tests for maps used in create event page/search page. Since these maps are mostly focused on view and their structure is defined by the library, I am planning to test the search bar I implemented for location search. This search bar is connected to an API; however, I can mock the response from the API and check if the response is formatted correctly before it is used in the search bar.
1.0
[Frontend] Unit tests for map functionalities - We have to create unit tests for maps used in create event page/search page. Since these maps are mostly focused on view and their structure is defined by the library, I am planning to test the search bar I implemented for location search. This search bar is connected to an API; however, I can mock the response from the API and check if the response is formatted correctly before it is used in the search bar.
priority
unit tests for map functionalities we have to create unit tests for maps used in create event page search page since these maps are mostly focused on view and their structure is defined by the library i am planning to test the search bar i implemented for location search this search bar is connected to an api however i can mock the response from the api and check if the response is formatted correctly before it is used in the search bar
1
307,813
9,422,548,672
IssuesEvent
2019-04-11 09:36:56
FundacionParaguaya/MentorApp
https://api.github.com/repos/FundacionParaguaya/MentorApp
closed
Sentry Reports for App crashes
bug client waiting... high priority
Please review if we have app crashing reports within Sentry (or anywhere!) for the following sierra leone related issue. See epic for details.
1.0
Sentry Reports for App crashes - Please review if we have app crashing reports within Sentry (or anywhere!) for the following sierra leone related issue. See epic for details.
priority
sentry reports for app crashes please review if we have app crashing reports within sentry or anywhere for the following sierra leone related issue see epic for details
1
306,303
9,383,546,916
IssuesEvent
2019-04-05 04:05:59
franceme/Cryptoguard
https://api.github.com/repos/franceme/Cryptoguard
reopened
Java 1.11 (Next LTS) Feature Changes
Needs: Review Priority: High Status: In Progress Type: Maintence
--To see a Jep:> https://openjdk.java.net/jeps/{JEP Number} Java 1.8 - [-] JEP126: Lambdas (?Breaks Source Scan (Java/Source)) Java 1.9 - [-] JEP261: Module System - [-] JEP238: Multi-Release JARs - [ ] JEP287: SHA3 Java 1.10 - [ ] JEP286: Var Java 1.11 - [ ] JEP329: ChaCha20/Poly1305 Crypto Algos - [ ] JEP323: Var within Lambdas - [ ] JEP332: TLS 1.3 - [ ] JEP181: Nest-Based Access Control - [ ] JEP336/335/320: Various deprecation
1.0
Java 1.11 (Next LTS) Feature Changes - --To see a Jep:> https://openjdk.java.net/jeps/{JEP Number} Java 1.8 - [-] JEP126: Lambdas (?Breaks Source Scan (Java/Source)) Java 1.9 - [-] JEP261: Module System - [-] JEP238: Multi-Release JARs - [ ] JEP287: SHA3 Java 1.10 - [ ] JEP286: Var Java 1.11 - [ ] JEP329: ChaCha20/Poly1305 Crypto Algos - [ ] JEP323: Var within Lambdas - [ ] JEP332: TLS 1.3 - [ ] JEP181: Nest-Based Access Control - [ ] JEP336/335/320: Various deprecation
priority
java next lts feature changes to see a jep number java lambdas breaks source scan java source java module system multi release jars java var java crypto algos var within lambdas tls nest based access control various deprecation
1
797,324
28,144,067,827
IssuesEvent
2023-04-02 09:21:47
KarinaKuznecova/Animal-Shelter-Game
https://api.github.com/repos/KarinaKuznecova/Animal-Shelter-Game
closed
Player has skills
new feature high priority complicated
player and skills should be saved not to lose progress skills should be visible in esc menu
1.0
Player has skills - player and skills should be saved not to lose progress skills should be visible in esc menu
priority
player has skills player and skills should be saved not to lose progress skills should be visible in esc menu
1
558,361
16,531,313,520
IssuesEvent
2021-05-27 06:28:52
opensrp/opensrp-client-reveal
https://api.github.com/repos/opensrp/opensrp-client-reveal
closed
NTD: Task generation failing at first then working the second time
NTD Priority: High bug
1. Log-in to Reveal with ntd11/Amani123 2. Select Lomahasha 3. Add a new child BUG: The task isn't generated for the new child ![image](https://user-images.githubusercontent.com/61143692/119664393-f5232080-be3b-11eb-96aa-faf79d584e36.png)
1.0
NTD: Task generation failing at first then working the second time - 1. Log-in to Reveal with ntd11/Amani123 2. Select Lomahasha 3. Add a new child BUG: The task isn't generated for the new child ![image](https://user-images.githubusercontent.com/61143692/119664393-f5232080-be3b-11eb-96aa-faf79d584e36.png)
priority
ntd task generation failing at first then working the second time log in to reveal with select lomahasha add a new child bug the task isn t generated for the new child
1
422,369
12,270,650,611
IssuesEvent
2020-05-07 15:49:18
ansible/galaxy_ng
https://api.github.com/repos/ansible/galaxy_ng
closed
Importer: ansible-test error inside pylint on openshift
area/importer priority/high status/fix-released type/bug
When `ansible-test sanity` runs during import, it runs in an openshift container and the `pylint` test is run. For collection `ibm.spectrum_virtualize`, `pylint` calls `multiprocessing.connection.recv` which hits an EOF error when unpickling an object. Environment(s) seen with issue: AH Prod and AH QA Environment(s) seen without issue: running importer in cli mode, running `ansible-test sanity -t pylint --docker` locally Traceback from AH QA import: ``` ... Running sanity test 'pep8' with Python 3.6 Running sanity test 'pylint' with Python 3.6 ERROR: Command "/usr/bin/python3.6 -m pylint --jobs 0 --reports n --max-line-length 160 --rcfile /usr/local/lib/python3.6/dist-packages/ansible_test/_data/sanity/pylint/config/collection.cfg --output-format json --load-plugins blacklist,deprecated,string_format tests/unit/plugins/module_utils/test_ibm_svc_utils.py tests/unit/plugins/modules/test_ibm_svc_host.py tests/unit/plugins/modules/test_ibm_svc_info.py tests/unit/plugins/modules/test_ibm_svc_mdisk.py tests/unit/plugins/modules/test_ibm_svc_mdiskgrp.py tests/unit/plugins/modules/test_ibm_svc_vdisk.py tests/unit/plugins/modules/test_ibm_svc_vol_map.py" returned exit status 1. >>> Standard Error Process ChildLinter-2: Traceback (most recent call last): File "/usr/lib/python3.6/multiprocessing/process.py", line 258, in _bootstrap self.run() File "/usr/local/lib/python3.6/dist-packages/pylint/lint.py", line 241, in run for file_or_module in iter(tasks_queue.get, "STOP"): File "<string>", line 2, in get File "/usr/lib/python3.6/multiprocessing/managers.py", line 757, in _callmethod kind, result = conn.recv() File "/usr/lib/python3.6/multiprocessing/connection.py", line 250, in recv buf = self._recv_bytes() File "/usr/lib/python3.6/multiprocessing/connection.py", line 407, in _recv_bytes buf = self._recv(4) File "/usr/lib/python3.6/multiprocessing/connection.py", line 383, in _recv raise EOFError EOFError Process ChildLinter-4: Traceback (most recent call last): File "/usr/lib/python3.6/multiprocessing/process.py", line 258, in _bootstrap self.run() File "/usr/local/lib/python3.6/dist-packages/pylint/lint.py", line 241, in run for file_or_module in iter(tasks_queue.get, "STOP"): File "<string>", line 2, in get File "/usr/lib/python3.6/multiprocessing/managers.py", line 757, in _callmethod kind, result = conn.recv() File "/usr/lib/python3.6/multiprocessing/connection.py", line 250, in recv buf = self._recv_bytes() File "/usr/lib/python3.6/multiprocessing/connection.py", line 407, in _recv_bytes buf = self._recv(4) File "/usr/lib/python3.6/multiprocessing/connection.py", line 383, in _recv raise EOFError EOFError internal error with sending report for module ['tests/unit/plugins/modules/test_ibm_svc_info.py'] [Errno 32] Broken pipe Process ChildLinter-3: Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/pylint/lint.py", line 244, in run results_queue.put(result) File "<string>", line 2, in put File "/usr/lib/python3.6/multiprocessing/managers.py", line 756, in _callmethod conn.send((self._id, methodname, args, kwds)) File "/usr/lib/python3.6/multiprocessing/connection.py", line 206, in send self._send_bytes(_ForkingPickler.dumps(obj)) File "/usr/lib/python3.6/multiprocessing/connection.py", line 404, in _send_bytes self._send(header + buf) File "/usr/lib/python3.6/multiprocessing/connection.py", line 368, in _send n = write(self._handle, buf) BrokenPipeError: [Errno 32] Broken pipe During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.6/multiprocessing/process.py", line 258, in _bootstrap self.run() File "/usr/local/lib/python3.6/dist-packages/pylint/lint.py", line 252, in run ... ```
1.0
Importer: ansible-test error inside pylint on openshift - When `ansible-test sanity` runs during import, it runs in an openshift container and the `pylint` test is run. For collection `ibm.spectrum_virtualize`, `pylint` calls `multiprocessing.connection.recv` which hits an EOF error when unpickling an object. Environment(s) seen with issue: AH Prod and AH QA Environment(s) seen without issue: running importer in cli mode, running `ansible-test sanity -t pylint --docker` locally Traceback from AH QA import: ``` ... Running sanity test 'pep8' with Python 3.6 Running sanity test 'pylint' with Python 3.6 ERROR: Command "/usr/bin/python3.6 -m pylint --jobs 0 --reports n --max-line-length 160 --rcfile /usr/local/lib/python3.6/dist-packages/ansible_test/_data/sanity/pylint/config/collection.cfg --output-format json --load-plugins blacklist,deprecated,string_format tests/unit/plugins/module_utils/test_ibm_svc_utils.py tests/unit/plugins/modules/test_ibm_svc_host.py tests/unit/plugins/modules/test_ibm_svc_info.py tests/unit/plugins/modules/test_ibm_svc_mdisk.py tests/unit/plugins/modules/test_ibm_svc_mdiskgrp.py tests/unit/plugins/modules/test_ibm_svc_vdisk.py tests/unit/plugins/modules/test_ibm_svc_vol_map.py" returned exit status 1. >>> Standard Error Process ChildLinter-2: Traceback (most recent call last): File "/usr/lib/python3.6/multiprocessing/process.py", line 258, in _bootstrap self.run() File "/usr/local/lib/python3.6/dist-packages/pylint/lint.py", line 241, in run for file_or_module in iter(tasks_queue.get, "STOP"): File "<string>", line 2, in get File "/usr/lib/python3.6/multiprocessing/managers.py", line 757, in _callmethod kind, result = conn.recv() File "/usr/lib/python3.6/multiprocessing/connection.py", line 250, in recv buf = self._recv_bytes() File "/usr/lib/python3.6/multiprocessing/connection.py", line 407, in _recv_bytes buf = self._recv(4) File "/usr/lib/python3.6/multiprocessing/connection.py", line 383, in _recv raise EOFError EOFError Process ChildLinter-4: Traceback (most recent call last): File "/usr/lib/python3.6/multiprocessing/process.py", line 258, in _bootstrap self.run() File "/usr/local/lib/python3.6/dist-packages/pylint/lint.py", line 241, in run for file_or_module in iter(tasks_queue.get, "STOP"): File "<string>", line 2, in get File "/usr/lib/python3.6/multiprocessing/managers.py", line 757, in _callmethod kind, result = conn.recv() File "/usr/lib/python3.6/multiprocessing/connection.py", line 250, in recv buf = self._recv_bytes() File "/usr/lib/python3.6/multiprocessing/connection.py", line 407, in _recv_bytes buf = self._recv(4) File "/usr/lib/python3.6/multiprocessing/connection.py", line 383, in _recv raise EOFError EOFError internal error with sending report for module ['tests/unit/plugins/modules/test_ibm_svc_info.py'] [Errno 32] Broken pipe Process ChildLinter-3: Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/pylint/lint.py", line 244, in run results_queue.put(result) File "<string>", line 2, in put File "/usr/lib/python3.6/multiprocessing/managers.py", line 756, in _callmethod conn.send((self._id, methodname, args, kwds)) File "/usr/lib/python3.6/multiprocessing/connection.py", line 206, in send self._send_bytes(_ForkingPickler.dumps(obj)) File "/usr/lib/python3.6/multiprocessing/connection.py", line 404, in _send_bytes self._send(header + buf) File "/usr/lib/python3.6/multiprocessing/connection.py", line 368, in _send n = write(self._handle, buf) BrokenPipeError: [Errno 32] Broken pipe During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.6/multiprocessing/process.py", line 258, in _bootstrap self.run() File "/usr/local/lib/python3.6/dist-packages/pylint/lint.py", line 252, in run ... ```
priority
importer ansible test error inside pylint on openshift when ansible test sanity runs during import it runs in an openshift container and the pylint test is run for collection ibm spectrum virtualize pylint calls multiprocessing connection recv which hits an eof error when unpickling an object environment s seen with issue ah prod and ah qa environment s seen without issue running importer in cli mode running ansible test sanity t pylint docker locally traceback from ah qa import running sanity test with python running sanity test pylint with python error command usr bin m pylint jobs reports n max line length rcfile usr local lib dist packages ansible test data sanity pylint config collection cfg output format json load plugins blacklist deprecated string format tests unit plugins module utils test ibm svc utils py tests unit plugins modules test ibm svc host py tests unit plugins modules test ibm svc info py tests unit plugins modules test ibm svc mdisk py tests unit plugins modules test ibm svc mdiskgrp py tests unit plugins modules test ibm svc vdisk py tests unit plugins modules test ibm svc vol map py returned exit status standard error process childlinter traceback most recent call last file usr lib multiprocessing process py line in bootstrap self run file usr local lib dist packages pylint lint py line in run for file or module in iter tasks queue get stop file line in get file usr lib multiprocessing managers py line in callmethod kind result conn recv file usr lib multiprocessing connection py line in recv buf self recv bytes file usr lib multiprocessing connection py line in recv bytes buf self recv file usr lib multiprocessing connection py line in recv raise eoferror eoferror process childlinter traceback most recent call last file usr lib multiprocessing process py line in bootstrap self run file usr local lib dist packages pylint lint py line in run for file or module in iter tasks queue get stop file line in get file usr lib multiprocessing managers py line in callmethod kind result conn recv file usr lib multiprocessing connection py line in recv buf self recv bytes file usr lib multiprocessing connection py line in recv bytes buf self recv file usr lib multiprocessing connection py line in recv raise eoferror eoferror internal error with sending report for module broken pipe process childlinter traceback most recent call last file usr local lib dist packages pylint lint py line in run results queue put result file line in put file usr lib multiprocessing managers py line in callmethod conn send self id methodname args kwds file usr lib multiprocessing connection py line in send self send bytes forkingpickler dumps obj file usr lib multiprocessing connection py line in send bytes self send header buf file usr lib multiprocessing connection py line in send n write self handle buf brokenpipeerror broken pipe during handling of the above exception another exception occurred traceback most recent call last file usr lib multiprocessing process py line in bootstrap self run file usr local lib dist packages pylint lint py line in run
1
781,200
27,427,269,791
IssuesEvent
2023-03-01 21:22:15
dtcenter/MET
https://api.github.com/repos/dtcenter/MET
closed
Upgrade to using Python 3.10.4
requestor: NOAA/EMC type: task alert: NEED ACCOUNT KEY component: external dependency requestor: METplus Team required: FOR OFFICIAL RELEASE priority: high
## Describe the Task ## Update any code necessary for upgrading to Python 3.10.4. We are taking this action because WCOSS2 is only allowing Python 3.10.4 to be used for operational installations. Therefore, we need to ensure that our next release uses Python 3.10.4. **Please also update .readthedocs.yaml with 3.10.4** Note that we decided NOT to upgrade to 3.10.4 since RTD supports up to 3.8. ### Time Estimate ### *Estimate the amount of work required here.* *Issues should represent approximately 1 to 3 days of work.* **(Assignee please fill this in once you know.)** ### Sub-Issues ### Consider breaking the task down into sub-issues. - [ ] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### Incorporate in a beta version as soon as possible to ensure there are no problems. ### Funding Source ### Please ask Tara before starting on this work. ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [x] Select **component(s)** - [x] Select **priority** - [x] Select **requestor(s)** ### Projects and Milestone ### - [x] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label - [x] Select **Milestone** as the next official version or **Future Versions** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Task Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Add any new Python packages to the [METplus Components Python Requirements](https://metplus.readthedocs.io/en/develop/Users_Guide/overview.html#metplus-components-python-requirements) table. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Development** issues Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
1.0
Upgrade to using Python 3.10.4 - ## Describe the Task ## Update any code necessary for upgrading to Python 3.10.4. We are taking this action because WCOSS2 is only allowing Python 3.10.4 to be used for operational installations. Therefore, we need to ensure that our next release uses Python 3.10.4. **Please also update .readthedocs.yaml with 3.10.4** Note that we decided NOT to upgrade to 3.10.4 since RTD supports up to 3.8. ### Time Estimate ### *Estimate the amount of work required here.* *Issues should represent approximately 1 to 3 days of work.* **(Assignee please fill this in once you know.)** ### Sub-Issues ### Consider breaking the task down into sub-issues. - [ ] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### Incorporate in a beta version as soon as possible to ensure there are no problems. ### Funding Source ### Please ask Tara before starting on this work. ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [x] Select **component(s)** - [x] Select **priority** - [x] Select **requestor(s)** ### Projects and Milestone ### - [x] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label - [x] Select **Milestone** as the next official version or **Future Versions** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Task Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Add any new Python packages to the [METplus Components Python Requirements](https://metplus.readthedocs.io/en/develop/Users_Guide/overview.html#metplus-components-python-requirements) table. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Development** issues Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
priority
upgrade to using python describe the task update any code necessary for upgrading to python we are taking this action because is only allowing python to be used for operational installations therefore we need to ensure that our next release uses python please also update readthedocs yaml with note that we decided not to upgrade to since rtd supports up to time estimate estimate the amount of work required here issues should represent approximately to days of work assignee please fill this in once you know sub issues consider breaking the task down into sub issues add a checkbox for each sub issue here relevant deadlines incorporate in a beta version as soon as possible to ensure there are no problems funding source please ask tara before starting on this work define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select repository and or organization level project s or add alert need project assignment label select milestone as the next official version or future versions define related issue s consider the impact to the other metplus components task checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation add any new python packages to the table push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s and development issues select repository level development cycle project for the next official release select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue
1
252,161
8,032,682,869
IssuesEvent
2018-07-28 18:11:37
cilium/cilium
https://api.github.com/repos/cilium/cilium
closed
Increase API client timeout
kind/bug priority/high
As we are increasing the build timeout in #4995. The API client timeout of 30 seconds as implemented by go-swagger has to be increased as well.
1.0
Increase API client timeout - As we are increasing the build timeout in #4995. The API client timeout of 30 seconds as implemented by go-swagger has to be increased as well.
priority
increase api client timeout as we are increasing the build timeout in the api client timeout of seconds as implemented by go swagger has to be increased as well
1
618,036
19,413,120,786
IssuesEvent
2021-12-20 11:49:42
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
closed
MSUnmerged: expose basic cleanup stats via REST API
New Feature High Priority MSUnmerged
**Impact of the new feature** MSUnmerged **Is your feature request related to a problem? Please describe.** It concerns a new microservice to carry out further cleanup of the sites' unmerged area (removal of files that are no longer needed/locked in the system). **Describe the solution you'd like** Since we are now having the remote deletions through `gfal` implemented, we would like to have some statistics exposed per RSE about the cleaning process. We are currently bound to the pooling cycle of the RucioConmon service, which is once per week, so the information and results of the consecutive reruns of our MSUnmerged service would be accumulated through out this period in the background database (MongoDB) and displayed through appropriate REST APIs developed to the service. In order to retrieve data for a given RSE, we need to call the `info` REST API and provide the RSE name via `rse=RSE_name` query string. If `detail=true` is provided, then it also returns directory related information (what is protected, what is deleted, etc). Detailed information is disabled by default. **Describe alternatives you've considered** Create the desired Methods and APIs to translate the supported REST calls to the correct MongoDB queries and display the information through the `info` and `status` interfaces to the service. **Additional context** Depends on https://github.com/dmwm/WMCore/issues/10411
1.0
MSUnmerged: expose basic cleanup stats via REST API - **Impact of the new feature** MSUnmerged **Is your feature request related to a problem? Please describe.** It concerns a new microservice to carry out further cleanup of the sites' unmerged area (removal of files that are no longer needed/locked in the system). **Describe the solution you'd like** Since we are now having the remote deletions through `gfal` implemented, we would like to have some statistics exposed per RSE about the cleaning process. We are currently bound to the pooling cycle of the RucioConmon service, which is once per week, so the information and results of the consecutive reruns of our MSUnmerged service would be accumulated through out this period in the background database (MongoDB) and displayed through appropriate REST APIs developed to the service. In order to retrieve data for a given RSE, we need to call the `info` REST API and provide the RSE name via `rse=RSE_name` query string. If `detail=true` is provided, then it also returns directory related information (what is protected, what is deleted, etc). Detailed information is disabled by default. **Describe alternatives you've considered** Create the desired Methods and APIs to translate the supported REST calls to the correct MongoDB queries and display the information through the `info` and `status` interfaces to the service. **Additional context** Depends on https://github.com/dmwm/WMCore/issues/10411
priority
msunmerged expose basic cleanup stats via rest api impact of the new feature msunmerged is your feature request related to a problem please describe it concerns a new microservice to carry out further cleanup of the sites unmerged area removal of files that are no longer needed locked in the system describe the solution you d like since we are now having the remote deletions through gfal implemented we would like to have some statistics exposed per rse about the cleaning process we are currently bound to the pooling cycle of the rucioconmon service which is once per week so the information and results of the consecutive reruns of our msunmerged service would be accumulated through out this period in the background database mongodb and displayed through appropriate rest apis developed to the service in order to retrieve data for a given rse we need to call the info rest api and provide the rse name via rse rse name query string if detail true is provided then it also returns directory related information what is protected what is deleted etc detailed information is disabled by default describe alternatives you ve considered create the desired methods and apis to translate the supported rest calls to the correct mongodb queries and display the information through the info and status interfaces to the service additional context depends on
1