id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
206609564
Test broken: predspec The predspec test fails to build. Please respond to this issue if you have any interest in keeping this test. I am trying to determine who has an interest in maintaining each test, and if no one is interested, I will delete the test. Build error: testmapper.cc:36:15: error: no matching function for call to ‘Legion::Mapping::DefaultMapper::DefaultMapper(LegionRuntime::HighLevel::Machine&, LegionRuntime::HighLevel::HighLevelRuntime*&, LegionRuntime::HighLevel::Processor&)’ , runtime(rt) ^ testmapper.cc:36:15: note: candidates are: In file included from testmapper.h:2:0, from testmapper.cc:1: /home/eslaught/legion/runtime/mappers/default_mapper.h:113:7: note: Legion::Mapping::DefaultMapper::DefaultMapper(const Legion::Mapping::DefaultMapper&) DefaultMapper(const DefaultMapper &rhs); ^ /home/eslaught/legion/runtime/mappers/default_mapper.h:113:7: note: candidate expects 1 argument, 3 provided /home/eslaught/legion/runtime/mappers/default_mapper.h:111:7: note: Legion::Mapping::DefaultMapper::DefaultMapper(Legion::Mapping::MapperRuntime*, Legion::Machine, Legion::Processor, const char*) DefaultMapper(MapperRuntime *rt, Machine machine, Processor local, ^ /home/eslaught/legion/runtime/mappers/default_mapper.h:111:7: note: no known conversion for argument 1 from ‘LegionRuntime::HighLevel::Machine {aka Realm::Machine}’ to ‘Legion::Mapping::MapperRuntime*’ testmapper.cc: In member function ‘virtual void TestMapper::select_task_options(LegionRuntime::HighLevel::Task*)’: testmapper.cc:109:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘inline_task’ task->inline_task = false; ^ testmapper.cc:110:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘spawn_task’ task->spawn_task = false; ^ testmapper.cc:111:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘map_locally’ task->map_locally = true; ^ testmapper.cc:112:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘profile_task’ task->profile_task = false; ^ testmapper.cc:113:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘task_priority’ task->task_priority = 0; ^ testmapper.cc:115:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘additional_procs’ task->additional_procs.insert(procs[node].begin(), procs[node].end()); ^ testmapper.cc: In member function ‘virtual void TestMapper::notify_mapping_result(const Mappable*)’: testmapper.cc:140:32: error: ‘const Mappable’ has no member named ‘as_mappable_task’ const Task *task = mappable->as_mappable_task(); ^ testmapper.cc:147:49: error: ‘const struct Legion::RegionRequirement’ has no member named ‘selected_memory’ << ") mapped on " << task->regions[idx].selected_memory ^ @streichler Have you decided whether it would be beneficial to keep this test? This test has been removed as of d0fb5b8d04cc6876612db969b67f7b4e66fef711. If you are interested in resurrecting this test, please reply to this issue.
gharchive/issue
2017-02-09T19:44:23
2025-04-01T04:33:06.245762
{ "authors": [ "elliottslaughter" ], "repo": "StanfordLegion/legion", "url": "https://github.com/StanfordLegion/legion/issues/223", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1172421012
Could you please update your tutorial for ros integration? the demo is using ros1, turtlebot1 and python 2.7, which are all out of date. By using miniconda env based on python2.7, you even can not properly install igibson2!! iGibson in its current form is not compatible with Python 2.7. We are looking into re-enabling backwards compatibility, but are you able to switch to a Python 3 version of ROS? Would updating the tutorial to work with that be helpful for you? that would be of great favor for me. @cgokmen and @MRWANG995 We have integrated ROS noetic (Python 3) with igibson 2. We are able to run the robot with move_base ROS navigation stack (Turtlebot and Fetch). Let us know if you need any help. We will happy to share our learnings with you :) . It may also help you to update the tutorial. Let us know :) Hi @vbalaji21 , please, try the new version we just released and tell us whether it presents any problem with you. Thanks! The release has not gone through yet - we'll push today. Will update here. @roberto-martinmartin and @cgokmen . Thanks a lot for letting me know. Please update me once it is pushed, I will try it out. This week, I had some problems with loading Pedestrians. I made some fixes and it is working now. I am not sure whether that is also taken into account in the new release. If not, I can share you the fixes I did. My aim (along with Prof. Rachid and Phanu Teja in LAAS, CNRS) is to study human aware social navigation and later extend with learning capabilities. My next 2 steps are - Check VR immersion of human and arm (head data) to conduct user studies with our planner Import a UAV into iGibson We are considering iGibson as our potential future platform for most of our development. We are hoping to do the above tasks as soon as possible and continue our development. We would be very happy to get any information or help on these 2 tasks. We definitely have well-maintained and functional VR capability right now (this is the feature we ourselves have used the most). You can try the VR demo in https://github.com/StanfordVL/iGibson/blob/master/igibson/examples/vr/vr_simple_demo.py using a SteamVR-installed computer (this demo was tested using a HTC Vive headset and Valve Index controller but I believe we support most devices). As for the UAV we also have https://github.com/StanfordVL/iGibson/blob/master/igibson/robots/legacy/quadrotor_robot.py but this is a not-so-realistic quadrotor that used to be functional but is currently not maintained (e.g. it will probably take some code editing to make it work with the new Robot interface). Should be simple enough - the above file should end up looking like this one: https://github.com/StanfordVL/iGibson/blob/master/igibson/robots/ant.py Oh, also https://github.com/utiasDSL/gym-pybullet-drones has some interesting pybullet-based UAV simulation stuff (dynamics, control) that perhaps you could try using on top of iGibson (most of the code should be compatible and not too hard to hook up) We have just cut the release for iGibson 2.2.0 - please let us know if you need anything else, and feel free to reopen this thread if you have any further questions. Thanks a lot for the update and suggestions regarding UAV and VR. I have 2 questions - How to make any robot visible in REGISTERED_ROBOTS? I haave attached a screenshot of this error. Please let me know. I tried importing UAV hector_quadcopter https://github.com/tu-darmstadt-ros-pkg/hector_quadrotor yesterday from here. It is on ROS kinetic but tried porting it for ROS noetic. I found that the controller classes for UAV needs to be ported. I did not find controller classes for UAV in iGibson 1.0 as well. I saw that you are using locomotion robot class and basecontroller for UAV if I am correct. Does your comment on the realistic/functionality of Quadrotor mean that it doesn't work properly in iGibson 1.0 or the controller is primitive to make the Quadrotor work realistically ? Thanks a lot for your comment about the Quadrorot/UAV suggestion. It will save a lot of time for us in the decision :) Hi @vbalaji21 , please, try the new version we just released and tell us whether it presents any problem with you. Thanks! @roberto-martinmartin and @cgokmen , I tried the new version by installing iGibson in the compiling from source method but it failed to build in pip install -e . com Let me know mand. I have attached the error log screenshot below. could it be that you didn't git clone recursively? @roberto-martinmartin Thank you the problem was indeed I didn't do git clone recursively. I have another doubt, I followed the instructions but I did not understand about the Python Path step as it has 4 paths and I don't know in which file I need to update them. Can you please let me know ? I think I am facing this problem with the path (Python path). I have attached the screenshot below.
gharchive/issue
2022-03-17T13:56:03
2025-04-01T04:33:06.259664
{ "authors": [ "MRWANG995", "cgokmen", "roberto-martinmartin", "vbalaji21" ], "repo": "StanfordVL/iGibson", "url": "https://github.com/StanfordVL/iGibson/issues/179", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
991667150
TilePos and relatives should use usize, not u32 This is an important perf and usability optimization when interfacing with external arrays (which are always indexed by usize slices). Currently, you must repeatedly waste operations performing this conversion and littering the code base with as usize. FYI there is no perf cost; integer type conversions compile to no-ops, there are no extra operations. There is no such thing as integer type conversion at the assembly / machine code level. This is just to make Rust code prettier and more correct. I think this is no longer relevant, but feel free to reopen if I am wrong.
gharchive/issue
2021-09-09T01:36:36
2025-04-01T04:33:06.261763
{ "authors": [ "StarArawn", "alice-i-cecile", "inodentry" ], "repo": "StarArawn/bevy_ecs_tilemap", "url": "https://github.com/StarArawn/bevy_ecs_tilemap/issues/95", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1591557504
[balancer] show tablet info based on be path Feature request Is your feature request related to a problem? Please describe. Describe the solution you'd like Describe alternatives you've considered Additional context fixed by #18457 will displayed all these info in information_schema db
gharchive/issue
2023-02-20T10:20:23
2025-04-01T04:33:06.280985
{ "authors": [ "nshangyiming" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/issues/18114", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1741225014
需求:新增对已经存在的表修改comment的功能 请问什么时候有计划增加对已经存在的表修改comment的功能,类似mysql:ALTER TABLE 表名称 COMMENT '***'; Already supported https://github.com/StarRocks/starrocks/pull/21035
gharchive/issue
2023-06-05T08:11:49
2025-04-01T04:33:06.282208
{ "authors": [ "Astralidea", "qy58644781" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/issues/24685", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1057271757
Support CTAS(create table as select) statement CTAS is used to simplify user to create table and insert data. The syntax is like the following. CREATE TABLE [IF NOT EXISTS] [database.]table_name [(column_name[, column_name, ...]] [COMMENT "table comment"] [partition_desc] [distribution_desc] [PROPERTIES ("key"="value", ...)] AS SELECT <query> [ ... ] The tips for the CTAS statement. The default model is a duplicate-key table. All char/varchar type will create as string in the new table. The replication_num of table will be equal the the maximum number of all tables selected. When choosing hash distributed key, If we have statistics cache, we pick the column with the highest cardinality according the column statistics. Otherwise, pick the first column. CTAS only support new CBO based planner and implement some createTable analyzer. run all run all
gharchive/pull-request
2021-11-18T11:57:58
2025-04-01T04:33:06.284689
{ "authors": [ "Astralidea" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/1392", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1617344913
[BugFix] Release workgroup token immediately when fragment is cancelled What type of PR is this: [x] BugFix [ ] Feature [ ] Enhancement [ ] Refactor [ ] UT [ ] Doc [ ] Tool Which issues of this PR fixes : Fixes # TODO. Problem Summary(Required) : Release workgroup token immediately when fragment is cancelled. Otherwise, it can only be released after query_delivery_timeout (5min by default) sometimes. Handle error results and cancel fragment instances, excluding TIMEOUT errors, until all the delivered fragment instances are completed. Otherwise, the cancellation RPC may arrive at BE before the delivery fragment instance RPC, causing the instances and drivers to become stale and only able to be released after query timeout. Checklist: [x] I have added test cases for my bug fix or my new feature [ ] This pr will affect users' behaviors [ ] This pr needs user documentation (for new or modified features or behaviors) [ ] I have added documentation for my new feature or new function Bugfix cherry-pick branch check: [x] I have checked the version labels which the pr will be auto backported to target branch [x] 3.0 [x] 2.5 [x] 2.4 [x] 2.3 run starrocks_be_unittest run starrocks_be_unittest run starrocks_be_unittest run starrocks_admit_test run starrocks_fe_unittest run starrocks_fe_unittest run starrocks_fe_unittest @Mergifyio backport branch-2.5 @Mergifyio backport branch-2.4 @Mergifyio backport branch-2.3
gharchive/pull-request
2023-03-09T14:16:53
2025-04-01T04:33:06.292898
{ "authors": [ "ZiheLiu", "wanpengfei-git" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/19310", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1674777115
[BugFix] fix query stuck issue in SpillableAggregateBlockingSinkOperator What type of PR is this: [x] BugFix [ ] Feature [ ] Enhancement [ ] Refactor [ ] UT [ ] Doc [ ] Tool Which issues of this PR fixes : Fixes # Problem Summary(Required) : Checklist: [ ] I have added test cases for my bug fix or my new feature [ ] This pr will affect users' behaviors [ ] This pr needs user documentation (for new or modified features or behaviors) [ ] I have added documentation for my new feature or new function Bugfix cherry-pick branch check: [x] I have checked the version labels which the pr will be auto backported to target branch [x] 3.0 [ ] 2.5 [ ] 2.4 [ ] 2.3 @Mergifyio backport branch-3.0
gharchive/pull-request
2023-04-19T12:11:23
2025-04-01T04:33:06.297832
{ "authors": [ "silverbullet233", "wanpengfei-git" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/21943", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1745118062
[BugFix] Fix bug drop not exists task NPE Fixes #24741 What type of PR is this: [x] BugFix [ ] Feature [ ] Enhancement [ ] Refactor [ ] UT [ ] Doc [ ] Tool Checklist: [x] I have added test cases for my bug fix or my new feature [ ] This pr will affect users' behaviors [ ] This pr needs user documentation (for new or modified features or behaviors) [ ] I have added documentation for my new feature or new function Bugfix cherry-pick branch check: [x] I have checked the version labels which the pr will be auto-backported to the target branch [x] 3.0 [x] 2.5 [ ] 2.4 [ ] 2.3 [FE PR Coverage Check] :disappointed: fail : 0 / 4 (00.00%) file detail path covered_line new_line coverage not_covered_line_detail :large_blue_circle: com/starrocks/qe/DDLStmtExecutor.java 0 4 00.00% [877, 878, 879, 881] @Mergifyio backport branch-3.0 @Mergifyio backport branch-2.5
gharchive/pull-request
2023-06-07T06:15:23
2025-04-01T04:33:06.304277
{ "authors": [ "Astralidea", "wanpengfei-git" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/24793", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1800541975
[Enhancement] generic range Fixes #issue Generic range so that it can be used for external file reader, such as parquet reader which row_index is uint64 What type of PR is this: [ ] BugFix [ ] Feature [x] Enhancement [ ] Refactor [ ] UT [ ] Doc [ ] Tool Checklist: [ ] I have added test cases for my bug fix or my new feature [ ] This pr will affect users' behaviors [ ] This pr needs user documentation (for new or modified features or behaviors) [ ] I have added documentation for my new feature or new function Bugfix cherry-pick branch check: [x] I have checked the version labels which the pr will be auto-backported to the target branch [ ] 3.1 [ ] 3.0 [ ] 2.5 [ ] 2.4 [FE PR Coverage Check] :heart_eyes: pass : 0 / 0 (0%) [FE PR Coverage Check] :heart_eyes: pass : 0 / 0 (0%) [FE PR Coverage Check] :heart_eyes: pass : 0 / 0 (0%)
gharchive/pull-request
2023-07-12T09:16:56
2025-04-01T04:33:06.310143
{ "authors": [ "wanpengfei-git", "zombee0" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/27056", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1823641658
[BugFix] primary key table inconsistent snapshot for full clone and backup caused by compaction(#28062) Problem: When we make snapshot for full clone and backup, _lock will not be holded during get_applied_rowsets and make_snapshot_on_tablet_meta. This cause a problem, If compaction happen between get_applied_rowsets and make_snapshot_on_tablet_meta, the delvector get from make_snapshot_on_tablet_meta maybe generated by a larger minor version which is inconsistent with the rowset get from get_applied_rowsets. Solution: hold _lock during get_applied_rowsets and make_snapshot_on_tablet_meta Fixes #issue What type of PR is this: [x] BugFix [ ] Feature [ ] Enhancement [ ] Refactor [ ] UT [ ] Doc [ ] Tool Checklist: [x] I have added test cases for my bug fix or my new feature [ ] This pr will affect users' behaviors [ ] This pr needs user documentation (for new or modified features or behaviors) [ ] I have added documentation for my new feature or new function Bugfix cherry-pick branch check: [x] I have checked the version labels which the pr will be auto-backported to the target branch [x] 3.1 [x] 3.0 [x] 2.5 [ ] 2.4 [FE PR Coverage Check] :heart_eyes: pass : 0 / 0 (0%) [FE PR Coverage Check] :heart_eyes: pass : 0 / 0 (0%) [FE PR Coverage Check] :heart_eyes: pass : 0 / 0 (0%) [FE PR Coverage Check] :heart_eyes: pass : 0 / 0 (0%) @Mergifyio backport branch-3.1 @Mergifyio backport branch-3.0 @Mergifyio backport branch-2.5 [FE PR Coverage Check] :heart_eyes: pass : 0 / 0 (0%)
gharchive/pull-request
2023-07-27T05:29:59
2025-04-01T04:33:06.318663
{ "authors": [ "srlch", "wanpengfei-git" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/28064", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1110438393
Put chunks of different driver_sequences in one request Background When enable pipeline level shuffle, chunks will hash to N groups (N is the dop of dest pipeline). Each group will build the request independently. When sending large amounts of data, the number of requests sent will not change significantly because each group will accumulate chunk until it exceeds a threshold before sending When sending small amounts of data, take the extreme case for example, assume the amount of data just fills one chunk, then each group will send one packet, the packets to be sent changed from 1 to N which may lead to significant performance drop Solution Add an auxiliary data structure PTransmitChunkParams::driver_sequences, indicating the driver_sequence to which the i-th chunk belongs Test 3be(64c/128G) 1fe(64c/128G) tpcds with scale factor=100 Almost all the querys have performance improvements, and a few are briefly listed as below. TPCDS Before After Q12 0.299s 0.19s Q15 0.378s 0.25s Q20 0.25s 0.16s Q45 0.442s 0.305s run starrocks_be_unittest create an issue and link it to this pr
gharchive/pull-request
2022-01-21T12:53:58
2025-04-01T04:33:06.324018
{ "authors": [ "liuyehcf", "satanson" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/3003", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1881963504
[BugFix] Fix source_scan_bytes compability with 3.0 Fixes #issue What type of PR is this: [x] BugFix [ ] Feature [ ] Enhancement [ ] Refactor [ ] UT [ ] Doc [ ] Tool Does this PR entail a change in behavior? [ ] Yes, this PR will result in a change in behavior. [x] No, this PR will not result in a change in behavior. If yes, please specify the type of change: [ ] Interface/UI changes: syntax, type conversion, expression evaluation, display information [ ] Parameter changes: default values, similar parameters but with different default values [ ] Policy changes: use new policy to replace old one, functionality automatically enabled [ ] Feature removed [ ] Miscellaneous: upgrade & downgrade compatibility, etc. Checklist: [ ] I have added test cases for my bug fix or my new feature [ ] This pr needs user documentation (for new or modified features or behaviors) [ ] I have added documentation for my new feature or new function Bugfix cherry-pick branch check: [x] I have checked the version labels which the pr will be auto-backported to the target branch [x] 3.1 [ ] 3.0 [ ] 2.5 [ ] 2.4 [BE Incremental Coverage Report] :heart_eyes: pass : 0 / 0 (0%) @Mergifyio backport branch-3.1
gharchive/pull-request
2023-09-05T13:16:17
2025-04-01T04:33:06.330366
{ "authors": [ "meegoo", "wanpengfei-git" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/30433", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1148043329
Fix Scalar Functions bugs (#3631) What type of PR is this: [ ] bug [ ] feature [ ] enhancement [ ] others Which issues of this PR fixes : Fixes # Problem Summary(Required) : run starrocks_fe_unittest
gharchive/pull-request
2022-02-23T12:41:35
2025-04-01T04:33:06.332446
{ "authors": [ "Seaven" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/3645", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2263042926
[BugFix] [UT] Distinguish TaskRun by using unique taskRunId Why I'm doing: submitTaskRegularTest is not stable mv sync mode refresh may hang until timeout What I'm doing: TaskRun's equals method only checks taskId and its definition but one task may contains many task runs. It may introduce bug if we don't distinguish those task runs. Use taskRunId to distinguish TaskRun @Override public boolean equals(Object o) { if (this == o) { return true; } if (o == null || getClass() != o.getClass()) { return false; } if (task.getDefinition() == null) { return false; } TaskRun taskRun = (TaskRun) o; return this.taskId == taskRun.getTaskId() && this.task.getDefinition().equals(taskRun.getTask().getDefinition()); } Fixes #issue What type of PR is this: [x] BugFix [ ] Feature [ ] Enhancement [ ] Refactor [x] UT [ ] Doc [ ] Tool Does this PR entail a change in behavior? [ ] Yes, this PR will result in a change in behavior. [x] No, this PR will not result in a change in behavior. If yes, please specify the type of change: [ ] Interface/UI changes: syntax, type conversion, expression evaluation, display information [ ] Parameter changes: default values, similar parameters but with different default values [ ] Policy changes: use new policy to replace old one, functionality automatically enabled [ ] Feature removed [ ] Miscellaneous: upgrade & downgrade compatibility, etc. Checklist: [x] I have added test cases for my bug fix or my new feature [ ] This pr needs user documentation (for new or modified features or behaviors) [ ] I have added documentation for my new feature or new function [ ] This is a backport pr Bugfix cherry-pick branch check: [x] I have checked the version labels which the pr will be auto-backported to the target branch [x] 3.3 [x] 3.2 [ ] 3.1 [ ] 3.0 [ ] 2.5 @mergify rebase
gharchive/pull-request
2024-04-25T08:50:52
2025-04-01T04:33:06.339798
{ "authors": [ "LiShuMing" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/44748", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2400604034
[BugFix] Fix possible dead lock in LoadJobStatsListener Why I'm doing: What I'm doing: Fix protential dead lock bug: try exception to release write lock finally in [InsertOverwriteJobRunner.java](https://github.com/StarRocks/starrocks/pull/48150/files#diff-9ba512541351f603a9bce0ba274658f639eafebe2fbf635d7ac26072f6ff4c81) No use lock in `LoadJobMVListener.java https://github.com/StarRocks/StarRocksTest/issues/8091 What type of PR is this: [x] BugFix [ ] Feature [ ] Enhancement [ ] Refactor [ ] UT [ ] Doc [ ] Tool Does this PR entail a change in behavior? [ ] Yes, this PR will result in a change in behavior. [x] No, this PR will not result in a change in behavior. If yes, please specify the type of change: [ ] Interface/UI changes: syntax, type conversion, expression evaluation, display information [ ] Parameter changes: default values, similar parameters but with different default values [ ] Policy changes: use new policy to replace old one, functionality automatically enabled [ ] Feature removed [ ] Miscellaneous: upgrade & downgrade compatibility, etc. Checklist: [x] I have added test cases for my bug fix or my new feature [ ] This pr needs user documentation (for new or modified features or behaviors) [ ] I have added documentation for my new feature or new function [ ] This is a backport pr Bugfix cherry-pick branch check: [x] I have checked the version labels which the pr will be auto-backported to the target branch [x] 3.3 [ ] 3.2 [ ] 3.1 [ ] 3.0 [ ] 2.5 ignore backport check 3.3.1
gharchive/pull-request
2024-07-10T12:16:25
2025-04-01T04:33:06.347609
{ "authors": [ "LiShuMing" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/48150", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1211849838
[Enhancement] Check user identity for show grants & support string literal role name for create role and drop role What type of PR is this: [x] bug [ ] feature [ ] enhancement [ ] others Which issues of this PR fixes : Fixes #5389 Problem Summary(Required) : So far, using show grants for a non-existent user will show a table with all cells 'null'. Fix it by raising an exception in this scenario. Allow usage of create role 'role_name' by changing role CUP type from ident to ident_or_text. Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it. [FE PR Coverage check] :heart_eyes: pass : 2 / 2 (100.00%) file detail path covered line new line coverage :large_blue_circle: com/starrocks/analysis/ShowGrantsStmt.java 2 2 100.00%
gharchive/pull-request
2022-04-22T05:44:24
2025-04-01T04:33:06.353862
{ "authors": [ "CLAassistant", "padmejin", "wanpengfei-git" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/5390", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1261821622
[cherry-pick] Get hive column statistics downgrade policy (#4409) (#6… …899) What type of PR is this: [ ] bug [ ] feature [ ] enhancement [ ] refactor [x] others Which issues of this PR fixes : Fixes # Problem Summary(Required) : run starrocks_fe_unittest run starrocks_fe_unittest
gharchive/pull-request
2022-06-06T13:06:27
2025-04-01T04:33:06.356517
{ "authors": [ "stephen-shelby", "wangruin" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/6926", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1274416996
[Bugfix] Avoid using released resources for ContextWithDependency after closing (backport #7363) What type of PR is this: [ ] bug [ ] feature [ ] enhancement [ ] refactor [ ] others Problem Summary(Required) : This is cherry-picked from #7363. run all run all run all
gharchive/pull-request
2022-06-17T02:28:27
2025-04-01T04:33:06.358776
{ "authors": [ "ZiheLiu", "mofeiatwork" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/7394", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2279226497
Use Localization for equipment on actors It's currently hardcoded. https://github.com/StarWarsFoundryVTT/StarWarsFFG/blob/dde154ea820d7c31d147b1402f38504222b60853/templates/parts/actor/ffg-weapon-armor-gear.html#L30 Translations already exist. SWFFG.Equipped SWFFG.Unequipped
gharchive/issue
2024-05-04T22:45:20
2025-04-01T04:33:06.360259
{ "authors": [ "TeddyBears", "wrycu" ], "repo": "StarWarsFoundryVTT/StarWarsFFG", "url": "https://github.com/StarWarsFoundryVTT/StarWarsFFG/issues/1469", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1859296128
Add MSRV to readme I think that good to put MSRV in the README. Fix somehow https://github.com/matklad/once_cell/issues/201 no idea how I end up doing the PR on the fork haha github betray me
gharchive/pull-request
2023-08-21T12:44:35
2025-04-01T04:33:06.381872
{ "authors": [ "Stargateur" ], "repo": "Stargateur/once_cell", "url": "https://github.com/Stargateur/once_cell/pull/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2140153899
Use transaction_connection_validator from Sequel Sounds like a good thing: https://github.com/jeremyevans/sequel/commit/c31dde28e79f5d67398c2ab849cedc84e45dd386 Needs https://github.com/Starkast/wikimum/pull/579
gharchive/issue
2024-02-17T14:44:12
2025-04-01T04:33:06.383307
{ "authors": [ "dentarg" ], "repo": "Starkast/wikimum", "url": "https://github.com/Starkast/wikimum/issues/589", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1104255038
Progress not properly displaying for explorer synced health check The data is coming back correctly but for some reason it's showing as 0.00%. {"bitcoinStatus":{"blocks":718729,"headers":718729,"verificationProgress":0.9999970594948908,"isSynched":true,"incrementalRelayFee":1.0,"minRelayTxFee":1.0,"externalAddresses":["<redacted>.onion:8333"],"capabilities":{"canScanTxoutSet":true,"canSupportSegwit":true,"canSupportTransactionCheck":true}},"repositoryPingTime":0.0,"isFullySynched":false,"chainHeight":718729,"syncHeight":715185,"instanceName":"","networkType":"Mainnet","cryptoCode":"BTC","supportedCryptoCodes":["BTC"],"version":"2.1.52.0"} Fixed in: https://github.com/Start9Labs/btcpayserver-wrapper/pull/9/commits/d6c0775848e9fcf151d85fe4c0c9fd0cc7f65c33
gharchive/issue
2022-01-15T01:15:59
2025-04-01T04:33:06.395176
{ "authors": [ "elvece" ], "repo": "Start9Labs/btcpayserver-wrapper", "url": "https://github.com/Start9Labs/btcpayserver-wrapper/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1263252264
[feat]: 🛑 Stop using root account for EmbassyOS Prerequisites [X] I have searched for existing issues that already suggest this feature, without success. Describe the Feature Request With Great Power Comes Great Responsibility. We should transition away from using root in EmbassyOS and instead, use the start9 user as a privileged account thus limiting the potential system damage. Describe the Use Case it is inevitable 😈 Describe Preferred Solution No response Describe Alternatives No response Anything else? Good security practices: Root account should be used to modify the underlying system. Using the system should be done by the user account. User can elevate his rights to become root. As an operating system, embassyd should be running as root. It pretty much has no other choice. We are already transitioning away from embassy-cli and embassy-sdk requiring root though. We are already transitioning away from embassy-cli and embassy-sdk requiring root though. That is good news. Filebrowser would be the natural next contender. Files should belong to start9 user so that you could mount filebrowser data directory via sftp and move those TB of images, movies and documents using systems file manager or simple SFTP client software. Good point, it probably should be the case that all files are owned by the start9 user and thats the user that the containers run as as well closed by #1227 This is complete, no?
gharchive/issue
2022-06-07T12:56:00
2025-04-01T04:33:06.400468
{ "authors": [ "ProofOfKeags", "dr-bonez", "k0gen", "kn0wmad" ], "repo": "Start9Labs/embassy-os", "url": "https://github.com/Start9Labs/embassy-os/issues/1503", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
857548807
Compatibility with Font Awesome Pro If I import FA pro the arrows on the left menu are replaced with the text "F105" (when pointing left/collapsed) and "F107" (when pointing down/expanded). If I bring in the free version of FA and remove the pro I get my arrows back but then I lose the pro icons. Is there a quick way around this that I'm missing? Are you using the same version of Font Awesome? The classes used in FA4 are not the same as FA5. The pro version also has a different font family than the free version in FA5, so you can't just swap the font files - you have to replace the CSS file as well. Font Awesome Pro uses a different font family, so if you want to use that, you'll need to edit either the Sass file or the compiled CSS, depending on which you're using. For the Sass version, the Font Awesome bits for the sidebar start here: https://github.com/StartBootstrap/startbootstrap-sb-admin-2/blob/master/scss/navs/_sidebar.scss#L104 For the CSS version, you'd start here: https://github.com/StartBootstrap/startbootstrap-sb-admin-2/blob/master/css/sb-admin-2.css#L10096 I imagine a find and replace would work for either. Looks like that did it. Thank you so much!
gharchive/issue
2021-04-14T05:57:54
2025-04-01T04:33:06.403539
{ "authors": [ "sean8300", "tiesont" ], "repo": "StartBootstrap/startbootstrap-sb-admin-2", "url": "https://github.com/StartBootstrap/startbootstrap-sb-admin-2/issues/391", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
803807727
VS Code config issue https://github.com/StatCan/StatCan.OrchardCore/blob/master/.vscode/launch.json#L40 This should never be set to Production while you are developping with VS Code. If you are using Azure Shells settings this could potentially trigger update of the prod database. @jptissot I think you should update these files with the latest ones from the main OC repo. I fixed them. Sure ! Thanks :)
gharchive/issue
2021-02-08T18:16:37
2025-04-01T04:33:06.406642
{ "authors": [ "Skrypt", "jptissot" ], "repo": "StatCan/StatCan.OrchardCore", "url": "https://github.com/StatCan/StatCan.OrchardCore/issues/188", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2247482575
Welcome to Charlie Welcome to our project! As we're just getting started, your input is crucial. Feel free to post your first issue, whether it’s a feature request, a bug you've noticed, or a suggestion for our documentation. This helps us prioritize next steps and understand who’s interested in contributing. Excited to see your ideas and get you onboard! Issue #2 is an example of what a good issue post entails. Should have a description or an idea and steps to get there and what to include. Also include any possible references or documentation that can help if available. Be creative with your post! It doesn't matter if your idea seems farfetched or unrealistic, all posts are appreciated and all active developers can comment and discuss for more clarification or interest in collaboration on your idea. This could be expanding on your issue or to help guide you away from a potential problem your idea could bring. Who know, maybe you have a great issue, but don't know exactly where to start. Make a post and we can have a discussion!
gharchive/issue
2024-04-17T06:26:07
2025-04-01T04:33:06.457774
{ "authors": [ "Stavdel" ], "repo": "Stavdel/CHARLIE-AI", "url": "https://github.com/Stavdel/CHARLIE-AI/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2630507138
[Bug] Linux version crashing at everytime the command line is used to execute it Before Reporting [X] I found no existing issues matching my bug [X] My issue is not caused by a theme [X] My issue is not caused by a plugin Describe the Bug using the command ~/.millennium/start.sh to turn millenium on results in a steam crash every single time, i have no idea why Expected Behavior everything has been installed properly so it should run fine, but... nope. Steps To Reproduce use the automatic install prompt from millennium website for linux install curl with pacman (i use arch btw) run the start command crash Operating System Linux Anything else? nothing else to say here I'll take a look! Give this a go: https://github.com/SteamClientHomebrew/Millennium/actions/runs/11645070964. just extract it to ~/.millennium replacing all fixed iam having the same issue startup crash . tried ur method but still its crashing
gharchive/issue
2024-11-02T15:01:42
2025-04-01T04:33:06.464990
{ "authors": [ "Xero646", "dougbug589", "shdwmtr" ], "repo": "SteamClientHomebrew/Millennium", "url": "https://github.com/SteamClientHomebrew/Millennium/issues/153", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
88755234
Graphic bug with multiple raining gold Abilities seems to work well, constantly re-using medic in the right way. Raining Gold, however, seems to have some sort of issue. Issue: Start the game with extra 100+ points to be able to spend on free abilities. Buy multiple (10x) raining gold to use when you get to bosses at level 60+. Load the script. Watch your raining gold go missing leaving a blank square that does nothing (but I assume is where raining gold would be). Reloading the game shows the remaining raining gold correctly (10x). Do not know if this is cosmetic or affects the way the script functions. This means it's being disabled, i noticed that too in 4.6.3 but then it was enabled again later permanently... Missing space means it's disabled.
gharchive/issue
2015-06-16T15:10:10
2025-04-01T04:33:06.467660
{ "authors": [ "Cavey", "K4r0qtuYNE5G4qgZ", "xPaw" ], "repo": "SteamDatabase/steamSummerMinigame", "url": "https://github.com/SteamDatabase/steamSummerMinigame/issues/225", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
983733681
Pixel-Change with Tab Key When having multiple pixel per player, it would be great, if the player could change the pixel with the tab key. Implemented in v2.4.0.
gharchive/issue
2021-08-31T11:07:07
2025-04-01T04:33:06.474280
{ "authors": [ "Steditor", "arneboockmeyer" ], "repo": "Steditor/yard", "url": "https://github.com/Steditor/yard/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
893270184
Add "Required Framework" to API help Add framework requirements to API help. Sample: The following section has a list of supported identifiers for the comma-separated list of "dependencies". +--------------------+--------------------------------------+---------------------------+--------------------+ | Id | Description | Required Steeltoe version | Required Framework | +--------------------+--------------------------------------+---------------------------+--------------------+ ... Resolved by 024efa3
gharchive/issue
2021-05-17T12:15:10
2025-04-01T04:33:06.475527
{ "authors": [ "ccheetham" ], "repo": "SteeltoeOSS/InitializrApi", "url": "https://github.com/SteeltoeOSS/InitializrApi/issues/86", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
238377716
Can't connect to Eureka for service registry I have a running instance of Eureka 1.7.0 running on Tomcat 8.5, when I surf to Eureka the management screen is working fine. I downloaded the Eureka directly from Netflix OSS. When I try to connect with a simple ASP.NET Core Web API project it does not work. I get the following error in the output screen: "Steeltoe.Discovery.Eureka.DiscoveryClient:Error: HeartBeat failed" Is there a way to get more info about why the heartbeat failed? or is steeltoe not compatible with Eureka 1.7.0? This is my appsettings.json: { "Logging": { "IncludeScopes": false, "LogLevel": { "Default": "Information", "Pivotal": "Debug", "Steeltoe": "Information" } }, "spring": { "application": { "name": "EurekaDiscoveryService" } }, "eureka": { "client": { "serviceUrl": "http://localhost:8080/", "shouldFetchRegistry": false }, "instance": { "port": 50612 } } } Change from "Information" to "Debug" for the Steeltoe category
gharchive/issue
2017-06-25T12:25:14
2025-04-01T04:33:06.478582
{ "authors": [ "dtillman", "pieterderycke" ], "repo": "SteeltoeOSS/Samples", "url": "https://github.com/SteeltoeOSS/Samples/issues/16", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
861218110
app.UseEndpoints should be called once in app Describe the bug app.UseEndpoints should not be called multiple times, see https://github.com/dotnet/aspnetcore/issues/17750#issuecomment-564508302 Expected behavior No additional calls to endpoints Environment (please complete the following information): Steeltoe Version 2.3.0 Platform: CF OS: Windows .NET Version .NET Core 3.1 Screenshots If applicable, add screenshots to help explain your problem. Additional context or links Add any other context about the problem here. Unless I'm forgetting something, UseEndpoints isn't used at all in any 2.x version of Steeltoe and I believe we've already corrected issues in 3.x, can you point to something specific please? Sorry, we are using 3.0.2. Nevermind, I see similar (closed) https://github.com/SteeltoeOSS/Steeltoe/issues/497 Do you have pland to release https://github.com/SteeltoeOSS/Steeltoe/pull/591 ?
gharchive/issue
2021-04-19T11:40:29
2025-04-01T04:33:06.483335
{ "authors": [ "Lonli-Lokli", "TimHess" ], "repo": "SteeltoeOSS/Steeltoe", "url": "https://github.com/SteeltoeOSS/Steeltoe/issues/637", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2762397554
Kollisionssensor erkennt keine Hindernisse. Wie gut zusehen ist im Video, reagiert der Kollisionssensor nicht auf sämtliche Hindernisse, erst wenn es das KFZ zum Stoppen bringt. log.txt https://drive.google.com/file/d/1q1weMumcAtRyIlQZUD6CJYo1CY3V4RQm/view?usp=drive_link Fehler gefunden, die Traktoren Fendt 900 und 1000 Vario von Giants haben ein Fehler, alle anderen getesteten Traktoren weisen nicht die Fehler auf.
gharchive/issue
2024-12-29T20:36:26
2025-04-01T04:33:06.533348
{ "authors": [ "regs1980" ], "repo": "Stephan-S/FS25_AutoDrive", "url": "https://github.com/Stephan-S/FS25_AutoDrive/issues/55", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1626528309
🛑 DNS (he.net) is down In 08b4234, DNS (he.net) ($HE_NS) was down: HTTP code: 0 Response time: 0 ms Resolved: DNS (he.net) is back up in 05ee866.
gharchive/issue
2023-03-16T01:22:45
2025-04-01T04:33:06.544911
{ "authors": [ "leitmori" ], "repo": "Sternwarte-St-Ottilien-e-V/status", "url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/2542", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1680480088
🛑 DNS (he.net) is down In e873d0a, DNS (he.net) ($HE_NS) was down: HTTP code: 0 Response time: 0 ms Resolved: DNS (he.net) is back up in 017dc17.
gharchive/issue
2023-04-24T05:22:44
2025-04-01T04:33:06.547595
{ "authors": [ "leitmori" ], "repo": "Sternwarte-St-Ottilien-e-V/status", "url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/5817", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1683985399
🛑 DNS (he.net) is down In 47dda78, DNS (he.net) ($HE_NS) was down: HTTP code: 0 Response time: 0 ms Resolved: DNS (he.net) is back up in 8a38f46.
gharchive/issue
2023-04-25T23:01:42
2025-04-01T04:33:06.550422
{ "authors": [ "leitmori" ], "repo": "Sternwarte-St-Ottilien-e-V/status", "url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/5949", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1698859004
🛑 DNS (he.net) is down In 7bf2ded, DNS (he.net) ($HE_NS) was down: HTTP code: 0 Response time: 0 ms Resolved: DNS (he.net) is back up in 0b0094b.
gharchive/issue
2023-05-07T02:34:43
2025-04-01T04:33:06.553137
{ "authors": [ "leitmori" ], "repo": "Sternwarte-St-Ottilien-e-V/status", "url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/6774", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1708413013
🛑 DNS (he.net) is down In 642fc03, DNS (he.net) ($HE_NS) was down: HTTP code: 0 Response time: 0 ms Resolved: DNS (he.net) is back up in 3790335.
gharchive/issue
2023-05-13T03:30:31
2025-04-01T04:33:06.555964
{ "authors": [ "leitmori" ], "repo": "Sternwarte-St-Ottilien-e-V/status", "url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/7247", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1671785707
TypeError: 'NoneType' object is not iterable The github action failed on pull_request with this error: TypeError: 'NoneType' object is not iterable Most recent callback: Traceback (most recent call last): File "/app/main.py", line 223, in <module> main() File "/app/main.py", line 159, in main most_recent_runs = get_recent_runs_for_job(base_url=base_dbt_cloud_api_url, headers=req_auth_headers, job_id=dbt_cloud_job_id, same_branch_flag=same_branch_flag, max_runs=max_runs) File "/app/main.py", line [12](https://github.com/trainual/trainual-data-dbt/actions/runs/4724221958/jobs/8381167543#step:3:13)3, in get_recent_runs_for_job recent_runs_info = extract_dbt_runs_info(recent_runs['data'], same_branch_flag) File "/app/main.py", line 68, in extract_dbt_runs_info for run in recent_runs_list: TypeError: 'NoneType' object is not iterable Questions: how would I troubleshoot or set to not run when this error, related to this, is encountered? Hey @normwarren! A couple things I would make sure of, to kick off trouble shooting: make sure that you have the dbt cloud token set and that the token has correct permissions on the given project dbt_cloud_token: ${{ secrets.DBT_CLOUD_TOKEN }} Make sure you all fill out account id and job id the following in the action yaml as well, example below: dbt_cloud_account_id: 12345 dbt_cloud_job_id: 130247 The NoneType error on this action tends to happen when it's trying to parse through the dbt Cloud API results but the results from dbt Cloud came back an error usually due to permissions Met with Steve and mentioned we had a team member left when github action broke. We created an access account in dbt-cloud and update the github secrets, and everything worked again ! Thank you Steve!
gharchive/issue
2023-04-17T18:53:56
2025-04-01T04:33:06.561573
{ "authors": [ "Stevedow99", "normwarren" ], "repo": "Stevedow99/dbt-cloud-cancel-running-ci-job-action", "url": "https://github.com/Stevedow99/dbt-cloud-cancel-running-ci-job-action/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
437884311
FBI Conflit with new super mario bros 2 Hello everyone I use fbi 2.6 and I found a beug with a particular game actually launch you new sueper mario bros 2 launch you a game and enjoy leaving you the game and launch you fbi and boom fbi crash fbi at launch it does its that with this game I have to test with 20 other games no problem is very annoying considering that when I leave the game I go directly to fbi and bim his crash here is it is you are all to bawl you can try to reproduce it you must fbi 2.6.0 and new super mario bros 2 thank you and good evening What This has been reported before and was determined to not be an FBI issue. See https://github.com/Steveice10/FBI/issues/382
gharchive/issue
2019-04-27T00:10:56
2025-04-01T04:33:06.564782
{ "authors": [ "InSimpleTermsJordan", "Steveice10", "StigiTT" ], "repo": "Steveice10/FBI", "url": "https://github.com/Steveice10/FBI/issues/501", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
135750128
Change menu order to mirror game difficulty As per the front page recommended order: Bandit Leviathan or Natas or Krypton Narnia Behemoth Utumno Maze … excellent suggestion, let's try this out
gharchive/pull-request
2016-02-23T14:23:18
2025-04-01T04:33:06.565806
{ "authors": [ "StevenVanAcker", "unfo" ], "repo": "StevenVanAcker/OverTheWire-website", "url": "https://github.com/StevenVanAcker/OverTheWire-website/pull/44", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
603253758
请问项目所需要的环境以及版本能否发下 我本地部署了一下,用的是jdk1.8+tomcat8.5+mysql5.7+maven3.5 但是提示无法加载db.properties文件 报错如上图所示,能否给看下 `20-Apr-2020 21:38:05.869 淇℃伅 [RMI TCP Connection(3)-127.0.0.1] org.apache.catalina.core.ApplicationContext.log No Spring WebApplicationInitializer types detected on classpath 20-Apr-2020 21:38:06.353 淇℃伅 [RMI TCP Connection(3)-127.0.0.1] org.apache.catalina.core.ApplicationContext.log Initializing Spring root WebApplicationContext 20-Apr-2020 21:38:07.458 涓ラ噸 [RMI TCP Connection(3)-127.0.0.1] org.apache.catalina.core.StandardContext.listenerStart Exception sending context initialized event to listener instance of class [org.springframework.web.context.ContextLoaderListener] org.springframework.beans.factory.BeanInitializationException: Could not load properties; nested exception is java.io.FileNotFoundException: class path resource [db.properties] cannot be opened because it does not exist at org.springframework.context.support.PropertySourcesPlaceholderConfigurer.postProcessBeanFactory(PropertySourcesPlaceholderConfigurer.java:151) at org.springframework.context.support.PostProcessorRegistrationDelegate.invokeBeanFactoryPostProcessors(PostProcessorRegistrationDelegate.java:265) at org.springframework.context.support.PostProcessorRegistrationDelegate.invokeBeanFactoryPostProcessors(PostProcessorRegistrationDelegate.java:162) at org.springframework.context.support.AbstractApplicationContext.invokeBeanFactoryPostProcessors(AbstractApplicationContext.java:606) at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:462) at org.springframework.web.context.ContextLoader.configureAndRefreshWebApplicationContext(ContextLoader.java:403) at org.springframework.web.context.ContextLoader.initWebApplicationContext(ContextLoader.java:306) at org.springframework.web.context.ContextLoaderListener.contextInitialized(ContextLoaderListener.java:106) at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:4770) at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5236) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150) at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:754) at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:730) at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:734) at org.apache.catalina.startup.HostConfig.manageApp(HostConfig.java:1731) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.tomcat.util.modeler.BaseModelMBean.invoke(BaseModelMBean.java:287) at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.invoke(DefaultMBeanServerInterceptor.java:819) at com.sun.jmx.mbeanserver.JmxMBeanServer.invoke(JmxMBeanServer.java:801) at org.apache.catalina.mbeans.MBeanFactory.createStandardContext(MBeanFactory.java:483) at org.apache.catalina.mbeans.MBeanFactory.createStandardContext(MBeanFactory.java:432) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.tomcat.util.modeler.BaseModelMBean.invoke(BaseModelMBean.java:287) at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.invoke(DefaultMBeanServerInterceptor.java:819) at com.sun.jmx.mbeanserver.JmxMBeanServer.invoke(JmxMBeanServer.java:801) at com.sun.jmx.remote.security.MBeanServerAccessController.invoke(MBeanServerAccessController.java:468) at javax.management.remote.rmi.RMIConnectionImpl.doOperation(RMIConnectionImpl.java:1468) at javax.management.remote.rmi.RMIConnectionImpl.access$300(RMIConnectionImpl.java:76) at javax.management.remote.rmi.RMIConnectionImpl$PrivilegedOperation.run(RMIConnectionImpl.java:1309) at java.security.AccessController.doPrivileged(Native Method) at javax.management.remote.rmi.RMIConnectionImpl.doPrivilegedOperation(RMIConnectionImpl.java:1408) at javax.management.remote.rmi.RMIConnectionImpl.invoke(RMIConnectionImpl.java:829) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at sun.rmi.server.UnicastServerRef.dispatch(UnicastServerRef.java:357) at sun.rmi.transport.Transport$1.run(Transport.java:200) at sun.rmi.transport.Transport$1.run(Transport.java:197) at java.security.AccessController.doPrivileged(Native Method) at sun.rmi.transport.Transport.serviceCall(Transport.java:196) at sun.rmi.transport.tcp.TCPTransport.handleMessages(TCPTransport.java:573) at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.run0(TCPTransport.java:834) at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.lambda$run$0(TCPTransport.java:688) at java.security.AccessController.doPrivileged(Native Method) at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.run(TCPTransport.java:687) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.io.FileNotFoundException: class path resource [db.properties] cannot be opened because it does not exist at org.springframework.core.io.ClassPathResource.getInputStream(ClassPathResource.java:172) at org.springframework.core.io.support.EncodedResource.getInputStream(EncodedResource.java:150) at org.springframework.core.io.support.PropertiesLoaderUtils.fillProperties(PropertiesLoaderUtils.java:98) at org.springframework.core.io.support.PropertiesLoaderSupport.loadProperties(PropertiesLoaderSupport.java:175) at org.springframework.core.io.support.PropertiesLoaderSupport.mergeProperties(PropertiesLoaderSupport.java:156) at org.springframework.context.support.PropertySourcesPlaceholderConfigurer.postProcessBeanFactory(PropertySourcesPlaceholderConfigurer.java:142) ... 54 more 20-Apr-2020 21:38:07.463 淇℃伅 [RMI TCP Connection(3)-127.0.0.1] org.apache.catalina.core.ApplicationContext.log Closing Spring root WebApplicationContext ` 我本地部署了一下,用的是jdk1.8+tomcat8.5+mysql5.7+maven3.5 但是提示无法加载db.properties文件 报错如上图所示,能否给看下 你好,请问你解决这个问题了吗 发生这个问题的原因是db.properties所在文件夹并未设置为resource目录 进行如下设置即可 在最后面的右边目录找到ResourcesRoot设置一下就行了,我这里因为设置过了,所以没有显示
gharchive/issue
2020-04-20T13:49:03
2025-04-01T04:33:06.582043
{ "authors": [ "AcRushGo", "AtomRun", "miaofengshi" ], "repo": "StevenWash/xxshop", "url": "https://github.com/StevenWash/xxshop/issues/7", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
99782355
There was an unhandled error - Sorry! Whenever I go to a page with this fishing rod in the screen, whether its on a character or in a bank tab, procurement crashes. https://i.gyazo.com/c323a1cc1a90a19b28521306edd352cf.png This is a known issue, if you can zip and send me all your .bin files I can take a look at what's causing this (I don't have a fishing rod) Hey im pretty shitty with techy stuff, so I'm not sure if this is what you mean, but this is a copy and paste of all the files within G;/Program Files/Path of Exile > Procurement > (my other email) Date: Sat, 8 Aug 2015 04:15:50 -0700 From: notifications@github.com To: Procurement@noreply.github.com CC: dannydc13@hotmail.com Subject: Re: [Procurement] There was an unhandled error - Sorry! (#157) This is a known issue, if you can zip and send me all your .bin files I can take a look at what's causing this (I don't have a fishing rod) — Reply to this email directly or view it on GitHub. This has, finally, been fixed in Procurement 1.24.2
gharchive/issue
2015-08-08T08:09:09
2025-04-01T04:33:06.614669
{ "authors": [ "DannyBooy", "Stickymaddness" ], "repo": "Stickymaddness/Procurement", "url": "https://github.com/Stickymaddness/Procurement/issues/157", "license": "Artistic-2.0", "license_type": "permissive", "license_source": "github-api" }
373219021
update Travis to Py3.7 I'm keeping Py3.4 around for the lab computers Coverage remained the same at 62.151% when pulling 9217f4198ef274bce9405ff1c8b5f38e0cde1ecf on python-3.7 into 1504998d68aab7e80f43292320516c4e766fc5a9 on master.
gharchive/pull-request
2018-10-23T21:53:46
2025-04-01T04:33:06.626887
{ "authors": [ "coveralls", "hawkrives" ], "repo": "StoDevX/cs251-toolkit", "url": "https://github.com/StoDevX/cs251-toolkit/pull/74", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
108556327
RuntimeError: maximum recursion depth exceeded in comparison Infinity loop: socket.gaierror: [Errno -3] Temporary failure in name resolution During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 45, in _url_query response = urllib.request.urlopen(req) File "/usr/lib64/python3.3/urllib/request.py", line 156, in urlopen return opener.open(url, data, timeout) File "/usr/lib64/python3.3/urllib/request.py", line 469, in open response = self._open(req, data) File "/usr/lib64/python3.3/urllib/request.py", line 487, in _open '_open', req) File "/usr/lib64/python3.3/urllib/request.py", line 447, in _call_chain result = func(*args) File "/usr/lib64/python3.3/urllib/request.py", line 1274, in http_open return self.do_open(http.client.HTTPConnection, req) File "/usr/lib64/python3.3/urllib/request.py", line 1251, in do_open raise URLError(err) urllib.error.URLError: <urlopen error [Errno -3] Temporary failure in name resolution> During handling of the above exception, another exception occurred: Traceback (most recent call last): File "dataserv-client", line 10, in cli.main(sys.argv[1:]) File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/cli.py", line 190, in main return getattr(client, command_name)(**arguments) File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/api.py", line 235, in farm self.poll(delay=delay, limit=limit) File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/api.py", line 161, in poll self.ping() File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/api.py", line 144, in ping self.messenger.ping() File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 109, in ping return self._url_query("/api/ping/%s" % self.auth_address()) File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 69, in _url_query return self._handle_connection_error(api_path, retries, authenticate) File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 81, in _handle_connection_error ............. File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 81, in _handle_connection_error return self._url_query(api_path, retries + 1, authenticate) File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 69, in _url_query return self._handle_connection_error(api_path, retries, authenticate) File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 81, in _handle_connection_error return self._url_query(api_path, retries + 1, authenticate) File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 40, in _url_query self.btctxstore, self._get_server_address(), self.wif File "/usr/lib64/python3.3/site-packages/storjcore-0.0.3-py3.3.egg/storjcore/auth.py", line 38, in create_headers signature = btctxstore.sign_unicode(sender_wif, msg) File "/usr/lib64/python3.3/site-packages/btctxstore-4.5.0-py3.3.egg/btctxstore/api.py", line 191, in sign_unicode return self.sign_data(wif, hexdata) File "/usr/lib64/python3.3/site-packages/btctxstore-4.5.0-py3.3.egg/btctxstore/api.py", line 174, in sign_data sigdata = control.sign_data(self.testnet, data, key) File "/usr/lib64/python3.3/site-packages/btctxstore-4.5.0-py3.3.egg/btctxstore/control.py", line 293, in sign_data pk = ecdsa.SigningKey.from_secret_exponent(secretexponent, curve=SECP256k1) File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/keys.py", line 142, in from_secret_exponent File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/keys.py", line 28, in from_public_point File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ecdsa.py", line 87, in init File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ellipticcurve.py", line 143, in rmul File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ellipticcurve.py", line 134, in mul File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ellipticcurve.py", line 107, in add File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ellipticcurve.py", line 73, in init File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ellipticcurve.py", line 59, in contains_point RuntimeError: maximum recursion depth exceeded in comparison Looks like the farm command? https://github.com/Storj/dataserv-client/blob/develop/dataserv_client/api.py#L241 I will fix it tomorrow. python max recursion is ~1000. For every 30 sec retry 2 recursions calls are needed. Max retry limit will be somewhere between 450 and 500. -> ~4h retry I will increase the retry delay for farm to 60 sec and the retry limit to 450-500. That way max retry will be ~8h without exception. Can you just drop recursion and use normal loop please? I did some unit tests and python recursion is realy a bad. Looks like your idea with the loop is the only solution. will be fixed in the next version (after pull request merge) Pull request merged. Unit test with 2000 retries passed.
gharchive/issue
2015-09-27T19:36:43
2025-04-01T04:33:06.654821
{ "authors": [ "alphaonex86", "littleskunk" ], "repo": "Storj/dataserv-client", "url": "https://github.com/Storj/dataserv-client/issues/160", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
194980653
from core.lib.utils migrate simpleDecrypt and simpleEncrypt of data f… …or id_ecdsa __________________________ PEP8-check(ignoring W503) ___________________________ [gw0] linux2 -- Python 2.7.9 /home/travis/build/Storj/storj-python-sdk/.tox/py27/bin/python2.7 /home/travis/build/Storj/storj-python-sdk/storj/model.py:265:38: E261 at least two spaces before inline comment padding = 16 - len(data) % 16# ^ /home/travis/build/Storj/storj-python-sdk/storj/model.py:271:45: E226 missing whitespace around arithmetic operator def decrypt_node(self, hex_data, key='0'*32, iv='0'*16): ^ /home/travis/build/Storj/storj-python-sdk/storj/model.py:271:56: E226 missing whitespace around arithmetic operator def decrypt_node(self, hex_data, key='0'*32, iv='0'*16): ^ /home/travis/build/Storj/storj-python-sdk/storj/model.py:276:41: E226 missing whitespace around arithmetic operator def encrypt_node(self, data, key='0'*32, iv='0'*16): ^ /home/travis/build/Storj/storj-python-sdk/storj/model.py:276:52: E226 missing whitespace around arithmetic operator def encrypt_node(self, data, key='0'*32, iv='0'*16): i hope i got it now @BlackDuck888 you can test on your workstation by using tox. ;) i will try it next time, step by step, so many new things, right now i am fighting with GitHub. i close this pr, and will start from scratch, this will help me to understand how everything ist working
gharchive/pull-request
2016-12-12T13:58:11
2025-04-01T04:33:06.665376
{ "authors": [ "BlackDuck888", "steenzout" ], "repo": "Storj/storj-python-sdk", "url": "https://github.com/Storj/storj-python-sdk/pull/69", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1523301009
Incomplete basic output When creating basic output sometimes the results look incomplete. For example: let schema = JSONSchema::compile(&json!({ "$id": "https://example.com/arrays.schema.json", "$schema": "https://json-schema.org/draft/2020-12/schema", "type": "object", "properties": { "vegetables": { "type": "array", "description": "Desc1", "items": { "$ref": "#/$defs/veggie" } } }, "$defs": { "veggie": { "type": "object", "required": [ "veggieName", "veggieLike" ], "properties": { "veggieName": { "description": "Desc2", "type": "string" }, "veggieLike": { "description": "Desc3", "type": "boolean" } } } } })) .unwrap(); if let BasicOutput::Valid(output) = schema .apply(&json!({ "vegetables":[{"veggieName": "carrot", "veggieLike": true}] })) .basic() { for unit in output { println!("{}", unit.instance_location()); } } Output: /vegetables /vegetables/0 Expected output: /vegetables /vegetables/0 /vegetables/0/veggieName /vegetables/0/veggieLike I updated the example and did a little investigation. I think the issue is that RefValidator doesn't contain a custom implementation of apply, so any PartialApplication information from down stream is discarded. Sorry for the delay, with the current approach, it is a bit complicated as annotations are bound to the schema. Not sure if it would be possible to keep the references, but maybe making annotations own their data would be a better way to go anyway. Fixed! However, keyword paths are incorrect in many cases. I am going to fix them separately
gharchive/issue
2023-01-06T23:24:40
2025-04-01T04:33:06.697176
{ "authors": [ "BrynCooke", "Stranger6667" ], "repo": "Stranger6667/jsonschema-rs", "url": "https://github.com/Stranger6667/jsonschema-rs/issues/403", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2684914660
🛑 Jellyseer is down In f0fd96f, Jellyseer (https://jellyseerr.stroyco.eu) was down: HTTP code: 502 Response time: 1272 ms Resolved: Jellyseer is back up in 6014d05 after 1 hour, 42 minutes.
gharchive/issue
2024-11-22T23:28:16
2025-04-01T04:33:06.771974
{ "authors": [ "DestroyCom" ], "repo": "StroyCo/stroy-up", "url": "https://github.com/StroyCo/stroy-up/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1674362970
Separate parsing APIs that don't depend on ANTLR See https://github.com/Strumenta/StarLasu/issues/27 Kolasu split them into a new module (kolasu-antlr) and a new package (com.strumenta.kolasu.antlr.parsing). Given the structure and choices made in Tylasu, we could introduce antlr/parsing.ts and export the current parsing.ts, which won't depend on ANTLR, from the core index.ts Just like Kolasu, we'll increase the version number to 1.3 since this is a breaking change. This looks completed
gharchive/issue
2023-04-19T07:43:03
2025-04-01T04:33:06.773565
{ "authors": [ "alessiostalla" ], "repo": "Strumenta/tylasu", "url": "https://github.com/Strumenta/tylasu/issues/40", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1012889034
🛑 Nextcloud Papa is down In 620138c, Nextcloud Papa ($NEXTCLOUD_2) was down: HTTP code: 0 Response time: 0 ms Resolved: Nextcloud Papa is back up in a102d8b.
gharchive/issue
2021-10-01T04:58:11
2025-04-01T04:33:06.776855
{ "authors": [ "StudFu-WordToMD" ], "repo": "StudFu-WordToMD/status", "url": "https://github.com/StudFu-WordToMD/status/issues/2967", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1045833926
🛑 Nextcloud Papa is down In 1a8c884, Nextcloud Papa ($NEXTCLOUD_2) was down: HTTP code: 0 Response time: 0 ms Resolved: Nextcloud Papa is back up in 8365fb1.
gharchive/issue
2021-11-05T13:10:42
2025-04-01T04:33:06.778903
{ "authors": [ "StudFu-WordToMD" ], "repo": "StudFu-WordToMD/status", "url": "https://github.com/StudFu-WordToMD/status/issues/3822", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
213611118
Tooltips at screen top are hidden/overlapped by top nav bar The addon/referenced Github addon themes include one to expand the top nav bar etc. and I know this is minor but just wanted to point out that the tooltips for the top right elements like WATCH/STAR/FORK when theme is enabled are hidden under the navbar as shown in these screenshots below. I am not sure if it is possible, but a really nice change just to ensure if in the future Github were to add other functions and corresponding tooltips would be just to MOVE those tooltips down UNDER the corresponding elements. The dumb placement of them above elements so close to the top of the screen of course were not even made by this theme but by the original site dev, but they got away with it (barely) only because the original site didn't have quite the height of the nav bar with all the functionality being added here. :_bow: FORK tooltip (the tiny bit of white if you look): https://1drv.ms/i/s!AhSqjTYUuHPKgcNbmBCOm-wDiU936w STAR tooltip (the tiny bit of white if you look): https://1drv.ms/i/s!AhSqjTYUuHPKgcNaLEhGWHrofnpBQg Again not even sure this is considered a bug in this theme but really a SITE bug that becomes a problem only when this theme is applied (and not a major one). Hi @CollinChaffin! GitHub has removed the tooltips from the star and fork buttons. So, it looks like the problem is in the github-script-make-tooltips.user.js code. It is currently set to make all tooltips point up, I'll update the code to make those point down... thanks for letting us know!
gharchive/issue
2017-03-12T15:19:56
2025-04-01T04:33:06.789033
{ "authors": [ "CollinChaffin", "Mottie" ], "repo": "StylishThemes/GitHub-Dark-Script", "url": "https://github.com/StylishThemes/GitHub-Dark-Script/issues/30", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1369466647
chore: aliasing substrapp django models To see the specific tasks where the Asana app for GitHub is being used, see below: https://app.asana.com/0/0/1202951505377093 /e2e --mode standalone End to end tests: :hourglass_flowing_sand: STARTING
gharchive/pull-request
2022-09-12T08:16:02
2025-04-01T04:33:06.823125
{ "authors": [ "Owlfred", "sergebouchut2" ], "repo": "Substra/substra-backend", "url": "https://github.com/Substra/substra-backend/pull/444", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2161274557
chore: add missing repo Description How has this been tested? Checklist [ ] changelog was updated with notable changes [ ] documentation was updated - ".github/workflows/helm.yml" Sounds good ! But the problem are having is detected when we package, which only called during "publish". Should we add a package in each PR to try that no dependency is missed ? - ".github/workflows/helm.yml" Sounds good ! But the problem are having is detected when we package, which only called during "publish". Should we add a package in each PR to try that no dependency is missed ? Yes, already did! https://github.com/Substra/substra-gha-workflows/pull/10
gharchive/pull-request
2024-02-29T13:32:24
2025-04-01T04:33:06.826410
{ "authors": [ "guilhem-barthes", "thbcmlowk" ], "repo": "Substra/substra-backend", "url": "https://github.com/Substra/substra-backend/pull/837", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1060170688
Make better label text for audio player Related to #336 and #339. Improve to words that describe these specific actions like "Read page", instead of generic ones like "play" FYI, references for this are things like the video aria-label in the code at https://developer.mozilla.org/en-US/docs/Learn/JavaScript/Client-side_web_APIs/Video_and_audio_APIs and https://www.w3.org/TR/wai-aria-practices-1.1/#no_aria_better_bad_aria.
gharchive/issue
2021-11-22T13:40:26
2025-04-01T04:33:06.860506
{ "authors": [ "plocket" ], "repo": "SuffolkLITLab/docassemble-AssemblyLine", "url": "https://github.com/SuffolkLITLab/docassemble-AssemblyLine/issues/342", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1358384526
🛑 Zinc 20 is down In f2e2fb4, Zinc 20 (https://zinc20.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 20 is back up in 7df4f4c.
gharchive/issue
2022-09-01T07:28:53
2025-04-01T04:33:06.862815
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/10196", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1402389583
🛑 Zinc 15 is down In cb5ca07, Zinc 15 (https://zinc15.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 15 is back up in e702afd.
gharchive/issue
2022-10-09T20:32:38
2025-04-01T04:33:06.865076
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/11594", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1451338012
🛑 Zinc 15 is down In 60dcae2, Zinc 15 (https://zinc15.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 15 is back up in c69183d.
gharchive/issue
2022-11-16T10:40:41
2025-04-01T04:33:06.867356
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/12953", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1555704920
🛑 Chem Exper is down In c16763a, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 2707037.
gharchive/issue
2023-01-24T21:32:38
2025-04-01T04:33:06.869834
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/14817", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1605798014
🛑 Chem Exper is down In 4df1c77, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 8e596ee.
gharchive/issue
2023-03-01T22:56:52
2025-04-01T04:33:06.872282
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/16306", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1635963858
🛑 Chem Exper is down In 517a95d, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 2cdd832.
gharchive/issue
2023-03-22T15:09:36
2025-04-01T04:33:06.874513
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/17143", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1656444109
🛑 Chem Exper is down In 931266c, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 4229061.
gharchive/issue
2023-04-05T23:38:19
2025-04-01T04:33:06.876787
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/17769", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1723376989
🛑 Zinc 15 is down In 9dc23b7, Zinc 15 (https://zinc15.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 15 is back up in e73e12c.
gharchive/issue
2023-05-24T07:42:26
2025-04-01T04:33:06.879146
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/20223", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1748725257
🛑 Zinc 20 is down In 275fb2c, Zinc 20 (https://zinc20.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 20 is back up in c63d372.
gharchive/issue
2023-06-08T22:18:46
2025-04-01T04:33:06.881455
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/21495", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1805257595
🛑 Chem Exper is down In abebb22, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 0fa3572.
gharchive/issue
2023-07-14T17:49:46
2025-04-01T04:33:06.883720
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/24359", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1807895131
🛑 Chem Exper is down In 92787e0, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 495fc4f.
gharchive/issue
2023-07-17T14:17:58
2025-04-01T04:33:06.886179
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/24483", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1824860679
🛑 Zinc 15 is down In 0f2572b, Zinc 15 (https://zinc15.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 15 is back up in 982e713.
gharchive/issue
2023-07-27T17:29:32
2025-04-01T04:33:06.888416
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/25099", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1843048420
🛑 Chem Exper is down In fe6189a, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 49c875e.
gharchive/issue
2023-08-09T11:35:41
2025-04-01T04:33:06.890666
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/26095", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1931473402
🛑 Chem Exper is down In da2d86e, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 9aa21cd after 16 minutes.
gharchive/issue
2023-10-07T19:24:41
2025-04-01T04:33:06.892954
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/30235", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1954794560
🛑 Chem Exper is down In 945aca4, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in b9f07b2 after 39 minutes.
gharchive/issue
2023-10-20T18:03:22
2025-04-01T04:33:06.895221
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/30831", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2117583702
🛑 Binding Database is down In 86481af, Binding Database (http://www.bindingdb.org/bind/index.jsp) was down: HTTP code: 0 Response time: 0 ms Resolved: Binding Database is back up in e697e69 after 11 minutes.
gharchive/issue
2024-02-05T03:08:31
2025-04-01T04:33:06.897730
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/36263", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2122529474
🛑 Chemical Resolver is down In 8db010a, Chemical Resolver (https://cactus.nci.nih.gov/chemical/structure) was down: HTTP code: 500 Response time: 76 ms Resolved: Chemical Resolver is back up in 488de03 after 10 minutes.
gharchive/issue
2024-02-07T09:09:11
2025-04-01T04:33:06.900084
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/36370", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2131574777
🛑 Chem Exper is down In 46b6631, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 411ea98 after 7 minutes.
gharchive/issue
2024-02-13T06:46:22
2025-04-01T04:33:06.902472
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/36668", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2246330663
🛑 Zinc 20 is down In ddf4a70, Zinc 20 (https://zinc20.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 20 is back up in fe7689a after 6 minutes.
gharchive/issue
2024-04-16T15:23:54
2025-04-01T04:33:06.904730
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/39172", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2363387936
🛑 Adverse Drug Reaction Classification System is down In 0811654, Adverse Drug Reaction Classification System (http://bioinf.xmu.edu.cn/ADReCS/) was down: HTTP code: 0 Response time: 0 ms Resolved: Adverse Drug Reaction Classification System is back up in 01f6e9a after 16 minutes.
gharchive/issue
2024-06-20T02:39:14
2025-04-01T04:33:06.907117
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/41815", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2509375431
🛑 Chem Exper is down In 23d7f37, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in f9e1ffd after 10 minutes.
gharchive/issue
2024-09-06T02:59:45
2025-04-01T04:33:06.909597
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/45263", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2547681351
🛑 Chem Exper is down In b20e1f0, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in b65a46b after 11 minutes.
gharchive/issue
2024-09-25T11:06:52
2025-04-01T04:33:06.911863
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/45883", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2620378024
🛑 Chem Exper is down In 46a92da, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 4f7b530 after 17 minutes.
gharchive/issue
2024-10-29T07:23:25
2025-04-01T04:33:06.914149
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/46998", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2684635845
🛑 Chem Exper is down In 87d8e38, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in d4d665c after 8 minutes.
gharchive/issue
2024-11-22T21:29:40
2025-04-01T04:33:06.916430
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/47794", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2725075008
🛑 Chem Exper is down In a4b77e6, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 8cbaf4e after 16 minutes.
gharchive/issue
2024-12-08T09:37:02
2025-04-01T04:33:06.918666
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/48295", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1268552256
🛑 Zinc 20 is down In 370e052, Zinc 20 (https://zinc20.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 20 is back up in 9785604.
gharchive/issue
2022-06-12T11:31:24
2025-04-01T04:33:06.921299
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/6808", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1286065540
🛑 Zinc 15 is down In baf39e6, Zinc 15 (https://zinc15.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 15 is back up in bb9985d.
gharchive/issue
2022-06-27T16:40:10
2025-04-01T04:33:06.923549
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/7456", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1288314531
🛑 Zinc 15 is down In e956fe1, Zinc 15 (https://zinc15.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 15 is back up in 04a9c0d.
gharchive/issue
2022-06-29T07:51:50
2025-04-01T04:33:06.925815
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/7515", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1342316538
🛑 Zinc 15 is down In a5d0e4f, Zinc 15 (https://zinc15.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 15 is back up in dba657b.
gharchive/issue
2022-08-17T21:38:46
2025-04-01T04:33:06.928277
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/9678", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1989059473
🛑 Stanford Tissue Microarray Database is down In e18f6af, Stanford Tissue Microarray Database (http://tma.im) was down: HTTP code: 0 Response time: 0 ms Resolved: Stanford Tissue Microarray Database is back up in d744d59 after 7 minutes.
gharchive/issue
2023-11-11T17:51:56
2025-04-01T04:33:06.930666
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Medical-Informatics", "url": "https://github.com/Sulstice/Uptime-Medical-Informatics/issues/1062", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1471488219
Get error as "Cannot read properties of null (reading 'once') When clicked on the first item in the dropdown menu, Getting error due to RUM javascript code as: Also nothing is rendered in screen so that application is broken. But, if the RUM javascript code is removed, there is no error. Even though the issue above is irrespective of the browser. Browser used: Chrome Version 107.0.5304.107 (Official Build) (64-bit) App uses: react : 17.0.2 and react-dom: 17.0.2 SumoLogic RUM used: https://rum.sumologic.com/sumologic-rum-v4.js Any help or idea would be appreciated 🙏 Hi @prajon84, thanks for letting us know. I prepared a fix for that issue: https://github.com/SumoLogic/sumologic-opentelemetry-js/pull/67 New version (4.1.1) with fix has been released: https://rum.sumologic.com/sumologic-rum-v4.1.1.js https://rum.sumologic.com/sumologic-rum-v4.1.js https://rum.sumologic.com/sumologic-rum-v4.js https://rum.sumologic.com/sumologic-rum.js In case of any problems, please let us know.
gharchive/issue
2022-12-01T15:11:43
2025-04-01T04:33:06.940149
{ "authors": [ "TRybakSUMO", "prajon84" ], "repo": "SumoLogic/sumologic-opentelemetry-js", "url": "https://github.com/SumoLogic/sumologic-opentelemetry-js/issues/66", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
784643168
ENH: Make ElastixMain database creation + loading components thread safe Ensures that only one thread may load the components into the ComponentDatabase, using using C++11 "magic statics". ElastixMain::GetComponentDatabase() now returns a "const reference" instead of a non-const pointer, to prevent one thread from modifying the database while another is reading from the database. Removed the two UnloadComponents() member functions (from both ElastixMain and ComponentLoader), as they appear not useful anymore. Aims to fix issue #174 "elastix 4.9 static libary version not thread safe" reported by @jiangliMED, August 16, 2019 With help from Konstantinos Ntatsis (@ntatsisk) @ntatsisk Hi Konstantinos! Can you please review this pull request? And could you possibly also try it out, to see if it does indeed fix a multi-threading issue that you encountered? You may check out the (temporary) branch of this pull request: https://github.com/SuperElastix/elastix/tree/ElastixMain-LoadComponents-thread-safe @ntatsisk Hi Konstantinos! Can you please review this pull request? And could you possibly also try it out, to see if it does indeed fix a multi-threading issue that you encountered? You may check out the (temporary) branch of this pull request: https://github.com/SuperElastix/elastix/tree/ElastixMain-LoadComponents-thread-safe Hello @N-Dekker, the solution looks great! I really liked the combination of magic statics together with a lambda function for the ComponentDatabase initialization. I merged this branch with my local branch where the logger is "deactivated" and managed to run some tests. It works with no issues! Hello @N-Dekker, the solution looks great! I really liked the combination of magic statics together with a lambda function for the ComponentDatabase initialization. I merged this branch with my local branch where the logger is "deactivated" and managed to run some tests. It works with no issues! @ntatsisk Thank you for your approval and your enthusiastic reply! I certainly enjoyed combining magic statics together with a lambda 😃 May I ask what your use case looks like? So you call RegisterImages multiple times simultaneously within the same process? How do you do that? Something like std::async? And what do you gain from doing so? @ntatsisk Thank you for your approval and your enthusiastic reply! I certainly enjoyed combining magic statics together with a lambda 😃 May I ask what your use case looks like? So you call RegisterImages multiple times simultaneously within the same process? How do you do that? Something like std::async? And what do you gain from doing so? @N-Dekker No, we are using OpenMP to parallelize the registrations. The idea is as follows: #pragma omp parallel for default(none) num_threads(24) for (int i = 0; i < 600; i++) { ElastixFilterType::Pointer elastixFilter = ElastixFilterType::New(); elastixFilter->SetFixedImage(images[i]); elastixFilter->SetMovingImage(images[i+1]); elastixFilter->SetParameterObject(m_elxParameterObjectPointer); elastixFilter->LogToConsoleOff(); elastixFilter->LogToFileOff(); // #pragma omp critical // (= not necessary if elastix is thread-safe) elastixFilter->Update(); } As you see, we are trying to do many (simple) registrations but in a short amount of time (~<10sec) which is a requirement. So, it is better to split the registrations among the threads instead of using many threads for a single registration, as ITK threading does. The gain is some saved seconds eg. from 12sec (with the critical section) to 7sec (in the thread-safe version). In another part, it goes from 30sec to 10sec. Generally, these times might not seem like a big deal, but they are for our application ;) @N-Dekker No, we are using OpenMP to parallelize the registrations. The idea is as follows: #pragma omp parallel for default(none) num_threads(24) for (int i = 0; i < 600; i++) { ElastixFilterType::Pointer elastixFilter = ElastixFilterType::New(); elastixFilter->SetFixedImage(images[i]); elastixFilter->SetMovingImage(images[i+1]); elastixFilter->SetParameterObject(m_elxParameterObjectPointer); elastixFilter->LogToConsoleOff(); elastixFilter->LogToFileOff(); // #pragma omp critical // (= not necessary if elastix is thread-safe) elastixFilter->Update(); } As you see, we are trying to do many (simple) registrations but in a short amount of time (~<10sec) which is a requirement. So, it is better to split the registrations among the threads instead of using many threads for a single registration, as ITK threading does. The gain is some saved seconds eg. from 12sec (with the critical section) to 7sec (in the thread-safe version). In another part, it goes from 30sec to 10sec. Generally, these times might not seem like a big deal, but they are for our application ;) @mstaring @stefanklein Hope it's OK to you that I'll merge this pull request this afternoon! @mstaring @stefanklein Hope it's OK to you that I'll merge this pull request this afternoon! I do not get all the details of this PR, but i assume i) elastix still works (all tests pass for the command line version and ITKElastix also still works), and ii) it was verified that it actually fixes a problem (a new test first failed but now passes). I do not get all the details of this PR, but i assume i) elastix still works (all tests pass for the command line version and ITKElastix also still works), and ii) it was verified that it actually fixes a problem (a new test first failed but now passes). @mstaring Thanks for your feedback i assume i) elastix still works (all tests pass for the command line version and ITKElastix also still works) All elastix test are passed successfully, at the CI. However, I did not test ITKElastix with this pull request. We currently do not test ITKElastix on each elastix pull request. But I don't think ITKElastix would have any problem with this pull request. Unless it would try to call UnloadComponents(), which is removed by this pull request. But that seems unlikely to me, especially because UnloadComponents() wasn't useful anymore anyway. @ViktorvdValk do you know if UnloadComponents() was wrapped to Python, by ITKElastix? ii) it was verified that it actually fixes a problem (a new test first failed but now passes). It is verified by Konstantinos (@ntatsisk) that it actually fixes a problem. I can still see if an automatic test can be added, based on his use case. @mstaring Thanks for your feedback i assume i) elastix still works (all tests pass for the command line version and ITKElastix also still works) All elastix test are passed successfully, at the CI. However, I did not test ITKElastix with this pull request. We currently do not test ITKElastix on each elastix pull request. But I don't think ITKElastix would have any problem with this pull request. Unless it would try to call UnloadComponents(), which is removed by this pull request. But that seems unlikely to me, especially because UnloadComponents() wasn't useful anymore anyway. @ViktorvdValk do you know if UnloadComponents() was wrapped to Python, by ITKElastix? ii) it was verified that it actually fixes a problem (a new test first failed but now passes). It is verified by Konstantinos (@ntatsisk) that it actually fixes a problem. I can still see if an automatic test can be added, based on his use case. sounds good then! sounds good then! @ntatsisk Sorry I did not yet merge, because I'm still trying to write a unit test (GoogleTest) for this pull request. But if you need it now, I can just merge this and write the test later. What do you think? Anyway, I'm just trying (locally on my laptop): GTEST_TEST(ElastixFilter, Parallel) { using ImageType = itk::Image<unsigned char>; std::vector<ImageType::Pointer> images(601); for (auto & image : images) { image = ImageType::New(); image->SetRegions({ 100, 100 }); image->Allocate(true); } #pragma omp parallel for for (int i{}; i < 600; ++i) { const auto elastixFilter = elx::ElastixFilter<ImageType, ImageType>::New(); elastixFilter->SetFixedImage(images[i]); elastixFilter->SetMovingImage(images[i + 1]); elastixFilter->LogToConsoleOff(); elastixFilter->LogToFileOff(); elastixFilter->Update(); } } But it says: User Error 1001: omp_set_num_threads should only be called in serial regions Do you have a suggestion? @ntatsisk Sorry I did not yet merge, because I'm still trying to write a unit test (GoogleTest) for this pull request. But if you need it now, I can just merge this and write the test later. What do you think? Anyway, I'm just trying (locally on my laptop): GTEST_TEST(ElastixFilter, Parallel) { using ImageType = itk::Image<unsigned char>; std::vector<ImageType::Pointer> images(601); for (auto & image : images) { image = ImageType::New(); image->SetRegions({ 100, 100 }); image->Allocate(true); } #pragma omp parallel for for (int i{}; i < 600; ++i) { const auto elastixFilter = elx::ElastixFilter<ImageType, ImageType>::New(); elastixFilter->SetFixedImage(images[i]); elastixFilter->SetMovingImage(images[i + 1]); elastixFilter->LogToConsoleOff(); elastixFilter->LogToFileOff(); elastixFilter->Update(); } } But it says: User Error 1001: omp_set_num_threads should only be called in serial regions Do you have a suggestion? Hi @N-Dekker, I am not in a hurry at all with the merging so it can wait. Regarding the error, I assume you are compiling with "warnings as errors" because I have got the same message but as a warning instead. I used to get it while running in Debug mode but it has been a while since then. Maybe, it is coming from here: https://github.com/SuperElastix/elastix/blob/af382ee4b598ac79fcbb7724bfad7bbcfe492fd6/Components/Optimizers/StandardGradientDescent/itkGradientDescentOptimizer2.cxx#L212-L224 Since this omp_set_num_threads is now called inside the user's omp parallel for directive (like the one that you have in your test script), the warning is emitted. To be honest, I haven't fully understood the internal elastix threading and whether this and the other few OpenMP clauses are necessary/helpful. If they do need to stay then I would try re-writing the clause in the above link as follows: #pragma omp parallel for num_threads(nthreads) Does this work? Btw, even if it does, I think your test will still crash due to the logger being thread-unsafe, right? Hi @N-Dekker, I am not in a hurry at all with the merging so it can wait. Regarding the error, I assume you are compiling with "warnings as errors" because I have got the same message but as a warning instead. I used to get it while running in Debug mode but it has been a while since then. Maybe, it is coming from here: https://github.com/SuperElastix/elastix/blob/af382ee4b598ac79fcbb7724bfad7bbcfe492fd6/Components/Optimizers/StandardGradientDescent/itkGradientDescentOptimizer2.cxx#L212-L224 Since this omp_set_num_threads is now called inside the user's omp parallel for directive (like the one that you have in your test script), the warning is emitted. To be honest, I haven't fully understood the internal elastix threading and whether this and the other few OpenMP clauses are necessary/helpful. If they do need to stay then I would try re-writing the clause in the above link as follows: #pragma omp parallel for num_threads(nthreads) Does this work? Btw, even if it does, I think your test will still crash due to the logger being thread-unsafe, right? Regarding the error, I assume you are compiling with "warnings as errors" because I have got the same message but as a warning instead. I used to get it while running in Debug mode but it has been a while since then. Maybe, it is coming from here: https://github.com/SuperElastix/elastix/blob/af382ee4b598ac79fcbb7724bfad7bbcfe492fd6/Components/Optimizers/StandardGradientDescent/itkGradientDescentOptimizer2.cxx#L212-L224 Thanks for the suggestion, Konstantinos, but as far as I can see, those lines of code are excluded from compilation, as they are part of the "else block" of an #if 1! Anyway, I just submitted a new "ElastixFilter.UpdateInParallel" test: https://github.com/SuperElastix/elastix/pull/393 My aim is to have a little test that just shows that this pull request (#389) really fixes a certain problem. But the test is still unstable! I see it crashing some times. Could you also please have a look? Regarding the error, I assume you are compiling with "warnings as errors" because I have got the same message but as a warning instead. I used to get it while running in Debug mode but it has been a while since then. Maybe, it is coming from here: https://github.com/SuperElastix/elastix/blob/af382ee4b598ac79fcbb7724bfad7bbcfe492fd6/Components/Optimizers/StandardGradientDescent/itkGradientDescentOptimizer2.cxx#L212-L224 Thanks for the suggestion, Konstantinos, but as far as I can see, those lines of code are excluded from compilation, as they are part of the "else block" of an #if 1! Anyway, I just submitted a new "ElastixFilter.UpdateInParallel" test: https://github.com/SuperElastix/elastix/pull/393 My aim is to have a little test that just shows that this pull request (#389) really fixes a certain problem. But the test is still unstable! I see it crashing some times. Could you also please have a look? Ops, my bad! So, it is clearly not the point that is called (maybe this instead?). Anyway, I also noticed that, actually, all the #pragma omp parallel directives are manually deactivated from the entire elastix codebase at this moment. Regarding the crashing, as I mentioned in my last comment, I think it is because of the logger ;) Ops, my bad! So, it is clearly not the point that is called (maybe this instead?). Anyway, I also noticed that, actually, all the #pragma omp parallel directives are manually deactivated from the entire elastix codebase at this moment. Regarding the crashing, as I mentioned in my last comment, I think it is because of the logger ;) Regarding the crashing, as I mentioned in my last comment, I think it is because of the logger ;) I believe so too. The logger is still troublesome. So then, is there still any possible test case for this specific pull request, which only improves the thread-safety of the component database? Regarding the crashing, as I mentioned in my last comment, I think it is because of the logger ;) I believe so too. The logger is still troublesome. So then, is there still any possible test case for this specific pull request, which only improves the thread-safety of the component database?
gharchive/pull-request
2021-01-12T22:33:23
2025-04-01T04:33:06.998357
{ "authors": [ "N-Dekker", "mstaring", "ntatsisk" ], "repo": "SuperElastix/elastix", "url": "https://github.com/SuperElastix/elastix/pull/389", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1686496285
🛑 Matrix redirection is down In 1205ae1, Matrix redirection (https://supersandro.de/_matrix/key/v2/server) was down: HTTP code: 504 Response time: 20158 ms Resolved: Matrix redirection is back up in d1bd924.
gharchive/issue
2023-04-27T09:42:38
2025-04-01T04:33:07.005132
{ "authors": [ "SuperSandro2000" ], "repo": "SuperSandro2000/upptime", "url": "https://github.com/SuperSandro2000/upptime/issues/196", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1832093961
🛑 Ping matrix.supersandro.de is down In 729f784, Ping matrix.supersandro.de (matrix.supersandro.de) was down: HTTP code: 0 Response time: 0 ms Resolved: Ping matrix.supersandro.de is back up in c61fb44.
gharchive/issue
2023-08-01T22:03:52
2025-04-01T04:33:07.008153
{ "authors": [ "SuperSandro2000" ], "repo": "SuperSandro2000/upptime", "url": "https://github.com/SuperSandro2000/upptime/issues/371", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1160422000
Development/algorithm Follow: #47 Done [x] Analysis of Complexion [x] Recursion [x] Divide & Conquer [ ] Hash [x] Backtracking [x] Greedy [ ] Dynamic Programming [x] Sort [ ] Search [ ] String Matching And I re-build the Basic part of the Algorithm. I'll add note for 8-queens in the future.
gharchive/pull-request
2022-03-05T18:51:58
2025-04-01T04:33:07.018737
{ "authors": [ "Jacob953" ], "repo": "Superego-CodeEngineer/GoGetit", "url": "https://github.com/Superego-CodeEngineer/GoGetit/pull/52", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
482149788
Windows 10 version 1903 - Original Windows 8 key not retrieved Hello. On the latest edition of Windows 10 (version 1903), for computers installed directly with Windows 8 keys in a clean install process, and which have been delivered a digital activation, the tool doesn't display the original Windows 8 key and edition (only the Windows 10 generic key is displayed). Is it a known issue? Are you working on it? or is there a limitation in this version of Windows 10 which prevents from obtaining this information? Thank you. Best regards. Hi geolives A clean install removes all traces of the original installaton and the key will not be there - only an upgrade will show the original key. Regards
gharchive/issue
2019-08-19T07:28:28
2025-04-01T04:33:07.020560
{ "authors": [ "Superfly-Inc", "geolives" ], "repo": "Superfly-Inc/ShowKeyPlus", "url": "https://github.com/Superfly-Inc/ShowKeyPlus/issues/23", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2623414555
Imagine ai update button doc Description I have made some changes inside button doc that are following 👉🏻 Front page of button changed totally 👉🏻Updates user experience with BG and hover and more 👉🏻updates toggles buttons color and BG 👉🏻made other changes Fixes #1361 Screenshots BEFORE AFTER Checklist [✅] Tests have been added or updated to cover the changes [✅] Documentation has been updated to reflect the changes [✅] Code follows the established coding style guidelines [✅] All tests are passing You need to show changes on the project you are working, this ain't my project So sorry for having glitches and misunderstanding with these project as soon as I will update you with changes ...
gharchive/pull-request
2024-10-30T09:31:25
2025-04-01T04:33:07.041167
{ "authors": [ "SurajPratap10", "ankitmodanwall" ], "repo": "SurajPratap10/Imagine_AI", "url": "https://github.com/SurajPratap10/Imagine_AI/pull/1362", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
296986237
Node strange issues Here are a few node issues which happened when running - not sure why they happened yet. But tracking them here so we can fix. Closed stream and invalid nonce on standalone node (when only running 1 miner with 2 nodes started up (1 standalone, 2 connected to standalone) Unhandled exception (I think this happened on the standalone node when a node connected to it was stopped) I'll try to recreate the situation. Thanks
gharchive/issue
2018-02-14T06:07:14
2025-04-01T04:33:07.056822
{ "authors": [ "kingsleyh", "tbrand" ], "repo": "SushiCoinHQ/SushiCoin", "url": "https://github.com/SushiCoinHQ/SushiCoin/issues/72", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
976843291
解压后的文件夹会被占用,无法进行其他操作 系统环境: Windows Server 2019 面板版本: 8.6.23 Node 版本: 16.7.0 出现概率: 100 出现的问题 (BUG): 解压后的文件夹会被占用,无法进行其他操作 复现操作步骤: 文件管理>解压 解压到到一半,未完成? 亦或防病毒软件扫描中? 解压到到一半,未完成? 亦或防病毒软件扫描中? 并不是
gharchive/issue
2021-08-23T09:49:37
2025-04-01T04:33:07.068185
{ "authors": [ "SeaIceNX", "huangsijun17" ], "repo": "Suwings/MCSManager", "url": "https://github.com/Suwings/MCSManager/issues/473", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
500742831
Joke type query parameter To only get jokes from a certain type (single or twopart) Example: https://sv443.net/jokeapi/category/Any?type=single Added in 7925adf468b5bf7bf62fc7e7c6f6e910ed9ca2d0
gharchive/issue
2019-10-01T08:30:08
2025-04-01T04:33:07.069540
{ "authors": [ "Sv443" ], "repo": "Sv443/JokeAPI", "url": "https://github.com/Sv443/JokeAPI/issues/16", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
377524998
Generate Table of Contents I'm trying out this tool and it is able generate both separate files and one single AsciiDoctor file, however I'm not seeing the Table of Contents as shown in the Swagger2Markup Example. I didn't see anything in the configuration section, so I feel like I'm missing something. It's an asciidoctor feature.
gharchive/issue
2018-11-05T18:26:23
2025-04-01T04:33:07.095012
{ "authors": [ "RobWin", "jvtrigueros" ], "repo": "Swagger2Markup/swagger2markup-cli", "url": "https://github.com/Swagger2Markup/swagger2markup-cli/issues/24", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2069364206
get memory size Description 增加获取内存大小的函数,为v0.1.4做准备 你在这分支顺带着把system.py里面的print改成swanlog吧
gharchive/pull-request
2024-01-08T00:08:21
2025-04-01T04:33:07.095973
{ "authors": [ "SAKURA-CAT", "xiaolin199912" ], "repo": "SwanHubX/SwanLab", "url": "https://github.com/SwanHubX/SwanLab/pull/169", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2337201943
🛑 swift-format.com is down In 65097b2, swift-format.com (https://swift-format.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: swift-format.com is back up in b996093 after 21 minutes.
gharchive/issue
2024-06-06T02:47:39
2025-04-01T04:33:07.109986
{ "authors": [ "kishikawakatsumi" ], "repo": "SwiftFiddle/status.swift-format.com", "url": "https://github.com/SwiftFiddle/status.swift-format.com/issues/81", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
222488764
Update to swift 3.1 / XCode 8.3 This updates the FileKit and Sourcekitten Dependencies to their current versions to make Cuckoo usable with Xcode 8.3 If I'm not mistaken travis still runs on Xcode 8, not even 8.2. So this would probably need to be updated to at least Xcode 8.2 or better 8.3. Hey, could you give stencil branch a try? It should work with Swift 3.1 and I'll be merging it by the end of this week (I just need more people to try it out). Thanks! Didn't get around to test the stencil branch yesterday. Release 0.9.0 looks good so far. My tests still run and it works with swift 3.1 /Xcode 8.3.2. I will look into the issues I posted to check whether they are fixed with this version in the next few days. So this PR is no longer needed.
gharchive/pull-request
2017-04-18T17:27:02
2025-04-01T04:33:07.119289
{ "authors": [ "TadeasKriz", "rbeeger" ], "repo": "SwiftKit/Cuckoo", "url": "https://github.com/SwiftKit/Cuckoo/pull/120", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }