id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
206609564 | Test broken: predspec
The predspec test fails to build. Please respond to this issue if you have any interest in keeping this test. I am trying to determine who has an interest in maintaining each test, and if no one is interested, I will delete the test.
Build error:
testmapper.cc:36:15: error: no matching function for call to ‘Legion::Mapping::DefaultMapper::DefaultMapper(LegionRuntime::HighLevel::Machine&, LegionRuntime::HighLevel::HighLevelRuntime*&, LegionRuntime::HighLevel::Processor&)’
, runtime(rt)
^
testmapper.cc:36:15: note: candidates are:
In file included from testmapper.h:2:0,
from testmapper.cc:1:
/home/eslaught/legion/runtime/mappers/default_mapper.h:113:7: note: Legion::Mapping::DefaultMapper::DefaultMapper(const Legion::Mapping::DefaultMapper&)
DefaultMapper(const DefaultMapper &rhs);
^
/home/eslaught/legion/runtime/mappers/default_mapper.h:113:7: note: candidate expects 1 argument, 3 provided
/home/eslaught/legion/runtime/mappers/default_mapper.h:111:7: note: Legion::Mapping::DefaultMapper::DefaultMapper(Legion::Mapping::MapperRuntime*, Legion::Machine, Legion::Processor, const char*)
DefaultMapper(MapperRuntime *rt, Machine machine, Processor local,
^
/home/eslaught/legion/runtime/mappers/default_mapper.h:111:7: note: no known conversion for argument 1 from ‘LegionRuntime::HighLevel::Machine {aka Realm::Machine}’ to ‘Legion::Mapping::MapperRuntime*’
testmapper.cc: In member function ‘virtual void TestMapper::select_task_options(LegionRuntime::HighLevel::Task*)’:
testmapper.cc:109:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘inline_task’
task->inline_task = false;
^
testmapper.cc:110:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘spawn_task’
task->spawn_task = false;
^
testmapper.cc:111:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘map_locally’
task->map_locally = true;
^
testmapper.cc:112:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘profile_task’
task->profile_task = false;
^
testmapper.cc:113:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘task_priority’
task->task_priority = 0;
^
testmapper.cc:115:9: error: ‘LegionRuntime::HighLevel::Task’ has no member named ‘additional_procs’
task->additional_procs.insert(procs[node].begin(), procs[node].end());
^
testmapper.cc: In member function ‘virtual void TestMapper::notify_mapping_result(const Mappable*)’:
testmapper.cc:140:32: error: ‘const Mappable’ has no member named ‘as_mappable_task’
const Task *task = mappable->as_mappable_task();
^
testmapper.cc:147:49: error: ‘const struct Legion::RegionRequirement’ has no member named ‘selected_memory’
<< ") mapped on " << task->regions[idx].selected_memory
^
@streichler Have you decided whether it would be beneficial to keep this test?
This test has been removed as of d0fb5b8d04cc6876612db969b67f7b4e66fef711. If you are interested in resurrecting this test, please reply to this issue.
| gharchive/issue | 2017-02-09T19:44:23 | 2025-04-01T04:33:06.245762 | {
"authors": [
"elliottslaughter"
],
"repo": "StanfordLegion/legion",
"url": "https://github.com/StanfordLegion/legion/issues/223",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1172421012 | Could you please update your tutorial for ros integration?
the demo is using ros1, turtlebot1 and python 2.7, which are all out of date.
By using miniconda env based on python2.7, you even can not properly install igibson2!!
iGibson in its current form is not compatible with Python 2.7. We are looking into re-enabling backwards compatibility, but are you able to switch to a Python 3 version of ROS? Would updating the tutorial to work with that be helpful for you?
that would be of great favor for me.
@cgokmen and @MRWANG995 We have integrated ROS noetic (Python 3) with igibson 2. We are able to run the robot with move_base ROS navigation stack (Turtlebot and Fetch). Let us know if you need any help. We will happy to share our learnings with you :) . It may also help you to update the tutorial. Let us know :)
Hi @vbalaji21 , please, try the new version we just released and tell us whether it presents any problem with you. Thanks!
The release has not gone through yet - we'll push today. Will update here.
@roberto-martinmartin and @cgokmen . Thanks a lot for letting me know. Please update me once it is pushed, I will try it out.
This week, I had some problems with loading Pedestrians. I made some fixes and it is working now. I am not sure whether that is also taken into account in the new release. If not, I can share you the fixes I did. My aim (along with Prof. Rachid and Phanu Teja in LAAS, CNRS) is to study human aware social navigation and later extend with learning capabilities.
My next 2 steps are -
Check VR immersion of human and arm (head data) to conduct user studies with our planner
Import a UAV into iGibson
We are considering iGibson as our potential future platform for most of our development. We are hoping to do the above tasks as soon as possible and continue our development. We would be very happy to get any information or help on these 2 tasks.
We definitely have well-maintained and functional VR capability right now (this is the feature we ourselves have used the most). You can try the VR demo in https://github.com/StanfordVL/iGibson/blob/master/igibson/examples/vr/vr_simple_demo.py using a SteamVR-installed computer (this demo was tested using a HTC Vive headset and Valve Index controller but I believe we support most devices).
As for the UAV we also have https://github.com/StanfordVL/iGibson/blob/master/igibson/robots/legacy/quadrotor_robot.py but this is a not-so-realistic quadrotor that used to be functional but is currently not maintained (e.g. it will probably take some code editing to make it work with the new Robot interface). Should be simple enough - the above file should end up looking like this one: https://github.com/StanfordVL/iGibson/blob/master/igibson/robots/ant.py
Oh, also https://github.com/utiasDSL/gym-pybullet-drones has some interesting pybullet-based UAV simulation stuff (dynamics, control) that perhaps you could try using on top of iGibson (most of the code should be compatible and not too hard to hook up)
We have just cut the release for iGibson 2.2.0 - please let us know if you need anything else, and feel free to reopen this thread if you have any further questions.
Thanks a lot for the update and suggestions regarding UAV and VR.
I have 2 questions -
How to make any robot visible in REGISTERED_ROBOTS? I haave attached a screenshot of this error. Please let me know.
I tried importing UAV hector_quadcopter https://github.com/tu-darmstadt-ros-pkg/hector_quadrotor yesterday from here.
It is on ROS kinetic but tried porting it for ROS noetic.
I found that the controller classes for UAV needs to be ported.
I did not find controller classes for UAV in iGibson 1.0 as well.
I saw that you are using locomotion robot class and basecontroller for UAV if I am correct.
Does your comment on the realistic/functionality of Quadrotor mean that it doesn't work properly in iGibson 1.0 or the controller is primitive to make the Quadrotor work realistically ?
Thanks a lot for your comment about the Quadrorot/UAV suggestion. It will save a lot of time for us in the decision :)
Hi @vbalaji21 , please, try the new version we just released and tell us whether it presents any problem with you. Thanks!
@roberto-martinmartin and @cgokmen , I tried the new version by installing iGibson in the compiling from source method but it failed to build in pip install -e . com
Let me know
mand.
I have attached the error log screenshot below.
could it be that you didn't git clone recursively?
@roberto-martinmartin Thank you the problem was indeed I didn't do git clone recursively.
I have another doubt, I followed the instructions but I did not understand about the Python Path step as it has 4 paths and I don't know in which file I need to update them. Can you please let me know ?
I think I am facing this problem with the path (Python path).
I have attached the screenshot below.
| gharchive/issue | 2022-03-17T13:56:03 | 2025-04-01T04:33:06.259664 | {
"authors": [
"MRWANG995",
"cgokmen",
"roberto-martinmartin",
"vbalaji21"
],
"repo": "StanfordVL/iGibson",
"url": "https://github.com/StanfordVL/iGibson/issues/179",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
991667150 | TilePos and relatives should use usize, not u32
This is an important perf and usability optimization when interfacing with external arrays (which are always indexed by usize slices).
Currently, you must repeatedly waste operations performing this conversion and littering the code base with as usize.
FYI there is no perf cost; integer type conversions compile to no-ops, there are no extra operations. There is no such thing as integer type conversion at the assembly / machine code level.
This is just to make Rust code prettier and more correct.
I think this is no longer relevant, but feel free to reopen if I am wrong.
| gharchive/issue | 2021-09-09T01:36:36 | 2025-04-01T04:33:06.261763 | {
"authors": [
"StarArawn",
"alice-i-cecile",
"inodentry"
],
"repo": "StarArawn/bevy_ecs_tilemap",
"url": "https://github.com/StarArawn/bevy_ecs_tilemap/issues/95",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1591557504 | [balancer] show tablet info based on be path
Feature request
Is your feature request related to a problem? Please describe.
Describe the solution you'd like
Describe alternatives you've considered
Additional context
fixed by #18457
will displayed all these info in information_schema db
| gharchive/issue | 2023-02-20T10:20:23 | 2025-04-01T04:33:06.280985 | {
"authors": [
"nshangyiming"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/issues/18114",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1741225014 | 需求:新增对已经存在的表修改comment的功能
请问什么时候有计划增加对已经存在的表修改comment的功能,类似mysql:ALTER TABLE 表名称 COMMENT '***';
Already supported
https://github.com/StarRocks/starrocks/pull/21035
| gharchive/issue | 2023-06-05T08:11:49 | 2025-04-01T04:33:06.282208 | {
"authors": [
"Astralidea",
"qy58644781"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/issues/24685",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1057271757 | Support CTAS(create table as select) statement
CTAS is used to simplify user to create table and insert data. The syntax is like the following.
CREATE TABLE [IF NOT EXISTS] [database.]table_name
[(column_name[, column_name, ...]]
[COMMENT "table comment"]
[partition_desc]
[distribution_desc]
[PROPERTIES ("key"="value", ...)]
AS SELECT <query>
[ ... ]
The tips for the CTAS statement.
The default model is a duplicate-key table.
All char/varchar type will create as string in the new table.
The replication_num of table will be equal the the maximum number of all tables selected.
When choosing hash distributed key, If we have statistics cache, we pick the column with the highest cardinality according the column statistics. Otherwise, pick the first column.
CTAS only support new CBO based planner and implement some createTable analyzer.
run all
run all
| gharchive/pull-request | 2021-11-18T11:57:58 | 2025-04-01T04:33:06.284689 | {
"authors": [
"Astralidea"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/1392",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1617344913 | [BugFix] Release workgroup token immediately when fragment is cancelled
What type of PR is this:
[x] BugFix
[ ] Feature
[ ] Enhancement
[ ] Refactor
[ ] UT
[ ] Doc
[ ] Tool
Which issues of this PR fixes :
Fixes #
TODO.
Problem Summary(Required) :
Release workgroup token immediately when fragment is cancelled.
Otherwise, it can only be released after query_delivery_timeout (5min by default) sometimes.
Handle error results and cancel fragment instances, excluding TIMEOUT errors, until all the delivered fragment instances are completed.
Otherwise, the cancellation RPC may arrive at BE before the delivery fragment instance RPC, causing the instances and drivers to become stale and only able to be released after query timeout.
Checklist:
[x] I have added test cases for my bug fix or my new feature
[ ] This pr will affect users' behaviors
[ ] This pr needs user documentation (for new or modified features or behaviors)
[ ] I have added documentation for my new feature or new function
Bugfix cherry-pick branch check:
[x] I have checked the version labels which the pr will be auto backported to target branch
[x] 3.0
[x] 2.5
[x] 2.4
[x] 2.3
run starrocks_be_unittest
run starrocks_be_unittest
run starrocks_be_unittest
run starrocks_admit_test
run starrocks_fe_unittest
run starrocks_fe_unittest
run starrocks_fe_unittest
@Mergifyio backport branch-2.5
@Mergifyio backport branch-2.4
@Mergifyio backport branch-2.3
| gharchive/pull-request | 2023-03-09T14:16:53 | 2025-04-01T04:33:06.292898 | {
"authors": [
"ZiheLiu",
"wanpengfei-git"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/19310",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1674777115 | [BugFix] fix query stuck issue in SpillableAggregateBlockingSinkOperator
What type of PR is this:
[x] BugFix
[ ] Feature
[ ] Enhancement
[ ] Refactor
[ ] UT
[ ] Doc
[ ] Tool
Which issues of this PR fixes :
Fixes #
Problem Summary(Required) :
Checklist:
[ ] I have added test cases for my bug fix or my new feature
[ ] This pr will affect users' behaviors
[ ] This pr needs user documentation (for new or modified features or behaviors)
[ ] I have added documentation for my new feature or new function
Bugfix cherry-pick branch check:
[x] I have checked the version labels which the pr will be auto backported to target branch
[x] 3.0
[ ] 2.5
[ ] 2.4
[ ] 2.3
@Mergifyio backport branch-3.0
| gharchive/pull-request | 2023-04-19T12:11:23 | 2025-04-01T04:33:06.297832 | {
"authors": [
"silverbullet233",
"wanpengfei-git"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/21943",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1745118062 | [BugFix] Fix bug drop not exists task NPE
Fixes #24741
What type of PR is this:
[x] BugFix
[ ] Feature
[ ] Enhancement
[ ] Refactor
[ ] UT
[ ] Doc
[ ] Tool
Checklist:
[x] I have added test cases for my bug fix or my new feature
[ ] This pr will affect users' behaviors
[ ] This pr needs user documentation (for new or modified features or behaviors)
[ ] I have added documentation for my new feature or new function
Bugfix cherry-pick branch check:
[x] I have checked the version labels which the pr will be auto-backported to the target branch
[x] 3.0
[x] 2.5
[ ] 2.4
[ ] 2.3
[FE PR Coverage Check]
:disappointed: fail : 0 / 4 (00.00%)
file detail
path
covered_line
new_line
coverage
not_covered_line_detail
:large_blue_circle:
com/starrocks/qe/DDLStmtExecutor.java
0
4
00.00%
[877, 878, 879, 881]
@Mergifyio backport branch-3.0
@Mergifyio backport branch-2.5
| gharchive/pull-request | 2023-06-07T06:15:23 | 2025-04-01T04:33:06.304277 | {
"authors": [
"Astralidea",
"wanpengfei-git"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/24793",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1800541975 | [Enhancement] generic range
Fixes #issue
Generic range so that it can be used for external file reader, such as parquet reader which row_index is uint64
What type of PR is this:
[ ] BugFix
[ ] Feature
[x] Enhancement
[ ] Refactor
[ ] UT
[ ] Doc
[ ] Tool
Checklist:
[ ] I have added test cases for my bug fix or my new feature
[ ] This pr will affect users' behaviors
[ ] This pr needs user documentation (for new or modified features or behaviors)
[ ] I have added documentation for my new feature or new function
Bugfix cherry-pick branch check:
[x] I have checked the version labels which the pr will be auto-backported to the target branch
[ ] 3.1
[ ] 3.0
[ ] 2.5
[ ] 2.4
[FE PR Coverage Check]
:heart_eyes: pass : 0 / 0 (0%)
[FE PR Coverage Check]
:heart_eyes: pass : 0 / 0 (0%)
[FE PR Coverage Check]
:heart_eyes: pass : 0 / 0 (0%)
| gharchive/pull-request | 2023-07-12T09:16:56 | 2025-04-01T04:33:06.310143 | {
"authors": [
"wanpengfei-git",
"zombee0"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/27056",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1823641658 | [BugFix] primary key table inconsistent snapshot for full clone and backup caused by compaction(#28062)
Problem:
When we make snapshot for full clone and backup, _lock will not be holded during get_applied_rowsets and make_snapshot_on_tablet_meta. This cause a problem, If compaction happen between get_applied_rowsets and make_snapshot_on_tablet_meta, the delvector get from make_snapshot_on_tablet_meta maybe generated by a larger minor version which is inconsistent with the rowset get from get_applied_rowsets.
Solution:
hold _lock during get_applied_rowsets and make_snapshot_on_tablet_meta
Fixes #issue
What type of PR is this:
[x] BugFix
[ ] Feature
[ ] Enhancement
[ ] Refactor
[ ] UT
[ ] Doc
[ ] Tool
Checklist:
[x] I have added test cases for my bug fix or my new feature
[ ] This pr will affect users' behaviors
[ ] This pr needs user documentation (for new or modified features or behaviors)
[ ] I have added documentation for my new feature or new function
Bugfix cherry-pick branch check:
[x] I have checked the version labels which the pr will be auto-backported to the target branch
[x] 3.1
[x] 3.0
[x] 2.5
[ ] 2.4
[FE PR Coverage Check]
:heart_eyes: pass : 0 / 0 (0%)
[FE PR Coverage Check]
:heart_eyes: pass : 0 / 0 (0%)
[FE PR Coverage Check]
:heart_eyes: pass : 0 / 0 (0%)
[FE PR Coverage Check]
:heart_eyes: pass : 0 / 0 (0%)
@Mergifyio backport branch-3.1
@Mergifyio backport branch-3.0
@Mergifyio backport branch-2.5
[FE PR Coverage Check]
:heart_eyes: pass : 0 / 0 (0%)
| gharchive/pull-request | 2023-07-27T05:29:59 | 2025-04-01T04:33:06.318663 | {
"authors": [
"srlch",
"wanpengfei-git"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/28064",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1110438393 | Put chunks of different driver_sequences in one request
Background
When enable pipeline level shuffle, chunks will hash to N groups (N is the dop of dest pipeline). Each group will build the request independently.
When sending large amounts of data, the number of requests sent will not change significantly because each group will accumulate chunk until it exceeds a threshold before sending
When sending small amounts of data, take the extreme case for example, assume the amount of data just fills one chunk, then each group will send one packet, the packets to be sent changed from 1 to N which may lead to significant performance drop
Solution
Add an auxiliary data structure PTransmitChunkParams::driver_sequences, indicating the driver_sequence to which the i-th chunk belongs
Test
3be(64c/128G)
1fe(64c/128G)
tpcds with scale factor=100
Almost all the querys have performance improvements, and a few are briefly listed as below.
TPCDS
Before
After
Q12
0.299s
0.19s
Q15
0.378s
0.25s
Q20
0.25s
0.16s
Q45
0.442s
0.305s
run starrocks_be_unittest
create an issue and link it to this pr
| gharchive/pull-request | 2022-01-21T12:53:58 | 2025-04-01T04:33:06.324018 | {
"authors": [
"liuyehcf",
"satanson"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/3003",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1881963504 | [BugFix] Fix source_scan_bytes compability with 3.0
Fixes #issue
What type of PR is this:
[x] BugFix
[ ] Feature
[ ] Enhancement
[ ] Refactor
[ ] UT
[ ] Doc
[ ] Tool
Does this PR entail a change in behavior?
[ ] Yes, this PR will result in a change in behavior.
[x] No, this PR will not result in a change in behavior.
If yes, please specify the type of change:
[ ] Interface/UI changes: syntax, type conversion, expression evaluation, display information
[ ] Parameter changes: default values, similar parameters but with different default values
[ ] Policy changes: use new policy to replace old one, functionality automatically enabled
[ ] Feature removed
[ ] Miscellaneous: upgrade & downgrade compatibility, etc.
Checklist:
[ ] I have added test cases for my bug fix or my new feature
[ ] This pr needs user documentation (for new or modified features or behaviors)
[ ] I have added documentation for my new feature or new function
Bugfix cherry-pick branch check:
[x] I have checked the version labels which the pr will be auto-backported to the target branch
[x] 3.1
[ ] 3.0
[ ] 2.5
[ ] 2.4
[BE Incremental Coverage Report]
:heart_eyes: pass : 0 / 0 (0%)
@Mergifyio backport branch-3.1
| gharchive/pull-request | 2023-09-05T13:16:17 | 2025-04-01T04:33:06.330366 | {
"authors": [
"meegoo",
"wanpengfei-git"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/30433",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1148043329 | Fix Scalar Functions bugs (#3631)
What type of PR is this:
[ ] bug
[ ] feature
[ ] enhancement
[ ] others
Which issues of this PR fixes :
Fixes #
Problem Summary(Required) :
run starrocks_fe_unittest
| gharchive/pull-request | 2022-02-23T12:41:35 | 2025-04-01T04:33:06.332446 | {
"authors": [
"Seaven"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/3645",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2263042926 | [BugFix] [UT] Distinguish TaskRun by using unique taskRunId
Why I'm doing:
submitTaskRegularTest is not stable
mv sync mode refresh may hang until timeout
What I'm doing:
TaskRun's equals method only checks taskId and its definition but one task may contains many task runs. It may introduce bug if we don't distinguish those task runs.
Use taskRunId to distinguish TaskRun
@Override
public boolean equals(Object o) {
if (this == o) {
return true;
}
if (o == null || getClass() != o.getClass()) {
return false;
}
if (task.getDefinition() == null) {
return false;
}
TaskRun taskRun = (TaskRun) o;
return this.taskId == taskRun.getTaskId() &&
this.task.getDefinition().equals(taskRun.getTask().getDefinition());
}
Fixes #issue
What type of PR is this:
[x] BugFix
[ ] Feature
[ ] Enhancement
[ ] Refactor
[x] UT
[ ] Doc
[ ] Tool
Does this PR entail a change in behavior?
[ ] Yes, this PR will result in a change in behavior.
[x] No, this PR will not result in a change in behavior.
If yes, please specify the type of change:
[ ] Interface/UI changes: syntax, type conversion, expression evaluation, display information
[ ] Parameter changes: default values, similar parameters but with different default values
[ ] Policy changes: use new policy to replace old one, functionality automatically enabled
[ ] Feature removed
[ ] Miscellaneous: upgrade & downgrade compatibility, etc.
Checklist:
[x] I have added test cases for my bug fix or my new feature
[ ] This pr needs user documentation (for new or modified features or behaviors)
[ ] I have added documentation for my new feature or new function
[ ] This is a backport pr
Bugfix cherry-pick branch check:
[x] I have checked the version labels which the pr will be auto-backported to the target branch
[x] 3.3
[x] 3.2
[ ] 3.1
[ ] 3.0
[ ] 2.5
@mergify rebase
| gharchive/pull-request | 2024-04-25T08:50:52 | 2025-04-01T04:33:06.339798 | {
"authors": [
"LiShuMing"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/44748",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2400604034 | [BugFix] Fix possible dead lock in LoadJobStatsListener
Why I'm doing:
What I'm doing:
Fix protential dead lock bug: try exception to release write lock finally in [InsertOverwriteJobRunner.java](https://github.com/StarRocks/starrocks/pull/48150/files#diff-9ba512541351f603a9bce0ba274658f639eafebe2fbf635d7ac26072f6ff4c81)
No use lock in `LoadJobMVListener.java
https://github.com/StarRocks/StarRocksTest/issues/8091
What type of PR is this:
[x] BugFix
[ ] Feature
[ ] Enhancement
[ ] Refactor
[ ] UT
[ ] Doc
[ ] Tool
Does this PR entail a change in behavior?
[ ] Yes, this PR will result in a change in behavior.
[x] No, this PR will not result in a change in behavior.
If yes, please specify the type of change:
[ ] Interface/UI changes: syntax, type conversion, expression evaluation, display information
[ ] Parameter changes: default values, similar parameters but with different default values
[ ] Policy changes: use new policy to replace old one, functionality automatically enabled
[ ] Feature removed
[ ] Miscellaneous: upgrade & downgrade compatibility, etc.
Checklist:
[x] I have added test cases for my bug fix or my new feature
[ ] This pr needs user documentation (for new or modified features or behaviors)
[ ] I have added documentation for my new feature or new function
[ ] This is a backport pr
Bugfix cherry-pick branch check:
[x] I have checked the version labels which the pr will be auto-backported to the target branch
[x] 3.3
[ ] 3.2
[ ] 3.1
[ ] 3.0
[ ] 2.5
ignore backport check 3.3.1
| gharchive/pull-request | 2024-07-10T12:16:25 | 2025-04-01T04:33:06.347609 | {
"authors": [
"LiShuMing"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/48150",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1211849838 | [Enhancement] Check user identity for show grants & support string literal role name for create role and drop role
What type of PR is this:
[x] bug
[ ] feature
[ ] enhancement
[ ] others
Which issues of this PR fixes :
Fixes #5389
Problem Summary(Required) :
So far, using show grants for a non-existent user will show a table with all cells 'null'. Fix it by raising an exception in this scenario.
Allow usage of create role 'role_name' by changing role CUP type from ident to ident_or_text.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
[FE PR Coverage check]
:heart_eyes: pass : 2 / 2 (100.00%)
file detail
path
covered line
new line
coverage
:large_blue_circle:
com/starrocks/analysis/ShowGrantsStmt.java
2
2
100.00%
| gharchive/pull-request | 2022-04-22T05:44:24 | 2025-04-01T04:33:06.353862 | {
"authors": [
"CLAassistant",
"padmejin",
"wanpengfei-git"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/5390",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1261821622 | [cherry-pick] Get hive column statistics downgrade policy (#4409) (#6…
…899)
What type of PR is this:
[ ] bug
[ ] feature
[ ] enhancement
[ ] refactor
[x] others
Which issues of this PR fixes :
Fixes #
Problem Summary(Required) :
run starrocks_fe_unittest
run starrocks_fe_unittest
| gharchive/pull-request | 2022-06-06T13:06:27 | 2025-04-01T04:33:06.356517 | {
"authors": [
"stephen-shelby",
"wangruin"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/6926",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1274416996 | [Bugfix] Avoid using released resources for ContextWithDependency after closing (backport #7363)
What type of PR is this:
[ ] bug
[ ] feature
[ ] enhancement
[ ] refactor
[ ] others
Problem Summary(Required) :
This is cherry-picked from #7363.
run all
run all
run all
| gharchive/pull-request | 2022-06-17T02:28:27 | 2025-04-01T04:33:06.358776 | {
"authors": [
"ZiheLiu",
"mofeiatwork"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/7394",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2279226497 | Use Localization for equipment on actors
It's currently hardcoded.
https://github.com/StarWarsFoundryVTT/StarWarsFFG/blob/dde154ea820d7c31d147b1402f38504222b60853/templates/parts/actor/ffg-weapon-armor-gear.html#L30
Translations already exist.
SWFFG.Equipped
SWFFG.Unequipped
| gharchive/issue | 2024-05-04T22:45:20 | 2025-04-01T04:33:06.360259 | {
"authors": [
"TeddyBears",
"wrycu"
],
"repo": "StarWarsFoundryVTT/StarWarsFFG",
"url": "https://github.com/StarWarsFoundryVTT/StarWarsFFG/issues/1469",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1859296128 | Add MSRV to readme
I think that good to put MSRV in the README.
Fix somehow https://github.com/matklad/once_cell/issues/201
no idea how I end up doing the PR on the fork haha github betray me
| gharchive/pull-request | 2023-08-21T12:44:35 | 2025-04-01T04:33:06.381872 | {
"authors": [
"Stargateur"
],
"repo": "Stargateur/once_cell",
"url": "https://github.com/Stargateur/once_cell/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2140153899 | Use transaction_connection_validator from Sequel
Sounds like a good thing: https://github.com/jeremyevans/sequel/commit/c31dde28e79f5d67398c2ab849cedc84e45dd386
Needs https://github.com/Starkast/wikimum/pull/579
| gharchive/issue | 2024-02-17T14:44:12 | 2025-04-01T04:33:06.383307 | {
"authors": [
"dentarg"
],
"repo": "Starkast/wikimum",
"url": "https://github.com/Starkast/wikimum/issues/589",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1104255038 | Progress not properly displaying for explorer synced health check
The data is coming back correctly but for some reason it's showing as 0.00%.
{"bitcoinStatus":{"blocks":718729,"headers":718729,"verificationProgress":0.9999970594948908,"isSynched":true,"incrementalRelayFee":1.0,"minRelayTxFee":1.0,"externalAddresses":["<redacted>.onion:8333"],"capabilities":{"canScanTxoutSet":true,"canSupportSegwit":true,"canSupportTransactionCheck":true}},"repositoryPingTime":0.0,"isFullySynched":false,"chainHeight":718729,"syncHeight":715185,"instanceName":"","networkType":"Mainnet","cryptoCode":"BTC","supportedCryptoCodes":["BTC"],"version":"2.1.52.0"}
Fixed in: https://github.com/Start9Labs/btcpayserver-wrapper/pull/9/commits/d6c0775848e9fcf151d85fe4c0c9fd0cc7f65c33
| gharchive/issue | 2022-01-15T01:15:59 | 2025-04-01T04:33:06.395176 | {
"authors": [
"elvece"
],
"repo": "Start9Labs/btcpayserver-wrapper",
"url": "https://github.com/Start9Labs/btcpayserver-wrapper/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1263252264 | [feat]: 🛑 Stop using root account for EmbassyOS
Prerequisites
[X] I have searched for existing issues that already suggest this feature, without success.
Describe the Feature Request
With Great Power Comes Great Responsibility.
We should transition away from using root in EmbassyOS and instead, use the start9 user as a privileged account thus limiting the potential system damage.
Describe the Use Case
it is inevitable 😈
Describe Preferred Solution
No response
Describe Alternatives
No response
Anything else?
Good security practices:
Root account should be used to modify the underlying system.
Using the system should be done by the user account.
User can elevate his rights to become root.
As an operating system, embassyd should be running as root. It pretty much has no other choice. We are already transitioning away from embassy-cli and embassy-sdk requiring root though.
We are already transitioning away from embassy-cli and embassy-sdk requiring root though.
That is good news.
Filebrowser would be the natural next contender. Files should belong to start9 user so that you could mount filebrowser data directory via sftp and move those TB of images, movies and documents using systems file manager or simple SFTP client software.
Good point, it probably should be the case that all files are owned by the start9 user and thats the user that the containers run as as well
closed by #1227
This is complete, no?
| gharchive/issue | 2022-06-07T12:56:00 | 2025-04-01T04:33:06.400468 | {
"authors": [
"ProofOfKeags",
"dr-bonez",
"k0gen",
"kn0wmad"
],
"repo": "Start9Labs/embassy-os",
"url": "https://github.com/Start9Labs/embassy-os/issues/1503",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
857548807 | Compatibility with Font Awesome Pro
If I import FA pro the arrows on the left menu are replaced with the text "F105" (when pointing left/collapsed) and "F107" (when pointing down/expanded). If I bring in the free version of FA and remove the pro I get my arrows back but then I lose the pro icons.
Is there a quick way around this that I'm missing?
Are you using the same version of Font Awesome? The classes used in FA4 are not the same as FA5. The pro version also has a different font family than the free version in FA5, so you can't just swap the font files - you have to replace the CSS file as well.
Font Awesome Pro uses a different font family, so if you want to use that, you'll need to edit either the Sass file or the compiled CSS, depending on which you're using.
For the Sass version, the Font Awesome bits for the sidebar start here: https://github.com/StartBootstrap/startbootstrap-sb-admin-2/blob/master/scss/navs/_sidebar.scss#L104
For the CSS version, you'd start here: https://github.com/StartBootstrap/startbootstrap-sb-admin-2/blob/master/css/sb-admin-2.css#L10096
I imagine a find and replace would work for either.
Looks like that did it. Thank you so much!
| gharchive/issue | 2021-04-14T05:57:54 | 2025-04-01T04:33:06.403539 | {
"authors": [
"sean8300",
"tiesont"
],
"repo": "StartBootstrap/startbootstrap-sb-admin-2",
"url": "https://github.com/StartBootstrap/startbootstrap-sb-admin-2/issues/391",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
803807727 | VS Code config issue
https://github.com/StatCan/StatCan.OrchardCore/blob/master/.vscode/launch.json#L40
This should never be set to Production while you are developping with VS Code.
If you are using Azure Shells settings this could potentially trigger update of the prod database.
@jptissot I think you should update these files with the latest ones from the main OC repo. I fixed them.
Sure ! Thanks :)
| gharchive/issue | 2021-02-08T18:16:37 | 2025-04-01T04:33:06.406642 | {
"authors": [
"Skrypt",
"jptissot"
],
"repo": "StatCan/StatCan.OrchardCore",
"url": "https://github.com/StatCan/StatCan.OrchardCore/issues/188",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2247482575 | Welcome to Charlie
Welcome to our project! As we're just getting started, your input is crucial. Feel free to post your first issue, whether it’s a feature request, a bug you've noticed, or a suggestion for our documentation. This helps us prioritize next steps and understand who’s interested in contributing. Excited to see your ideas and get you onboard!
Issue #2 is an example of what a good issue post entails. Should have a description or an idea and steps to get there and what to include. Also include any possible references or documentation that can help if available.
Be creative with your post! It doesn't matter if your idea seems farfetched or unrealistic, all posts are appreciated and all active developers can comment and discuss for more clarification or interest in collaboration on your idea. This could be expanding on your issue or to help guide you away from a potential problem your idea could bring.
Who know, maybe you have a great issue, but don't know exactly where to start. Make a post and we can have a discussion!
| gharchive/issue | 2024-04-17T06:26:07 | 2025-04-01T04:33:06.457774 | {
"authors": [
"Stavdel"
],
"repo": "Stavdel/CHARLIE-AI",
"url": "https://github.com/Stavdel/CHARLIE-AI/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2630507138 | [Bug] Linux version crashing at everytime the command line is used to execute it
Before Reporting
[X] I found no existing issues matching my bug
[X] My issue is not caused by a theme
[X] My issue is not caused by a plugin
Describe the Bug
using the command ~/.millennium/start.sh
to turn millenium on results in a steam crash every single time, i have no idea why
Expected Behavior
everything has been installed properly so it should run fine, but... nope.
Steps To Reproduce
use the automatic install prompt from millennium website for linux
install curl with pacman (i use arch btw)
run the start command
crash
Operating System
Linux
Anything else?
nothing else to say here
I'll take a look!
Give this a go: https://github.com/SteamClientHomebrew/Millennium/actions/runs/11645070964. just extract it to ~/.millennium replacing all
fixed
iam having the same issue startup crash . tried ur method but still its crashing
| gharchive/issue | 2024-11-02T15:01:42 | 2025-04-01T04:33:06.464990 | {
"authors": [
"Xero646",
"dougbug589",
"shdwmtr"
],
"repo": "SteamClientHomebrew/Millennium",
"url": "https://github.com/SteamClientHomebrew/Millennium/issues/153",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
88755234 | Graphic bug with multiple raining gold
Abilities seems to work well, constantly re-using medic in the right way. Raining Gold, however, seems to have some sort of issue.
Issue:
Start the game with extra 100+ points to be able to spend on free abilities.
Buy multiple (10x) raining gold to use when you get to bosses at level 60+.
Load the script.
Watch your raining gold go missing leaving a blank square that does nothing (but I assume is where raining gold would be).
Reloading the game shows the remaining raining gold correctly (10x).
Do not know if this is cosmetic or affects the way the script functions.
This means it's being disabled, i noticed that too in 4.6.3 but then it was enabled again later permanently...
Missing space means it's disabled.
| gharchive/issue | 2015-06-16T15:10:10 | 2025-04-01T04:33:06.467660 | {
"authors": [
"Cavey",
"K4r0qtuYNE5G4qgZ",
"xPaw"
],
"repo": "SteamDatabase/steamSummerMinigame",
"url": "https://github.com/SteamDatabase/steamSummerMinigame/issues/225",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
983733681 | Pixel-Change with Tab Key
When having multiple pixel per player, it would be great, if the player could change the pixel with the tab key.
Implemented in v2.4.0.
| gharchive/issue | 2021-08-31T11:07:07 | 2025-04-01T04:33:06.474280 | {
"authors": [
"Steditor",
"arneboockmeyer"
],
"repo": "Steditor/yard",
"url": "https://github.com/Steditor/yard/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
893270184 | Add "Required Framework" to API help
Add framework requirements to API help.
Sample:
The following section has a list of supported identifiers for the comma-separated
list of "dependencies".
+--------------------+--------------------------------------+---------------------------+--------------------+
| Id | Description | Required Steeltoe version | Required Framework |
+--------------------+--------------------------------------+---------------------------+--------------------+
...
Resolved by 024efa3
| gharchive/issue | 2021-05-17T12:15:10 | 2025-04-01T04:33:06.475527 | {
"authors": [
"ccheetham"
],
"repo": "SteeltoeOSS/InitializrApi",
"url": "https://github.com/SteeltoeOSS/InitializrApi/issues/86",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
238377716 | Can't connect to Eureka for service registry
I have a running instance of Eureka 1.7.0 running on Tomcat 8.5, when I surf to Eureka the management screen is working fine. I downloaded the Eureka directly from Netflix OSS.
When I try to connect with a simple ASP.NET Core Web API project it does not work. I get the following error in the output screen: "Steeltoe.Discovery.Eureka.DiscoveryClient:Error: HeartBeat failed"
Is there a way to get more info about why the heartbeat failed? or is steeltoe not compatible with Eureka 1.7.0?
This is my appsettings.json:
{ "Logging": { "IncludeScopes": false, "LogLevel": { "Default": "Information", "Pivotal": "Debug", "Steeltoe": "Information" } }, "spring": { "application": { "name": "EurekaDiscoveryService" } }, "eureka": { "client": { "serviceUrl": "http://localhost:8080/", "shouldFetchRegistry": false }, "instance": { "port": 50612 } } }
Change from "Information" to "Debug" for the Steeltoe category
| gharchive/issue | 2017-06-25T12:25:14 | 2025-04-01T04:33:06.478582 | {
"authors": [
"dtillman",
"pieterderycke"
],
"repo": "SteeltoeOSS/Samples",
"url": "https://github.com/SteeltoeOSS/Samples/issues/16",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
861218110 | app.UseEndpoints should be called once in app
Describe the bug
app.UseEndpoints should not be called multiple times, see https://github.com/dotnet/aspnetcore/issues/17750#issuecomment-564508302
Expected behavior
No additional calls to endpoints
Environment (please complete the following information):
Steeltoe Version 2.3.0
Platform: CF
OS: Windows
.NET Version .NET Core 3.1
Screenshots
If applicable, add screenshots to help explain your problem.
Additional context or links
Add any other context about the problem here.
Unless I'm forgetting something, UseEndpoints isn't used at all in any 2.x version of Steeltoe and I believe we've already corrected issues in 3.x, can you point to something specific please?
Sorry, we are using 3.0.2.
Nevermind, I see similar (closed) https://github.com/SteeltoeOSS/Steeltoe/issues/497
Do you have pland to release https://github.com/SteeltoeOSS/Steeltoe/pull/591 ?
| gharchive/issue | 2021-04-19T11:40:29 | 2025-04-01T04:33:06.483335 | {
"authors": [
"Lonli-Lokli",
"TimHess"
],
"repo": "SteeltoeOSS/Steeltoe",
"url": "https://github.com/SteeltoeOSS/Steeltoe/issues/637",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2762397554 | Kollisionssensor erkennt keine Hindernisse.
Wie gut zusehen ist im Video, reagiert der Kollisionssensor nicht auf sämtliche Hindernisse, erst wenn es das KFZ zum Stoppen bringt.
log.txt
https://drive.google.com/file/d/1q1weMumcAtRyIlQZUD6CJYo1CY3V4RQm/view?usp=drive_link
Fehler gefunden, die Traktoren Fendt 900 und 1000 Vario von Giants haben ein Fehler, alle anderen getesteten Traktoren weisen nicht die Fehler auf.
| gharchive/issue | 2024-12-29T20:36:26 | 2025-04-01T04:33:06.533348 | {
"authors": [
"regs1980"
],
"repo": "Stephan-S/FS25_AutoDrive",
"url": "https://github.com/Stephan-S/FS25_AutoDrive/issues/55",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1626528309 | 🛑 DNS (he.net) is down
In 08b4234, DNS (he.net) ($HE_NS) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS (he.net) is back up in 05ee866.
| gharchive/issue | 2023-03-16T01:22:45 | 2025-04-01T04:33:06.544911 | {
"authors": [
"leitmori"
],
"repo": "Sternwarte-St-Ottilien-e-V/status",
"url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/2542",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1680480088 | 🛑 DNS (he.net) is down
In e873d0a, DNS (he.net) ($HE_NS) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS (he.net) is back up in 017dc17.
| gharchive/issue | 2023-04-24T05:22:44 | 2025-04-01T04:33:06.547595 | {
"authors": [
"leitmori"
],
"repo": "Sternwarte-St-Ottilien-e-V/status",
"url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/5817",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1683985399 | 🛑 DNS (he.net) is down
In 47dda78, DNS (he.net) ($HE_NS) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS (he.net) is back up in 8a38f46.
| gharchive/issue | 2023-04-25T23:01:42 | 2025-04-01T04:33:06.550422 | {
"authors": [
"leitmori"
],
"repo": "Sternwarte-St-Ottilien-e-V/status",
"url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/5949",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1698859004 | 🛑 DNS (he.net) is down
In 7bf2ded, DNS (he.net) ($HE_NS) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS (he.net) is back up in 0b0094b.
| gharchive/issue | 2023-05-07T02:34:43 | 2025-04-01T04:33:06.553137 | {
"authors": [
"leitmori"
],
"repo": "Sternwarte-St-Ottilien-e-V/status",
"url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/6774",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1708413013 | 🛑 DNS (he.net) is down
In 642fc03, DNS (he.net) ($HE_NS) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS (he.net) is back up in 3790335.
| gharchive/issue | 2023-05-13T03:30:31 | 2025-04-01T04:33:06.555964 | {
"authors": [
"leitmori"
],
"repo": "Sternwarte-St-Ottilien-e-V/status",
"url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/7247",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1671785707 | TypeError: 'NoneType' object is not iterable
The github action failed on pull_request with this error:
TypeError: 'NoneType' object is not iterable
Most recent callback:
Traceback (most recent call last):
File "/app/main.py", line 223, in <module>
main()
File "/app/main.py", line 159, in main
most_recent_runs = get_recent_runs_for_job(base_url=base_dbt_cloud_api_url, headers=req_auth_headers, job_id=dbt_cloud_job_id, same_branch_flag=same_branch_flag, max_runs=max_runs)
File "/app/main.py", line [12](https://github.com/trainual/trainual-data-dbt/actions/runs/4724221958/jobs/8381167543#step:3:13)3, in get_recent_runs_for_job
recent_runs_info = extract_dbt_runs_info(recent_runs['data'], same_branch_flag)
File "/app/main.py", line 68, in extract_dbt_runs_info
for run in recent_runs_list:
TypeError: 'NoneType' object is not iterable
Questions:
how would I troubleshoot or set to not run when this error, related to this, is encountered?
Hey @normwarren!
A couple things I would make sure of, to kick off trouble shooting:
make sure that you have the dbt cloud token set and that the token has correct permissions on the given project
dbt_cloud_token: ${{ secrets.DBT_CLOUD_TOKEN }}
Make sure you all fill out account id and job id the following in the action yaml as well, example below:
dbt_cloud_account_id: 12345
dbt_cloud_job_id: 130247
The NoneType error on this action tends to happen when it's trying to parse through the dbt Cloud API results but the results from dbt Cloud came back an error usually due to permissions
Met with Steve and mentioned we had a team member left when github action broke. We created an access account in dbt-cloud and update the github secrets, and everything worked again ! Thank you Steve!
| gharchive/issue | 2023-04-17T18:53:56 | 2025-04-01T04:33:06.561573 | {
"authors": [
"Stevedow99",
"normwarren"
],
"repo": "Stevedow99/dbt-cloud-cancel-running-ci-job-action",
"url": "https://github.com/Stevedow99/dbt-cloud-cancel-running-ci-job-action/issues/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
437884311 | FBI Conflit with new super mario bros 2
Hello everyone I use fbi 2.6 and I found a beug with a particular game actually launch you new sueper mario bros 2 launch you a game and enjoy leaving you the game and launch you fbi and boom fbi crash fbi at launch it does its that with this game I have to test with 20 other games no problem is very annoying considering that when I leave the game I go directly to fbi and bim his crash here is it is you are all to bawl you can try to reproduce it you must fbi 2.6.0 and new super mario bros 2 thank you and good evening
What
This has been reported before and was determined to not be an FBI issue. See https://github.com/Steveice10/FBI/issues/382
| gharchive/issue | 2019-04-27T00:10:56 | 2025-04-01T04:33:06.564782 | {
"authors": [
"InSimpleTermsJordan",
"Steveice10",
"StigiTT"
],
"repo": "Steveice10/FBI",
"url": "https://github.com/Steveice10/FBI/issues/501",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
135750128 | Change menu order to mirror game difficulty
As per the front page recommended order:
Bandit
Leviathan or Natas or Krypton
Narnia
Behemoth
Utumno
Maze
…
excellent suggestion, let's try this out
| gharchive/pull-request | 2016-02-23T14:23:18 | 2025-04-01T04:33:06.565806 | {
"authors": [
"StevenVanAcker",
"unfo"
],
"repo": "StevenVanAcker/OverTheWire-website",
"url": "https://github.com/StevenVanAcker/OverTheWire-website/pull/44",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
603253758 | 请问项目所需要的环境以及版本能否发下
我本地部署了一下,用的是jdk1.8+tomcat8.5+mysql5.7+maven3.5
但是提示无法加载db.properties文件
报错如上图所示,能否给看下
`20-Apr-2020 21:38:05.869 淇℃伅 [RMI TCP Connection(3)-127.0.0.1] org.apache.catalina.core.ApplicationContext.log No Spring WebApplicationInitializer types detected on classpath
20-Apr-2020 21:38:06.353 淇℃伅 [RMI TCP Connection(3)-127.0.0.1] org.apache.catalina.core.ApplicationContext.log Initializing Spring root WebApplicationContext
20-Apr-2020 21:38:07.458 涓ラ噸 [RMI TCP Connection(3)-127.0.0.1] org.apache.catalina.core.StandardContext.listenerStart Exception sending context initialized event to listener instance of class [org.springframework.web.context.ContextLoaderListener]
org.springframework.beans.factory.BeanInitializationException: Could not load properties; nested exception is java.io.FileNotFoundException: class path resource [db.properties] cannot be opened because it does not exist
at org.springframework.context.support.PropertySourcesPlaceholderConfigurer.postProcessBeanFactory(PropertySourcesPlaceholderConfigurer.java:151)
at org.springframework.context.support.PostProcessorRegistrationDelegate.invokeBeanFactoryPostProcessors(PostProcessorRegistrationDelegate.java:265)
at org.springframework.context.support.PostProcessorRegistrationDelegate.invokeBeanFactoryPostProcessors(PostProcessorRegistrationDelegate.java:162)
at org.springframework.context.support.AbstractApplicationContext.invokeBeanFactoryPostProcessors(AbstractApplicationContext.java:606)
at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:462)
at org.springframework.web.context.ContextLoader.configureAndRefreshWebApplicationContext(ContextLoader.java:403)
at org.springframework.web.context.ContextLoader.initWebApplicationContext(ContextLoader.java:306)
at org.springframework.web.context.ContextLoaderListener.contextInitialized(ContextLoaderListener.java:106)
at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:4770)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5236)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:754)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:730)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:734)
at org.apache.catalina.startup.HostConfig.manageApp(HostConfig.java:1731)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.tomcat.util.modeler.BaseModelMBean.invoke(BaseModelMBean.java:287)
at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.invoke(DefaultMBeanServerInterceptor.java:819)
at com.sun.jmx.mbeanserver.JmxMBeanServer.invoke(JmxMBeanServer.java:801)
at org.apache.catalina.mbeans.MBeanFactory.createStandardContext(MBeanFactory.java:483)
at org.apache.catalina.mbeans.MBeanFactory.createStandardContext(MBeanFactory.java:432)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.tomcat.util.modeler.BaseModelMBean.invoke(BaseModelMBean.java:287)
at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.invoke(DefaultMBeanServerInterceptor.java:819)
at com.sun.jmx.mbeanserver.JmxMBeanServer.invoke(JmxMBeanServer.java:801)
at com.sun.jmx.remote.security.MBeanServerAccessController.invoke(MBeanServerAccessController.java:468)
at javax.management.remote.rmi.RMIConnectionImpl.doOperation(RMIConnectionImpl.java:1468)
at javax.management.remote.rmi.RMIConnectionImpl.access$300(RMIConnectionImpl.java:76)
at javax.management.remote.rmi.RMIConnectionImpl$PrivilegedOperation.run(RMIConnectionImpl.java:1309)
at java.security.AccessController.doPrivileged(Native Method)
at javax.management.remote.rmi.RMIConnectionImpl.doPrivilegedOperation(RMIConnectionImpl.java:1408)
at javax.management.remote.rmi.RMIConnectionImpl.invoke(RMIConnectionImpl.java:829)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at sun.rmi.server.UnicastServerRef.dispatch(UnicastServerRef.java:357)
at sun.rmi.transport.Transport$1.run(Transport.java:200)
at sun.rmi.transport.Transport$1.run(Transport.java:197)
at java.security.AccessController.doPrivileged(Native Method)
at sun.rmi.transport.Transport.serviceCall(Transport.java:196)
at sun.rmi.transport.tcp.TCPTransport.handleMessages(TCPTransport.java:573)
at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.run0(TCPTransport.java:834)
at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.lambda$run$0(TCPTransport.java:688)
at java.security.AccessController.doPrivileged(Native Method)
at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.run(TCPTransport.java:687)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.io.FileNotFoundException: class path resource [db.properties] cannot be opened because it does not exist
at org.springframework.core.io.ClassPathResource.getInputStream(ClassPathResource.java:172)
at org.springframework.core.io.support.EncodedResource.getInputStream(EncodedResource.java:150)
at org.springframework.core.io.support.PropertiesLoaderUtils.fillProperties(PropertiesLoaderUtils.java:98)
at org.springframework.core.io.support.PropertiesLoaderSupport.loadProperties(PropertiesLoaderSupport.java:175)
at org.springframework.core.io.support.PropertiesLoaderSupport.mergeProperties(PropertiesLoaderSupport.java:156)
at org.springframework.context.support.PropertySourcesPlaceholderConfigurer.postProcessBeanFactory(PropertySourcesPlaceholderConfigurer.java:142)
... 54 more
20-Apr-2020 21:38:07.463 淇℃伅 [RMI TCP Connection(3)-127.0.0.1] org.apache.catalina.core.ApplicationContext.log Closing Spring root WebApplicationContext
`
我本地部署了一下,用的是jdk1.8+tomcat8.5+mysql5.7+maven3.5
但是提示无法加载db.properties文件
报错如上图所示,能否给看下
你好,请问你解决这个问题了吗
发生这个问题的原因是db.properties所在文件夹并未设置为resource目录
进行如下设置即可
在最后面的右边目录找到ResourcesRoot设置一下就行了,我这里因为设置过了,所以没有显示
| gharchive/issue | 2020-04-20T13:49:03 | 2025-04-01T04:33:06.582043 | {
"authors": [
"AcRushGo",
"AtomRun",
"miaofengshi"
],
"repo": "StevenWash/xxshop",
"url": "https://github.com/StevenWash/xxshop/issues/7",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
99782355 | There was an unhandled error - Sorry!
Whenever I go to a page with this fishing rod in the screen, whether its on a character or in a bank tab, procurement crashes.
https://i.gyazo.com/c323a1cc1a90a19b28521306edd352cf.png
This is a known issue, if you can zip and send me all your .bin files I can take a look at what's causing this (I don't have a fishing rod)
Hey im pretty shitty with techy stuff, so I'm not sure if this is what you mean, but this is a copy and paste of all the files within G;/Program Files/Path of Exile > Procurement > (my other email)
Date: Sat, 8 Aug 2015 04:15:50 -0700
From: notifications@github.com
To: Procurement@noreply.github.com
CC: dannydc13@hotmail.com
Subject: Re: [Procurement] There was an unhandled error - Sorry! (#157)
This is a known issue, if you can zip and send me all your .bin files I can take a look at what's causing this (I don't have a fishing rod)
—
Reply to this email directly or view it on GitHub.
This has, finally, been fixed in Procurement 1.24.2
| gharchive/issue | 2015-08-08T08:09:09 | 2025-04-01T04:33:06.614669 | {
"authors": [
"DannyBooy",
"Stickymaddness"
],
"repo": "Stickymaddness/Procurement",
"url": "https://github.com/Stickymaddness/Procurement/issues/157",
"license": "Artistic-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
373219021 | update Travis to Py3.7
I'm keeping Py3.4 around for the lab computers
Coverage remained the same at 62.151% when pulling 9217f4198ef274bce9405ff1c8b5f38e0cde1ecf on python-3.7 into 1504998d68aab7e80f43292320516c4e766fc5a9 on master.
| gharchive/pull-request | 2018-10-23T21:53:46 | 2025-04-01T04:33:06.626887 | {
"authors": [
"coveralls",
"hawkrives"
],
"repo": "StoDevX/cs251-toolkit",
"url": "https://github.com/StoDevX/cs251-toolkit/pull/74",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
108556327 | RuntimeError: maximum recursion depth exceeded in comparison
Infinity loop:
socket.gaierror: [Errno -3] Temporary failure in name resolution
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 45, in _url_query
response = urllib.request.urlopen(req)
File "/usr/lib64/python3.3/urllib/request.py", line 156, in urlopen
return opener.open(url, data, timeout)
File "/usr/lib64/python3.3/urllib/request.py", line 469, in open
response = self._open(req, data)
File "/usr/lib64/python3.3/urllib/request.py", line 487, in _open
'_open', req)
File "/usr/lib64/python3.3/urllib/request.py", line 447, in _call_chain
result = func(*args)
File "/usr/lib64/python3.3/urllib/request.py", line 1274, in http_open
return self.do_open(http.client.HTTPConnection, req)
File "/usr/lib64/python3.3/urllib/request.py", line 1251, in do_open
raise URLError(err)
urllib.error.URLError: <urlopen error [Errno -3] Temporary failure in name resolution>
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "dataserv-client", line 10, in
cli.main(sys.argv[1:])
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/cli.py", line 190, in main
return getattr(client, command_name)(**arguments)
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/api.py", line 235, in farm
self.poll(delay=delay, limit=limit)
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/api.py", line 161, in poll
self.ping()
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/api.py", line 144, in ping
self.messenger.ping()
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 109, in ping
return self._url_query("/api/ping/%s" % self.auth_address())
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 69, in _url_query
return self._handle_connection_error(api_path, retries, authenticate)
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 81, in _handle_connection_error
.............
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 81, in _handle_connection_error
return self._url_query(api_path, retries + 1, authenticate)
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 69, in _url_query
return self._handle_connection_error(api_path, retries, authenticate)
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 81, in _handle_connection_error
return self._url_query(api_path, retries + 1, authenticate)
File "/usr/lib64/python3.3/site-packages/dataserv_client-2.1.4-py3.3.egg/dataserv_client/messaging.py", line 40, in _url_query
self.btctxstore, self._get_server_address(), self.wif
File "/usr/lib64/python3.3/site-packages/storjcore-0.0.3-py3.3.egg/storjcore/auth.py", line 38, in create_headers
signature = btctxstore.sign_unicode(sender_wif, msg)
File "/usr/lib64/python3.3/site-packages/btctxstore-4.5.0-py3.3.egg/btctxstore/api.py", line 191, in sign_unicode
return self.sign_data(wif, hexdata)
File "/usr/lib64/python3.3/site-packages/btctxstore-4.5.0-py3.3.egg/btctxstore/api.py", line 174, in sign_data
sigdata = control.sign_data(self.testnet, data, key)
File "/usr/lib64/python3.3/site-packages/btctxstore-4.5.0-py3.3.egg/btctxstore/control.py", line 293, in sign_data
pk = ecdsa.SigningKey.from_secret_exponent(secretexponent, curve=SECP256k1)
File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/keys.py", line 142, in from_secret_exponent
File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/keys.py", line 28, in from_public_point
File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ecdsa.py", line 87, in init
File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ellipticcurve.py", line 143, in rmul
File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ellipticcurve.py", line 134, in mul
File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ellipticcurve.py", line 107, in add
File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ellipticcurve.py", line 73, in init
File "/usr/lib64/python3.3/site-packages/ecdsa-0.13-py3.3.egg/ecdsa/ellipticcurve.py", line 59, in contains_point
RuntimeError: maximum recursion depth exceeded in comparison
Looks like the farm command? https://github.com/Storj/dataserv-client/blob/develop/dataserv_client/api.py#L241
I will fix it tomorrow.
python max recursion is ~1000. For every 30 sec retry 2 recursions calls are needed. Max retry limit will be somewhere between 450 and 500. -> ~4h retry
I will increase the retry delay for farm to 60 sec and the retry limit to 450-500. That way max retry will be ~8h without exception.
Can you just drop recursion and use normal loop please?
I did some unit tests and python recursion is realy a bad. Looks like your idea with the loop is the only solution.
will be fixed in the next version (after pull request merge)
Pull request merged. Unit test with 2000 retries passed.
| gharchive/issue | 2015-09-27T19:36:43 | 2025-04-01T04:33:06.654821 | {
"authors": [
"alphaonex86",
"littleskunk"
],
"repo": "Storj/dataserv-client",
"url": "https://github.com/Storj/dataserv-client/issues/160",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
194980653 | from core.lib.utils migrate simpleDecrypt and simpleEncrypt of data f…
…or id_ecdsa
__________________________ PEP8-check(ignoring W503) ___________________________
[gw0] linux2 -- Python 2.7.9 /home/travis/build/Storj/storj-python-sdk/.tox/py27/bin/python2.7
/home/travis/build/Storj/storj-python-sdk/storj/model.py:265:38: E261 at least two spaces before inline comment
padding = 16 - len(data) % 16#
^
/home/travis/build/Storj/storj-python-sdk/storj/model.py:271:45: E226 missing whitespace around arithmetic operator
def decrypt_node(self, hex_data, key='0'*32, iv='0'*16):
^
/home/travis/build/Storj/storj-python-sdk/storj/model.py:271:56: E226 missing whitespace around arithmetic operator
def decrypt_node(self, hex_data, key='0'*32, iv='0'*16):
^
/home/travis/build/Storj/storj-python-sdk/storj/model.py:276:41: E226 missing whitespace around arithmetic operator
def encrypt_node(self, data, key='0'*32, iv='0'*16):
^
/home/travis/build/Storj/storj-python-sdk/storj/model.py:276:52: E226 missing whitespace around arithmetic operator
def encrypt_node(self, data, key='0'*32, iv='0'*16):
i hope i got it now
@BlackDuck888 you can test on your workstation by using tox. ;)
i will try it next time, step by step, so many new things, right now i am fighting with GitHub.
i close this pr, and will start from scratch, this will help me to understand how everything ist working
| gharchive/pull-request | 2016-12-12T13:58:11 | 2025-04-01T04:33:06.665376 | {
"authors": [
"BlackDuck888",
"steenzout"
],
"repo": "Storj/storj-python-sdk",
"url": "https://github.com/Storj/storj-python-sdk/pull/69",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1523301009 | Incomplete basic output
When creating basic output sometimes the results look incomplete. For example:
let schema = JSONSchema::compile(&json!({
"$id": "https://example.com/arrays.schema.json",
"$schema": "https://json-schema.org/draft/2020-12/schema",
"type": "object",
"properties": {
"vegetables": {
"type": "array",
"description": "Desc1",
"items": { "$ref": "#/$defs/veggie" }
}
},
"$defs": {
"veggie": {
"type": "object",
"required": [ "veggieName", "veggieLike" ],
"properties": {
"veggieName": {
"description": "Desc2",
"type": "string"
},
"veggieLike": {
"description": "Desc3",
"type": "boolean"
}
}
}
}
}))
.unwrap();
if let BasicOutput::Valid(output) = schema
.apply(&json!({
"vegetables":[{"veggieName": "carrot", "veggieLike": true}]
}))
.basic()
{
for unit in output {
println!("{}", unit.instance_location());
}
}
Output:
/vegetables
/vegetables/0
Expected output:
/vegetables
/vegetables/0
/vegetables/0/veggieName
/vegetables/0/veggieLike
I updated the example and did a little investigation.
I think the issue is that RefValidator doesn't contain a custom implementation of apply, so any PartialApplication information from down stream is discarded.
Sorry for the delay, with the current approach, it is a bit complicated as annotations are bound to the schema. Not sure if it would be possible to keep the references, but maybe making annotations own their data would be a better way to go anyway.
Fixed! However, keyword paths are incorrect in many cases. I am going to fix them separately
| gharchive/issue | 2023-01-06T23:24:40 | 2025-04-01T04:33:06.697176 | {
"authors": [
"BrynCooke",
"Stranger6667"
],
"repo": "Stranger6667/jsonschema-rs",
"url": "https://github.com/Stranger6667/jsonschema-rs/issues/403",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2684914660 | 🛑 Jellyseer is down
In f0fd96f, Jellyseer (https://jellyseerr.stroyco.eu) was down:
HTTP code: 502
Response time: 1272 ms
Resolved: Jellyseer is back up in 6014d05 after 1 hour, 42 minutes.
| gharchive/issue | 2024-11-22T23:28:16 | 2025-04-01T04:33:06.771974 | {
"authors": [
"DestroyCom"
],
"repo": "StroyCo/stroy-up",
"url": "https://github.com/StroyCo/stroy-up/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1674362970 | Separate parsing APIs that don't depend on ANTLR
See https://github.com/Strumenta/StarLasu/issues/27
Kolasu split them into a new module (kolasu-antlr) and a new package (com.strumenta.kolasu.antlr.parsing).
Given the structure and choices made in Tylasu, we could introduce antlr/parsing.ts and export the current parsing.ts, which won't depend on ANTLR, from the core index.ts
Just like Kolasu, we'll increase the version number to 1.3 since this is a breaking change.
This looks completed
| gharchive/issue | 2023-04-19T07:43:03 | 2025-04-01T04:33:06.773565 | {
"authors": [
"alessiostalla"
],
"repo": "Strumenta/tylasu",
"url": "https://github.com/Strumenta/tylasu/issues/40",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1012889034 | 🛑 Nextcloud Papa is down
In 620138c, Nextcloud Papa ($NEXTCLOUD_2) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Nextcloud Papa is back up in a102d8b.
| gharchive/issue | 2021-10-01T04:58:11 | 2025-04-01T04:33:06.776855 | {
"authors": [
"StudFu-WordToMD"
],
"repo": "StudFu-WordToMD/status",
"url": "https://github.com/StudFu-WordToMD/status/issues/2967",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1045833926 | 🛑 Nextcloud Papa is down
In 1a8c884, Nextcloud Papa ($NEXTCLOUD_2) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Nextcloud Papa is back up in 8365fb1.
| gharchive/issue | 2021-11-05T13:10:42 | 2025-04-01T04:33:06.778903 | {
"authors": [
"StudFu-WordToMD"
],
"repo": "StudFu-WordToMD/status",
"url": "https://github.com/StudFu-WordToMD/status/issues/3822",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
213611118 | Tooltips at screen top are hidden/overlapped by top nav bar
The addon/referenced Github addon themes include one to expand the top nav bar etc. and I know this is minor but just wanted to point out that the tooltips for the top right elements like WATCH/STAR/FORK when theme is enabled are hidden under the navbar as shown in these screenshots below. I am not sure if it is possible, but a really nice change just to ensure if in the future Github were to add other functions and corresponding tooltips would be just to MOVE those tooltips down UNDER the corresponding elements. The dumb placement of them above elements so close to the top of the screen of course were not even made by this theme but by the original site dev, but they got away with it (barely) only because the original site didn't have quite the height of the nav bar with all the functionality being added here. :_bow:
FORK tooltip (the tiny bit of white if you look):
https://1drv.ms/i/s!AhSqjTYUuHPKgcNbmBCOm-wDiU936w
STAR tooltip (the tiny bit of white if you look):
https://1drv.ms/i/s!AhSqjTYUuHPKgcNaLEhGWHrofnpBQg
Again not even sure this is considered a bug in this theme but really a SITE bug that becomes a problem only when this theme is applied (and not a major one).
Hi @CollinChaffin!
GitHub has removed the tooltips from the star and fork buttons. So, it looks like the problem is in the github-script-make-tooltips.user.js code. It is currently set to make all tooltips point up, I'll update the code to make those point down... thanks for letting us know!
| gharchive/issue | 2017-03-12T15:19:56 | 2025-04-01T04:33:06.789033 | {
"authors": [
"CollinChaffin",
"Mottie"
],
"repo": "StylishThemes/GitHub-Dark-Script",
"url": "https://github.com/StylishThemes/GitHub-Dark-Script/issues/30",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1369466647 | chore: aliasing substrapp django models
To see the specific tasks where the Asana app for GitHub is being used, see below:
https://app.asana.com/0/0/1202951505377093
/e2e --mode standalone
End to end tests: :hourglass_flowing_sand: STARTING
| gharchive/pull-request | 2022-09-12T08:16:02 | 2025-04-01T04:33:06.823125 | {
"authors": [
"Owlfred",
"sergebouchut2"
],
"repo": "Substra/substra-backend",
"url": "https://github.com/Substra/substra-backend/pull/444",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2161274557 | chore: add missing repo
Description
How has this been tested?
Checklist
[ ] changelog was updated with notable changes
[ ] documentation was updated
- ".github/workflows/helm.yml"
Sounds good ! But the problem are having is detected when we package, which only called during "publish". Should we add a package in each PR to try that no dependency is missed ?
- ".github/workflows/helm.yml"
Sounds good ! But the problem are having is detected when we package, which only called during "publish". Should we add a package in each PR to try that no dependency is missed ?
Yes, already did! https://github.com/Substra/substra-gha-workflows/pull/10
| gharchive/pull-request | 2024-02-29T13:32:24 | 2025-04-01T04:33:06.826410 | {
"authors": [
"guilhem-barthes",
"thbcmlowk"
],
"repo": "Substra/substra-backend",
"url": "https://github.com/Substra/substra-backend/pull/837",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1060170688 | Make better label text for audio player
Related to #336 and #339. Improve to words that describe these specific actions like "Read page", instead of generic ones like "play"
FYI, references for this are things like the video aria-label in the code at https://developer.mozilla.org/en-US/docs/Learn/JavaScript/Client-side_web_APIs/Video_and_audio_APIs and https://www.w3.org/TR/wai-aria-practices-1.1/#no_aria_better_bad_aria.
| gharchive/issue | 2021-11-22T13:40:26 | 2025-04-01T04:33:06.860506 | {
"authors": [
"plocket"
],
"repo": "SuffolkLITLab/docassemble-AssemblyLine",
"url": "https://github.com/SuffolkLITLab/docassemble-AssemblyLine/issues/342",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1358384526 | 🛑 Zinc 20 is down
In f2e2fb4, Zinc 20 (https://zinc20.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 20 is back up in 7df4f4c.
| gharchive/issue | 2022-09-01T07:28:53 | 2025-04-01T04:33:06.862815 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/10196",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1402389583 | 🛑 Zinc 15 is down
In cb5ca07, Zinc 15 (https://zinc15.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 15 is back up in e702afd.
| gharchive/issue | 2022-10-09T20:32:38 | 2025-04-01T04:33:06.865076 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/11594",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1451338012 | 🛑 Zinc 15 is down
In 60dcae2, Zinc 15 (https://zinc15.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 15 is back up in c69183d.
| gharchive/issue | 2022-11-16T10:40:41 | 2025-04-01T04:33:06.867356 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/12953",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1555704920 | 🛑 Chem Exper is down
In c16763a, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 2707037.
| gharchive/issue | 2023-01-24T21:32:38 | 2025-04-01T04:33:06.869834 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/14817",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1605798014 | 🛑 Chem Exper is down
In 4df1c77, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 8e596ee.
| gharchive/issue | 2023-03-01T22:56:52 | 2025-04-01T04:33:06.872282 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/16306",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1635963858 | 🛑 Chem Exper is down
In 517a95d, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 2cdd832.
| gharchive/issue | 2023-03-22T15:09:36 | 2025-04-01T04:33:06.874513 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/17143",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1656444109 | 🛑 Chem Exper is down
In 931266c, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 4229061.
| gharchive/issue | 2023-04-05T23:38:19 | 2025-04-01T04:33:06.876787 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/17769",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1723376989 | 🛑 Zinc 15 is down
In 9dc23b7, Zinc 15 (https://zinc15.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 15 is back up in e73e12c.
| gharchive/issue | 2023-05-24T07:42:26 | 2025-04-01T04:33:06.879146 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/20223",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1748725257 | 🛑 Zinc 20 is down
In 275fb2c, Zinc 20 (https://zinc20.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 20 is back up in c63d372.
| gharchive/issue | 2023-06-08T22:18:46 | 2025-04-01T04:33:06.881455 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/21495",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1805257595 | 🛑 Chem Exper is down
In abebb22, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 0fa3572.
| gharchive/issue | 2023-07-14T17:49:46 | 2025-04-01T04:33:06.883720 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/24359",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1807895131 | 🛑 Chem Exper is down
In 92787e0, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 495fc4f.
| gharchive/issue | 2023-07-17T14:17:58 | 2025-04-01T04:33:06.886179 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/24483",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1824860679 | 🛑 Zinc 15 is down
In 0f2572b, Zinc 15 (https://zinc15.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 15 is back up in 982e713.
| gharchive/issue | 2023-07-27T17:29:32 | 2025-04-01T04:33:06.888416 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/25099",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1843048420 | 🛑 Chem Exper is down
In fe6189a, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 49c875e.
| gharchive/issue | 2023-08-09T11:35:41 | 2025-04-01T04:33:06.890666 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/26095",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1931473402 | 🛑 Chem Exper is down
In da2d86e, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 9aa21cd after 16 minutes.
| gharchive/issue | 2023-10-07T19:24:41 | 2025-04-01T04:33:06.892954 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/30235",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1954794560 | 🛑 Chem Exper is down
In 945aca4, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in b9f07b2 after 39 minutes.
| gharchive/issue | 2023-10-20T18:03:22 | 2025-04-01T04:33:06.895221 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/30831",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2117583702 | 🛑 Binding Database is down
In 86481af, Binding Database (http://www.bindingdb.org/bind/index.jsp) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Binding Database is back up in e697e69 after 11 minutes.
| gharchive/issue | 2024-02-05T03:08:31 | 2025-04-01T04:33:06.897730 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/36263",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2122529474 | 🛑 Chemical Resolver is down
In 8db010a, Chemical Resolver (https://cactus.nci.nih.gov/chemical/structure) was down:
HTTP code: 500
Response time: 76 ms
Resolved: Chemical Resolver is back up in 488de03 after 10 minutes.
| gharchive/issue | 2024-02-07T09:09:11 | 2025-04-01T04:33:06.900084 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/36370",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2131574777 | 🛑 Chem Exper is down
In 46b6631, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 411ea98 after 7 minutes.
| gharchive/issue | 2024-02-13T06:46:22 | 2025-04-01T04:33:06.902472 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/36668",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2246330663 | 🛑 Zinc 20 is down
In ddf4a70, Zinc 20 (https://zinc20.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 20 is back up in fe7689a after 6 minutes.
| gharchive/issue | 2024-04-16T15:23:54 | 2025-04-01T04:33:06.904730 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/39172",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2363387936 | 🛑 Adverse Drug Reaction Classification System is down
In 0811654, Adverse Drug Reaction Classification System (http://bioinf.xmu.edu.cn/ADReCS/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Adverse Drug Reaction Classification System is back up in 01f6e9a after 16 minutes.
| gharchive/issue | 2024-06-20T02:39:14 | 2025-04-01T04:33:06.907117 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/41815",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2509375431 | 🛑 Chem Exper is down
In 23d7f37, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in f9e1ffd after 10 minutes.
| gharchive/issue | 2024-09-06T02:59:45 | 2025-04-01T04:33:06.909597 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/45263",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2547681351 | 🛑 Chem Exper is down
In b20e1f0, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in b65a46b after 11 minutes.
| gharchive/issue | 2024-09-25T11:06:52 | 2025-04-01T04:33:06.911863 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/45883",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2620378024 | 🛑 Chem Exper is down
In 46a92da, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 4f7b530 after 17 minutes.
| gharchive/issue | 2024-10-29T07:23:25 | 2025-04-01T04:33:06.914149 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/46998",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2684635845 | 🛑 Chem Exper is down
In 87d8e38, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in d4d665c after 8 minutes.
| gharchive/issue | 2024-11-22T21:29:40 | 2025-04-01T04:33:06.916430 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/47794",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2725075008 | 🛑 Chem Exper is down
In a4b77e6, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 8cbaf4e after 16 minutes.
| gharchive/issue | 2024-12-08T09:37:02 | 2025-04-01T04:33:06.918666 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/48295",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1268552256 | 🛑 Zinc 20 is down
In 370e052, Zinc 20 (https://zinc20.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 20 is back up in 9785604.
| gharchive/issue | 2022-06-12T11:31:24 | 2025-04-01T04:33:06.921299 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/6808",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1286065540 | 🛑 Zinc 15 is down
In baf39e6, Zinc 15 (https://zinc15.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 15 is back up in bb9985d.
| gharchive/issue | 2022-06-27T16:40:10 | 2025-04-01T04:33:06.923549 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/7456",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1288314531 | 🛑 Zinc 15 is down
In e956fe1, Zinc 15 (https://zinc15.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 15 is back up in 04a9c0d.
| gharchive/issue | 2022-06-29T07:51:50 | 2025-04-01T04:33:06.925815 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/7515",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1342316538 | 🛑 Zinc 15 is down
In a5d0e4f, Zinc 15 (https://zinc15.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 15 is back up in dba657b.
| gharchive/issue | 2022-08-17T21:38:46 | 2025-04-01T04:33:06.928277 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/9678",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1989059473 | 🛑 Stanford Tissue Microarray Database is down
In e18f6af, Stanford Tissue Microarray Database (http://tma.im) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Stanford Tissue Microarray Database is back up in d744d59 after 7 minutes.
| gharchive/issue | 2023-11-11T17:51:56 | 2025-04-01T04:33:06.930666 | {
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Medical-Informatics",
"url": "https://github.com/Sulstice/Uptime-Medical-Informatics/issues/1062",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1471488219 | Get error as "Cannot read properties of null (reading 'once')
When clicked on the first item in the dropdown menu,
Getting error due to RUM javascript code as:
Also nothing is rendered in screen so that application is broken.
But, if the RUM javascript code is removed, there is no error.
Even though the issue above is irrespective of the browser.
Browser used: Chrome Version 107.0.5304.107 (Official Build) (64-bit)
App uses:
react : 17.0.2 and react-dom: 17.0.2
SumoLogic RUM used: https://rum.sumologic.com/sumologic-rum-v4.js
Any help or idea would be appreciated 🙏
Hi @prajon84, thanks for letting us know.
I prepared a fix for that issue: https://github.com/SumoLogic/sumologic-opentelemetry-js/pull/67
New version (4.1.1) with fix has been released:
https://rum.sumologic.com/sumologic-rum-v4.1.1.js
https://rum.sumologic.com/sumologic-rum-v4.1.js
https://rum.sumologic.com/sumologic-rum-v4.js
https://rum.sumologic.com/sumologic-rum.js
In case of any problems, please let us know.
| gharchive/issue | 2022-12-01T15:11:43 | 2025-04-01T04:33:06.940149 | {
"authors": [
"TRybakSUMO",
"prajon84"
],
"repo": "SumoLogic/sumologic-opentelemetry-js",
"url": "https://github.com/SumoLogic/sumologic-opentelemetry-js/issues/66",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
784643168 | ENH: Make ElastixMain database creation + loading components thread safe
Ensures that only one thread may load the components into the ComponentDatabase, using using C++11 "magic statics".
ElastixMain::GetComponentDatabase() now returns a "const reference" instead of a non-const pointer, to prevent one thread from modifying the database while another is reading from the database.
Removed the two UnloadComponents() member functions (from both ElastixMain and ComponentLoader), as they appear not useful anymore.
Aims to fix issue #174
"elastix 4.9 static libary version not thread safe"
reported by @jiangliMED, August 16, 2019
With help from Konstantinos Ntatsis (@ntatsisk)
@ntatsisk Hi Konstantinos! Can you please review this pull request? And could you possibly also try it out, to see if it does indeed fix a multi-threading issue that you encountered? You may check out the (temporary) branch of this pull request: https://github.com/SuperElastix/elastix/tree/ElastixMain-LoadComponents-thread-safe
@ntatsisk Hi Konstantinos! Can you please review this pull request? And could you possibly also try it out, to see if it does indeed fix a multi-threading issue that you encountered? You may check out the (temporary) branch of this pull request: https://github.com/SuperElastix/elastix/tree/ElastixMain-LoadComponents-thread-safe
Hello @N-Dekker, the solution looks great! I really liked the combination of magic statics together with a lambda function for the ComponentDatabase initialization. I merged this branch with my local branch where the logger is "deactivated" and managed to run some tests. It works with no issues!
Hello @N-Dekker, the solution looks great! I really liked the combination of magic statics together with a lambda function for the ComponentDatabase initialization. I merged this branch with my local branch where the logger is "deactivated" and managed to run some tests. It works with no issues!
@ntatsisk Thank you for your approval and your enthusiastic reply! I certainly enjoyed combining magic statics together with a lambda 😃 May I ask what your use case looks like? So you call RegisterImages multiple times simultaneously within the same process? How do you do that? Something like std::async? And what do you gain from doing so?
@ntatsisk Thank you for your approval and your enthusiastic reply! I certainly enjoyed combining magic statics together with a lambda 😃 May I ask what your use case looks like? So you call RegisterImages multiple times simultaneously within the same process? How do you do that? Something like std::async? And what do you gain from doing so?
@N-Dekker No, we are using OpenMP to parallelize the registrations. The idea is as follows:
#pragma omp parallel for default(none) num_threads(24)
for (int i = 0; i < 600; i++)
{
ElastixFilterType::Pointer elastixFilter = ElastixFilterType::New();
elastixFilter->SetFixedImage(images[i]);
elastixFilter->SetMovingImage(images[i+1]);
elastixFilter->SetParameterObject(m_elxParameterObjectPointer);
elastixFilter->LogToConsoleOff();
elastixFilter->LogToFileOff();
// #pragma omp critical // (= not necessary if elastix is thread-safe)
elastixFilter->Update();
}
As you see, we are trying to do many (simple) registrations but in a short amount of time (~<10sec) which is a requirement. So, it is better to split the registrations among the threads instead of using many threads for a single registration, as ITK threading does. The gain is some saved seconds eg. from 12sec (with the critical section) to 7sec (in the thread-safe version). In another part, it goes from 30sec to 10sec. Generally, these times might not seem like a big deal, but they are for our application ;)
@N-Dekker No, we are using OpenMP to parallelize the registrations. The idea is as follows:
#pragma omp parallel for default(none) num_threads(24)
for (int i = 0; i < 600; i++)
{
ElastixFilterType::Pointer elastixFilter = ElastixFilterType::New();
elastixFilter->SetFixedImage(images[i]);
elastixFilter->SetMovingImage(images[i+1]);
elastixFilter->SetParameterObject(m_elxParameterObjectPointer);
elastixFilter->LogToConsoleOff();
elastixFilter->LogToFileOff();
// #pragma omp critical // (= not necessary if elastix is thread-safe)
elastixFilter->Update();
}
As you see, we are trying to do many (simple) registrations but in a short amount of time (~<10sec) which is a requirement. So, it is better to split the registrations among the threads instead of using many threads for a single registration, as ITK threading does. The gain is some saved seconds eg. from 12sec (with the critical section) to 7sec (in the thread-safe version). In another part, it goes from 30sec to 10sec. Generally, these times might not seem like a big deal, but they are for our application ;)
@mstaring @stefanklein Hope it's OK to you that I'll merge this pull request this afternoon!
@mstaring @stefanklein Hope it's OK to you that I'll merge this pull request this afternoon!
I do not get all the details of this PR, but i assume i) elastix still works (all tests pass for the command line version and ITKElastix also still works), and ii) it was verified that it actually fixes a problem (a new test first failed but now passes).
I do not get all the details of this PR, but i assume i) elastix still works (all tests pass for the command line version and ITKElastix also still works), and ii) it was verified that it actually fixes a problem (a new test first failed but now passes).
@mstaring Thanks for your feedback
i assume i) elastix still works (all tests pass for the command line version and ITKElastix also still works)
All elastix test are passed successfully, at the CI. However, I did not test ITKElastix with this pull request. We currently do not test ITKElastix on each elastix pull request.
But I don't think ITKElastix would have any problem with this pull request. Unless it would try to call UnloadComponents(), which is removed by this pull request. But that seems unlikely to me, especially because UnloadComponents() wasn't useful anymore anyway. @ViktorvdValk do you know if UnloadComponents() was wrapped to Python, by ITKElastix?
ii) it was verified that it actually fixes a problem (a new test first failed but now passes).
It is verified by Konstantinos (@ntatsisk) that it actually fixes a problem. I can still see if an automatic test can be added, based on his use case.
@mstaring Thanks for your feedback
i assume i) elastix still works (all tests pass for the command line version and ITKElastix also still works)
All elastix test are passed successfully, at the CI. However, I did not test ITKElastix with this pull request. We currently do not test ITKElastix on each elastix pull request.
But I don't think ITKElastix would have any problem with this pull request. Unless it would try to call UnloadComponents(), which is removed by this pull request. But that seems unlikely to me, especially because UnloadComponents() wasn't useful anymore anyway. @ViktorvdValk do you know if UnloadComponents() was wrapped to Python, by ITKElastix?
ii) it was verified that it actually fixes a problem (a new test first failed but now passes).
It is verified by Konstantinos (@ntatsisk) that it actually fixes a problem. I can still see if an automatic test can be added, based on his use case.
sounds good then!
sounds good then!
@ntatsisk Sorry I did not yet merge, because I'm still trying to write a unit test (GoogleTest) for this pull request. But if you need it now, I can just merge this and write the test later. What do you think? Anyway, I'm just trying (locally on my laptop):
GTEST_TEST(ElastixFilter, Parallel)
{
using ImageType = itk::Image<unsigned char>;
std::vector<ImageType::Pointer> images(601);
for (auto & image : images)
{
image = ImageType::New();
image->SetRegions({ 100, 100 });
image->Allocate(true);
}
#pragma omp parallel for
for (int i{}; i < 600; ++i)
{
const auto elastixFilter = elx::ElastixFilter<ImageType, ImageType>::New();
elastixFilter->SetFixedImage(images[i]);
elastixFilter->SetMovingImage(images[i + 1]);
elastixFilter->LogToConsoleOff();
elastixFilter->LogToFileOff();
elastixFilter->Update();
}
}
But it says:
User Error 1001: omp_set_num_threads should only be called in serial regions
Do you have a suggestion?
@ntatsisk Sorry I did not yet merge, because I'm still trying to write a unit test (GoogleTest) for this pull request. But if you need it now, I can just merge this and write the test later. What do you think? Anyway, I'm just trying (locally on my laptop):
GTEST_TEST(ElastixFilter, Parallel)
{
using ImageType = itk::Image<unsigned char>;
std::vector<ImageType::Pointer> images(601);
for (auto & image : images)
{
image = ImageType::New();
image->SetRegions({ 100, 100 });
image->Allocate(true);
}
#pragma omp parallel for
for (int i{}; i < 600; ++i)
{
const auto elastixFilter = elx::ElastixFilter<ImageType, ImageType>::New();
elastixFilter->SetFixedImage(images[i]);
elastixFilter->SetMovingImage(images[i + 1]);
elastixFilter->LogToConsoleOff();
elastixFilter->LogToFileOff();
elastixFilter->Update();
}
}
But it says:
User Error 1001: omp_set_num_threads should only be called in serial regions
Do you have a suggestion?
Hi @N-Dekker, I am not in a hurry at all with the merging so it can wait. Regarding the error, I assume you are compiling with "warnings as errors" because I have got the same message but as a warning instead. I used to get it while running in Debug mode but it has been a while since then. Maybe, it is coming from here:
https://github.com/SuperElastix/elastix/blob/af382ee4b598ac79fcbb7724bfad7bbcfe492fd6/Components/Optimizers/StandardGradientDescent/itkGradientDescentOptimizer2.cxx#L212-L224
Since this omp_set_num_threads is now called inside the user's omp parallel for directive (like the one that you have in your test script), the warning is emitted. To be honest, I haven't fully understood the internal elastix threading and whether this and the other few OpenMP clauses are necessary/helpful. If they do need to stay then I would try re-writing the clause in the above link as follows:
#pragma omp parallel for num_threads(nthreads)
Does this work? Btw, even if it does, I think your test will still crash due to the logger being thread-unsafe, right?
Hi @N-Dekker, I am not in a hurry at all with the merging so it can wait. Regarding the error, I assume you are compiling with "warnings as errors" because I have got the same message but as a warning instead. I used to get it while running in Debug mode but it has been a while since then. Maybe, it is coming from here:
https://github.com/SuperElastix/elastix/blob/af382ee4b598ac79fcbb7724bfad7bbcfe492fd6/Components/Optimizers/StandardGradientDescent/itkGradientDescentOptimizer2.cxx#L212-L224
Since this omp_set_num_threads is now called inside the user's omp parallel for directive (like the one that you have in your test script), the warning is emitted. To be honest, I haven't fully understood the internal elastix threading and whether this and the other few OpenMP clauses are necessary/helpful. If they do need to stay then I would try re-writing the clause in the above link as follows:
#pragma omp parallel for num_threads(nthreads)
Does this work? Btw, even if it does, I think your test will still crash due to the logger being thread-unsafe, right?
Regarding the error, I assume you are compiling with "warnings as errors" because I have got the same message but as a warning instead. I used to get it while running in Debug mode but it has been a while since then. Maybe, it is coming from here:
https://github.com/SuperElastix/elastix/blob/af382ee4b598ac79fcbb7724bfad7bbcfe492fd6/Components/Optimizers/StandardGradientDescent/itkGradientDescentOptimizer2.cxx#L212-L224
Thanks for the suggestion, Konstantinos, but as far as I can see, those lines of code are excluded from compilation, as they are part of the "else block" of an #if 1!
Anyway, I just submitted a new "ElastixFilter.UpdateInParallel" test: https://github.com/SuperElastix/elastix/pull/393 My aim is to have a little test that just shows that this pull request (#389) really fixes a certain problem. But the test is still unstable! I see it crashing some times. Could you also please have a look?
Regarding the error, I assume you are compiling with "warnings as errors" because I have got the same message but as a warning instead. I used to get it while running in Debug mode but it has been a while since then. Maybe, it is coming from here:
https://github.com/SuperElastix/elastix/blob/af382ee4b598ac79fcbb7724bfad7bbcfe492fd6/Components/Optimizers/StandardGradientDescent/itkGradientDescentOptimizer2.cxx#L212-L224
Thanks for the suggestion, Konstantinos, but as far as I can see, those lines of code are excluded from compilation, as they are part of the "else block" of an #if 1!
Anyway, I just submitted a new "ElastixFilter.UpdateInParallel" test: https://github.com/SuperElastix/elastix/pull/393 My aim is to have a little test that just shows that this pull request (#389) really fixes a certain problem. But the test is still unstable! I see it crashing some times. Could you also please have a look?
Ops, my bad! So, it is clearly not the point that is called (maybe this instead?). Anyway, I also noticed that, actually, all the #pragma omp parallel directives are manually deactivated from the entire elastix codebase at this moment.
Regarding the crashing, as I mentioned in my last comment, I think it is because of the logger ;)
Ops, my bad! So, it is clearly not the point that is called (maybe this instead?). Anyway, I also noticed that, actually, all the #pragma omp parallel directives are manually deactivated from the entire elastix codebase at this moment.
Regarding the crashing, as I mentioned in my last comment, I think it is because of the logger ;)
Regarding the crashing, as I mentioned in my last comment, I think it is because of the logger ;)
I believe so too. The logger is still troublesome. So then, is there still any possible test case for this specific pull request, which only improves the thread-safety of the component database?
Regarding the crashing, as I mentioned in my last comment, I think it is because of the logger ;)
I believe so too. The logger is still troublesome. So then, is there still any possible test case for this specific pull request, which only improves the thread-safety of the component database?
| gharchive/pull-request | 2021-01-12T22:33:23 | 2025-04-01T04:33:06.998357 | {
"authors": [
"N-Dekker",
"mstaring",
"ntatsisk"
],
"repo": "SuperElastix/elastix",
"url": "https://github.com/SuperElastix/elastix/pull/389",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1686496285 | 🛑 Matrix redirection is down
In 1205ae1, Matrix redirection (https://supersandro.de/_matrix/key/v2/server) was down:
HTTP code: 504
Response time: 20158 ms
Resolved: Matrix redirection is back up in d1bd924.
| gharchive/issue | 2023-04-27T09:42:38 | 2025-04-01T04:33:07.005132 | {
"authors": [
"SuperSandro2000"
],
"repo": "SuperSandro2000/upptime",
"url": "https://github.com/SuperSandro2000/upptime/issues/196",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1832093961 | 🛑 Ping matrix.supersandro.de is down
In 729f784, Ping matrix.supersandro.de (matrix.supersandro.de) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Ping matrix.supersandro.de is back up in c61fb44.
| gharchive/issue | 2023-08-01T22:03:52 | 2025-04-01T04:33:07.008153 | {
"authors": [
"SuperSandro2000"
],
"repo": "SuperSandro2000/upptime",
"url": "https://github.com/SuperSandro2000/upptime/issues/371",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1160422000 | Development/algorithm
Follow: #47
Done
[x] Analysis of Complexion
[x] Recursion
[x] Divide & Conquer
[ ] Hash
[x] Backtracking
[x] Greedy
[ ] Dynamic Programming
[x] Sort
[ ] Search
[ ] String Matching
And I re-build the Basic part of the Algorithm.
I'll add note for 8-queens in the future.
| gharchive/pull-request | 2022-03-05T18:51:58 | 2025-04-01T04:33:07.018737 | {
"authors": [
"Jacob953"
],
"repo": "Superego-CodeEngineer/GoGetit",
"url": "https://github.com/Superego-CodeEngineer/GoGetit/pull/52",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
482149788 | Windows 10 version 1903 - Original Windows 8 key not retrieved
Hello.
On the latest edition of Windows 10 (version 1903), for computers installed directly with Windows 8 keys in a clean install process, and which have been delivered a digital activation, the tool doesn't display the original Windows 8 key and edition (only the Windows 10 generic key is displayed).
Is it a known issue? Are you working on it? or is there a limitation in this version of Windows 10 which prevents from obtaining this information?
Thank you.
Best regards.
Hi geolives
A clean install removes all traces of the original installaton and the key will not be there - only an upgrade will show the original key.
Regards
| gharchive/issue | 2019-08-19T07:28:28 | 2025-04-01T04:33:07.020560 | {
"authors": [
"Superfly-Inc",
"geolives"
],
"repo": "Superfly-Inc/ShowKeyPlus",
"url": "https://github.com/Superfly-Inc/ShowKeyPlus/issues/23",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2623414555 | Imagine ai update button doc
Description
I have made some changes inside button doc that are following
👉🏻 Front page of button changed totally
👉🏻Updates user experience with BG and hover and more
👉🏻updates toggles buttons color and BG
👉🏻made other changes
Fixes #1361
Screenshots
BEFORE
AFTER
Checklist
[✅] Tests have been added or updated to cover the changes
[✅] Documentation has been updated to reflect the changes
[✅] Code follows the established coding style guidelines
[✅] All tests are passing
You need to show changes on the project you are working, this ain't my project
So sorry for having glitches and misunderstanding with these project as soon as I will update you with changes ...
| gharchive/pull-request | 2024-10-30T09:31:25 | 2025-04-01T04:33:07.041167 | {
"authors": [
"SurajPratap10",
"ankitmodanwall"
],
"repo": "SurajPratap10/Imagine_AI",
"url": "https://github.com/SurajPratap10/Imagine_AI/pull/1362",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
296986237 | Node strange issues
Here are a few node issues which happened when running - not sure why they happened yet. But tracking them here so we can fix.
Closed stream and invalid nonce on standalone node (when only running 1 miner with 2 nodes started up (1 standalone, 2 connected to standalone)
Unhandled exception (I think this happened on the standalone node when a node connected to it was stopped)
I'll try to recreate the situation. Thanks
| gharchive/issue | 2018-02-14T06:07:14 | 2025-04-01T04:33:07.056822 | {
"authors": [
"kingsleyh",
"tbrand"
],
"repo": "SushiCoinHQ/SushiCoin",
"url": "https://github.com/SushiCoinHQ/SushiCoin/issues/72",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
976843291 | 解压后的文件夹会被占用,无法进行其他操作
系统环境: Windows Server 2019
面板版本: 8.6.23
Node 版本: 16.7.0
出现概率: 100
出现的问题 (BUG): 解压后的文件夹会被占用,无法进行其他操作
复现操作步骤: 文件管理>解压
解压到到一半,未完成?
亦或防病毒软件扫描中?
解压到到一半,未完成?
亦或防病毒软件扫描中?
并不是
| gharchive/issue | 2021-08-23T09:49:37 | 2025-04-01T04:33:07.068185 | {
"authors": [
"SeaIceNX",
"huangsijun17"
],
"repo": "Suwings/MCSManager",
"url": "https://github.com/Suwings/MCSManager/issues/473",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
500742831 | Joke type query parameter
To only get jokes from a certain type (single or twopart)
Example: https://sv443.net/jokeapi/category/Any?type=single
Added in 7925adf468b5bf7bf62fc7e7c6f6e910ed9ca2d0
| gharchive/issue | 2019-10-01T08:30:08 | 2025-04-01T04:33:07.069540 | {
"authors": [
"Sv443"
],
"repo": "Sv443/JokeAPI",
"url": "https://github.com/Sv443/JokeAPI/issues/16",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
377524998 | Generate Table of Contents
I'm trying out this tool and it is able generate both separate files and one single AsciiDoctor file, however I'm not seeing the Table of Contents as shown in the Swagger2Markup Example.
I didn't see anything in the configuration section, so I feel like I'm missing something.
It's an asciidoctor feature.
| gharchive/issue | 2018-11-05T18:26:23 | 2025-04-01T04:33:07.095012 | {
"authors": [
"RobWin",
"jvtrigueros"
],
"repo": "Swagger2Markup/swagger2markup-cli",
"url": "https://github.com/Swagger2Markup/swagger2markup-cli/issues/24",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2069364206 | get memory size
Description
增加获取内存大小的函数,为v0.1.4做准备
你在这分支顺带着把system.py里面的print改成swanlog吧
| gharchive/pull-request | 2024-01-08T00:08:21 | 2025-04-01T04:33:07.095973 | {
"authors": [
"SAKURA-CAT",
"xiaolin199912"
],
"repo": "SwanHubX/SwanLab",
"url": "https://github.com/SwanHubX/SwanLab/pull/169",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2337201943 | 🛑 swift-format.com is down
In 65097b2, swift-format.com (https://swift-format.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: swift-format.com is back up in b996093 after 21 minutes.
| gharchive/issue | 2024-06-06T02:47:39 | 2025-04-01T04:33:07.109986 | {
"authors": [
"kishikawakatsumi"
],
"repo": "SwiftFiddle/status.swift-format.com",
"url": "https://github.com/SwiftFiddle/status.swift-format.com/issues/81",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
222488764 | Update to swift 3.1 / XCode 8.3
This updates the FileKit and Sourcekitten Dependencies to their current versions to make Cuckoo usable with Xcode 8.3
If I'm not mistaken travis still runs on Xcode 8, not even 8.2. So this would probably need to be updated to at least Xcode 8.2 or better 8.3.
Hey, could you give stencil branch a try? It should work with Swift 3.1 and I'll be merging it by the end of this week (I just need more people to try it out). Thanks!
Didn't get around to test the stencil branch yesterday.
Release 0.9.0 looks good so far. My tests still run and it works with swift 3.1 /Xcode 8.3.2.
I will look into the issues I posted to check whether they are fixed with this version in the next few days.
So this PR is no longer needed.
| gharchive/pull-request | 2017-04-18T17:27:02 | 2025-04-01T04:33:07.119289 | {
"authors": [
"TadeasKriz",
"rbeeger"
],
"repo": "SwiftKit/Cuckoo",
"url": "https://github.com/SwiftKit/Cuckoo/pull/120",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.